FFMPEG ffmpeg + cocos2d video test

Posted yzfhkms-x

tags:

篇首语:本文由小常识网(cha138.com)小编为大家整理,主要介绍了FFMPEG ffmpeg + cocos2d video test相关的知识,希望对你有一定的参考价值。

//核心

{

void AVdio::update(float delta)
{
	auto video = (Sprite*)(this->getChildByName("video"));
	if (video != nullptr)
	{
		double      tims = fl->_pts * fl->_timeBase * 1000;
		double      elsped = _timestamp.getElapsedTimeInMilliSec();
		double      sleeps = (tims - elsped);

		s = sleeps + elsped;

		if (elsped - s > 0.0001)
		{
			is = ffReader.readFrame(*fl);
			if (is)
			{
				texture->initWithData((const unsigned char*)fl->_data, fl->_dataSize, Texture2D::PixelFormat::RGB888,
					fl->_width, fl->_height,
					Size(fl->_width, fl->_height));

				video->initWithTexture(texture);
				video->setContentSize(Size(displayex.width - 10, displayex.height - 50));
			}
			else
			{
				video->removeFromParent();
			}
		}
	}

}

 

}

 

.h

{

#ifndef __AVDIO_H__
#define __AVDIO_H__

#include "cocos2d.h"
using namespace cocos2d;
#include <string.h>
#include <iostream>
#include "cocos/ui/CocosGUI.h"
#include "socket/Private/ThreadPool.h"
#include "Tools/Tools.h"
#include "xml/XML.h"
#include "iconv/UTF8.h"
#ifdef _WIN32
#include "ATBAudioEngine/ATBAudioEngine.h"
#endif
using namespace std;
using namespace ui;


class AVdio :public LayerColor, EditBoxDelegate
{
	Tools m_Tools;
public:

	static cocos2d::LayerColor* createAVdio();

	~AVdio();

	virtual bool init();

	CREATE_FUNC(AVdio);



	void OnCallback(cocos2d::Ref* pSender);
	virtual bool onTouchBegan(Touch *touch, Event *unused_event) override;
	virtual void update(float delta);

	virtual void editBoxEditingDidBegin(EditBox* editBox)override;
	CC_DEPRECATED_ATTRIBUTE virtual void editBoxEditingDidEnd(EditBox* editBox)override;
	virtual void editBoxTextChanged(EditBox* editBox, const std::string& text)override;
	virtual void editBoxReturn(EditBox* editBox)override;
	virtual void editBoxEditingDidEndWithAction(EditBox* editBox, EditBoxDelegate::EditBoxEndAction action)override;
private:
	void initVideoStream(string filename);
	void initInput();
};









extern "C"
{
#include <libavutil/imgutils.h>
#include <libavutil/parseutils.h>
#include <libswscale/swscale.h>
#include <libavcodec/avcodec.h>
#include <libavformat/avformat.h>
#include <libavformat/avio.h>
#include <libavutil/file.h>
#include <libavdevice/avdevice.h>
}

struct  FrameInfor
{
	void*   _data;
	int     _dataSize;
	int     _width;
	int     _height;
	int64_t _pts;
	double  _timeBase;

};
class   FFVideoReader
{
public:
	AVFormatContext*_formatCtx;
	int             _videoIndex;
	AVCodecContext* _codecCtx;
	AVCodec*        _codec;
	AVFrame*        _frame;
	AVFrame*        _frameRGB;
	SwsContext*     _convertCtx;
public:
	int             _screenW;
	int             _screenH;

	int             _imageSize;
public:
	FFVideoReader()
	{
		_formatCtx = 0;
		_videoIndex = -1;
		_codecCtx = 0;
		_codec = 0;
		_frame = 0;
		_frameRGB = 0;
		_convertCtx = 0;
		_screenW = 0;
		_screenH = 0;

	}

	~FFVideoReader()
	{
		sws_freeContext(_convertCtx);
		av_free(_frameRGB);
		av_free(_frame);
		avcodec_close(_codecCtx);
		avformat_close_input(&_formatCtx);
	}

	void    setup()
	{
		av_register_all();
		_formatCtx = avformat_alloc_context();
	}
	int     load(const char* filepath = "11.flv")
	{
		int     ret = 0;

		//! 打开文件
		if (avformat_open_input(&_formatCtx, filepath, NULL, NULL) != 0)
		{
			return -1;
		}
		//! 检测文件中是否存在数据流
		if (avformat_find_stream_info(_formatCtx, NULL) < 0)
		{
			printf("检测文件中是否存在数据流");
			return -1;
		}
		//! 获取视频流索引
		_videoIndex = -1;
		for (int i = 0; i < _formatCtx->nb_streams; i++)
		{
			if (_formatCtx->streams[i]->codec->codec_type == AVMEDIA_TYPE_VIDEO)
			{
				_videoIndex = i;
				break;
			}
		}
		/**
		*   没有视频流,则返回
		*/
		if (_videoIndex == -1)
		{
			return -1;
		}
		_codecCtx = _formatCtx->streams[_videoIndex]->codec;

		double dur = _formatCtx->duration / double(AV_TIME_BASE);
		_codec = avcodec_find_decoder(_codecCtx->codec_id);
		if (_codec == NULL)
		{
			printf("find decoder faild !!
");
			return -1;
		}
		/**
		*   打开解码器
		*/
		if (avcodec_open2(_codecCtx, _codec, NULL) < 0)
		{
			return -1;
		}
		_frame = av_frame_alloc();
		_frameRGB = av_frame_alloc();

		_screenW = _codecCtx->width;
		_screenH = _codecCtx->height;

		_convertCtx = sws_getContext(
			_codecCtx->width
			, _codecCtx->height
			, _codecCtx->pix_fmt
			, _codecCtx->width
			, _codecCtx->height
			, AV_PIX_FMT_RGB24
			, SWS_BICUBIC
			, NULL
			, NULL
			, NULL
		);

		int     numBytes = avpicture_get_size(AV_PIX_FMT_RGB24, _codecCtx->width, _codecCtx->height);
		uint8_t*buffer = (uint8_t *)av_malloc(numBytes * sizeof(uint8_t));
		avpicture_fill((AVPicture *)_frameRGB, buffer, AV_PIX_FMT_RGB24, _codecCtx->width, _codecCtx->height);
		_imageSize = numBytes;
		return  0;
	}

	bool    readFrame(FrameInfor& infor)
	{
		AVPacket packet;
		av_init_packet(&packet);
		for (;;)
		{
			if (av_read_frame(_formatCtx, &packet))
			{
				av_free_packet(&packet);
				return false;
			}
			if (packet.stream_index != _videoIndex)
			{
				continue;
			}
			int frame_finished = 0;

			int res = avcodec_decode_video2(_codecCtx, _frame, &frame_finished, &packet);

			if (frame_finished)
			{
				AVStream*   streams = _formatCtx->streams[_videoIndex];
				double      tmbase = av_q2d(streams->time_base);
				int64_t     pts = _frame->pts;

				char        buf[128];
				sprintf(buf, "pts = %I64d     dts =  %I64d
", packet.pts, packet.dts);
				int res = sws_scale(
					_convertCtx
					, (const uint8_t* const*)_frame->data
					, _frame->linesize
					, 0
					, _codecCtx->height
					, _frameRGB->data
					, _frameRGB->linesize
				);
				av_packet_unref(&packet);

				infor._data = _frameRGB->data[0];
				infor._dataSize = _imageSize;
				infor._width = _screenW;
				infor._height = _screenH;
				infor._pts = _frame->pts;
				infor._timeBase = av_q2d(streams->time_base);

				return  true;
			}
		}
		return  false;
	}
	void*   readFrame()
	{
		AVPacket packet;
		av_init_packet(&packet);
		for (;;)
		{
			if (av_read_frame(_formatCtx, &packet))
			{
				av_free_packet(&packet);
				return 0;
			}
			if (packet.stream_index != _videoIndex)
			{
				continue;
			}
			int frame_finished = 0;

			int res = avcodec_decode_video2(_codecCtx, _frame, &frame_finished, &packet);

			if (frame_finished)
			{
				AVStream*   streams = _formatCtx->streams[_videoIndex];
				double      tmbase = av_q2d(streams->time_base);
				int64_t     pts = _frame->pts;

				char        buf[128];
				sprintf(buf, "pts = %I64d     dts =  %I64d
", packet.pts, packet.dts);
				int res = sws_scale(
					_convertCtx
					, (const uint8_t* const*)_frame->data
					, _frame->linesize
					, 0
					, _codecCtx->height
					, _frameRGB->data
					, _frameRGB->linesize
				);
				av_packet_unref(&packet);

				return  _frameRGB->data[0];
			}
		}
		return  0;
	}
};



#include <windows.h>

class Timestamp
{
public:
	Timestamp()
	{
		QueryPerformanceFrequency(&_frequency);
		QueryPerformanceCounter(&_startCount);
	}
	~Timestamp()
	{}

	void    update()
	{
		QueryPerformanceCounter(&_startCount);
	}
	/**
	*   获取当前秒
	*/
	double getElapsedSecond()
	{
		return  getElapsedTimeInMicroSec() * 0.000001;
	}
	/**
	*   获取毫秒
	*/
	double getElapsedTimeInMilliSec()
	{
		return this->getElapsedTimeInMicroSec() * 0.001;
	}
	/**
	*   获取微妙
	*/
	double getElapsedTimeInMicroSec()
	{
		LARGE_INTEGER endCount;
		QueryPerformanceCounter(&endCount);

		double  startTimeInMicroSec = _startCount.QuadPart * (1000000.0 / _frequency.QuadPart);
		double  endTimeInMicroSec = endCount.QuadPart * (1000000.0 / _frequency.QuadPart);

		return  endTimeInMicroSec - startTimeInMicroSec;
	}
protected:
	LARGE_INTEGER   _frequency;
	LARGE_INTEGER   _startCount;
};





#endif

 

}

 

.cpp

{

#include "AVdio.h"

#ifndef DISPLAY
#define DISPLAY Director::getInstance()->getVisibleSize()
#endif
#ifndef displayex
#define displayex Director::getInstance()->getVisibleSize()
#endif


FFVideoReader   ffReader;
FrameInfor *fl = new (std::nothrow) FrameInfor;
auto texture = new (std::nothrow) Texture2D();
bool is = false;
Timestamp       _timestamp;
double s = 0;


cocos2d::LayerColor* AVdio::createAVdio()
{
	auto LC = AVdio::create();
	if (LC)
	{
		return LC;
	}
	else
	{
		return nullptr;
	}
}

AVdio::~AVdio()
{

}

void AVdio::update(float delta)
{
	auto video = (Sprite*)(this->getChildByName("video"));
	if (video != nullptr)
	{
		double      tims = fl->_pts * fl->_timeBase * 1000;
		double      elsped = _timestamp.getElapsedTimeInMilliSec();
		double      sleeps = (tims - elsped);

		s = sleeps + elsped;

		if (elsped - s > 0.0001)
		{
			is = ffReader.readFrame(*fl);
			if (is)
			{
				texture->initWithData((const unsigned char*)fl->_data, fl->_dataSize, Texture2D::PixelFormat::RGB888,
					fl->_width, fl->_height,
					Size(fl->_width, fl->_height));

				video->initWithTexture(texture);
				video->setContentSize(Size(displayex.width - 10, displayex.height - 50));
			}
			else
			{
				video->removeFromParent();
			}
		}
	}

}

bool AVdio::init()
{
	if (!LayerColor::init())
	{
		return false;
	}

	auto display = Director::getInstance()->getVisibleSize();
	Vec2 origin = Director::getInstance()->getVisibleOrigin();

	this->setTouchEnabled(true);
	auto ELTOBO = EventListenerTouchOneByOne::create();
	ELTOBO->setSwallowTouches(true);
	ELTOBO->onTouchBegan = std::move(std::bind(&AVdio::onTouchBegan, this, std::placeholders::_1, std::placeholders::_2));
	this->getEventDispatcher()->addEventListenerWithSceneGraphpriority(ELTOBO, this);




	auto background = cocos2d::LayerColor::create(cocos2d::Color4B(0, 0, 0, 255));
	this->addChild(background);

	//close button
	string file("res/Button.png");
	auto btn = cocos2d::ui::Button::create(file, file, file);
	btn->setColor(Color3B(24, 48, 64));
	btn->setPressedActionEnabled(true);
	btn->setScale9Enabled(true);
	btn->setContentSize(Size(100, 50));
	btn->setPosition(Vec2(displayex.width - btn->getContentSize().width / 2,
		displayex.height - btn->getContentSize().height / 2));
	btn->setTitleColor(Color3B::RED);
	btn->setTitleFontSize(50);
	btn->setName("X");
	btn->setTitleText("X");
	btn->addClickEventListener(std::bind(&AVdio::OnCallback, this, std::placeholders::_1));
	this->addChild(btn);


	//this->initVideoStream();

	this->initInput();
}

void AVdio::initInput()
{
	auto edit = ui::EditBox::create(Size(800, 50), "res/input.png");
	edit->setPosition(Vec2(0, displayex.height - edit->getContentSize().height / 2));
	edit->setAnchorPoint(Vec2(0.0f, 0.5f));
	edit->setName("videoPath");
	edit->setFontColor(Color4B::YELLOW);
	edit->setDelegate(this);
	this->addChild(edit);

	string file("res/Button.png");
	auto btn = cocos2d::ui::Button::create(file, file, file);
	btn->setColor(Color3B(24, 48, 64));
	btn->setPressedActionEnabled(true);
	btn->setScale9Enabled(true);
	btn->setContentSize(Size(150, 50));
	btn->setPosition(Vec2(displayex.width - (displayex.width - edit->getContentSize().width) + btn->getContentSize().width / 2,
		edit->getPosition().y));
	btn->setTitleColor(Color3B::RED);
	btn->setTitleFontSize(50);
	btn->setName("PLAY");
	btn->setTitleText("PLAY");
	btn->addClickEventListener(std::bind(&AVdio::OnCallback, this, std::placeholders::_1));
	this->addChild(btn);
}

void AVdio::initVideoStream(string filename)
{
	this->unscheduleUpdate();

	_timestamp.update();
	
	

	ATBAE::GetInstance()->PauseAllMusicAndEffects();
	auto video = (Sprite*)(this->getChildByName("video"));
	if (video != nullptr)
	{
		video->removeFromParent();
	}

	s = 0;

	auto display = Director::getInstance()->getVisibleSize();
	Vec2 origin = Director::getInstance()->getVisibleOrigin();

	auto ret = avdevice_version();
	std::cout << ret << std::endl;

	ffReader.~FFVideoReader();
	ffReader.setup();
	ffReader.load(filename.c_str());

	ATBAE::GetInstance()->LoadMusicsAndPlay(filename.c_str());


	is = ffReader.readFrame(*fl);

	texture->initWithData((const unsigned char*)fl->_data, fl->_dataSize, Texture2D::PixelFormat::RGB888,
		fl->_width - 1, fl->_height - 1,
		Size(500, 500));

	auto s = Sprite::create("06a03.jpg");
	s->initWithTexture(texture);
	s->setName("video");
	s->setPosition(Vec2(display.width / 2, display.height / 2 - 25));
	this->addChild(s);

	this->scheduleUpdate();
}


void AVdio::OnCallback(cocos2d::Ref* pSender)
{
	string name = ((Node*)(pSender))->getName();
	if (name == "X")
	{
		auto action = Sequence::create(MoveTo::create(0.2f, Vec3(-(displayex.width), 0, 0)),
			DelayTime::create(0.1f),
			CallFunc::create([=]()
		{
			this->removeFromParent();
			ATBAE::GetInstance()->PauseAllMusicAndEffects();
		}), nullptr);
		this->runAction(action);
	}
	else if (name == "PLAY")
	{
		auto edit = (ui::EditBox*)this->getChildByName("videoPath");
		string path = edit->getText();
		if (path.length() > 0)
		{
			this->initVideoStream(path);
		}
		
	}
}
bool AVdio::onTouchBegan(Touch *touch, Event *unused_event)
{
	return true;
}


void AVdio::editBoxEditingDidBegin(EditBox* editBox)
{

}
void AVdio::editBoxEditingDidEnd(EditBox* editBox)
{

}
void AVdio::editBoxTextChanged(EditBox* editBox, const std::string& text)
{

}
void AVdio::editBoxReturn(EditBox* editBox)
{

}
void AVdio::editBoxEditingDidEndWithAction(EditBox* editBox, EditBoxDelegate::EditBoxEndAction action)
{

}

 

}

以上是关于FFMPEG ffmpeg + cocos2d video test的主要内容,如果未能解决你的问题,请参考以下文章

将视频转换为 H264 时出现无法识别的选项“c:v”ffmpeg 错误

ffmpeg命令行示例

ffmpeg实践

ffmpeg overlay图片异常

ffmpeg 复制以保留时间戳

我想使用 ffmpeg 添加水印和文字