ffmpeg转码常用容器编码格式与解编码器的搭配总结,-c与-f参数的运用
Posted 戴国进
tags:
篇首语:本文由小常识网(cha138.com)小编为大家整理,主要介绍了ffmpeg转码常用容器编码格式与解编码器的搭配总结,-c与-f参数的运用相关的知识,希望对你有一定的参考价值。
转码
现今常用的一些编码格式以及它们所支持的容器格式如下:
特别提一点,mkv是个万能的容器格式,下面说的所有编码格式,mkv几乎都能“装”,就不再列出了。
视频编码格式:
h264(又称mpeg-4 avc、mpeg-4 part 10):mp4、flv、avi、mov、wmv、m4v、f4v、3gp、ts
mpeg4(不只一种,这里指mpeg-4 part 2、divx、xvid):mp4、avi、mov、wmv、m4v、3gp、ts
h265(又称hevc、mpeg-h part 2):mp4、avi、mov、ts
vp8:avi、wmv、ts、webm
vp9:mp4、avi、wmv、ts、webm
音频编码格式:(【】左边是视频容器格式,【】右边是音频容器格式,*和^后面再说明)
aac:mp4、flv、avi、mov、wmv、3gp、m4v、f4v、ts【】aac、m4a、wma、ac3
mp3:mp4、avi、mov、wmv、f4v、ts【】mp3、wma、ac3
ac-3:mp4、avi*、mov、wmv、m4v、ts【】ac3、m4a、wma
flac:mp4*、avi*、wmv、ts【】flac、m4a、wma、ac3^
vorbis:mp4、avi*、mov、wmv、ts、webm【】ogg、wma、ac3^
opus:mp4*、ts、webm【】ogg、ac3^
(*的意思是需要进一步使用相应参数,小白不建议)
(^的意思是能转码成功,但很可能播放器不能播放)
(一般听感来说,上述中opus编码格式是有损编码里最好的,其次是vorbis,之后是aac、mp3之类)
(ffmpeg似乎不支持ape容器格式)
对新人来说看到这些庞杂的格式可能很懵,通俗说明一下:“一样容器能装不同菜,一样菜能盛在不同容器”的道理,我们只要找到菜和容器都互相能“接受”的组合就行了。只是视频是一个容器两道菜(画面和声音),音频就一个容器一道菜。
反之,上面列的这些,也可以理解为视音频容器格式所支持的编码格式。上期说一个视频有画面和声音,那么画和声就有它俩各自的编码格式,所以对一个现存的视频来说,它的容器格式要同时支持画和声的编码格式;音频就只有声音的编码格式,和它的容器格式互相支持就可以。
这里指出两点前提:以原文件的编码为参考,ffmpeg的支持。
虽然有着许多的格式,但目前大多数视频的编码格式是h264和aac,音频的编码格式是mp3或aac。因此,只要不同的容器格式之间能同时支持一样的编码格式,就能够进行“高效转码”。
例如同时支持h264和aac的有mp4、flv、avi、mov、wmv……满足这个前提后,这些容器两两之间都可以任意“高效转码”。可如果像上期说的,只知道容器格式mp4和flv,并不能直接认为它俩可以用“-c copy”“高效转码”。
对于需要转换编码格式的音视频,用ffmpeg的前提是ffmpeg要能支持我们想转的编码格式。ffmpeg就像一个批发中心,但它的货总归还是有限的。
不过对于无需转换编码格式的情况,即便是ffmpeg不支持的编码格式,也能用“-c copy”复制照抄(例如将原视频转码成mkv,可理解为原视频中的编码格式给ffmpeg提供了个“模板”),但是如果ffmpeg不支持最后输出文件的容器格式的话,就不能这样转码。
除了音视频,ffmpeg还可以转码字幕和图像文件:
(直接按“输入→输出”流程即可,例如ffmpeg -i .\\input.png output.jpg)
(srt转码ass要注意srt先以记事本打开,另存为将编码改成UTF-8或unicode)
字幕容器格式:srt、ass、ssa之间两两任意互转
图像容器格式:jpg、jpeg、png、bmp、gif、webp之间两两任意互转
-c与-f参数的运用
若要转换编码格式,我们将运用“-c”来使用ffmpeg支持的编解码器。
“-c”参数包括了音视频的全部编解码器,所以以前的“-c copy”是复制音视频编码的意思。当然我们也可以直接用“-c copy”转码音频,没有视频也没关系。
要只对视频里的画面或者声音进行操作的话,我们可以用“-c:v”来限定只处理视频画面,“-c:a”来限定只处理视频里的音频声音,“-c:s”来限定只处理字幕。
经常用的比如“-c:v libx264”,表示将编码格式转换成h264;“-c:a libmp3lame”,表示将编码格式转换成mp3。
这里的“libx264”、“libmp3lame”是编码器的名字(文末再说编码器名怎么得知)。简单形象地说,“容器格式、编码格式、编码器”的一个可行搭配是“avi、(h264、mp3)、(libx264、libmp3lame)”。
我们也可以用“-c:v h264”、“-c:a mp3”来进行上述操作,这样的话不用知道编码器名也可以。如下图:
ffmpeg里的“stream”可以理解为“流”、“轨道”、“通道”。红圈里的是编码格式转换的指示说明,从原来的什么转换成什么。例如上图是,视频编码格式从mpeg4转换成h264,视频里的音频编码格式从aac转换成mp3。
我们还可以通过“-f”参数来进行转码。“-f”,“format,格式”的意思,强制输出什么格式的文件,让ffmpeg自行挑选编解码器进行转码输出。但只是一般情况来说,就像前面说过ffmpeg的支持也是有限的,不支持的话当然也转码输出不了。
这个“-f”可以用来“查询”自己想转的容器格式是有什么编码格式或编码器可以支持。它的用法例如“-f webm”。如下图:
这说明vp9和opus是webm容器格式支持的编码格式。
(在一些时候,用“-f”参数转码完的视频质量会不好,这个以后再说)
转码思路流程
①先用命令行ffprobe .\\input.mp4
查看文件的编码格式,如下图:
红圈表示该视频的画面(video)编码格式是h264,声音(audio)编码格式是aac
另外,一般查看文件属性也可以知道其编码格式
②确定自己想转的容器或编码格式
③用“ffmpeg -formats”命令行
查询ffmpeg支持哪些容器格式。如下图
demuxing是“解封装”的意思,muxing是“封装”的意思,
就像容器格式又称封装格式
转码也可以看作是封装
解封装就可以看作是转码的相反操作
还有,用“ffmpeg -codecs”命令行
查询ffmpeg支持哪些编码格式及编解码器名。如下图
decoder指解码器名,encoder指编码器名,以此类推
④输入相应命令行进行转码
FFMPEG实现的转码程序
本例子是由FFEMPG的doc/example例子transcode.c修改而来,可以根据需求任意转换音视频的编码。
原来的例子的作用更类似于remux,并没有实现转码的功能,只是实现了格式转换,比如ts转avi等。并不能实现音视频编码格式的转换,比如将h264转为mpeg2。
FFMPEG转码的实现有多种方式:
一种方式是:流解复用->视频+音频流->解码->YUV/PCM等->视音频编码->重新生成的音视频流->复用->流
另一种方式依赖AVFilter,这一部分在另外的几篇文章中解释怎么用。虽然AVFilter学习起来可能比较困难,但是在实际的编程应用中,依靠AVFilter做转码效率比第一种方式高,并且解码的CPU和时间消耗也少的多。所以,还是建议好好学习这部分的,毕竟我一直觉得FFMPEG的强项就是解码和转码。
本例子是视频mpeg2转h264,音频mpegaudio转g711。
- <span style="font-family:SimHei;font-size:18px;">/*
- * based on FFMPEG transcode.c
- * modified by tongli
- */
- #include <stdio.h>
- #include "snprintf.h"
- extern "C"
- {
- #include <libavcodec/avcodec.h>
- #include <libavformat/avformat.h>
- #include <libavfilter/avfiltergraph.h>
- #include <libavfilter/avcodec.h>
- #include <libavfilter/buffersink.h>
- #include <libavfilter/buffersrc.h>
- #include <libavutil/opt.h>
- #include <libavutil/pixdesc.h>
- }
- static AVFormatContext *ifmt_ctx;
- static AVFormatContext *ofmt_ctx;
- typedef struct FilteringContext {
- AVFilterContext *buffersink_ctx;
- AVFilterContext *buffersrc_ctx;
- AVFilterGraph *filter_graph;
- } FilteringContext;
- static FilteringContext *filter_ctx;
- static int open_input_file(const char *filename)
- {
- int ret;
- unsigned int i;
- ifmt_ctx = NULL;
- if ((ret = avformat_open_input(&ifmt_ctx, filename, NULL, NULL)) < 0) {
- av_log(NULL, AV_LOG_ERROR, "Cannot open input file\n");
- return ret;
- }
- if ((ret = avformat_find_stream_info(ifmt_ctx, NULL)) < 0) {
- av_log(NULL, AV_LOG_ERROR, "Cannot find stream information\n");
- return ret;
- }
- for (i = 0; i < ifmt_ctx->nb_streams; i++) {
- AVStream *stream;
- AVCodecContext *codec_ctx;
- stream = ifmt_ctx->streams[i];
- codec_ctx = stream->codec;
- /* Reencode video & audio and remux subtitles etc. */
- if (codec_ctx->codec_type == AVMEDIA_TYPE_VIDEO
- || codec_ctx->codec_type == AVMEDIA_TYPE_AUDIO) {
- /* Open decoder */
- ret = avcodec_open2(codec_ctx,
- avcodec_find_decoder(codec_ctx->codec_id), NULL);
- if (ret < 0) {
- av_log(NULL, AV_LOG_ERROR, "Failed to open decoder for stream #%u\n", i);
- return ret;
- }
- }
- }
- av_dump_format(ifmt_ctx, 0, filename, 0);
- return 0;
- }
- static int open_output_file(const char *filename)
- {
- AVStream *out_stream;
- AVStream *in_stream;
- AVCodecContext *dec_ctx, *enc_ctx;
- AVCodec *encoder;
- int ret;
- unsigned int i;
- ofmt_ctx = NULL;
- avformat_alloc_output_context2(&ofmt_ctx, NULL, NULL, filename);
- if (!ofmt_ctx) {
- av_log(NULL, AV_LOG_ERROR, "Could not create output context\n");
- return AVERROR_UNKNOWN;
- }
- for (i = 0; i < ifmt_ctx->nb_streams; i++) {
- out_stream = avformat_new_stream(ofmt_ctx, NULL);
- if (!out_stream) {
- av_log(NULL, AV_LOG_ERROR, "Failed allocating output stream\n");
- return AVERROR_UNKNOWN;
- }
- in_stream = ifmt_ctx->streams[i];
- dec_ctx = in_stream->codec;
- enc_ctx = out_stream->codec;
- if (dec_ctx->codec_type == AVMEDIA_TYPE_VIDEO)
- {
- encoder = avcodec_find_encoder(AV_CODEC_ID_H264);
- if (!encoder) {
- av_log(NULL, AV_LOG_FATAL, "Neccessary encoder not found\n");
- return AVERROR_INVALIDDATA;
- }
- enc_ctx->height = dec_ctx->height;
- enc_ctx->width = dec_ctx->width;
- enc_ctx->sample_aspect_ratio = dec_ctx->sample_aspect_ratio;
- enc_ctx->pix_fmt = encoder->pix_fmts[0];
- enc_ctx->time_base = dec_ctx->time_base;
- enc_ctx->me_range = 16;
- enc_ctx->max_qdiff = 4;
- enc_ctx->qmin = 10;
- enc_ctx->qmax = 51;
- enc_ctx->qcompress = 0.6;
- enc_ctx->refs = 3;
- enc_ctx->bit_rate = 500000;
- ret = avcodec_open2(enc_ctx, encoder, NULL);
- if (ret < 0) {
- av_log(NULL, AV_LOG_ERROR, "Cannot open video encoder for stream #%u\n", i);
- return ret;
- }
- }
- else if (dec_ctx->codec_type == AVMEDIA_TYPE_UNKNOWN) {
- av_log(NULL, AV_LOG_FATAL, "Elementary stream #%d is of unknown type, cannot proceed\n", i);
- return AVERROR_INVALIDDATA;
- }
- else if (dec_ctx->codec_type == AVMEDIA_TYPE_AUDIO)
- {
- encoder = avcodec_find_encoder(AV_CODEC_ID_PCM_ALAW);
- enc_ctx->sample_rate = dec_ctx->sample_rate;
- enc_ctx->channel_layout = dec_ctx->channel_layout;
- enc_ctx->channels = av_get_channel_layout_nb_channels(enc_ctx->channel_layout);
- enc_ctx->sample_fmt = encoder->sample_fmts[0];
- AVRational ar = { 1, enc_ctx->sample_rate };
- enc_ctx->time_base = ar;
- ret = avcodec_open2(enc_ctx, encoder, NULL);
- if (ret < 0) {
- av_log(NULL, AV_LOG_ERROR, "Cannot open video encoder for stream #%u\n", i);
- return ret;
- }
- }
- else {
- ret = avcodec_copy_context(ofmt_ctx->streams[i]->codec,
- ifmt_ctx->streams[i]->codec);
- if (ret < 0) {
- av_log(NULL, AV_LOG_ERROR, "Copying stream context failed\n");
- return ret;
- }
- }
- if (ofmt_ctx->oformat->flags & AVFMT_GLOBALHEADER)
- enc_ctx->flags |= CODEC_FLAG_GLOBAL_HEADER;
- }
- av_dump_format(ofmt_ctx, 0, filename, 1);
- if (!(ofmt_ctx->oformat->flags & AVFMT_NOFILE)) {
- ret = avio_open(&ofmt_ctx->pb, filename, AVIO_FLAG_WRITE);
- if (ret < 0) {
- av_log(NULL, AV_LOG_ERROR, "Could not open output file ‘%s‘", filename);
- return ret;
- }
- }
- /* init muxer, write output file header */
- ret = avformat_write_header(ofmt_ctx, NULL);
- if (ret < 0) {
- av_log(NULL, AV_LOG_ERROR, "Error occurred when opening output file\n");
- return ret;
- }
- return 0;
- }
- static int init_filter(FilteringContext* fctx, AVCodecContext *dec_ctx,
- AVCodecContext *enc_ctx, const char *filter_spec)
- {
- char args[512];
- int ret = 0;
- AVFilter *buffersrc = NULL;
- AVFilter *buffersink = NULL;
- AVFilterContext *buffersrc_ctx = NULL;
- AVFilterContext *buffersink_ctx = NULL;
- AVFilterInOut *outputs = avfilter_inout_alloc();
- AVFilterInOut *inputs = avfilter_inout_alloc();
- AVFilterGraph *filter_graph = avfilter_graph_alloc();
- if (!outputs || !inputs || !filter_graph) {
- ret = AVERROR(ENOMEM);
- goto end;
- }
- if (dec_ctx->codec_type == AVMEDIA_TYPE_VIDEO) {
- buffersrc = avfilter_get_by_name("buffer");
- buffersink = avfilter_get_by_name("buffersink");
- if (!buffersrc || !buffersink) {
- av_log(NULL, AV_LOG_ERROR, "filtering source or sink element not found\n");
- ret = AVERROR_UNKNOWN;
- goto end;
- }
- snprintf(args, sizeof(args),
- "video_size=%dx%d:pix_fmt=%d:time_base=%d/%d:pixel_aspect=%d/%d",
- dec_ctx->width, dec_ctx->height, dec_ctx->pix_fmt,
- dec_ctx->time_base.num, dec_ctx->time_base.den,
- dec_ctx->sample_aspect_ratio.num,
- dec_ctx->sample_aspect_ratio.den);
- ret = avfilter_graph_create_filter(&buffersrc_ctx, buffersrc, "in",
- args, NULL, filter_graph);
- if (ret < 0) {
- av_log(NULL, AV_LOG_ERROR, "Cannot create buffer source\n");
- goto end;
- }
- ret = avfilter_graph_create_filter(&buffersink_ctx, buffersink, "out",
- NULL, NULL, filter_graph);
- if (ret < 0) {
- av_log(NULL, AV_LOG_ERROR, "Cannot create buffer sink\n");
- goto end;
- }
- ret = av_opt_set_bin(buffersink_ctx, "pix_fmts",
- (uint8_t*)&enc_ctx->pix_fmt, sizeof(enc_ctx->pix_fmt),
- AV_OPT_SEARCH_CHILDREN);
- if (ret < 0) {
- av_log(NULL, AV_LOG_ERROR, "Cannot set output pixel format\n");
- goto end;
- }
- }
- else if (dec_ctx->codec_type == AVMEDIA_TYPE_AUDIO) {
- buffersrc = avfilter_get_by_name("abuffer");
- buffersink = avfilter_get_by_name("abuffersink");
- if (!buffersrc || !buffersink) {
- av_log(NULL, AV_LOG_ERROR, "filtering source or sink element not found\n");
- ret = AVERROR_UNKNOWN;
- goto end;
- }
- if (!dec_ctx->channel_layout)
- dec_ctx->channel_layout =
- av_get_default_channel_layout(dec_ctx->channels);
- snprintf(args, sizeof(args),
- "time_base=%d/%d:sample_rate=%d:sample_fmt=%s:channel_layout=0x%"PRIx64,
- dec_ctx->time_base.num, dec_ctx->time_base.den, dec_ctx->sample_rate,
- av_get_sample_fmt_name(dec_ctx->sample_fmt),
- dec_ctx->channel_layout);
- ret = avfilter_graph_create_filter(&buffersrc_ctx, buffersrc, "in",
- args, NULL, filter_graph);
- if (ret < 0) {
- av_log(NULL, AV_LOG_ERROR, "Cannot create audio buffer source\n");
- goto end;
- }
- ret = avfilter_graph_create_filter(&buffersink_ctx, buffersink, "out",
- NULL, NULL, filter_graph);
- if (ret < 0) {
- av_log(NULL, AV_LOG_ERROR, "Cannot create audio buffer sink\n");
- goto end;
- }
- ret = av_opt_set_bin(buffersink_ctx, "sample_fmts",
- (uint8_t*)&enc_ctx->sample_fmt, sizeof(enc_ctx->sample_fmt),
- AV_OPT_SEARCH_CHILDREN);
- if (ret < 0) {
- av_log(NULL, AV_LOG_ERROR, "Cannot set output sample format\n");
- goto end;
- }
- ret = av_opt_set_bin(buffersink_ctx, "channel_layouts",
- (uint8_t*)&enc_ctx->channel_layout,
- sizeof(enc_ctx->channel_layout), AV_OPT_SEARCH_CHILDREN);
- if (ret < 0) {
- av_log(NULL, AV_LOG_ERROR, "Cannot set output channel layout\n");
- goto end;
- }
- ret = av_opt_set_bin(buffersink_ctx, "sample_rates",
- (uint8_t*)&enc_ctx->sample_rate, sizeof(enc_ctx->sample_rate),
- AV_OPT_SEARCH_CHILDREN);
- if (ret < 0) {
- av_log(NULL, AV_LOG_ERROR, "Cannot set output sample rate\n");
- goto end;
- }
- }
- else {
- ret = AVERROR_UNKNOWN;
- goto end;
- }
- /* Endpoints for the filter graph. */
- outputs->name = av_strdup("in");
- outputs->filter_ctx = buffersrc_ctx;
- outputs->pad_idx = 0;
- outputs->next = NULL;
- inputs->name = av_strdup("out");
- inputs->filter_ctx = buffersink_ctx;
- inputs->pad_idx = 0;
- inputs->next = NULL;
- if (!outputs->name || !inputs->name) {
- ret = AVERROR(ENOMEM);
- goto end;
- }
- if ((ret = avfilter_graph_parse_ptr(filter_graph, filter_spec,
- &inputs, &outputs, NULL)) < 0)
- goto end;
- if ((ret = avfilter_graph_config(filter_graph, NULL)) < 0)
- goto end;
- /* Fill FilteringContext */
- fctx->buffersrc_ctx = buffersrc_ctx;
- fctx->buffersink_ctx = buffersink_ctx;
- fctx->filter_graph = filter_graph;
- end:
- avfilter_inout_free(&inputs);
- avfilter_inout_free(&outputs);
- return ret;
- }
- static int init_filters(void)
- {
- const char *filter_spec;
- unsigned int i;
- int ret;
- filter_ctx = (FilteringContext*)av_malloc_array(ifmt_ctx->nb_streams, sizeof(*filter_ctx));
- if (!filter_ctx)
- return AVERROR(ENOMEM);
- for (i = 0; i < ifmt_ctx->nb_streams; i++) {
- filter_ctx[i].buffersrc_ctx = NULL;
- filter_ctx[i].buffersink_ctx = NULL;
- filter_ctx[i].filter_graph = NULL;
- if (!(ifmt_ctx->streams[i]->codec->codec_type == AVMEDIA_TYPE_AUDIO
- || ifmt_ctx->streams[i]->codec->codec_type == AVMEDIA_TYPE_VIDEO))
- continue;
- if (ifmt_ctx->streams[i]->codec->codec_type == AVMEDIA_TYPE_VIDEO)
- filter_spec = "null"; /* passthrough (dummy) filter for video */
- else
- filter_spec = "anull"; /* passthrough (dummy) filter for audio */
- ret = init_filter(&filter_ctx[i], ifmt_ctx->streams[i]->codec,
- ofmt_ctx->streams[i]->codec, filter_spec);
- if (ret)
- return ret;
- }
- return 0;
- }
- static int encode_write_frame(AVFrame *filt_frame, unsigned int stream_index, int *got_frame) {
- int ret;
- int got_frame_local;
- AVPacket enc_pkt;
- int(*enc_func)(AVCodecContext *, AVPacket *, const AVFrame *, int *) =
- (ifmt_ctx->streams[stream_index]->codec->codec_type ==
- AVMEDIA_TYPE_VIDEO) ? avcodec_encode_video2 : avcodec_encode_audio2;
- if (!got_frame)
- got_frame = &got_frame_local;
- av_log(NULL, AV_LOG_INFO, "Encoding frame\n");
- /* encode filtered frame */
- enc_pkt.data = NULL;
- enc_pkt.size = 0;
- av_init_packet(&enc_pkt);
- ret = enc_func(ofmt_ctx->streams[stream_index]->codec, &enc_pkt,
- filt_frame, got_frame);
- av_frame_free(&filt_frame);
- if (ret < 0)
- return ret;
- if (!(*got_frame))
- return 0;
- /* prepare packet for muxing */
- enc_pkt.stream_index = stream_index;
- av_packet_rescale_ts(&enc_pkt,
- ofmt_ctx->streams[stream_index]->codec->time_base,
- ofmt_ctx->streams[stream_index]->time_base);
- av_log(NULL, AV_LOG_DEBUG, "Muxing frame\n");
- /* mux encoded frame */
- ret = av_interleaved_write_frame(ofmt_ctx, &enc_pkt);
- return ret;
- }
- static int filter_encode_write_frame(AVFrame *frame, unsigned int stream_index)
- {
- int ret;
- AVFrame *filt_frame;
- av_log(NULL, AV_LOG_INFO, "Pushing decoded frame to filters\n");
- /* push the decoded frame into the filtergraph */
- ret = av_buffersrc_add_frame_flags(filter_ctx[stream_index].buffersrc_ctx,
- frame, 0);
- if (ret < 0) {
- av_log(NULL, AV_LOG_ERROR, "Error while feeding the filtergraph\n");
- return ret;
- }
- /* pull filtered frames from the filtergraph */
- while (1) {
- filt_frame = av_frame_alloc();
- if (!filt_frame) {
- ret = AVERROR(ENOMEM);
- break;
- }
- av_log(NULL, AV_LOG_INFO, "Pulling filtered frame from filters\n");
- ret = av_buffersink_get_frame(filter_ctx[stream_index].buffersink_ctx,
- filt_frame);
- if (ret < 0) {
- /* if no more frames for output - returns AVERROR(EAGAIN)
- * if flushed and no more frames for output - returns AVERROR_EOF
- * rewrite retcode to 0 to show it as normal procedure completion
- */
- if (ret == AVERROR(EAGAIN) || ret == AVERROR_EOF)
- ret = 0;
- av_frame_free(&filt_frame);
- break;
- }
- filt_frame->pict_type = AV_PICTURE_TYPE_NONE;
- ret = encode_write_frame(filt_frame, stream_index, NULL);
- if (ret < 0)
- break;
- }
- return ret;
- }
- static int flush_encoder(unsigned int stream_index)
- {
- int ret;
- int got_frame;
- if (!(ofmt_ctx->streams[stream_index]->codec->codec->capabilities &
- CODEC_CAP_DELAY))
- return 0;
- while (1) {
- av_log(NULL, AV_LOG_INFO, "Flushing stream #%u encoder\n", stream_index);
- ret = encode_write_frame(NULL, stream_index, &got_frame);
- if (ret < 0)
- break;
- if (!got_frame)
- return 0;
- }
- return ret;
- }
- int main(int argc, char **argv)
- {
- int ret;
- AVPacket packet; //= { .data = NULL, .size = 0 };
- packet.data = NULL;
- packet.size = 0;
- AVFrame *frame = NULL;
- enum AVMediaType type;
- unsigned int stream_index;
- unsigned int i;
- int got_frame;
- int(*dec_func)(AVCodecContext *, AVFrame *, int *, const AVPacket *);
- av_register_all();
- avfilter_register_all();
- if ((ret = open_input_file("test.ts")) < 0)
- goto end;
- if ((ret = open_output_file("test.avi")) < 0)
- goto end;
- if ((ret = init_filters()) < 0)
- goto end;
- /* read all packets */
- while (1) {
- if ((ret = av_read_frame(ifmt_ctx, &packet)) < 0)
- break;
- stream_index = packet.stream_index;
- type = ifmt_ctx->streams[packet.stream_index]->codec->codec_type;
- av_log(NULL, AV_LOG_DEBUG, "Demuxer gave frame of stream_index %u\n",
- stream_index);
- if (filter_ctx[stream_index].filter_graph) {
- av_log(NULL, AV_LOG_DEBUG, "Going to reencode&filter the frame\n");
- frame = av_frame_alloc();
- if (!frame) {
- ret = AVERROR(ENOMEM);
- break;
- }
- av_packet_rescale_ts(&packet,
- ifmt_ctx->streams[stream_index]->time_base,
- ifmt_ctx->streams[stream_index]->codec->time_base);
- dec_func = (type == AVMEDIA_TYPE_VIDEO) ? avcodec_decode_video2 :
- avcodec_decode_audio4;
- ret = dec_func(ifmt_ctx->streams[stream_index]->codec, frame,
- &got_frame, &packet);
- if (ret < 0) {
- av_frame_free(&frame);
- av_log(NULL, AV_LOG_ERROR, "Decoding failed\n");
- break;
- }
- if (got_frame) {
- frame->pts = av_frame_get_best_effort_timestamp(frame);
- ret = filter_encode_write_frame(frame, stream_index);
- av_frame_free(&frame);
- if (ret < 0)
- goto end;
- }
- else {
- av_frame_free(&frame);
- }
- }
- else {
- /* remux this frame without reencoding */
- av_packet_rescale_ts(&packet,
- ifmt_ctx->streams[stream_index]->time_base,
- ofmt_ctx->streams[stream_index]->time_base);
- ret = av_interleaved_write_frame(ofmt_ctx, &packet);
- if (ret < 0)
- goto end;
- }
- av_free_packet(&packet);
- }
- /* flush filters and encoders */
- for (i = 0; i < ifmt_ctx->nb_streams; i++) {
- /* flush filter */
- if (!filter_ctx[i].filter_graph)
- continue;
- ret = filter_encode_write_frame(NULL, i);
- if (ret < 0) {
- av_log(NULL, AV_LOG_ERROR, "Flushing filter failed\n");
- goto end;
- }
- /* flush encoder */
- ret = flush_encoder(i);
- if (ret < 0) {
- av_log(NULL, AV_LOG_ERROR, "Flushing encoder failed\n");
- goto end;
- }
- }
- av_write_trailer(ofmt_ctx);
- end:
- av_free_packet(&packet);
- av_frame_free(&frame);
- for (i = 0; i < ifmt_ctx->nb_streams; i++) {
- avcodec_close(ifmt_ctx->streams[i]->codec);
- if (ofmt_ctx && ofmt_ctx->nb_streams > i && ofmt_ctx->streams[i] && ofmt_ctx->streams[i]->codec)
- avcodec_close(ofmt_ctx->streams[i]->codec);
- if (filter_ctx && filter_ctx[i].filter_graph)
- avfilter_graph_free(&filter_ctx[i].filter_graph);
- }
- av_free(filter_ctx);
- avformat_close_input(&ifmt_ctx);
- if (ofmt_ctx && !(ofmt_ctx->oformat->flags & AVFMT_NOFILE))
- avio_closep(&ofmt_ctx->pb);
- avformat_free_context(ofmt_ctx);
- if (ret < 0)
- av_log(NULL, AV_LOG_ERROR, "Error occurred: %s\n"); //av_err2str(ret));
- return ret ? 1 : 0;
- }
- </span>
源代码下载:
csdn工程:http://download.csdn.NET/detail/rootusers/8425619
from:http://blog.csdn.net/rootusers/article/details/43488827
以上是关于ffmpeg转码常用容器编码格式与解编码器的搭配总结,-c与-f参数的运用的主要内容,如果未能解决你的问题,请参考以下文章