原文出自:http://3xin2yi.info/wwwroot/tech/doku.php/tech:multimedia:ffmpeg
转自:FFMPEG详解 [C.L's Tech Notes]
根据多处转载的记录,原文写于2012年12月,原文出处已经访问不到,这里做一个完整版的备份,供参考。
FFMPEG堪称自由软件中最完备的一套多媒体支持库,它几乎实现了所有当下常见的数据封装格式、多媒体传输协议以及音视频编解码器,堪称多媒体业界的瑞士军刀。因此,对于从事多媒体技术开发的工程师来说,深入研究FFMPEG成为一门必不可少的工作,可以这样说,FFMPEG之于多媒体开发工程师的重要性正如kernel之于嵌入式系统工程师一般。
几个小知识:
关于耻辱厅(Hall of Shame):FFMPEG大部分代码遵循LGPL许可证,如果使用者对FFMpeg进行了修改,要求公布修改的源代码;有少部分代码遵循GPL许可证,要求使用者同时公开使用FFMpeg的软件的源代码。实际上,除去部分具备系统软件开发能力的大型公司(Microsoft、Apple等)以及某些著名的音视频技术提供商(Divx、Real等)提供的自有播放器之外,绝大部分第三方开发的播放器都离不开FFMpeg的支持,像Linux桌面环境中的开源播放器VLC、MPlayer,Windows下的KMPlayer、暴风影音以及Android下几乎全部第三方播放器都是基于FFMPEG的。也有许多看似具备自主技术的播放器,其实也都不声不响地使用了FFMPEG,这种行为被称为“盗窃”,参与“盗窃”的公司则被请入耻辱厅,如于2009年上榜的国产播放器暴风影音、QQ影音。
关于FFMPEG的商业应用:与其他开源软件不同的是,FFMPEG所触及的多媒体编解码算法中有相当一部分处于大量的专利涵盖范围之内,因此,在商业软件中使用FFMPEG必须考虑可能造成的对专利所有者的权利侵犯,这一点在FFMPEG的官方网站也有所提及,所涉及的风险需使用者自行评估应对。
FFMPEG从功能上划分为几个模块,分别为核心工具(libutils)、媒体格式(libavformat)、编解码(libavcodec)、设备(libavdevice)和后处理(libavfilter, libswscale, libpostproc),分别负责提供公用的功能函数、实现多媒体文件的读包和写包、完成音视频的编解码、管理音视频设备的操作以及进行音视频后处理。
这里指FFMPEG提供的命令行(CLI)工具ffmpeg,其使用方法如下(方括号表示可选项,花括号表示必选项目):
- ffmpeg [global options] {[infile options]['-i' 'infile'] ...} {[outfile options] 'outfile' ...}
参数选项由三部分组成:可选的一组全局参数、一组或多组输入文件参数、一组或多组输出文件参数,其中,每组输入文件参数以‘-i’为结束标记;每组输出文件参数以输出文件名为结束标记。
能力集列表
常用输入选项
对于输入,以下选项通常是自动识别的,但也可以强制设定。
常用输出选项
FFMPEG的某些选项可以对一个特定的媒体流起作用,这种情况下需要在选项后面增加一个流标识。流标识允许以下几种格式:
假如要设定第二个音频流为copy,则需要指定-codec:a:1 copy
1、将一个老式的avi文件转成mp4
- ffmpeg -i final.avi -acodec copy -vcodec copy final.mp4
2、从一个视频文件中抽取一帧图像:
- ffmpeg -y -i test.mp4 -ss 00:03:22.000 -vframes 1 -an test.jpg
3、
- ffmpeg -i final.avi -vf scale=640:640 square.avi
4、使用alsa接口录制一段音频存放到某个wav文件中
- ffmpeg -f alsa -i hw:0 -t 100 alsaout.wav
5、使用alsa接口搭建一个个人网络电台
- ffmpeg -f alsa -i default -acodec aac -strict -2 -b:a 128k -r 44100 /var/www/data/main.m3u8
6、将一个mp4文件的音视频流实时转码之后发送给某个远程设备,远程设备可以通过http获取的sdp文件来接收rtp媒体数据。
- ffmpeg -re -i example.mp4 -acodec copy -vcodec libx264 -s 480x270 -map 0:0 -f rtp rtp://10.131.202.62:1234 -map 0:1 -f rtp rtp://10.131.202.62:1238 > /var/www/live.sdp
FFMpeg与大部分GNU软件的编译方式类似,是通过configure脚本来实现编译前定制的。这种途径允许用户在编译前对软件进行裁剪,同时通过对宿主系统和目标系统的自动检测来筛选参与编译的模块并为其设定合适的配置。但是,FFMpeg的configure脚本并非如通常的GNU软件一样通过配置工具自动生成,而是完全由人工编写的。configure脚本生成的config.mak和config.h分别在Makefile和源代码的层次上实现编译的控制。
通过运行“./configure –help”可以了解到脚本支持的参数,这些参数大体分为下面几类:
对于–disable、–enable类的控制选项,如果以–disable为前缀,则缺省是enable的,反之亦然。
总之,无论从商业角度还是技术角度出发,使用configure脚本对FFMpeg进行裁剪是最安全的方式,只有针对于某些configure无法满足的定制要求,才需要考虑修改configure脚本——甚至修改configure生成的配置文件。
以下是一个配置实例,实现运行与Android系统中的ffmpeg库的编译:
- ./configure --prefix=. --cross-prefix=$NDK_TOOLCHAIN_PREFIX --enable-cross-compile --arch=arm --target-os=linux --cpu=cortex-a8 \
- --disable-static --enable-shared --enable-pic --disable-ffmpeg --disable-ffplay --disable-ffserver --disable-ffprobe \
- --extra-cflags="-I$NDK_PLATFORM/usr/include" \
- --extra-ldflags="-nostdlib -Wl,-T,$NDK_PREBUILT/arm-linux-androideabi/lib/ldscripts/armelf_linux_eabi.x \
- -L$NDK_PLATFORM/usr/lib \
- $NDK_PREBUILT/lib/gcc/arm-linux-androideabi/4.4.3/crtbegin.o $NDK_PREBUILT/lib/gcc/arm-linux-androideabi/4.4.3/crtend.o \
- -lc -lm -ldl"
缺省的编译会生成4个可执行文件和9个库:可执行文件包括用于转码的ffmpeg、用于获取媒体信息的ffprobe、用于播放媒体的ffplay和用于推送媒体流的ffserver;库包括avutil、avformat、avcodec、avfilter、avdevice、swresample、swscale、postproc及avresample,其中,核心库有5个,分别为基础库avutil、文件格式及协议库avformat、编解码库avcodec、输入输出设备库avdevice和过滤器avfilter。
解码
- #include <stdio.h>
- #include <string.h>
- #include <stdlib.h>
-
- #include <sys/time.h>
-
- #include "libavutil/avstring.h"
- #include "libavformat/avformat.h"
- #include "libavdevice/avdevice.h"
- #include "libavutil/opt.h"
- #include "libswscale/swscale.h"
-
- #define DECODED_AUDIO_BUFFER_SIZE 192000
-
- struct options
- {
- int streamId;
- int frames;
- int nodec;
- int bplay;
- int thread_count;
- int64_t lstart;
- char finput[256];
- char foutput1[256];
- char foutput2[256];
- };
-
- int parse_options(struct options *opts, int argc, char** argv)
- {
- int optidx;
- char *optstr;
-
- if (argc < 2) return -1;
-
- opts->streamId = -1;
- opts->lstart = -1;
- opts->frames = -1;
- opts->foutput1[0] = 0;
- opts->foutput2[0] = 0;
- opts->nodec = 0;
- opts->bplay = 0;
- opts->thread_count = 0;
- strcpy(opts->finput, argv[1]);
-
- optidx = 2;
- while (optidx < argc)
- {
- optstr = argv[optidx++];
- if (*optstr++ != '-') return -1;
- switch (*optstr++)
- {
- case 's': //< stream id
- opts->streamId = atoi(optstr);
- break;
- case 'f': //< frames
- opts->frames = atoi(optstr);
- break;
- case 'k': //< skipped
- opts->lstart = atoll(optstr);
- break;
- case 'o': //< output
- strcpy(opts->foutput1, optstr);
- strcat(opts->foutput1, ".mpg");
- strcpy(opts->foutput2, optstr);
- strcat(opts->foutput2, ".raw");
- break;
- case 'n': //decoding and output options
- if (strcmp("dec", optstr) == 0)
- opts->nodec = 1;
- break;
- case 'p':
- opts->bplay = 1;
- break;
- case 't':
- opts->thread_count = atoi(optstr);
- break;
- default:
- return -1;
- }
- }
-
- return 0;
- }
-
- void show_help(char* program)
- {
- printf("Simple FFMPEG test program\n");
- printf("Usage: %s inputfile [-sstreamid [-fframes] [-kskipped] [-ooutput_filename(without extension)] [-ndec] [-p] [-tthread_count]]\n",
- program);
- return;
- }
-
- static void log_callback(void* ptr, int level, const char* fmt, va_list vl)
- {
- vfprintf(stdout, fmt, vl);
- }
-
- /*
- * audio renderer code (oss)
- */
- #include <sys/ioctl.h>
- #include <unistd.h>
- #include <fcntl.h>
- #include <sys/soundcard.h>
-
- #define OSS_DEVICE "/dev/dsp0"
-
- struct audio_dsp
- {
- int audio_fd;
- int channels;
- int format;
- int speed;
- };
- int map_formats(enum AVSampleFormat format)
- {
- switch(format)
- {
- case AV_SAMPLE_FMT_U8:
- return AFMT_U8;
- case AV_SAMPLE_FMT_S16:
- return AFMT_S16_LE;
- default:
- return AFMT_U8;
- }
- }
- int set_audio(struct audio_dsp* dsp)
- {
- if (dsp->audio_fd == -1)
- {
- printf("Invalid audio dsp id!\n");
- return -1;
- }
-
- if (-1 == ioctl(dsp->audio_fd, SNDCTL_DSP_SETFMT, &dsp->format))
- {
- printf("Failed to set dsp format!\n");
- return -1;
- }
-
- if (-1 == ioctl(dsp->audio_fd, SNDCTL_DSP_CHANNELS, &dsp->channels))
- {
- printf("Failed to set dsp format!\n");
- return -1;
- }
-
- if (-1 == ioctl(dsp->audio_fd, SNDCTL_DSP_SPEED, &dsp->speed))
- {
- printf("Failed to set dsp format!\n");
- return -1;
- }
- return 0;
- }
-
- int play_pcm(struct audio_dsp* dsp, unsigned char *buf, int size)
- {
- if (dsp->audio_fd == -1)
- {
- printf("Invalid audio dsp id!\n");
- return -1;
- }
-
- if (-1 == write(dsp->audio_fd, buf, size))
- {
- printf("Failed to write audio dsp!\n");
- return -1;
- }
-
- return 0;
- }
- /* audio renderer code end */
-
- /* video renderer code*/
- #include <linux/fb.h>
- #include <sys/mman.h>
-
- #define FB_DEVICE "/dev/fb0"
-
- enum pic_format
- {
- eYUV_420_Planer,
- };
- struct video_fb
- {
- int video_fd;
- struct fb_var_screeninfo vinfo;
- struct fb_fix_screeninfo finfo;
- unsigned char *fbp;
- AVFrame *frameRGB;
- struct
- {
- int x;
- int y;
- } video_pos;
- };
-
- int open_video(struct video_fb *fb, int x, int y)
- {
- int screensize;
- fb->video_fd = open(FB_DEVICE, O_WRONLY);
- if (fb->video_fd == -1) return -1;
-
- if (ioctl(fb->video_fd, FBIOGET_FSCREENINFO, &fb->finfo)) return -2;
- if (ioctl(fb->video_fd, FBIOGET_VSCREENINFO, &fb->vinfo)) return -2;
-
-
- printf("video device: resolution %dx%d, %dbpp\n", fb->vinfo.xres, fb->vinfo.yres, fb->vinfo.bits_per_pixel);
- screensize = fb->vinfo.xres * fb->vinfo.yres * fb->vinfo.bits_per_pixel / 8;
- fb->fbp = (unsigned char *) mmap(0, screensize, PROT_READ|PROT_WRITE, MAP_SHARED, fb->video_fd, 0);
- if (fb->fbp == -1) return -3;
-
- if (x >= fb->vinfo.xres || y >= fb->vinfo.yres)
- {
- return -4;
- }
- else
- {
- fb->video_pos.x = x;
- fb->video_pos.y = y;
- }
-
- fb->frameRGB = avcodec_alloc_frame();
- if (!fb->frameRGB) return -5;
-
- return 0;
- }
-
- /* only 420P supported now */
- int show_picture(struct video_fb *fb, AVFrame *frame, int width, int height, enum pic_format format)
- {
- struct SwsContext *sws;
- int i;
- unsigned char *dest;
- unsigned char *src;
-
- if (fb->video_fd == -1) return -1;
- if ((fb->video_pos.x >= fb->vinfo.xres) || (fb->video_pos.y >= fb->vinfo.yres)) return -2;
-
- if (fb->video_pos.x + width > fb->vinfo.xres)
- {
- width = fb->vinfo.xres - fb->video_pos.x;
- }
- if (fb->video_pos.y + height > fb->vinfo.yres)
- {
- height = fb->vinfo.yres - fb->video_pos.y;
- }
-
- if (format == PIX_FMT_YUV420P)
- {
- sws = sws_getContext(width, height, format, width, height, PIX_FMT_RGB32, SWS_FAST_BILINEAR, NULL, NULL, NULL);
- if (sws == 0)
- {
- return -3;
- }
- if (sws_scale(sws, frame->data, frame->linesize, 0, height, fb->frameRGB->data, fb->frameRGB->linesize))
- {
- return -3;
- }
-
- dest = fb->fbp + (fb->video_pos.x+fb->vinfo.xoffset) * (fb->vinfo.bits_per_pixel/8) +(fb->video_pos.y+fb->vinfo.yoffset) * fb->finfo.line_length;
- for (i = 0; i < height; i++)
- {
- memcpy(dest, src, width*4);
- src += fb->frameRGB->linesize[0];
- dest += fb->finfo.line_length;
- }
- }
- return 0;
- }
-
- void close_video(struct video_fb *fb)
- {
- if (fb->video_fd != -1)
- {
- munmap(fb->fbp, fb->vinfo.xres * fb->vinfo.yres * fb->vinfo.bits_per_pixel / 8);
- close(fb->video_fd);
- fb->video_fd = -1;
- }
- }
- /* video renderer code end */
-
- int main(int argc, char **argv)
- {
- AVFormatContext* pCtx = 0;
- AVCodecContext *pCodecCtx = 0;
- AVCodec *pCodec = 0;
- AVPacket packet;
- AVFrame *pFrame = 0;
- FILE *fpo1 = NULL;
- FILE *fpo2 = NULL;
- int nframe;
- int err;
- int got_picture;
- int picwidth, picheight, linesize;
- unsigned char *pBuf;
- int i;
- int64_t timestamp;
- struct options opt;
- int usefo = 0;
- struct audio_dsp dsp;
- struct video_fb fb;
- int dusecs;
- float usecs1 = 0;
- float usecs2 = 0;
- struct timeval elapsed1, elapsed2;
- int decoded = 0;
-
- av_register_all();
-
- av_log_set_callback(log_callback);
- av_log_set_level(50);
-
- if (parse_options(&opt, argc, argv) < 0 || (strlen(opt.finput) == 0))
- {
- show_help(argv[0]);
- return 0;
- }
-
-
- err = avformat_open_input(&pCtx, opt.finput, 0, 0);
- if (err < 0)
- {
- printf("\n->(avformat_open_input)\tERROR:\t%d\n", err);
- goto fail;
- }
- err = avformat_find_stream_info(pCtx, 0);
- if (err < 0)
- {
- printf("\n->(avformat_find_stream_info)\tERROR:\t%d\n", err);
- goto fail;
- }
- if (opt.streamId < 0)
- {
- av_dump_format(pCtx, 0, pCtx->filename, 0);
- goto fail;
- }
- else
- {
- printf("\n extra data in Stream %d (%dB):", opt.streamId, pCtx->streams[opt.streamId]->codec->extradata_size);
- for (i = 0; i < pCtx->streams[opt.streamId]->codec->extradata_size; i++)
- {
- if (i%16 == 0) printf("\n");
- printf("%2x ", pCtx->streams[opt.streamId]->codec->extradata[i]);
- }
- }
- /* try opening output files */
- if (strlen(opt.foutput1) && strlen(opt.foutput2))
- {
- fpo1 = fopen(opt.foutput1, "wb");
- fpo2 = fopen(opt.foutput2, "wb");
- if (!fpo1 || !fpo2)
- {
- printf("\n->error opening output files\n");
- goto fail;
- }
- usefo = 1;
- }
- else
- {
- usefo = 0;
- }
-
- if (opt.streamId >= pCtx->nb_streams)
- {
- printf("\n->StreamId\tERROR\n");
- goto fail;
- }
-
- if (opt.lstart > 0)
- {
- err = av_seek_frame(pCtx, opt.streamId, opt.lstart, AVSEEK_FLAG_ANY);
- if (err < 0)
- {
- printf("\n->(av_seek_frame)\tERROR:\t%d\n", err);
- goto fail;
- }
- }
-
- /* for decoder configuration */
- if (!opt.nodec)
- {
- /* prepare codec */
- pCodecCtx = pCtx->streams[opt.streamId]->codec;
-
- if (opt.thread_count <= 16 && opt.thread_count > 0 )
- {
- pCodecCtx->thread_count = opt.thread_count;
- pCodecCtx->thread_type = FF_THREAD_FRAME;
- }
- pCodec = avcodec_find_decoder(pCodecCtx->codec_id);
- if (!pCodec)
- {
- printf("\n->can not find codec!\n");
- goto fail;
- }
- err = avcodec_open2(pCodecCtx, pCodec, 0);
- if (err < 0)
- {
- printf("\n->(avcodec_open)\tERROR:\t%d\n", err);
- goto fail;
- }
- pFrame = avcodec_alloc_frame();
-
- /* prepare device */
- if (opt.bplay)
- {
- /* audio devices */
- dsp.audio_fd = open(OSS_DEVICE, O_WRONLY);
- if (dsp.audio_fd == -1)
- {
- printf("\n-> can not open audio device\n");
- goto fail;
- }
- dsp.channels = pCodecCtx->channels;
- dsp.speed = pCodecCtx->sample_rate;
- dsp.format = map_formats(pCodecCtx->sample_fmt);
- if (set_audio(&dsp) < 0)
- {
- printf("\n-> can not set audio device\n");
- goto fail;
- }
- /* video devices */
- if (open_video(&fb, 0, 0) != 0)
- {
- printf("\n-> can not open video device\n");
- goto fail;
- }
- }
- }
-
- nframe = 0;
- while(nframe < opt.frames || opt.frames == -1)
- {
- gettimeofday(&elapsed1, NULL);
- err = av_read_frame(pCtx, &packet);
- if (err < 0)
- {
- printf("\n->(av_read_frame)\tERROR:\t%d\n", err);
- break;
- }
- gettimeofday(&elapsed2, NULL);
- dusecs = (elapsed2.tv_sec - elapsed1.tv_sec)*1000000 + (elapsed2.tv_usec - elapsed1.tv_usec);
- usecs2 += dusecs;
- timestamp = av_rescale_q(packet.dts, pCtx->streams[packet.stream_index]->time_base, (AVRational){1, AV_TIME_BASE});
- printf("\nFrame No %5d stream#%d\tsize %6dB, timestamp:%6lld, dts:%6lld, pts:%6lld, ", nframe++, packet.stream_index, packet.size,
- timestamp, packet.dts, packet.pts);
-
- if (packet.stream_index == opt.streamId)
- {
- #if 0
- for (i = 0; i < 16; /*packet.size;*/ i++)
- {
- if (i%16 == 0) printf("\n pktdata: ");
- printf("%2x ", packet.data[i]);
- }
- printf("\n");
- #endif
- if (usefo)
- {
- fwrite(packet.data, packet.size, 1, fpo1);
- fflush(fpo1);
- }
-
- if (pCtx->streams[opt.streamId]->codec->codec_type == AVMEDIA_TYPE_VIDEO && !opt.nodec)
- {
- picheight = pCtx->streams[opt.streamId]->codec->height;
- picwidth = pCtx->streams[opt.streamId]->codec->width;
-
- gettimeofday(&elapsed1, NULL);
- avcodec_decode_video2(pCodecCtx, pFrame, &got_picture, &packet);
- decoded++;
- gettimeofday(&elapsed2, NULL);
- dusecs = (elapsed2.tv_sec - elapsed1.tv_sec)*1000000 + (elapsed2.tv_usec - elapsed1.tv_usec);
- usecs1 += dusecs;
-
- if (got_picture)
- {
- printf("[Video: type %d, ref %d, pts %lld, pkt_pts %lld, pkt_dts %lld]",
- pFrame->pict_type, pFrame->reference, pFrame->pts, pFrame->pkt_pts, pFrame->pkt_dts);
-
- if (pCtx->streams[opt.streamId]->codec->pix_fmt == PIX_FMT_YUV420P)
- {
- if (usefo)
- {
- linesize = pFrame->linesize[0];
- pBuf = pFrame->data[0];
- for (i = 0; i < picheight; i++)
- {
- fwrite(pBuf, picwidth, 1, fpo2);
- pBuf += linesize;
- }
-
- linesize = pFrame->linesize[1];
- pBuf = pFrame->data[1];
- for (i = 0; i < picheight/2; i++)
- {
- fwrite(pBuf, picwidth/2, 1, fpo2);
- pBuf += linesize;
- }
-
- linesize = pFrame->linesize[2];
- pBuf = pFrame->data[2];
- for (i = 0; i < picheight/2; i++)
- {
- fwrite(pBuf, picwidth/2, 1, fpo2);
- pBuf += linesize;
- }
- fflush(fpo2);
- }
-
- if (opt.bplay)
- {
- /* show picture */
- show_picture(&fb, pFrame, picheight, picwidth, PIX_FMT_YUV420P);
- }
- }
- }
- av_free_packet(&packet);
- }
- else if (pCtx->streams[opt.streamId]->codec->codec_type == AVMEDIA_TYPE_AUDIO && !opt.nodec)
- {
- int got;
-
- gettimeofday(&elapsed1, NULL);
- avcodec_decode_audio4(pCodecCtx, pFrame, &got, &packet);
- decoded++;
- gettimeofday(&elapsed2, NULL);
- dusecs = (elapsed2.tv_sec - elapsed1.tv_sec)*1000000 + (elapsed2.tv_usec - elapsed1.tv_usec);
- usecs1 += dusecs;
-
- if (got)
- {
- printf("[Audio: %5dB raw data, decoding time: %d]", pFrame->linesize[0], dusecs);
- if (usefo)
- {
- fwrite(pFrame->data[0], pFrame->linesize[0], 1, fpo2);
- fflush(fpo2);
- }
- if (opt.bplay)
- {
- play_pcm(&dsp, pFrame->data[0], pFrame->linesize[0]);
- }
- }
- }
- }
- }
-
- if (!opt.nodec && pCodecCtx)
- {
- avcodec_close(pCodecCtx);
- }
-
- printf("\n%d frames parsed, average %.2f us per frame\n", nframe, usecs2/nframe);
- printf("%d frames decoded, average %.2f us per frame\n", decoded, usecs1/decoded);
-
- fail:
- if (pCtx)
- {
- avformat_close_input(&pCtx);
- }
- if (fpo1)
- {
- fclose(fpo1);
- }
- if (fpo2)
- {
- fclose(fpo2);
- }
- if (!pFrame)
- {
- av_free(pFrame);
- }
- if (!usefo && (dsp.audio_fd != -1))
- {
- close(dsp.audio_fd);
- }
- if (!usefo && (fb.video_fd != -1))
- {
- close_video(&fb);
- }
- return 0;
- }
这一小段代码可以实现的功能包括:
这些功能足以支持一个功能强大的多媒体播放器,因为最复杂的解复用、解码、数据分析过程已经在FFMpeg内部实现了,需要关注的仅剩同步问题。
数据结构
基本概念
编解码器、数据帧、媒体流和容器是数字媒体处理系统的四个基本概念。
首先需要统一术语:
在FFMPEG中,使用AVFormatContext、AVStream、AVCodecContext、AVCodec及AVPacket等结构来抽象这些基本要素,它们的关系如下图所示:
AVCodecContext
这是一个描述编解码器上下文的数据结构,包含了众多编解码器需要的参数信息,如下列出了部分比较重要的域:
- typedef struct AVCodecContext {
-
- ......
-
- /**
- * some codecs need / can use extradata like Huffman tables.
- * mjpeg: Huffman tables
- * rv10: additional flags
- * mpeg4: global headers (they can be in the bitstream or here)
- * The allocated memory should be FF_INPUT_BUFFER_PADDING_SIZE bytes larger
- * than extradata_size to avoid prolems if it is read with the bitstream reader.
- * The bytewise contents of extradata must not depend on the architecture or CPU endianness.
- * - encoding: Set/allocated/freed by libavcodec.
- * - decoding: Set/allocated/freed by user.
- */
- uint8_t *extradata;
- int extradata_size;
- /**
- * This is the fundamental unit of time (in seconds) in terms
- * of which frame timestamps are represented. For fixed-fps content,
- * timebase should be 1/framerate and timestamp increments should be
- * identically 1.
- * - encoding: MUST be set by user.
- * - decoding: Set by libavcodec.
- */
- AVRational time_base;
-
- /* video only */
- /**
- * picture width / height.
- * - encoding: MUST be set by user.
- * - decoding: Set by libavcodec.
- * Note: For compatibility it is possible to set this instead of
- * coded_width/height before decoding.
- */
- int width, height;
-
- ......
-
- /* audio only */
- int sample_rate; ///< samples per second
- int channels; ///< number of audio channels
-
- /**
- * audio sample format
- * - encoding: Set by user.
- * - decoding: Set by libavcodec.
- */
- enum SampleFormat sample_fmt; ///< sample format
-
- /* The following data should not be initialized. */
- /**
- * Samples per packet, initialized when calling 'init'.
- */
- int frame_size;
- int frame_number; ///< audio or video frame number
-
- ......
-
- char codec_name[32];
- enum AVMediaType codec_type; /* see AVMEDIA_TYPE_xxx */
- enum CodecID codec_id; /* see CODEC_ID_xxx */
-
- /**
- * fourcc (LSB first, so "ABCD" -> ('D'<<24) + ('C'<<16) + ('B'<<8) + 'A').
- * This is used to work around some encoder bugs.
- * A demuxer should set this to what is stored in the field used to identify the codec.
- * If there are multiple such fields in a container then the demuxer should choose the one
- * which maximizes the information about the used codec.
- * If the codec tag field in a container is larger then 32 bits then the demuxer should
- * remap the longer ID to 32 bits with a table or other structure. Alternatively a new
- * extra_codec_tag + size could be added but for this a clear advantage must be demonstrated
- * first.
- * - encoding: Set by user, if not then the default based on codec_id will be used.
- * - decoding: Set by user, will be converted to uppercase by libavcodec during init.
- */
- unsigned int codec_tag;
-
- ......
-
- /**
- * Size of the frame reordering buffer in the decoder.
- * For MPEG-2 it is 1 IPB or 0 low delay IP.
- * - encoding: Set by libavcodec.
- * - decoding: Set by libavcodec.
- */
- int has_b_frames;
-
- /**
- * number of bytes per packet if constant and known or 0
- * Used by some WAV based audio codecs.
- */
- int block_align;
-
- ......
-
- /**
- * bits per sample/pixel from the demuxer (needed for huffyuv).
- * - encoding: Set by libavcodec.
- * - decoding: Set by user.
- */
- int bits_per_coded_sample;
-
- ......
-
- } AVCodecContext;
如果是单纯使用libavcodec,这部分信息需要调用者进行初始化;如果是使用整个FFMPEG库,这部分信息在调用avformat_open_input和avformat_find_stream_info的过程中根据文件的头信息及媒体流内的头部信息完成初始化。其中几个主要域的释义如下:
AVStream
该结构体描述一个媒体流,定义如下:
- typedef struct AVStream {
- int index; /**< stream index in AVFormatContext */
- int id; /**< format-specific stream ID */
- AVCodecContext *codec; /**< codec context */
- /**
- * Real base framerate of the stream.
- * This is the lowest framerate with which all timestamps can be
- * represented accurately (it is the least common multiple of all
- * framerates in the stream). Note, this value is just a guess!
- * For example, if the time base is 1/90000 and all frames have either
- * approximately 3600 or 1800 timer ticks, then r_frame_rate will be 50/1.
- */
- AVRational r_frame_rate;
-
- ......
-
- /**
- * This is the fundamental unit of time (in seconds) in terms
- * of which frame timestamps are represented. For fixed-fps content,
- * time base should be 1/framerate and timestamp increments should be 1.
- */
- AVRational time_base;
-
- ......
-
- /**
- * Decoding: pts of the first frame of the stream, in stream time base.
- * Only set this if you are absolutely 100% sure that the value you set
- * it to really is the pts of the first frame.
- * This may be undefined (AV_NOPTS_VALUE).
- * @note The ASF header does NOT contain a correct start_time the ASF
- * demuxer must NOT set this.
- */
- int64_t start_time;
- /**
- * Decoding: duration of the stream, in stream time base.
- * If a source file does not specify a duration, but does specify
- * a bitrate, this value will be estimated from bitrate and file size.
- */
- int64_t duration;
-
- #if LIBAVFORMAT_VERSION_INT < (53<<16)
- char language[4]; /** ISO 639-2/B 3-letter language code (empty string if undefined) */
- #endif
-
- /* av_read_frame() support */
- enum AVStreamParseType need_parsing;
- struct AVCodecParserContext *parser;
-
- ......
-
- /* av_seek_frame() support */
- AVIndexEntry *index_entries; /**< Only used if the format does not
- support seeking natively. */
- int nb_index_entries;
- unsigned int index_entries_allocated_size;
-
- int64_t nb_frames; ///< number of frames in this stream if known or 0
-
- ......
-
- /**
- * Average framerate
- */
- AVRational avg_frame_rate;
- ......
- } AVStream;
主要域的释义如下,其中大部分域的值可以由avformat_open_input根据文件头的信息确定,缺少的信息需要通过调用avformat_find_stream_info读帧及软解码进一步获取:
AVFormatContext
这个结构体描述了一个媒体文件或媒体流的构成和基本信息,定义如下:
- typedef struct AVFormatContext {
- const AVClass *av_class; /**< Set by avformat_alloc_context. */
- /* Can only be iformat or oformat, not both at the same time. */
- struct AVInputFormat *iformat;
- struct AVOutputFormat *oformat;
- void *priv_data;
- ByteIOContext *pb;
- unsigned int nb_streams;
- AVStream *streams[MAX_STREAMS];
- char filename[1024]; /**< input or output filename */
- /* stream info */
- int64_t timestamp;
- #if LIBAVFORMAT_VERSION_INT < (53<<16)
- char title[512];
- char author[512];
- char copyright[512];
- char comment[512];
- char album[512];
- int year; /**< ID3 year, 0 if none */
- int track; /**< track number, 0 if none */
- char genre[32]; /**< ID3 genre */
- #endif
-
- int ctx_flags; /**< Format-specific flags, see AVFMTCTX_xx */
- /* private data for pts handling (do not modify directly). */
- /** This buffer is only needed when packets were already buffered but
- not decoded, for example to get the codec parameters in MPEG
- streams. */
- struct AVPacketList *packet_buffer;
-
- /** Decoding: position of the first frame of the component, in
- AV_TIME_BASE fractional seconds. NEVER set this value directly:
- It is deduced from the AVStream values. */
- int64_t start_time;
- /** Decoding: duration of the stream, in AV_TIME_BASE fractional
- seconds. Only set this value if you know none of the individual stream
- durations and also dont set any of them. This is deduced from the
- AVStream values if not set. */
- int64_t duration;
- /** decoding: total file size, 0 if unknown */
- int64_t file_size;
- /** Decoding: total stream bitrate in bit/s, 0 if not
- available. Never set it directly if the file_size and the
- duration are known as FFmpeg can compute it automatically. */
- int bit_rate;
-
- /* av_read_frame() support */
- AVStream *cur_st;
- #if LIBAVFORMAT_VERSION_INT < (53<<16)
- const uint8_t *cur_ptr_deprecated;
- int cur_len_deprecated;
- AVPacket cur_pkt_deprecated;
- #endif
-
- /* av_seek_frame() support */
- int64_t data_offset; /** offset of the first packet */
- int index_built;
-
- int mux_rate;
- unsigned int packet_size;
- int preload;
- int max_delay;
-
- #define AVFMT_NOOUTPUTLOOP -1
- #define AVFMT_INFINITEOUTPUTLOOP 0
- /** number of times to loop output in formats that support it */
- int loop_output;
-
- int flags;
- #define AVFMT_FLAG_GENPTS 0x0001 ///< Generate missing pts even if it requires parsing future frames.
- #define AVFMT_FLAG_IGNIDX 0x0002 ///< Ignore index.
- #define AVFMT_FLAG_NONBLOCK 0x0004 ///< Do not block when reading packets from input.
- #define AVFMT_FLAG_IGNDTS 0x0008 ///< Ignore DTS on frames that contain both DTS & PTS
- #define AVFMT_FLAG_NOFILLIN 0x0010 ///< Do not infer any values from other values, just return what is stored in the container
- #define AVFMT_FLAG_NOPARSE 0x0020 ///< Do not use AVParsers, you also must set AVFMT_FLAG_NOFILLIN as the fillin code works on frames and no parsing -> no frames. Also seeking to frames can not work if parsing to find frame boundaries has been disabled
- #define AVFMT_FLAG_RTP_HINT 0x0040 ///< Add RTP hinting to the output file
-
- int loop_input;
- /** decoding: size of data to probe; encoding: unused. */
- unsigned int probesize;
-
- /**
- * Maximum time (in AV_TIME_BASE units) during which the input should
- * be analyzed in avformat_find_stream_info().
- */
- int max_analyze_duration;
-
- const uint8_t *key;
- int keylen;
-
- unsigned int nb_programs;
- AVProgram **programs;
-
- /**
- * Forced video codec_id.
- * Demuxing: Set by user.
- */
- enum CodecID video_codec_id;
- /**
- * Forced audio codec_id.
- * Demuxing: Set by user.
- */
- enum CodecID audio_codec_id;
- /**
- * Forced subtitle codec_id.
- * Demuxing: Set by user.
- */
- enum CodecID subtitle_codec_id;
-
- /**
- * Maximum amount of memory in bytes to use for the index of each stream.
- * If the index exceeds this size, entries will be discarded as
- * needed to maintain a smaller size. This can lead to slower or less
- * accurate seeking (depends on demuxer).
- * Demuxers for which a full in-memory index is mandatory will ignore
- * this.
- * muxing : unused
- * demuxing: set by user
- */
- unsigned int max_index_size;
-
- /**
- * Maximum amount of memory in bytes to use for buffering frames
- * obtained from realtime capture devices.
- */
- unsigned int max_picture_buffer;
-
- unsigned int nb_chapters;
- AVChapter **chapters;
-
- /**
- * Flags to enable debugging.
- */
- int debug;
- #define FF_FDEBUG_TS 0x0001
-
- /**
- * Raw packets from the demuxer, prior to parsing and decoding.
- * This buffer is used for buffering packets until the codec can
- * be identified, as parsing cannot be done without knowing the
- * codec.
- */
- struct AVPacketList *raw_packet_buffer;
- struct AVPacketList *raw_packet_buffer_end;
-
- struct AVPacketList *packet_buffer_end;
-
- AVMetadata *metadata;
-
- /**
- * Remaining size available for raw_packet_buffer, in bytes.
- * NOT PART OF PUBLIC API
- */
- #define RAW_PACKET_BUFFER_SIZE 2500000
- int raw_packet_buffer_remaining_size;
-
- /**
- * Start time of the stream in real world time, in microseconds
- * since the unix epoch (00:00 1st January 1970). That is, pts=0
- * in the stream was captured at this real world time.
- * - encoding: Set by user.
- * - decoding: Unused.
- */
- int64_t start_time_realtime;
- } AVFormatContext;
这是FFMpeg中最为基本的一个结构,是其他所有结构的根,是一个多媒体文件或流的根本抽象。其中:
通常,这个结构由avformat_open_input在内部创建并以缺省值初始化部分成员。但是,如果调用者希望自己创建该结构,则需要显式为该结构的一些成员置缺省值——如果没有缺省值的话,会导致之后的动作产生异常。以下成员需要被关注:
AVPacket
AVPacket定义在avcodec.h中,如下:
- typedef struct AVPacket {
- /**
- * Presentation timestamp in AVStream->time_base units; the time at which
- * the decompressed packet will be presented to the user.
- * Can be AV_NOPTS_VALUE if it is not stored in the file.
- * pts MUST be larger or equal to dts as presentation cannot happen before
- * decompression, unless one wants to view hex dumps. Some formats misuse
- * the terms dts and pts/cts to mean something different. Such timestamps
- * must be converted to true pts/dts before they are stored in AVPacket.
- */
- int64_t pts;
- /**
- * Decompression timestamp in AVStream->time_base units; the time at which
- * the packet is decompressed.
- * Can be AV_NOPTS_VALUE if it is not stored in the file.
- */
- int64_t dts;
- uint8_t *data;
- int size;
- int stream_index;
- int flags;
- /**
- * Duration of this packet in AVStream->time_base units, 0 if unknown.
- * Equals next_pts - this_pts in presentation order.
- */
- int duration;
- void (*destruct)(struct AVPacket *);
- void *priv;
- int64_t pos; ///< byte position in stream, -1 if unknown
-
- /**
- * Time difference in AVStream->time_base units from the pts of this
- * packet to the point at which the output from the decoder has converged
- * independent from the availability of previous frames. That is, the
- * frames are virtually identical no matter if decoding started from
- * the very first frame or from this keyframe.
- * Is AV_NOPTS_VALUE if unknown.
- * This field is not the display duration of the current packet.
- *
- * The purpose of this field is to allow seeking in streams that have no
- * keyframes in the conventional sense. It corresponds to the
- * recovery point SEI in H.264 and match_time_delta in NUT. It is also
- * essential for some types of subtitle streams to ensure that all
- * subtitles are correctly displayed after seeking.
- */
- int64_t convergence_duration;
- } AVPacket;
FFMPEG使用AVPacket来暂存媒体数据包及附加信息(解码时间戳、显示时间戳、时长等),这样的媒体数据包所承载的往往不是原始格式的音视频数据,而是以某种方式编码后的数据,编码信息由对应的媒体流结构AVStream给出。AVPacket包含以下数据域:
AVPacket结构本身只是个容器,它使用data成员引用实际的数据缓冲区。这个缓冲区的管理方式有两种,其一是通过调用av_new_packet直接创建缓冲区,其二是引用已经存在的缓冲区。缓冲区的释放通过调用av_free_packet实现,其内部实现也采用了两种不同的释放方式,第一种方式是调用AVPacket的destruct函数,这个destruct函数可能是缺省的av_destruct_packet,对应av_new_packet或av_dup_packet创建的缓冲区,也可能是某个自定义的释放函数,表示缓冲区的提供者希望使用者在结束缓冲区的时候按照提供者期望的方式将其释放,第二种方式是仅仅将data和size的值清0,这种情况下往往是引用了一个已经存在的缓冲区,AVPacket的destruct指针为空。
在使用AVPacket时,对于缓冲区的提供者,必须注意通过设置destruct函数指针指定正确的释放方式,如果缓冲区提供者打算自己释放缓冲区,则切记将destruct置空;而对于缓冲区的使用者,务必在使用结束时调用av_free_packet释放缓冲区(虽然释放操作可能只是一个假动作)。如果某个使用者打算较长时间内占用一个AVPacket——比如不打算在函数返回之前释放它——最好调用av_dup_packet进行缓冲区的克隆,将其转化为自有分配的缓冲区,以免对缓冲区的不当占用造成异常错误。av_dup_packet会为destruct指针为空的AVPacket新建一个缓冲区,然后将原缓冲区的数据拷贝至新缓冲区,置data的值为新缓冲区的地址,同时设destruct指针为av_destruct_packet。
上述媒体结构可以通过FFMPEG提供的av_dump_format方法直观展示出来,以下例子以一个MPEG-TS文件为输入,其展示结果为:
- Input #0, mpegts, from '/Videos/suite/ts/H.264_High_L3.1_720x480_23.976fps_AAC-LC.ts':
- Duration: 00:01:43.29, start: 599.958300, bitrate: 20934 kb/s
- Program 1
- Stream #0.0[0x1011]: Video: h264 (High), yuv420p, 720x480 [PAR 32:27 DAR 16:9], 23.98 fps, 23.98 tbr, 90k tbn, 47.95 tbc
- Stream #0.1[0x1100]: Audio: aac, 48000 Hz, stereo, s16, 159 kb/s
从中可以找到媒体的格式、路径、时长、开始时间、全局比特率。此外,列出了媒体包含的一个节目,由两个媒体流组成,第一个媒体流是视频流,id为0x1011;第二个为音频流,id为0x1100。视频流是h264的High Profile编码,色彩空间为420p,大小为720×480,平均帧率23.98,参考帧率23.98,流时间基准是90000,编码的时间基准为47.95;音频流是aac编码,48kHz的采样,立体声,每个样点16比特,流的比特率是159kbps。
时间信息
时间信息用于实现多媒体同步。
同步的目的在于展示多媒体信息时,能够保持媒体对象之间固有的时间关系。同步有两类,一类是流内同步,其主要任务是保证单个媒体流内的时间关系,以满足感知要求,如按照规定的帧率播放一段视频;另一类是流间同步,主要任务是保证不同媒体流之间的时间关系,如音频和视频之间的关系(lipsync)。
对于固定速率的媒体,如固定帧率的视频或固定比特率的音频,可以将时间信息(帧率或比特率)置于文件首部(header),如AVI的hdrl List、MP4的moov box,还有一种相对复杂的方案是将时间信息嵌入媒体流的内部,如MPEG TS和Real video,这种方案可以处理变速率的媒体,亦可有效避免同步过程中的时间漂移。
FFMPEG会为每一个数据包打上时间标签,以更有效地支持上层应用的同步机制。时间标签有两种,一种是DTS,称为解码时间标签,另一种是PTS,称为显示时间标签。对于声音来说 ,这两个时间标签是相同的,但对于某些视频编码格式,由于采用了双向预测技术,会造成DTS和PTS的不一致。
无双向预测帧的情况:
- 图像类型: I P P P P P P ... I P P
- DTS: 0 1 2 3 4 5 6... 100 101 102
- PTS: 0 1 2 3 4 5 6... 100 101 102
有双向预测帧的情况:
- 图像类型: I P B B P B B ... I P B
- DTS: 0 1 2 3 4 5 6 ... 100 101 102
- PTS: 0 3 1 2 6 4 5 ... 100 104 102
对于存在双向预测帧的情况,通常要求解码器对图像重排序,以保证输出的图像顺序为显示顺序:
- 解码器输入:I P B B P B B
- (DTS) 0 1 2 3 4 5 6
- (PTS) 0 3 1 2 6 4 5
- 解码器输出:X I B B P B B P
- (PTS) X 0 1 2 3 4 5 6
时间信息的获取:
通过调用avformat_find_stream_info,多媒体应用可以从AVFormatContext对象中拿到媒体文件的时间信息:主要是总时间长度和开始时间,此外还有与时间信息相关的比特率和文件大小。其中时间信息的单位是AV_TIME_BASE:微秒。
- typedef struct AVFormatContext {
-
- ......
-
- /** Decoding: position of the first frame of the component, in
- AV_TIME_BASE fractional seconds. NEVER set this value directly:
- It is deduced from the AVStream values. */
- int64_t start_time;
- /** Decoding: duration of the stream, in AV_TIME_BASE fractional
- seconds. Only set this value if you know none of the individual stream
- durations and also dont set any of them. This is deduced from the
- AVStream values if not set. */
- int64_t duration;
- /** decoding: total file size, 0 if unknown */
- int64_t file_size;
- /** Decoding: total stream bitrate in bit/s, 0 if not
- available. Never set it directly if the file_size and the
- duration are known as FFmpeg can compute it automatically. */
- int bit_rate;
-
- ......
-
- } AVFormatContext;
以上4个成员变量都是只读的,基于FFMpeg的中间件需要将其封装到某个接口中,如:
- LONG GetDuratioin(IntfX*);
- LONG GetStartTime(IntfX*);
- LONG GetFileSize(IntfX*);
- LONG GetBitRate(IntfX*);
APIs
FFMpeg的API大部分以0作为成功返回值而一个负数作为错误码。
读系列
读系列API的主要功能是根据某个指定的源获取媒体数据包,这个源可以是一个本地文件、一个RTSP或HTTP源、一个摄像头驱动或者其它。
avformat_open_input
- int avformat_open_input(AVFormatContext **ic_ptr, const char *filename, AVInputFormat *fmt, AVDictionary **options);
avformat_open_input完成两个任务:
一个多媒体文件或多媒体流与其包含的原始流的关系如下:
- 多媒体文件/多媒体流 (movie.mkv)
- 原始流 1 (h.264 video)
- 原始流 2 (aac audio for Chinese)
- 原始流 3 (aac audio for english)
- 原始流 4 (Chinese Subtitle)
- 原始流 5 (English Subtitle)
- ...
关于输入参数:
这个函数通过解析多媒体文件或流的头信息及其他辅助数据,能够获取足够多的关于文件、流和编解码器的信息,但由于任何一种多媒体格式提供的信息都是有限的,而且不同的多媒体内容制作软件对头信息的设置不尽相同,此外这些软件在产生多媒体内容时难免会引入一些错误,因此这个函数并不保证能够获取所有需要的信息,在这种情况下,则需要考虑另一个函数:avformat_find_stream_info。
avformat_find_stream_info
- int avformat_find_stream_info(AVFormatContext *ic, AVDictionary **options);
这个函数主要用于获取必要的编解码器参数,设置到ic→streams[i]→codec中。
首先必须得到各媒体流对应编解码器的类型和id,这是两个定义在avutils.h和avcodec.h中的枚举:
- enum AVMediaType {
- AVMEDIA_TYPE_UNKNOWN = -1,
- AVMEDIA_TYPE_VIDEO,
- AVMEDIA_TYPE_AUDIO,
- AVMEDIA_TYPE_DATA,
- AVMEDIA_TYPE_SUBTITLE,
- AVMEDIA_TYPE_ATTACHMENT,
- AVMEDIA_TYPE_NB
- };
- enum CodecID {
- CODEC_ID_NONE,
-
- /* video codecs */
- CODEC_ID_MPEG1VIDEO,
- CODEC_ID_MPEG2VIDEO, ///< preferred ID for MPEG-1/2 video decoding
- CODEC_ID_MPEG2VIDEO_XVMC,
- CODEC_ID_H261,
- CODEC_ID_H263,
- ...
- };
通常,如果某种媒体格式具备完备而正确的头信息,调用avformat_open_input即可以得到这两个参数,但若是因某种原因avformat_open_input无法获取它们,这一任务将由avformat_find_stream_info完成。
其次还要获取各媒体流对应编解码器的时间基准。
此外,对于音频编解码器,还需要得到:
对于视频编解码器,则是:
av_read_frame
- int av_read_frame(AVFormatContext *s, AVPacket *pkt);
这个函数用于从多媒体文件或多媒体流中读取媒体数据,获取的数据由AVPacket结构pkt来存放。对于音频数据,如果是固定比特率,则pkt中装载着一个或多个音频帧;如果是可变比特率,则pkt中装载有一个音频帧。对于视频数据,pkt中装载有一个视频帧。需要注意的是:再次调用本函数之前,必须使用av_free_packet释放pkt所占用的资源。
通过pkt→stream_index可以查到获取的媒体数据的类型,从而将数据送交相应的解码器进行后续处理。
av_seek_frame
- int av_seek_frame(AVFormatContext *s, int stream_index, int64_t timestamp, int flags);
这个函数通过改变媒体文件的读写指针来实现对媒体文件的随机访问,支持以下三种方式:
关于参数:
av_close_input_file
- void av_close_input_file(AVFormatContext *s);
关闭一个媒体文件:释放资源,关闭物理IO。
编解码系列
编解码API负责实现媒体原始数据的编码和媒体压缩数据的解码。
avcodec_find_decoder
- AVCodec *avcodec_find_decoder(enum CodecID id);
- AVCodec *avcodec_find_decoder_by_name(const char *name);
根据给定的codec id或解码器名称从系统中搜寻并返回一个AVCodec结构的指针。
avcodec_open2
- int avcodec_open2(AVCodecContext *avctx, const AVCodec *codec, AVDictionary **options);
此函数根据输入的AVCodec指针具体化AVCodecContext结构。在调用该函数之前,需要首先调用avcodec_alloc_context分配一个AVCodecContext结构,或调用avformat_open_input获取媒体文件中对应媒体流的AVCodecContext结构;此外还需要通过avcodec_find_decoder获取AVCodec结构。
这一函数还将初始化对应的解码器。
avcodec_decode_video2
- int avcodec_decode_video2(AVCodecContext *avctx, AVFrame *picture, int *got_picture_ptr, AVPacket *avpkt);
解码一个视频帧。got_picture_ptr指示是否有解码数据输出。
输入数据在AVPacket结构中,输出数据在AVFrame结构中。AVFrame是定义在avcodec.h中的一个数据结构:
- typedef struct AVFrame {
- FF_COMMON_FRAME
- } AVFrame;
FF_COMMON_FRAME定义了诸多数据域,大部分由FFMpeg内部使用,对于用户来说,比较重要的主要包括:
- #define FF_COMMON_FRAME \
- ......
- uint8_t *data[4];\
- int linesize[4];\
- int key_frame;\
- int pict_type;\
- int64_t pts;\
- int reference;\
- ......
FFMpeg内部以planar的方式存储原始图像数据,即将图像像素分为多个平面(R/G/B或Y/U/V),data数组内的指针分别指向四个像素平面的起始位置,linesize数组则存放各个存贮各个平面的缓冲区的行宽:
- +++++++++++++++++++++++++++++++++++++++++++++++++++++++++
- +++data[0]->#################################++++++++++++
- ++++++++++++###########picture data##########++++++++++++
- ++++++++++++#################################++++++++++++
- ++++++++++++#################################++++++++++++
- ........................
- ++++++++++++#################################++++++++++++
- |<-------------------line_size[0]---------------------->|
此外,key_frame标识该图像是否是关键帧;pict_type表示该图像的编码类型:I(1)/P(2)/B(3)……;pts是以time_base为单位的时间标签,对于部分解码器如H.261、H.263和MPEG4,可以从头信息中获取;reference表示该图像是否被用作参考。
avcodec_decode_audio4
- int avcodec_decode_audio4(AVCodecContext *avctx, AVFrame *frame, int *got_frame_ptr, AVPacket *avpkt);
解码一个音频帧。输入数据在AVPacket结构中,输出数据在frame中,got_frame_ptr表示是否有数据输出。
- int avcodec_close(AVCodecContext *avctx);
关闭解码器,释放avcodec_open中分配的资源。
写系列
写系列API负责将媒体数据以包的形式分发到指定的目标,这个目标可以是一个本地文件、一个RTSP或HTTP流或者其他。
avformat_alloc_output_context2
- int avformat_alloc_output_context2(AVFormatContext **ctx, AVOutputFormat *oformat,
- const char *format_name, const char *filename);
这个函数负责分配一个用于输出目的的AVFormatContext,输出格式由oformat、format_name和filename决定,oformat优先级最高,如果oformat为空则依据format_name,如果format_name为空则依据filename。
avformat_write_header
- int avformat_write_header(AVFormatContext *s, AVDictionary **options);
此函数负责产生一个文件头并写入到输出目标中,调用之前,AVFormatContext结构中的oformat、pb必须正确设置,并且streams列表不能为空,列表中各stream的codec参数需要配置正确。
av_write_frame
- int av_write_frame(AVFormatContext *s, AVPacket *pkt);
此函数负责输出一个媒体包。AVFormatContext参数给出输出目标及媒体流的格式设置,pkt是一个包含媒体数据的结构,其内部成员如dts/pts、stream_index等必须被正确设置。
av_interleaved_write_frame
- int av_interleaved_write_frame(AVFormatContext *s, AVPacket *pkt);
此函数负责交错地输出一个媒体包。如果调用者无法保证来自各个媒体流的包正确交错,则最好调用此函数输出媒体包,反之,可以调用av_write_frame以提高性能。
av_write_trailer
- int av_write_trailer(AVFormatContext *s);
其他
avformat_new_stream
- AVStream *avformat_new_stream(AVFormatContext *s, AVCodec *c);
这个函数负责创建一个AVStream结构,并将其添加到指定的AVFormatContext中,此时,AVStream的大部分域处于非法状态。如果输入参数c不为空,AVStream的codec域根据c实现初始化。
FFMPEG支持由多个过滤器结构组成的过滤链,以实现对视频帧和音频采样数据的后续处理,如图像缩放、图像增强、声音的重采样等。
数据结构
FFMPEG使用了的四个主要的数据结构来构造过滤链,如上图所示。其中,AVFilter是一个过滤器的抽象,它拥有若干AVFilterPad,分别实现过滤器数据的输入和输出。而AVFilterLink表示各个过滤器之间的连接,这种连接的实现是基于AVFilterPad的,每个连接都有一个实现输入的AVFilterPad和一个实现输出的AVFilterPad以及对应的源过滤器和目的过滤器。
此外,还有两个数据结构AVFilterBuffer和AVFilterBufferRef来表示过滤器的输入输出数据,其中AVFilterBufferRef是AVFilterBuffer的引用,这种设计的目的是避免不必要的数据拷贝。通过avfilter_get_buffer_ref_from_frame可以获得一个AVFrame结构的引用,从而实现过滤器对来自AVFrame的图像原数据或声音采样序列的处理。
APIs
avfilter_open
avfilter_free
avfilter_init_filter
avfilter_link
avfilter_link_free
avfilter_ref_buffer
运作原理