PC机操作系统:Redhat 6.3 (32bit)
交叉编译器: 4.5.1 (该编译器是开发板光盘自带的)
2.1 下载X264库
下载地址: https://www.videolan.org/developers/x264.html
2.2 解压、配置X264库
[wbyq@wbyq tiny4412]$ tar xvf /mnt/hgfs/linux-share-dir/x264-master.tar.bz2
[wbyq@wbyq tiny4412]$ cd x264-master/
[wbyq@wbyq x264-master]$ ./configure --prefix=$PWD/_install --disable-asm --enable-shared --enable-static --host=arm-none-linux-gnueabi
2.3 修改配置文件
打开当前目录下的config.mak文件,将里面的编译器相关的命令加上arm-linux-前缀,因为默认是使用gcc
[wbyq@wbyq x264-master]$ gedit config.mak
2.4 编译安装X264
[wbyq@wbyq x264-master]$ make && make install
查看编译安装的文件:
[wbyq@wbyq x264-master]$ tree _install/
_install/
├── bin
│ └── x264
├── include
│ ├── x264_config.h
│ └── x264.h
└── lib
├── libx264.a
├── libx264.so -> libx264.so.160
├── libx264.so.160
└── pkgconfig
└── x264.pc
4 directories, 7 files
[wbyq@wbyq x264-master]$
3.1 下载FFMPEG
下载地址: http://www.ffmpeg.org/download.html
3.2 编译FFMPEG
[wbyq@wbyq ffmpeg-4.2.2]$ ./configure --enable-shared --enable-static --prefix=$PWD/_install --cross-prefix=arm-linux- --arch=arm --target-os=linux --enable-gpl --extra-cflags=-I/home/wbyq/work/tiny4412/x264-master/_install/include --extra-ldflags=-L/home/wbyq/work/tiny4412/x264-master/_install/lib --enable-ffmpeg --enable-libx264
[wbyq@wbyq ffmpeg-4.2.2]$ make && make install
3.3 查看编译成功的文件
[wbyq@wbyq ffmpeg-4.2.2]$ tree _install/
_install/
├── bin
│ ├── ffmpeg
│ └── ffprobe
├── include
│ ├── libavcodec
│ │ ├── ac3_parser.h
│ │ ├── adts_parser.h
│ │ ├── avcodec.h
│ │ ├── avdct.h
│ │ ├── avfft.h
│ │ ├── d3d11va.h
│ │ ├── dirac.h
│ │ ├── dv_profile.h
│ │ ├── dxva2.h
│ │ ├── jni.h
│ │ ├── mediacodec.h
│ │ ├── qsv.h
│ │ ├── vaapi.h
│ │ ├── vdpau.h
│ │ ├── version.h
│ │ ├── videotoolbox.h
│ │ ├── vorbis_parser.h
│ │ └── xvmc.h
│ ├── libavdevice
│ │ ├── avdevice.h
│ │ └── version.h
│ ├── libavfilter
│ │ ├── avfilter.h
│ │ ├── buffersink.h
│ │ ├── buffersrc.h
│ │ └── version.h
│ ├── libavformat
│ │ ├── avformat.h
│ │ ├── avio.h
│ │ └── version.h
│ ├── libavutil
│ │ ├── adler32.h
│ │ ├── aes_ctr.h
│ │ ├── aes.h
│ │ ├── attributes.h
│ │ ├── audio_fifo.h
│ │ ├── avassert.h
│ │ ├── avconfig.h
│ │ ├── avstring.h
│ │ ├── avutil.h
│ │ ├── base64.h
│ │ ├── blowfish.h
│ │ ├── bprint.h
│ │ ├── bswap.h
│ │ ├── buffer.h
│ │ ├── camellia.h
│ │ ├── cast5.h
│ │ ├── channel_layout.h
│ │ ├── common.h
│ │ ├── cpu.h
│ │ ├── crc.h
│ │ ├── des.h
│ │ ├── dict.h
│ │ ├── display.h
│ │ ├── downmix_info.h
│ │ ├── encryption_info.h
│ │ ├── error.h
│ │ ├── eval.h
│ │ ├── ffversion.h
│ │ ├── fifo.h
│ │ ├── file.h
│ │ ├── frame.h
│ │ ├── hash.h
│ │ ├── hdr_dynamic_metadata.h
│ │ ├── hmac.h
│ │ ├── hwcontext_cuda.h
│ │ ├── hwcontext_d3d11va.h
│ │ ├── hwcontext_drm.h
│ │ ├── hwcontext_dxva2.h
│ │ ├── hwcontext.h
│ │ ├── hwcontext_mediacodec.h
│ │ ├── hwcontext_qsv.h
│ │ ├── hwcontext_vaapi.h
│ │ ├── hwcontext_vdpau.h
│ │ ├── hwcontext_videotoolbox.h
│ │ ├── imgutils.h
│ │ ├── intfloat.h
│ │ ├── intreadwrite.h
│ │ ├── lfg.h
│ │ ├── log.h
│ │ ├── lzo.h
│ │ ├── macros.h
│ │ ├── mastering_display_metadata.h
│ │ ├── mathematics.h
│ │ ├── md5.h
│ │ ├── mem.h
│ │ ├── motion_vector.h
│ │ ├── murmur3.h
│ │ ├── opt.h
│ │ ├── parseutils.h
│ │ ├── pixdesc.h
│ │ ├── pixelutils.h
│ │ ├── pixfmt.h
│ │ ├── random_seed.h
│ │ ├── rational.h
│ │ ├── rc4.h
│ │ ├── replaygain.h
│ │ ├── ripemd.h
│ │ ├── samplefmt.h
│ │ ├── sha512.h
│ │ ├── sha.h
│ │ ├── spherical.h
│ │ ├── stereo3d.h
│ │ ├── tea.h
│ │ ├── threadmessage.h
│ │ ├── timecode.h
│ │ ├── time.h
│ │ ├── timestamp.h
│ │ ├── tree.h
│ │ ├── twofish.h
│ │ ├── tx.h
│ │ ├── version.h
│ │ └── xtea.h
│ ├── libpostproc
│ │ ├── postprocess.h
│ │ └── version.h
│ ├── libswresample
│ │ ├── swresample.h
│ │ └── version.h
│ └── libswscale
│ ├── swscale.h
│ └── version.h
├── lib
│ ├── libavcodec.a
│ ├── libavcodec.so -> libavcodec.so.58.54.100
│ ├── libavcodec.so.58 -> libavcodec.so.58.54.100
│ ├── libavcodec.so.58.54.100
│ ├── libavdevice.a
│ ├── libavdevice.so -> libavdevice.so.58.8.100
│ ├── libavdevice.so.58 -> libavdevice.so.58.8.100
│ ├── libavdevice.so.58.8.100
│ ├── libavfilter.a
│ ├── libavfilter.so -> libavfilter.so.7.57.100
│ ├── libavfilter.so.7 -> libavfilter.so.7.57.100
│ ├── libavfilter.so.7.57.100
│ ├── libavformat.a
│ ├── libavformat.so -> libavformat.so.58.29.100
│ ├── libavformat.so.58 -> libavformat.so.58.29.100
│ ├── libavformat.so.58.29.100
│ ├── libavutil.a
│ ├── libavutil.so -> libavutil.so.56.31.100
│ ├── libavutil.so.56 -> libavutil.so.56.31.100
│ ├── libavutil.so.56.31.100
│ ├── libpostproc.a
│ ├── libpostproc.so -> libpostproc.so.55.5.100
│ ├── libpostproc.so.55 -> libpostproc.so.55.5.100
│ ├── libpostproc.so.55.5.100
│ ├── libswresample.a
│ ├── libswresample.so -> libswresample.so.3.5.100
│ ├── libswresample.so.3 -> libswresample.so.3.5.100
│ ├── libswresample.so.3.5.100
│ ├── libswscale.a
│ ├── libswscale.so -> libswscale.so.5.5.100
│ ├── libswscale.so.5 -> libswscale.so.5.5.100
│ ├── libswscale.so.5.5.100
│ └── pkgconfig
│ ├── libavcodec.pc
│ ├── libavdevice.pc
│ ├── libavfilter.pc
│ ├── libavformat.pc
│ ├── libavutil.pc
│ ├── libpostproc.pc
│ ├── libswresample.pc
│ └── libswscale.pc
└── share
├── ffmpeg
│ ├── examples
│ │ ├── avio_dir_cmd.c
│ │ ├── avio_reading.c
│ │ ├── decode_audio.c
│ │ ├── decode_video.c
│ │ ├── demuxing_decoding.c
│ │ ├── encode_audio.c
│ │ ├── encode_video.c
│ │ ├── extract_mvs.c
│ │ ├── filter_audio.c
│ │ ├── filtering_audio.c
│ │ ├── filtering_video.c
│ │ ├── http_multiclient.c
│ │ ├── hw_decode.c
│ │ ├── Makefile
│ │ ├── metadata.c
│ │ ├── muxing.c
│ │ ├── qsvdec.c
│ │ ├── README
│ │ ├── remuxing.c
│ │ ├── resampling_audio.c
│ │ ├── scaling_video.c
│ │ ├── transcode_aac.c
│ │ ├── transcoding.c
│ │ ├── vaapi_encode.c
│ │ └── vaapi_transcode.c
│ ├── ffprobe.xsd
│ ├── libvpx-1080p50_60.ffpreset
│ ├── libvpx-1080p.ffpreset
│ ├── libvpx-360p.ffpreset
│ ├── libvpx-720p50_60.ffpreset
│ └── libvpx-720p.ffpreset
└── man
├── man1
│ ├── ffmpeg.1
│ ├── ffmpeg-all.1
│ ├── ffmpeg-bitstream-filters.1
│ ├── ffmpeg-codecs.1
│ ├── ffmpeg-devices.1
│ ├── ffmpeg-filters.1
│ ├── ffmpeg-formats.1
│ ├── ffmpeg-protocols.1
│ ├── ffmpeg-resampler.1
│ ├── ffmpeg-scaler.1
│ ├── ffmpeg-utils.1
│ ├── ffprobe.1
│ └── ffprobe-all.1
└── man3
├── libavcodec.3
├── libavdevice.3
├── libavfilter.3
├── libavformat.3
├── libavutil.3
├── libswresample.3
└── libswscale.3
18 directories, 208 files
[wbyq@wbyq ffmpeg-4.2.2]$
4.1 拷贝库到开发板
为了方便部署到开发板,可以先创建一个“run_lib”目录,然后将X264和FFMPEG编译出来的动态库拷贝到“run_lib”目录下,再通过NFS服务器将“run_lib”目录拷贝到目标开发板上。 将库拷贝到开发板上之后,可以选择直接把“run_lib”目录下的库拷贝到/lib目录下,也可以将“”run_lib“”目录的路径添加到系统库的环境变量里。反正最终目的,只要让程序在运行时,能找到使用的库就对了。
拷贝过程如下:
[wbyq@wbyq tiny4412]$ ls
ffmpeg-4.2.2 x264-master
[wbyq@wbyq tiny4412]$ mkdir run_lib
[wbyq@wbyq tiny4412]$ cd run_lib/
[wbyq@wbyq run_lib]$ cp ../ffmpeg-4.2.2/_install/lib/*.so* ./
[wbyq@wbyq run_lib]$ cp ../x264-master/_install/lib/*.so* ./
[wbyq@wbyq run_lib]$ ls
libavcodec.so libavfilter.so libavutil.so libswresample.so libx264.so
libavcodec.so.58 libavfilter.so.7 libavutil.so.56 libswresample.so.3 libx264.so.160
libavcodec.so.58.54.100 libavfilter.so.7.57.100 libavutil.so.56.31.100 libswresample.so.3.5.100
libavdevice.so libavformat.so libpostproc.so libswscale.so
libavdevice.so.58 libavformat.so.58 libpostproc.so.55 libswscale.so.5
libavdevice.so.58.8.100 libavformat.so.58.29.100 libpostproc.so.55.5.100 libswscale.so.5.5.100
4.2 拷贝可执行命令到开发板进行测试
将库拷贝到开发板之后,再将“ffmpeg-4.2.2/_install/bin/ffmpeg” 可执行文件拷贝到开发板的bin目录下。
下面是在开发板的终端运行ffmpeg命令效果:
[root@tiny4412 ]# ls
bin etc linuxrc nfs.sh root tmp work
code home lost+found opt sbin usr
dev lib mnt proc sys var
[root@tiny4412 ]# ffmpeg
ffmpeg version 4.2.2 Copyright (c) 2000-2019 the FFmpeg developers
built with gcc 4.5.1 (ctng-1.8.1-FA)
configuration: --enable-shared --enable-static --prefix=/home/wbyq/work/tiny4412/ffmpeg-4.2.2/_install --cross-prefix=arm-linux- --arch=arm --target-os=linux --enable-gpl --extra-cflags=-I/home/wbyq/work/tiny4412/x264-master/_install/include --extra-ldflags=-L/home/wbyq/work/tiny4412/x264-master/_install/lib --enable-ffmpeg --enable-libx264
libavutil 56. 31.100 / 56. 31.100
libavcodec 58. 54.100 / 58. 54.100
libavformat 58. 29.100 / 58. 29.100
libavdevice 58. 8.100 / 58. 8.100
libavfilter 7. 57.100 / 7. 57.100
libswscale 5. 5.100 / 5. 5.100
libswresample 3. 5.100 / 3. 5.100
libpostproc 55. 5.100 / 55. 5.100
Hyper fast Audio and Video encoder
usage: ffmpeg [options] [[infile options] -i infile]... {[outfile options] outfile}...
Use -h to get full help or, even better, run 'man ffmpeg'
[root@tiny4412 ]#
命令能运行,接来下就可以进行音频视频相关的程序开发了。
5.1 完整代码
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#include
#define STREAM_DURATION 50.0 /*录制视频的持续时间 秒*/
#define STREAM_FRAME_RATE 5 /* images/s 这里可以根据摄像头的采集速度来设置帧率 */
#define STREAM_PIX_FMT AV_PIX_FMT_YUV420P /* default pix_fmt */
#define SCALE_FLAGS SWS_BICUBIC
//存放视频的宽度和高度
int video_width;
int video_height;
// 单个输出AVStream的包装器
typedef struct OutputStream
{
AVStream *st;
AVCodecContext *enc;
/*下一帧的点数*/
int64_t next_pts;
int samples_count;
AVFrame *frame;
AVFrame *tmp_frame;
float t, tincr, tincr2;
struct SwsContext *sws_ctx;
struct SwrContext *swr_ctx;
}OutputStream;
typedef struct IntputDev
{
AVCodecContext *pCodecCtx;
AVCodec *pCodec;
AVFormatContext *v_ifmtCtx;
int videoindex;
struct SwsContext *img_convert_ctx;
AVPacket *in_packet;
AVFrame *pFrame,*pFrameYUV;
}IntputDev;
static void log_packet(const AVFormatContext *fmt_ctx, const AVPacket *pkt)
{
AVRational *time_base = &fmt_ctx->streams[pkt->stream_index]->time_base;
printf("pts:%s pts_time:%s dts:%s dts_time:%s duration:%s duration_time:%s stream_index:%d\n",
av_ts2str(pkt->pts), av_ts2timestr(pkt->pts, time_base),
av_ts2str(pkt->dts), av_ts2timestr(pkt->dts, time_base),
av_ts2str(pkt->duration), av_ts2timestr(pkt->duration, time_base),
pkt->stream_index);
}
static int write_frame(AVFormatContext *fmt_ctx, const AVRational *time_base, AVStream *st, AVPacket *pkt)
{
/* 将输出数据包时间戳值从编解码器重新调整为流时基 */
av_packet_rescale_ts(pkt, *time_base, st->time_base);
pkt->stream_index = st->index;
/*将压缩的帧写入媒体文件。*/
log_packet(fmt_ctx, pkt);
return av_interleaved_write_frame(fmt_ctx, pkt);
}
/*添加输出流。 */
static void add_stream(OutputStream *ost, AVFormatContext *oc,AVCodec **codec,enum AVCodecID codec_id)
{
AVCodecContext *c;
int i;
/* find the encoder */
*codec = avcodec_find_encoder(codec_id);
if (!(*codec))
{
fprintf(stderr, "Could not find encoder for '%s'\n",
avcodec_get_name(codec_id));
exit(1);
}
ost->st = avformat_new_stream(oc, NULL);
if (!ost->st) {
fprintf(stderr, "Could not allocate stream\n");
exit(1);
}
ost->st->id = oc->nb_streams-1;
c = avcodec_alloc_context3(*codec);
if (!c) {
fprintf(stderr, "Could not alloc an encoding context\n");
exit(1);
}
ost->enc = c;
switch((*codec)->type)
{
case AVMEDIA_TYPE_AUDIO:
c->sample_fmt = (*codec)->sample_fmts ?
(*codec)->sample_fmts[0] : AV_SAMPLE_FMT_FLTP;
c->bit_rate = 64000;
c->sample_rate = 44100;
if ((*codec)->supported_samplerates) {
c->sample_rate = (*codec)->supported_samplerates[0];
for (i = 0; (*codec)->supported_samplerates[i]; i++) {
if ((*codec)->supported_samplerates[i] == 44100)
c->sample_rate = 44100;
}
}
c->channels = av_get_channel_layout_nb_channels(c->channel_layout);
c->channel_layout = AV_CH_LAYOUT_STEREO;
if ((*codec)->channel_layouts) {
c->channel_layout = (*codec)->channel_layouts[0];
for (i = 0; (*codec)->channel_layouts[i]; i++) {
if ((*codec)->channel_layouts[i] == AV_CH_LAYOUT_STEREO)
c->channel_layout = AV_CH_LAYOUT_STEREO;
}
}
c->channels = av_get_channel_layout_nb_channels(c->channel_layout);
ost->st->time_base = (AVRational){ 1, c->sample_rate };
break;
case AVMEDIA_TYPE_VIDEO:
c->codec_id = codec_id;
c->bit_rate = 2500000; //平均比特率,例子代码默认值是400000
/* 分辨率必须是2的倍数。*/
c->width=video_width;
c->height=video_height;
/*时基:这是基本的时间单位(以秒为单位)
*表示其中的帧时间戳。 对于固定fps内容,
*时基应为1 /framerate,时间戳增量应为
*等于1。*/
ost->st->time_base = (AVRational){1,STREAM_FRAME_RATE}; //帧率设置
c->time_base = ost->st->time_base;
c->gop_size = 12; /* 最多每十二帧发射一帧内帧 */
c->pix_fmt = STREAM_PIX_FMT;
if(c->codec_id == AV_CODEC_ID_MPEG2VIDEO)
{
/* 只是为了测试,我们还添加了B帧 */
c->max_b_frames = 2;
}
if(c->codec_id == AV_CODEC_ID_MPEG1VIDEO)
{
/*需要避免使用其中一些系数溢出的宏块。
*普通视频不会发生这种情况,因为
*色度平面的运动与亮度平面不匹配。 */
c->mb_decision = 2;
}
break;
default:
break;
}
/* 某些格式希望流头分开。 */
if (oc->oformat->flags & AVFMT_GLOBALHEADER)
c->flags |= AV_CODEC_FLAG_GLOBAL_HEADER;
}
static AVFrame *alloc_picture(enum AVPixelFormat pix_fmt, int width, int height)
{
AVFrame *picture;
int ret;
picture = av_frame_alloc();
if (!picture)
return NULL;
picture->format = pix_fmt;
picture->width = width;
picture->height = height;
/* 为帧数据分配缓冲区 */
ret = av_frame_get_buffer(picture, 32);
if(ret<0)
{
fprintf(stderr, "Could not allocate frame data.\n");
exit(1);
}
return picture;
}
static void open_video(AVFormatContext *oc, AVCodec *codec, OutputStream *ost, AVDictionary *opt_arg)
{
int ret;
AVCodecContext *c = ost->enc;
AVDictionary *opt = NULL;
av_dict_copy(&opt, opt_arg, 0);
/* open the codec */
ret = avcodec_open2(c, codec, &opt);
av_dict_free(&opt);
if (ret < 0)
{
fprintf(stderr, "Could not open video codec: %s\n", av_err2str(ret));
exit(1);
}
/* 分配并初始化可重用框架 */
ost->frame = alloc_picture(c->pix_fmt, c->width, c->height);
if (!ost->frame)
{
fprintf(stderr, "Could not allocate video frame\n");
exit(1);
}
printf("ost->frame alloc success fmt=%d w=%d h=%d\n",c->pix_fmt,c->width, c->height);
/*如果输出格式不是YUV420P,则为临时YUV420P
*也需要图片。 然后将其转换为所需的
*输出格式。 */
ost->tmp_frame = NULL;
if(c->pix_fmt != AV_PIX_FMT_YUV420P)
{
ost->tmp_frame = alloc_picture(AV_PIX_FMT_YUV420P, c->width, c->height);
if (!ost->tmp_frame)
{
fprintf(stderr, "Could not allocate temporary picture\n");
exit(1);
}
}
/* 将流参数复制到多路复用器*/
ret=avcodec_parameters_from_context(ost->st->codecpar, c);
if(ret<0)
{
fprintf(stderr, "Could not copy the stream parameters\n");
exit(1);
}
}
/*
*编码一个视频帧
*编码完成后返回1,否则返回0
*/
static int write_video_frame(AVFormatContext *oc, OutputStream *ost,AVFrame *frame)
{
int ret;
AVCodecContext *c;
int got_packet=0;
AVPacket pkt={0};
if(frame==NULL)
return 1;
c = ost->enc;
av_init_packet(&pkt);
/* 编码图像*/
ret = avcodec_encode_video2(c, &pkt, frame, &got_packet);
if(ret<0)
{
fprintf(stderr, "Error encoding video frame: %s\n", av_err2str(ret));
exit(1);
}
printf("--------------video- pkt.pts=%s\n",av_ts2str(pkt.pts));
printf("----st.num=%d st.den=%d codec.num=%d codec.den=%d---------\n",ost->st->time_base.num,ost->st->time_base.den,
c->time_base.num,c->time_base.den);
if(got_packet)
{
ret = write_frame(oc, &c->time_base, ost->st, &pkt);
}else
{
ret = 0;
}
if(ret<0)
{
fprintf(stderr, "Error while writing video frame: %s\n", av_err2str(ret));
exit(1);
}
return (frame || got_packet) ? 0 : 1;
}
static AVFrame *get_video_frame(OutputStream *ost,IntputDev* input,int *got_pic)
{
int ret, got_picture;
AVCodecContext *c = ost->enc;
AVFrame * ret_frame=NULL;
if(av_compare_ts(ost->next_pts, c->time_base,STREAM_DURATION, (AVRational){1,1})>=0)
return NULL;
/*当我们将帧传递给编码器时,它可能会保留对它的引用
*内部,确保我们在这里不覆盖它*/
if (av_frame_make_writable(ost->frame)<0)
exit(1);
if(av_read_frame(input->v_ifmtCtx, input->in_packet)>=0)
{
if(input->in_packet->stream_index==input->videoindex)
{
ret = avcodec_decode_video2(input->pCodecCtx, input->pFrame, &got_picture, input->in_packet);
*got_pic=got_picture;
if(ret<0)
{
printf("Decode Error.\n");
av_packet_unref(input->in_packet);
return NULL;
}
if(got_picture)
{
sws_scale(input->img_convert_ctx, (const unsigned char* const*)input->pFrame->data, input->pFrame->linesize, 0, input->pCodecCtx->height, ost->frame->data, ost->frame->linesize);
ost->frame->pts =ost->next_pts++;
ret_frame= ost->frame;
}
}
av_packet_unref(input->in_packet);
}
return ret_frame;
}
static void close_stream(AVFormatContext *oc, OutputStream *ost)
{
avcodec_free_context(&ost->enc);
av_frame_free(&ost->frame);
av_frame_free(&ost->tmp_frame);
sws_freeContext(ost->sws_ctx);
swr_free(&ost->swr_ctx);
}
/*
采集摄像头数据编码成MP4视频
*/
int main(int argc, char **argv)
{
OutputStream video_st = { 0 }, audio_st = { 0 };
const char *filename;
AVOutputFormat *fmt;
AVFormatContext *oc;
AVCodec *audio_codec, *video_codec;
int ret;
int have_video = 0, have_audio = 0;
int encode_video = 0, encode_audio = 0;
AVDictionary *opt = NULL;
int i;
if(argc<3)
{
//./app /dev/video0 123.mp4
printf("用法: %s <摄像头设备节点> <文件名称> \n", argv[0]);
return 1;
}
filename = argv[2];
printf("当前存储的视频文件名称:%s\n",filename);
/*分配输出媒体环境*/
avformat_alloc_output_context2(&oc, NULL, NULL, filename);
if(!oc)
{
printf("无法从文件扩展名推断出输出格式:使用MPEG。\n");
avformat_alloc_output_context2(&oc, NULL, "mpeg", filename);
}
if(!oc)return 1;
//添加摄像头----------------------------------
IntputDev video_input={0};
AVCodecContext *pCodecCtx;
AVCodec *pCodec;
AVFormatContext *v_ifmtCtx;
avdevice_register_all();
v_ifmtCtx = avformat_alloc_context();
//Linux下指定摄像头信息
AVInputFormat *ifmt=av_find_input_format("video4linux2");
if(avformat_open_input(&v_ifmtCtx,argv[1],ifmt,NULL)!=0)
{
printf("无法打开输入流.%s\n",argv[1]);
return -1;
}
if(avformat_find_stream_info(v_ifmtCtx,NULL)<0)
{
printf("找不到流信息.\n");
return -1;
}
int videoindex=-1;
for(i=0; inb_streams; i++)
if(v_ifmtCtx->streams[i]->codec->codec_type==AVMEDIA_TYPE_VIDEO)
{
videoindex=i;
printf("videoindex=%d\n",videoindex);
break;
}
if(videoindex==-1)
{
printf("找不到视频流。\n");
return -1;
}
pCodecCtx=v_ifmtCtx->streams[videoindex]->codec;
pCodec=avcodec_find_decoder(pCodecCtx->codec_id);
if(pCodec==NULL)
{
printf("找不到编解码器。\n");
return -1;
}
if(avcodec_open2(pCodecCtx, pCodec,NULL)<0)
{
printf("无法打开编解码器。\n");
return -1;
}
AVFrame *pFrame,*pFrameYUV;
pFrame=av_frame_alloc();
pFrameYUV=av_frame_alloc();
unsigned char *out_buffer=(unsigned char *)av_malloc(av_image_get_buffer_size(AV_PIX_FMT_YUV420P, pCodecCtx->width, pCodecCtx->height,16));
av_image_fill_arrays((AVPicture *)pFrameYUV->data,(AVPicture *)pFrameYUV->linesize, out_buffer, AV_PIX_FMT_YUV420P, pCodecCtx->width, pCodecCtx->height,16);
printf("摄像头尺寸(WxH): %d x %d \n",pCodecCtx->width, pCodecCtx->height);
video_width=pCodecCtx->width;
video_height=pCodecCtx->height;
struct SwsContext *img_convert_ctx;
img_convert_ctx = sws_getContext(pCodecCtx->width, pCodecCtx->height, pCodecCtx->pix_fmt, pCodecCtx->width, pCodecCtx->height, AV_PIX_FMT_YUV420P, SWS_BICUBIC, NULL, NULL, NULL);
AVPacket *in_packet=(AVPacket *)av_malloc(sizeof(AVPacket));
video_input.img_convert_ctx=img_convert_ctx;
video_input.in_packet=in_packet;
video_input.pCodecCtx=pCodecCtx;
video_input.pCodec=pCodec;
video_input.v_ifmtCtx=v_ifmtCtx;
video_input.videoindex=videoindex;
video_input.pFrame=pFrame;
video_input.pFrameYUV=pFrameYUV;
//-----------------------------添加摄像头结束
fmt=oc->oformat;
/*使用默认格式的编解码器添加音频和视频流并初始化编解码器。*/
printf("fmt->video_codec = %d\n", fmt->video_codec);
if(fmt->video_codec != AV_CODEC_ID_NONE)
{
add_stream(&video_st,oc,&video_codec,fmt->video_codec);
have_video=1;
encode_video=1;
}
/*现在已经设置了所有参数,可以打开音频并视频编解码器,并分配必要的编码缓冲区。*/
if(have_video)open_video(oc, video_codec, &video_st, opt);
av_dump_format(oc,0,filename,1);
/* 打开输出文件(如果需要) */
if(!(fmt->flags & AVFMT_NOFILE))
{
ret=avio_open(&oc->pb,filename,AVIO_FLAG_WRITE);
if(ret<0)
{
fprintf(stderr, "打不开'%s': %s\n", filename,av_err2str(ret));
return 1;
}
}
/* 编写流头(如果有)*/
ret=avformat_write_header(oc, &opt);
if(ret<0)
{
fprintf(stderr, "打开输出文件时发生错误: %s\n",av_err2str(ret));
return 1;
}
int got_pic;
while(encode_video)
{
/*选择要编码的流*/
AVFrame *frame=get_video_frame(&video_st,&video_input,&got_pic);
if(!got_pic)
{
usleep(10000);
continue;
}
encode_video=!write_video_frame(oc,&video_st,frame);
}
av_write_trailer(oc);
sws_freeContext(video_input.img_convert_ctx);
avcodec_close(video_input.pCodecCtx);
av_free(video_input.pFrameYUV);
av_free(video_input.pFrame);
avformat_close_input(&video_input.v_ifmtCtx);
/*关闭每个编解码器*/
if (have_video)close_stream(oc, &video_st);
/*关闭输出文件*/
if (!(fmt->flags & AVFMT_NOFILE))avio_closep(&oc->pb);
/*释放流*/
avformat_free_context(oc);
return 0;
}
5.2 Makefile文件
app1:
arm-linux-gcc ffmpeg_video.c -I /home/wbyq/work/tiny4412/ffmpeg-4.2.2/_install/include -L /home/wbyq/work/tiny4412/ffmpeg-4.2.2/_install/lib -I/home/wbyq/work/tiny4412/x264-master/_install/include -L/home/wbyq/work/tiny4412/x264-master/_install/lib -lavcodec -lavfilter -lavutil -lswresample -lavdevice -lavformat -lpostproc -lswscale -lx264
5.3 运行测试
[root@tiny4412 code]# ./a.out /dev/video15 123.mp4
由于目前开发板上没有安装播放器,下面是将123.MP4文件拷贝到PC机上播放的效果。
下面公众号有全套的C/C++/QT的基础教程,欢迎关注: