1.先启动流服务器 ./mediamtx
2.开始推流: ffmpeg -re -stream_loop -1 -i /Users/hackerx/Desktop/test.mp4 -c copy -rtsp_transport tcp -f rtsp rtsp://127.0.0.1:8554/stream
3. 安装ffmpeg 4.4
brew install ffmpeg@4
4.添加ffmpeg头文件目录与库目录
5.链接ffmpeg相关库
6.设计界面
7.拉流
ffmpegmanager.cpp:
#include "ffmpegmananger.h"
#include
//构造
ffmpegMananger::ffmpegMananger(QObject *parent) : QObject{parent}
{
m_pInFmtCtx = nullptr;//输入流格式上下文
m_pTsFmtCtx = nullptr;//输出流格式上下文
m_strInputStreamUrl = "";//输入流地址
m_strOutputStreamPath = "";//输出流地址
}
//拆构
ffmpegMananger::~ffmpegMananger()
{
avformat_free_context(m_pInFmtCtx);//释放输入流格式上下文
avformat_free_context(m_pTsFmtCtx);//释放输出流格式上下文
}
//取输入流地址
void ffmpegMananger::getRtspAddress(QString url)
{
this->m_strInputStreamUrl = url;
}
//取输出流地址
void ffmpegMananger::getOutputAddress(QString path)
{
this->m_strOutputStreamPath = path;
printf("输出流地址: %s\n",m_strOutputStreamPath.toStdString().c_str());
}
void ffmpegMananger::setOutputCtx(AVCodecContext *encCtx, AVFormatContext **pTsFmtCtx,int &nVideoIdx_out)
{
avformat_alloc_output_context2(pTsFmtCtx , nullptr, nullptr, m_strOutputStreamPath.toStdString().c_str());
if (!pTsFmtCtx ) {
printf("创建输出上下文失败: avformat_alloc_output_context2\n");
return;
}
if (avio_open(&((*pTsFmtCtx)->pb), m_strOutputStreamPath.toStdString().c_str(), AVIO_FLAG_READ_WRITE) < 0)
{
avformat_free_context(*pTsFmtCtx);
printf("打开输出流失败: avio_open\n");
return;
}
AVStream *out_stream = avformat_new_stream(*pTsFmtCtx, encCtx->codec);
nVideoIdx_out = out_stream->index;
avcodec_parameters_from_context(out_stream->codecpar, encCtx);
printf("输出流信息:\n");
av_dump_format(*pTsFmtCtx, 0, m_strOutputStreamPath.toStdString().c_str(), 1);
printf("----------------------------\n");
}
//拉流并播放
int ffmpegMananger::ffmepgInput()
{
int nRet = 0;
AVCodecContext *encCtx = nullptr;
std::string temp = m_strInputStreamUrl.toStdString();
const char *pUrl = temp.c_str();
printf("输入流地址: %s\n",pUrl);
//设置选项
AVDictionary *dict = nullptr;
av_dict_set(&dict,"rtsp_transport", "tcp", 0);
av_dict_set(&dict,"stimeout","10000000",0);
av_dict_set(&dict, "buffer_size", "1024000", 0);
//打开输入流
nRet = avformat_open_input(&m_pInFmtCtx,pUrl,nullptr,&dict);
if( nRet < 0)
{
printf("打开输入流失败\n");
return nRet;
}
avformat_find_stream_info(m_pInFmtCtx, nullptr);
printf("输入流信息:\n");
av_dump_format(m_pInFmtCtx, 0, pUrl, 0);
printf("---------------------------\n");
//视频流索引
int nVideo_indx = av_find_best_stream(m_pInFmtCtx,AVMEDIA_TYPE_VIDEO,-1,-1,nullptr,0);
if(nVideo_indx < 0)
{
avformat_free_context(m_pInFmtCtx);
printf("查找视频流索引失败: av_find_best_stream\n");
return -1;
}
//查找解码器
auto pInCodec = avcodec_find_decoder(m_pInFmtCtx->streams[nVideo_indx]->codecpar->codec_id);
if(nullptr == pInCodec)
{
printf("查找解码器失败: avcodec_find_decoder fail.");
return -1;
}
//解码器上下文
AVCodecContext* pInCodecCtx = avcodec_alloc_context3(pInCodec);
//设置解码器参数
nRet = avcodec_parameters_to_context(pInCodecCtx, m_pInFmtCtx->streams[nVideo_indx]->codecpar);
if(nRet < 0)
{
avcodec_free_context(&pInCodecCtx);
printf("设置解码器参数失败: avcodec_parameters_to_context");
return -1;
}
//打开解码器
if(avcodec_open2(pInCodecCtx, pInCodec, nullptr) < 0)
{
avcodec_free_context(&pInCodecCtx);
printf("打开解码器失败: avcodec_open2\n");
return -1;
}
//输出视频分辨率
printf("视频宽:%d\n", pInCodecCtx->width);
printf("视频高:%d\n", pInCodecCtx->height);
int frame_index = 0;//帧索引
int got_picture = 0;//帧解码结果
//输出输入流
AVStream *in_stream =nullptr;
AVStream *out_stream =nullptr;
//分配内存
AVFrame *pFrame= av_frame_alloc();
AVFrame *pFrameRGB = av_frame_alloc();
AVPacket *newpkt = av_packet_alloc();
AVPacket *packet = av_packet_alloc();
//初始化视频包
av_init_packet(newpkt);
av_init_packet(packet);
//图像色彩空间转换、分辨率缩放、前后图像滤波处理
SwsContext *m_SwsContext = sws_getContext(
pInCodecCtx->width,
pInCodecCtx->height,
pInCodecCtx->pix_fmt,
pInCodecCtx->width,
pInCodecCtx->height,
AV_PIX_FMT_RGB32,
SWS_BICUBIC,
nullptr, nullptr, nullptr);
int bytes = av_image_get_buffer_size(
AV_PIX_FMT_RGB32,
pInCodecCtx->width,
pInCodecCtx->height,
4);
uint8_t *m_OutBuffer = (uint8_t *)av_malloc(bytes * sizeof(uint8_t));
//将分配的内存空间给pFrameRGB使用
avpicture_fill((AVPicture *)pFrameRGB,
m_OutBuffer,
AV_PIX_FMT_RGB32,
pInCodecCtx->width,
pInCodecCtx->height);
if(encCtx == nullptr)
{
//打开编码器
openEncoder(pInCodecCtx->width, pInCodecCtx->height,&encCtx);
}
//视频索引
int videoindex_out = 0;
//设置输出文件上下文
setOutputCtx(encCtx,&m_pTsFmtCtx,videoindex_out);
//写文件头
if (avformat_write_header(m_pTsFmtCtx, nullptr) < 0)
{
avformat_free_context(m_pTsFmtCtx);
printf("写文件头失败\n");
return -1;
}
printf("写文件头成功.\n");
int count = 0;//已解码帧数量
nRet = 0;//读取帧结果
//从pInFmtCtx读H264数据到packet;
while(av_read_frame(m_pInFmtCtx, packet) >= 0)
{
if(packet->stream_index != nVideo_indx)//仅保留图像
{
continue;
}
//送packet中H264数据给解码器码器进行解码,解码好的YUV数据放在pInCodecCtx,
if(avcodec_send_packet(pInCodecCtx, packet)<0)
{
break;
}
//释放已解码帧引用
av_packet_unref(packet);
//把解码好的YUV数据放到pFrame中
got_picture = avcodec_receive_frame(pInCodecCtx, pFrame);
//解码好一帧数据
if(0 == got_picture)
{
//发送显示图像的信号
// 对解码视频帧进行缩放、格式转换等操作
sws_scale(m_SwsContext,
(uint8_t const * const *)pFrame->data,
pFrame->linesize,
0,
pInCodecCtx->height,
pFrameRGB->data,
pFrameRGB->linesize);
// 转换到QImage
QImage tmmImage((uchar *)m_OutBuffer, pInCodecCtx->width, pInCodecCtx->height, QImage::Format_RGB32);
//复制图像
QImage image = tmmImage.copy();
//发送图像帧解码完成信息
emit Sig_GetOneFrame(image);
//设置解码器PTS
setDecoderPts(newpkt->stream_index,count, pFrame);
count++;//已解码计数
//送原始数据给编码器进行编码
nRet = avcodec_send_frame(encCtx,pFrame);
if(nRet < 0)
{
continue;
}
//从编码器获取编号的数据
while(nRet >= 0)
{
//接收已编码包
nRet = avcodec_receive_packet(encCtx,newpkt);
if(nRet < 0)
{
break;
}
//设置编码包PTS
setEncoderPts(nVideo_indx,frame_index,videoindex_out,newpkt);
int _count = 1;
printf("写%d包,大小:%5d,PTS:%lld\n", _count,newpkt->size, newpkt->pts);
if (av_interleaved_write_frame(m_pTsFmtCtx, newpkt) < 0)
{
printf("写帧失败: av_interleaved_write_frame\n");
goto end;
}
_count++;
av_packet_unref(newpkt);//释放已编码包
}
}
}
while(1)//从pInFmtCtx读H264数据到packet;
{
if(packet->stream_index != nVideo_indx)//仅保留图像
{
continue;
}
//送packet中H264数据给解码器码器进行解码,解码好的YUV数据放在pInCodecCtx,
if(avcodec_send_packet(pInCodecCtx, packet)<0)
{
continue;
}
//释放已解码包
av_packet_unref(packet);
//把解码好的YUV数据放到pFrame中
got_picture = avcodec_receive_frame(pInCodecCtx, pFrame);
//解码好一帧数据
if(!got_picture)
{
AVRational in_time_base1 = in_stream->time_base;
in_stream = m_pInFmtCtx->streams[newpkt->stream_index];
//PTS
int64_t in_duration = (double)AV_TIME_BASE / av_q2d(in_stream->r_frame_rate);
pFrame->pts = (double)(count*in_duration) / (double)(av_q2d(in_time_base1)*AV_TIME_BASE);
count++;
//送原始数据给编码器进行编码
nRet = avcodec_send_frame(encCtx,pFrame);
if(nRet < 0)
{
break;
}
//从编码器获取编号的数据
while(nRet >= 0)
{
nRet = avcodec_receive_packet(encCtx,newpkt);
if(nRet < 0)
{
continue;
}
in_stream = m_pInFmtCtx->streams[newpkt->stream_index];
out_stream = m_pTsFmtCtx->streams[videoindex_out];
if (newpkt->stream_index == nVideo_indx)
{
if (newpkt->pts == AV_NOPTS_VALUE)
{
//写入PTS
AVRational time_base1 = in_stream->time_base;
int64_t calc_duration = (double)AV_TIME_BASE / av_q2d(in_stream->r_frame_rate);
//设置包参数
newpkt->pts = (double)(frame_index*calc_duration) / (double)(av_q2d(time_base1)*AV_TIME_BASE);
newpkt->dts = newpkt->pts;
newpkt->duration = (double)calc_duration / (double)(av_q2d(time_base1)*AV_TIME_BASE);
frame_index++;
}
}
//转换PTS/DTS
newpkt->pts = av_rescale_q_rnd(newpkt->pts, in_stream->time_base, out_stream->time_base, (AVRounding)(AV_ROUND_NEAR_INF | AV_ROUND_PASS_MINMAX));
newpkt->dts = av_rescale_q_rnd(newpkt->dts, in_stream->time_base, out_stream->time_base, (AVRounding)(AV_ROUND_NEAR_INF | AV_ROUND_PASS_MINMAX));
newpkt->duration = av_rescale_q(newpkt->duration, in_stream->time_base, out_stream->time_base);
newpkt->pos = -1;
newpkt->stream_index = videoindex_out;
int count = 1;
printf("写%d包,大小:%5dPTS:%lld\n", count,newpkt->size, newpkt->pts);
if (av_interleaved_write_frame(m_pTsFmtCtx, newpkt) < 0)
{
printf("写帧失败: av_interleaved_write_frame\n");
goto end;
}
count++;
av_packet_unref(newpkt);
}
}
}
//Write file trailer
av_write_trailer(m_pTsFmtCtx);
end:
av_frame_free(&pFrame);
av_frame_free(&pFrameRGB);
av_packet_unref(newpkt);
av_packet_unref(packet);
std::cout<<"拉流完成";
return 0;
}
void ffmpegMananger::setDecoderPts(int idx,int count,AVFrame *pFrame)
{
AVStream* in_stream = m_pInFmtCtx->streams[idx];
AVRational in_time_base1 = in_stream->time_base;
//Duration between 2 frames (us)
int64_t in_duration = (double)AV_TIME_BASE / av_q2d(in_stream->r_frame_rate);
pFrame->pts = (double)(count*in_duration) / (double)(av_q2d(in_time_base1)*AV_TIME_BASE);
}
void ffmpegMananger::setEncoderPts(int nVideo_indx,int frame_index,int videoindex_out,AVPacket *newpkt)
{
AVStream*in_stream = m_pInFmtCtx->streams[newpkt->stream_index];
AVStream*out_stream = m_pTsFmtCtx->streams[videoindex_out];
if (newpkt->stream_index == nVideo_indx)
{
//FIX:No PTS (Example: Raw H.264)
//Simple Write PTS
if (newpkt->pts == AV_NOPTS_VALUE)
{
//Write PTS
AVRational time_base1 = in_stream->time_base;
//Duration between 2 frames (us)
int64_t calc_duration = (double)AV_TIME_BASE / av_q2d(in_stream->r_frame_rate);
//Parameters
newpkt->pts = (double)(frame_index*calc_duration) / (double)(av_q2d(time_base1)*AV_TIME_BASE);
newpkt->dts = newpkt->pts;
newpkt->duration = (double)calc_duration / (double)(av_q2d(time_base1)*AV_TIME_BASE);
frame_index++;
}
}
//Convert PTS/DTS
newpkt->pts = av_rescale_q_rnd(newpkt->pts, in_stream->time_base, out_stream->time_base, (AVRounding)(AV_ROUND_NEAR_INF | AV_ROUND_PASS_MINMAX));
newpkt->dts = av_rescale_q_rnd(newpkt->dts, in_stream->time_base, out_stream->time_base, (AVRounding)(AV_ROUND_NEAR_INF | AV_ROUND_PASS_MINMAX));
newpkt->duration = av_rescale_q(newpkt->duration, in_stream->time_base, out_stream->time_base);
newpkt->pos = -1;
newpkt->stream_index = videoindex_out;
}
void ffmpegMananger::writeTail()
{
//Write file trailer
av_write_trailer(m_pTsFmtCtx);
}
void ffmpegMananger::openEncoder(int width, int height, AVCodecContext** enc_ctx)
{
//使用libx264编码器
auto pCodec = avcodec_find_encoder_by_name("libx264");
if(nullptr == pCodec)
{
printf("avcodec_find_encoder_by_name fail.\n");
return;
}
//获取编码器上下文
*enc_ctx = avcodec_alloc_context3(pCodec);
if(nullptr == enc_ctx)
{
printf("avcodec_alloc_context3(pCodec) fail.\n");
return;
}
//sps/pps
(*enc_ctx)->profile = FF_PROFILE_H264_MAIN;
(*enc_ctx)->level = 30;//表示level是5.0
//分辨率
(*enc_ctx)->width = width;
(*enc_ctx)->height = height;
//gop
(*enc_ctx)->gop_size = 25;//i帧间隔
(*enc_ctx)->keyint_min = 20;//设置最小自动插入i帧的间隔.OPTION
//B帧
(*enc_ctx)->max_b_frames = 0;//不要B帧
(*enc_ctx)->has_b_frames = 0;//
//参考帧
(*enc_ctx)->refs = 3;//OPTION
//设置输入的yuv格式
(*enc_ctx)->pix_fmt = AV_PIX_FMT_YUV420P;
//设置码率
(*enc_ctx)->bit_rate = 3000000;
//设置帧率
(*enc_ctx)->time_base = (AVRational){1,25};//帧与帧之间的间隔
(*enc_ctx)->framerate = (AVRational){25,1};//帧率 25帧每秒
if(avcodec_open2((*enc_ctx),pCodec,nullptr) < 0)
{
printf("avcodec_open2 fail.\n");
}
return;
}
ffmpegmanager.h
#ifndef FFMPEGMANANGER_H
#define FFMPEGMANANGER_H
#pragma execution_character_set("utf-8")
//QT头
#include
#include
#include
//C标准头
#include
#include
//FFmpeg头
extern "C"
{
#include "libswscale/swscale.h"
#include "libavdevice/avdevice.h"
#include "libavcodec/avcodec.h"
#include "libavcodec/bsf.h"
#include "libavformat/avformat.h"
#include "libavutil/avutil.h"
#include "libavutil/imgutils.h"
#include "libavutil/log.h"
#include "libavutil/time.h"
#include
}
class ffmpegMananger : public QObject
{
Q_OBJECT
public:
//构造
explicit ffmpegMananger(QObject *parent = nullptr);
//拆构
~ffmpegMananger();
//取输入流地址
void getRtspAddress(QString url);
//取输出流地址
void getOutputAddress(QString path);
//ffmpeg拉流播放
int ffmepgInput();
//打开解码器
void openEncoder(int width, int height, AVCodecContext** enc_ctx);
//设置输出上下文
void setOutputCtx(AVCodecContext *encCtx, AVFormatContext **pTsFmtCtx,int &nVideoIdx_out);
//写文件尾
void writeTail();
//设置解码的pts
void setDecoderPts(int idx,int count,AVFrame *pFrame);
//设置编码的pts
void setEncoderPts(int nVideo_indx,int frame_index,int videoindex_out,AVPacket *newpkt);
signals:
//取一帧图像信号
void Sig_GetOneFrame(QImage img);
private:
//输入流地址
QString m_strInputStreamUrl;
//输出流地址
QString m_strOutputStreamPath;
//输入流动格式上下文
AVFormatContext *m_pInFmtCtx;
//输出流动格式上下文
AVFormatContext *m_pTsFmtCtx;
bool m_ifRec;
};
#endif // FFMPEGMANANGER_H