ffmpeg h264文件转mp4

h264文件不能直接在网页上播放,比如在浏览器上输入http://10.0.0.2/2022-01-08T22-32-58.h264,变成了下载。

若在浏览器上输入http://10.0.0.2/2022-01-08T22-32-58.mp4,则可以播放。

本文讲解用ffmpeg将h264文件转换成mp4。

首先,准备h264文件,这个可以用ffmpeg将一个mp4的视频部分转成h264,命令如下:

ffmpeg -i 2022-01-08T22-32-58.mp4 -an -vcodec copy 2022-01-08T22-32-58.h264

注意,我这里mp4里面的视频编码格式是h264,故用的vcodec copy,否则用vcodec libx264。

下面是将这个2022-01-08T22-32-58.h264封装成mp4,注意,没有解码编码过程,速度会很快。
大概过程如下:
1.启动两个线程
2.第一个线程av_read_frame读到packet,然后av_packet_clone拷贝出一个packet,塞入队列
3.第二个线程从队列中读取packet,进行pts,dts,duration设置,然后调用av_interleaved_write_frame写入。

这里面需要注意的地方在于pts,dts的设置,从ffmpeg源码上看,对于pts,dts的设置,也是很麻烦的,源码中,很多地方,动不动就看pts,dts是否需要设置。

本人之前采取下列方式进行dts和pts的设置,即由于读的context和写入的context的时间基(time_base)不一致,需要转换成目标文件context的时间基。

pPacketVideoA->dts = av_rescale_q_rnd(pPacketVideoA->dts, m_pFormatCtx_File->streams[m_iVideoIndex]->time_base, m_pFormatCtx_Out->streams[m_iVideoIndex]->time_base, AVRounding(1));

但实际运行时,出现了问题,读取出来的packet的pts和dts都是AV_NOPTS_VALUE,经过转换还是AV_NOPTS_VALUE,然后调用av_interleaved_write_frame出错。

最后调试ffmpeg源码,得到pts和dts的处理如下:

pPacketVideoA->duration = av_rescale_q_rnd(pPacketVideoA->duration, m_pFormatCtx_File->streams[m_iVideoIndex]->time_base, m_pFormatCtx_Out->streams[m_iVideoIndex]->time_base, AVRounding(1));
if (pPacketVideoA->dts != AV_NOPTS_VALUE)
{
	pPacketVideoA->dts = av_rescale_q_rnd(pPacketVideoA->dts, m_pFormatCtx_File->streams[m_iVideoIndex]->time_base, m_pFormatCtx_Out->streams[m_iVideoIndex]->time_base, AVRounding(1));
	pPacketVideoA->pts = pPacketVideoA->dts;
}
else
{
	pPacketVideoA->pts = pPacketVideoA->dts = m_iTotalDuration;
}
m_iTotalDuration += pPacketVideoA->duration;

m_iTotalDuration的初始值,本人设置的是0,即第一帧的pts和dts都设置成0,下一帧的dts就是前面总帧的duration和,这个是ffmpeg里面的逻辑。

当然,ffmpeg里面判断复杂多了,比如参考了是否有B帧,本人所写的没考虑B帧,只考虑一个文件里面只有I帧,P帧的情况。

下面是代码结构
ffmpeg h264文件转mp4_第1张图片
其中FfmpegH264ToMp4.cpp的内容如下:

// FfmpegH264ToMp4.cpp : 此文件包含 "main" 函数。程序执行将在此处开始并结束。
//

#include 
#include "H264ToMp4.h"

int main()
{
	CVideoConvert cVideoConvert;

	std::string strVideoFile = "2022-01-08T22-32-58.h264";
	std::string strOutFile = "out.mp4";
	cVideoConvert.StartConvert(strVideoFile, strOutFile);
	cVideoConvert.WaitFinish();
    return 0;
}



H264ToMp4.h里面的内容如下:

#pragma once
#include 

#include 
#include 

#define MAX_VIDEOPACKET_NUM 100

#ifdef	__cplusplus
extern "C"
{
#endif
#include "libavcodec/avcodec.h"
#include "libavformat/avformat.h"
#include "libswscale/swscale.h"
#include "libswresample/swresample.h"
#include "libavdevice/avdevice.h"
#include "libavutil/audio_fifo.h"
#include "libavutil/avutil.h"
#include "libavutil/fifo.h"
#include "libavutil/frame.h"
#include "libavutil/imgutils.h"

#include "libavfilter/avfilter.h"
#include "libavfilter/buffersink.h"
#include "libavfilter/buffersrc.h"


#ifdef __cplusplus
};
#endif

class CVideoConvert
{
public:
	CVideoConvert();
	~CVideoConvert();
public:
	int StartConvert(std::string strFileName, std::string strFileOut);
	void StopConvert();
public:
	int OpenFile(const char *pFile);
	int OpenOutPut(const char *pFileOut);
private:
	static DWORD WINAPI FileReadProc(LPVOID lpParam);
	void FileRead();

	static DWORD WINAPI FileWriteProc(LPVOID lpParam);
	void FileWrite();

public:
	void WaitFinish();
public:
	AVFormatContext *m_pFormatCtx_File = NULL;
	AVFormatContext *m_pFormatCtx_Out = NULL;

	std::deque<AVPacket *> m_vecReadVideo;


	int m_iVideoWidth = 1920;
	int m_iVideoHeight = 1080;
	int m_iYuv420FrameSize = 0;
	int m_iVideoFrameNum = 0;

	int m_iVideoIndex = -1;
	bool m_bStartConvert = false;
	int64_t m_iTotalDuration = 0;
private:
	CRITICAL_SECTION m_csVideoReadSection;
	HANDLE m_hFileReadThread = NULL;
	HANDLE m_hFileWriteThread = NULL;
};


H264ToMp4.cpp里面的内容如下:

#include "H264ToMp4.h"


#ifdef	__cplusplus
extern "C"
{
#endif

#pragma comment(lib, "avcodec.lib")
#pragma comment(lib, "avformat.lib")
#pragma comment(lib, "avutil.lib")
#pragma comment(lib, "avdevice.lib")
#pragma comment(lib, "avfilter.lib")
#pragma comment(lib, "postproc.lib")
#pragma comment(lib, "swresample.lib")
#pragma comment(lib, "swscale.lib")


#ifdef __cplusplus
};
#endif

CVideoConvert::CVideoConvert()
{
	InitializeCriticalSection(&m_csVideoReadSection);
}

CVideoConvert::~CVideoConvert()
{
	DeleteCriticalSection(&m_csVideoReadSection);
}

int CVideoConvert::StartConvert(std::string strFileName, std::string strFileOut)
{
	int ret = -1;
	do
	{
		ret = OpenFile(strFileName.c_str());
		if (ret != 0)
		{
			break;
		}

		ret = OpenOutPut(strFileOut.c_str());
		if (ret != 0)
		{
			break;
		}

		if (ret < 0)
		{
			break;
		}

		m_bStartConvert = true;
		m_hFileReadThread = CreateThread(NULL, 0, FileReadProc, this, 0, NULL);
		m_hFileWriteThread = CreateThread(NULL, 0, FileWriteProc, this, 0, NULL);
	} while (0);

	return ret;
}

void CVideoConvert::StopConvert()
{
	m_bStartConvert = false;
}

int CVideoConvert::OpenFile(const char *pFile)
{
	int ret = -1;

	do
	{
		if ((ret = avformat_open_input(&m_pFormatCtx_File, pFile, 0, 0)) < 0) {
			printf("Could not open input file.");
			break;
		}
		if ((ret = avformat_find_stream_info(m_pFormatCtx_File, 0)) < 0) {
			printf("Failed to retrieve input stream information");
			break;
		}
		if (m_pFormatCtx_File->nb_streams > 2)
		{
			break;
		}
		for (int i = 0; i < m_pFormatCtx_File->nb_streams; i++)
		{
			if (m_pFormatCtx_File->streams[i]->codecpar->codec_type == AVMEDIA_TYPE_VIDEO)
			{
				m_iVideoIndex = i;
			}
		}

		if (m_iVideoIndex == -1)
		{
			break;
		}

		ret = 0;
	} while (0);


	return ret;
}


int CVideoConvert::OpenOutPut(const char *pFileOut)
{
	int iRet = -1;

	AVStream *pVideoStream = NULL;
	do
	{
		avformat_alloc_output_context2(&m_pFormatCtx_Out, NULL, NULL, pFileOut);
		if (m_iVideoIndex != -1)
		{
			AVCodecID eTargetVideoCodecId = m_pFormatCtx_File->streams[m_iVideoIndex]->codecpar->codec_id;
			AVCodec* pCodecEncode_Video = (AVCodec *)avcodec_find_encoder(eTargetVideoCodecId);
			pVideoStream = avformat_new_stream(m_pFormatCtx_Out, pCodecEncode_Video);
			if (!pVideoStream)
			{
				break;
			}

			avcodec_parameters_copy(pVideoStream->codecpar, m_pFormatCtx_File->streams[m_iVideoIndex]->codecpar);
			pVideoStream->codecpar->codec_tag = 0;
		}

		if (!(m_pFormatCtx_Out->oformat->flags & AVFMT_NOFILE))
		{
			if (avio_open(&m_pFormatCtx_Out->pb, pFileOut, AVIO_FLAG_WRITE) < 0)
			{
				break;
			}
		}

		if (avformat_write_header(m_pFormatCtx_Out, NULL) < 0)
		{
			break;
		}

		iRet = 0;
	} while (0);


	if (iRet != 0)
	{
		if (m_pFormatCtx_Out != NULL)
		{
			avformat_free_context(m_pFormatCtx_Out);
			m_pFormatCtx_Out = NULL;
		}
	}

	return iRet;
}

DWORD WINAPI CVideoConvert::FileReadProc(LPVOID lpParam)
{
	CVideoConvert *pVideoConvert = (CVideoConvert *)lpParam;
	if (pVideoConvert != NULL)
	{
		pVideoConvert->FileRead();
	}
	return 0;
}

void CVideoConvert::FileRead()
{
	AVPacket packet = { 0 };
	int ret = 0;
	int iPicCount = 0;
	const int genpts = m_pFormatCtx_File->flags & AVFMT_FLAG_GENPTS;
	while (m_bStartConvert)
	{
		av_packet_unref(&packet);

		ret = av_read_frame(m_pFormatCtx_File, &packet);
		if (ret == AVERROR(EAGAIN))
		{
			continue;
		}
		else if (ret == AVERROR_EOF)
		{
			break;
		}
		else if (ret < 0)
		{
			break;
		}

		if (packet.stream_index == m_iVideoIndex)
		{
			while (1)
			{
				int iAudioPacketNum = m_vecReadVideo.size();
				if (iAudioPacketNum >= MAX_VIDEOPACKET_NUM)
				{
					Sleep(10);
					continue;
				}
				else
				{
					AVPacket *pAudioPacket = av_packet_clone(&packet);
					if (pAudioPacket != NULL)
					{
						EnterCriticalSection(&m_csVideoReadSection);
						m_vecReadVideo.push_back(pAudioPacket);
						LeaveCriticalSection(&m_csVideoReadSection);
					}
				}
				break;
			}
		}
		
	}
	//FlushVideoDecoder(iPicCount);
}



DWORD WINAPI CVideoConvert::FileWriteProc(LPVOID lpParam)
{
	CVideoConvert *pVideoConvert = (CVideoConvert *)lpParam;
	if (pVideoConvert != NULL)
	{
		pVideoConvert->FileWrite();
	}
	return 0;
}


void CVideoConvert::FileWrite()
{
	int ret = 0;


	AVPacket packet = { 0 };
	int iRealPicCount = 0;

	while (m_bStartConvert)
	{

		int iVideoSize = m_vecReadVideo.size();

		if (iVideoSize > 0)
		{
			AVPacket *pPacketVideoA = NULL;
			EnterCriticalSection(&m_csVideoReadSection);
			if (!m_vecReadVideo.empty())
			{
				pPacketVideoA = m_vecReadVideo.front();
				m_vecReadVideo.pop_front();
			}
			LeaveCriticalSection(&m_csVideoReadSection);

			pPacketVideoA->stream_index = m_iVideoIndex;
			pPacketVideoA->duration = av_rescale_q_rnd(pPacketVideoA->duration, m_pFormatCtx_File->streams[m_iVideoIndex]->time_base, m_pFormatCtx_Out->streams[m_iVideoIndex]->time_base, AVRounding(1));
			if (pPacketVideoA->dts != AV_NOPTS_VALUE)
			{
				pPacketVideoA->dts = av_rescale_q_rnd(pPacketVideoA->dts, m_pFormatCtx_File->streams[m_iVideoIndex]->time_base, m_pFormatCtx_Out->streams[m_iVideoIndex]->time_base, AVRounding(1));
				pPacketVideoA->pts = pPacketVideoA->dts;
			}
			else
			{
				pPacketVideoA->pts = pPacketVideoA->dts = m_iTotalDuration;
			}
			m_iTotalDuration += pPacketVideoA->duration;

			//ret = av_write_frame(m_pFormatCtx_Out, pPacketVideoA);
			ret = av_interleaved_write_frame(m_pFormatCtx_Out, pPacketVideoA);
			av_packet_free(&pPacketVideoA);
			if (ret == AVERROR(EAGAIN))
			{
				continue;
			}
			else if (ret == AVERROR_EOF)
			{
				break;
			}
			iRealPicCount++;
		}
		else
		{
			Sleep(10);
		}
	}
	///FlushEncoder(iRealPicCount);
	av_write_trailer(m_pFormatCtx_Out);
	avio_close(m_pFormatCtx_Out->pb);
}

void CVideoConvert::WaitFinish()
{
	int ret = 0;
	do
	{
		if (NULL == m_hFileReadThread)
		{
			break;
		}
		WaitForSingleObject(m_hFileReadThread, INFINITE);

		CloseHandle(m_hFileReadThread);
		m_hFileReadThread = NULL;

		while (m_vecReadVideo.size() > 0)
		{
			Sleep(1000);
		}
		m_bStartConvert = false;
		WaitForSingleObject(m_hFileWriteThread, INFINITE);
		CloseHandle(m_hFileWriteThread);
		m_hFileWriteThread = NULL;
	} while (0);
}



你可能感兴趣的:(ffmpeg,音视频)