困扰我有一段日子的这个东东今晚终于搞好了。
先声明一下,我用的ffmpeg的版本是ffmpeg-20140227-git-b5005de
版本升级以后,跟以前主要的区别是ffmpeg对音频解码后的帧格式不再是AV_SAMPLE_FMT_S16了,跟视频一样,多了平面格式
枚举类型AVSampleFormat声明于libavutil/samplefmt.h中,如下:
/**
* Audio Sample Formats
*
* @par
* The data described by the sample format is always in native-endian order.
* Sample values can be expressed by native C types, hence the lack of a signed
* 24-bit sample format even though it is a common raw audio data format.
*
* @par
* The floating-point formats are based on full volume being in the range
* [-1.0, 1.0]. Any values outside this range are beyond full volume level.
*
* @par
* The data layout as used in av_samples_fill_arrays() and elsewhere in FFmpeg
* (such as AVFrame in libavcodec) is as follows:
*
* For planar sample formats, each audio channel is in a separate data plane,
* and linesize is the buffer size, in bytes, for a single plane. All data
* planes must be the same size. For packed sample formats, only the first data
* plane is used, and samples for each channel are interleaved. In this case,
* linesize is the buffer size, in bytes, for the 1 plane.
*/
enum AVSampleFormat {
AV_SAMPLE_FMT_NONE = -1,
AV_SAMPLE_FMT_U8, ///< unsigned 8 bits
AV_SAMPLE_FMT_S16, ///< signed 16 bits
AV_SAMPLE_FMT_S32, ///< signed 32 bits
AV_SAMPLE_FMT_FLT, ///< float
AV_SAMPLE_FMT_DBL, ///< double
AV_SAMPLE_FMT_U8P, ///< unsigned 8 bits, planar
AV_SAMPLE_FMT_S16P, ///< signed 16 bits, planar
AV_SAMPLE_FMT_S32P, ///< signed 32 bits, planar
AV_SAMPLE_FMT_FLTP, ///< float, planar
AV_SAMPLE_FMT_DBLP, ///< double, planar
AV_SAMPLE_FMT_NB ///< Number of sample formats. DO NOT USE if linking dynamically
};
#include "stdafx.h"
#include
using namespace std;
#include
#include
#include
#include
extern "C"
{
#include "libavutil/opt.h"
#include "libavcodec/avcodec.h"
#include "libavformat/avformat.h"
#include "libswscale/swscale.h"
#include "libswresample/swresample.h"
#include "SDL.h"
#include "SDL_thread.h"
};
#pragma warning(disable: 4996)
#pragma comment(lib,"avutil.lib")
#pragma comment(lib,"avcodec.lib")
#pragma comment(lib,"avformat.lib")
#pragma comment(lib,"swscale.lib")
#pragma comment(lib,"swresample.lib")
#pragma comment(lib,"sdl2.lib")
#define NB_SAMPLE 1152//这里如果是播放aac要改成1024
//全局变量---------------------
static Uint8 *audio_chunk;
static Uint32 audio_len;
static Uint8 *audio_pos;
int AudioResampling(AVCodecContext * audio_dec_ctx,
AVFrame * pAudioDecodeFrame,
int out_sample_fmt,
int out_channels,
int out_sample_rate)
{
SwrContext * swr_ctx = NULL;
int data_size = 0;
int ret = 0;
int64_t src_ch_layout = audio_dec_ctx->channel_layout;
int64_t dst_ch_layout = AV_CH_LAYOUT_STEREO;
int dst_nb_channels = 0;
int dst_linesize = 0;
int src_nb_samples = 0;
int dst_nb_samples = 0;
int max_dst_nb_samples = 0;
uint8_t **dst_data = NULL;
int resampled_data_size = 0;
swr_ctx = swr_alloc();
if (!swr_ctx)
{
printf("swr_alloc error \n");
return -1;
}
src_ch_layout = (audio_dec_ctx->channels ==
av_get_channel_layout_nb_channels(audio_dec_ctx->channel_layout)) ?
audio_dec_ctx->channel_layout :
av_get_default_channel_layout(audio_dec_ctx->channels);
if (out_channels == 1)
{
dst_ch_layout = AV_CH_LAYOUT_MONO;
printf("dst_ch_layout: AV_CH_LAYOUT_MONO\n");
}
else if (out_channels == 2)
{
dst_ch_layout = AV_CH_LAYOUT_STEREO;
printf("dst_ch_layout: AV_CH_LAYOUT_STEREO\n");
}
else
{
dst_ch_layout = AV_CH_LAYOUT_SURROUND;
printf("dst_ch_layout: AV_CH_LAYOUT_SURROUND\n");
}
if (src_ch_layout <= 0)
{
printf("src_ch_layout error \n");
return -1;
}
src_nb_samples = pAudioDecodeFrame->nb_samples;
if (src_nb_samples <= 0)
{
printf("src_nb_samples error \n");
return -1;
}
av_opt_set_int(swr_ctx, "in_channel_layout", src_ch_layout, 0);
av_opt_set_int(swr_ctx, "in_sample_rate", audio_dec_ctx->sample_rate, 0);
av_opt_set_sample_fmt(swr_ctx, "in_sample_fmt", audio_dec_ctx->sample_fmt, 0);
av_opt_set_int(swr_ctx, "out_channel_layout", dst_ch_layout, 0);
av_opt_set_int(swr_ctx, "out_sample_rate", out_sample_rate, 0);
av_opt_set_sample_fmt(swr_ctx, "out_sample_fmt", (AVSampleFormat)out_sample_fmt, 0);
if ((ret = swr_init(swr_ctx)) < 0) {
printf("Failed to initialize the resampling context\n");
return -1;
}
max_dst_nb_samples = dst_nb_samples = av_rescale_rnd(src_nb_samples,
out_sample_rate, audio_dec_ctx->sample_rate, AV_ROUND_UP);
if (max_dst_nb_samples <= 0)
{
printf("av_rescale_rnd error \n");
return -1;
}
dst_nb_channels = av_get_channel_layout_nb_channels(dst_ch_layout);
ret = av_samples_alloc_array_and_samples(&dst_data, &dst_linesize, dst_nb_channels,
dst_nb_samples, (AVSampleFormat)out_sample_fmt, 0);
if (ret < 0)
{
printf("av_samples_alloc_array_and_samples error \n");
return -1;
}
dst_nb_samples = av_rescale_rnd(swr_get_delay(swr_ctx, audio_dec_ctx->sample_rate) +
src_nb_samples, out_sample_rate, audio_dec_ctx->sample_rate, AV_ROUND_UP);
if (dst_nb_samples <= 0)
{
printf("av_rescale_rnd error \n");
return -1;
}
if (dst_nb_samples > max_dst_nb_samples)
{
av_free(dst_data[0]);
ret = av_samples_alloc(dst_data, &dst_linesize, dst_nb_channels,
dst_nb_samples, (AVSampleFormat)out_sample_fmt, 1);
max_dst_nb_samples = dst_nb_samples;
}
if (swr_ctx)
{
ret = swr_convert(swr_ctx, dst_data, dst_nb_samples,
(const uint8_t **)pAudioDecodeFrame->data, pAudioDecodeFrame->nb_samples);
if (ret < 0)
{
printf("swr_convert error \n");
return -1;
}
resampled_data_size = av_samples_get_buffer_size(&dst_linesize, dst_nb_channels,
ret, (AVSampleFormat)out_sample_fmt, 1);
if (resampled_data_size < 0)
{
printf("av_samples_get_buffer_size error \n");
return -1;
}
}
else
{
printf("swr_ctx null error \n");
return -1;
}
if (!audio_chunk){
audio_chunk = (uint8_t *)av_malloc(resampled_data_size*sizeof(uint8_t));
}
memcpy(audio_chunk, dst_data[0], resampled_data_size);
if (dst_data)
{
av_freep(&dst_data[0]);
}
av_freep(&dst_data);
dst_data = NULL;
if (swr_ctx)
{
swr_free(&swr_ctx);
}
return resampled_data_size;
}
//-----------------
/* The audio function callback takes the following parameters:
stream: A pointer to the audio buffer to be filled
len: The length (in bytes) of the audio buffer (这是固定的4096?)
回调函数
注意:mp3为什么播放不顺畅?
len=4096;audio_len=4608;两个相差512!为了这512,还得再调用一次回调函数。。。
m4a,aac就不存在此问题(都是4096)!
*/
void fill_audio(void *udata, Uint8 *stream, int len){
/* Only play if we have data left */
if (audio_len == 0)
return;
/* Mix as much data as possible */
len = (len>audio_len ? audio_len : len);
SDL_memset(stream, 0, len);// make sure this is silence.
SDL_MixAudio(stream, audio_pos, len, SDL_MIX_MAXVOLUME);
audio_pos += len;
audio_len -= len;
}
//-----------------
int decode_audio(char* no_use)
{
AVFormatContext *pFormatCtx;
int i, audioStream;
AVCodecContext *pCodecCtx;
AVCodec *pCodec;
char url[300] = { 0 };
strcpy(url, no_use);
//Register all available file formats and codecs
av_register_all();
//支持网络流输入
avformat_network_init();
//初始化
pFormatCtx = avformat_alloc_context();
//有参数avdic
//if(avformat_open_input(&pFormatCtx,url,NULL,&avdic)!=0){
if (avformat_open_input(&pFormatCtx, url, NULL, NULL) != 0){
printf("Couldn't open file.\n");
return -1;
}
// Retrieve stream information
if (av_find_stream_info(pFormatCtx)<0)
{
printf("Couldn't find stream information.\n");
return -1;
}
// Dump valid information onto standard error
av_dump_format(pFormatCtx, 0, url, false);
// Find the first audio stream
audioStream = -1;
for (i = 0; i < pFormatCtx->nb_streams; i++)
//原为codec_type==CODEC_TYPE_AUDIO
if (pFormatCtx->streams[i]->codec->codec_type == AVMEDIA_TYPE_AUDIO)
{
audioStream = i;
break;
}
if (audioStream == -1)
{
printf("Didn't find a audio stream.\n");
return -1;
}
// Get a pointer to the codec context for the audio stream
pCodecCtx = pFormatCtx->streams[audioStream]->codec;
// Find the decoder for the audio stream
pCodec = avcodec_find_decoder(pCodecCtx->codec_id);
if (pCodec == NULL)
{
printf("Codec not found.\n");
return -1;
}
// Open codec
if (avcodec_open2(pCodecCtx, pCodec, NULL)<0)
{
printf("Could not open codec.\n");
return -1;
}
/********* For output file ******************/
FILE *pFile;
#ifdef _WAVE_
pFile = fopen("output.wav", "wb");
fseek(pFile, 44, SEEK_SET); //预留文件头的位置
#else
pFile = fopen("output.pcm", "wb");
#endif
/*** Write audio into file ******/
//把结构体改为指针
AVPacket *packet = (AVPacket *)malloc(sizeof(AVPacket));
av_init_packet(packet);
//音频和视频解码更加统一!
//新加
AVFrame *pFrame;
pFrame = av_frame_alloc();
//---------SDL--------------------------------------
//初始化
if (SDL_Init(SDL_INIT_EVERYTHING)) {
printf("Could not initialize SDL - %s\n", SDL_GetError());
exit(1);
}
//结构体,包含PCM数据的相关信息
SDL_AudioSpec wanted_spec;
wanted_spec.freq = pCodecCtx->sample_rate;
wanted_spec.format = AUDIO_S16SYS;
wanted_spec.channels = pCodecCtx->channels;
wanted_spec.silence = 0;
wanted_spec.samples = NB_SAMPLE;
wanted_spec.callback = fill_audio;
wanted_spec.userdata = pCodecCtx;
if (SDL_OpenAudio(&wanted_spec, NULL)<0)//步骤(2)打开音频设备
{
printf("can't open audio.\n");
return 0;
}
//-----------------------------------------------------
printf("比特率 %3d\n", pFormatCtx->bit_rate);
printf("解码器名称 %s\n", pCodecCtx->codec->long_name);
printf("time_base %d \n", pCodecCtx->time_base);
printf("声道数 %d \n", pCodecCtx->channels);
printf("sample per second %d \n", pCodecCtx->sample_rate);
//新版不再需要
// short decompressed_audio_buf[(AVCODEC_MAX_AUDIO_FRAME_SIZE * 3) / 2];
// int decompressed_audio_buf_size;
uint32_t ret, len = 0;
int got_picture;
int index = 0;
while (av_read_frame(pFormatCtx, packet) >= 0)
{
if (packet->stream_index == audioStream)
{
//decompressed_audio_buf_size = (AVCODEC_MAX_AUDIO_FRAME_SIZE * 3) / 2;
//原为avcodec_decode_audio2
//ret = avcodec_decode_audio2( pCodecCtx, decompressed_audio_buf,
//&decompressed_audio_buf_size, packet.data, packet.size );
//改为
ret = avcodec_decode_audio4(pCodecCtx, pFrame,
&got_picture, packet);
if (ret < 0) // if error len = -1
{
printf("Error in decoding audio frame.\n");
exit(0);
}
//这里输出解码后的格式
cout <<"Decoded format:"<< pFrame->format << endl;
if (got_picture > 0)
{
printf("index %3d\n", index);
printf("pts %5d\n", packet->pts);
printf("dts %5d\n", packet->dts);
printf("packet_size %5d\n", packet->size);
audio_len = AudioResampling(pCodecCtx, pFrame, AV_SAMPLE_FMT_S16, 2, 44100);
//直接写入PCM格式数据
fwrite(audio_chunk, 1, audio_len, pFile);
}
#if 1
//---------------------------------------
//printf("begin....\n");
//设置音频数据缓冲,PCM数据
//audio_chunk = (Uint8*)pFrame->data[0];
//设置音频数据长度
//audio_len = pFrame->linesize[0];
//audio_len = 4096;
//播放mp3的时候改为audio_len = 4096
//则会比较流畅,但是声音会变调!MP3一帧长度4608
//使用一次回调函数(4096字节缓冲)播放不完,所以还要使用一次回调函数,导致播放缓慢。。。
//设置初始播放位置
audio_pos = audio_chunk;
//回放音频数据
SDL_PauseAudio(0);
//printf("don't close, audio playing...\n");
while (audio_len>0)//等待直到音频数据播放完毕!
SDL_Delay(1);
//---------------------------------------
#endif
}
av_free_packet(packet);
}
//printf("The length of PCM data is %d bytes.\n", len);
av_free_packet(packet);
SDL_CloseAudio();//关闭音频设备
// Close file
fclose(pFile);
// Close the codec
avcodec_close(pCodecCtx);
// Close the video file
av_close_input_file(pFormatCtx);
return 0;
}
int main(int argc, char* argv[])
{
//char filename[]="nxn.m4a";
//char filename[]="nxn.wma";
char filename[] = "../../resource/test.mp3";
//char filename[]="nwn.flv";
if (decode_audio(filename) == 0)
printf("Decode audio successfully.\n");
return 0;
}
SDL_memset(stream, 0, len);// make sure this is silence.
否则的话播放出来的声音仍然会有噪声或者说变调,估计是SDL_MixAudio这个函数内部处理不当,而且这个函数现在已经是deprecated的了
下面再给出根据另一个教程的代码,我主要是重写了audio_decode_frame函数,这里ffmpeg新版本中用的avcodec_decode_audio4的注释里面写道:
* Some decoders may support multiple frames in a single AVPacket. Such
* decoders would then just decode the first frame and the return value would be
* less than the packet size. In this case, avcodec_decode_audio4 has to be
* called again with an AVPacket containing the remaining data in order to
* decode the second frame, etc... Even if no frames are returned, the packet
* needs to be fed to the decoder with remaining data until it is completely
* consumed or an error occurs.
意思就是ffmpeg里面一个包可能封装了多个帧,谁能给我一个这样的音频文件测试一下?
还有一个问题就是这份代码运行起来音频跟视频一起播放的时候音频不流畅,因为没同步的问题?
这些问题有待后续解决。。。
#include "stdafx.h"
extern "C"{
#include "libavutil/opt.h"
#include "libavcodec/avcodec.h"
#include "libavformat/avformat.h"
#include "libswscale/swscale.h"
#include "libswresample/swresample.h"
}
#include "SDL.h"
#include "SDL_image.h"
#include "SDL_thread.h"
#include
using namespace std;
#pragma warning(disable: 4996)
#pragma comment(lib,"avutil.lib")
#pragma comment(lib,"avcodec.lib")
#pragma comment(lib,"avformat.lib")
#pragma comment(lib,"swscale.lib")
#pragma comment(lib,"swresample.lib")
#pragma comment(lib,"sdl2.lib")
#define SDL_AUDIO_BUFFER_SIZE 1152
#define AVCODEC_MAX_AUDIO_FRAME_SIZE 192000
static Uint8 *audio_chunk;
static Uint32 audio_len;
static Uint8 *audio_pos;
void fill_audio(void *udata, Uint8 *stream, int len){
if (audio_len == 0)
return;
len = (len > audio_len ? audio_len : len);
SDL_MixAudio(stream, audio_pos, len, SDL_MIX_MAXVOLUME);
audio_pos += len;
audio_len -= len;
}
int AudioResampling(AVCodecContext * audio_dec_ctx,
AVFrame * pAudioDecodeFrame,
int out_sample_fmt,
int out_channels,
int out_sample_rate,
uint8_t* out_buf)
{
SwrContext * swr_ctx = NULL;
int data_size = 0;
int ret = 0;
int64_t src_ch_layout = audio_dec_ctx->channel_layout;
int64_t dst_ch_layout = AV_CH_LAYOUT_STEREO;
int dst_nb_channels = 0;
int dst_linesize = 0;
int src_nb_samples = 0;
int dst_nb_samples = 0;
int max_dst_nb_samples = 0;
uint8_t **dst_data = NULL;
int resampled_data_size = 0;
swr_ctx = swr_alloc();
if (!swr_ctx)
{
printf("swr_alloc error \n");
return -1;
}
src_ch_layout = (audio_dec_ctx->channels ==
av_get_channel_layout_nb_channels(audio_dec_ctx->channel_layout)) ?
audio_dec_ctx->channel_layout :
av_get_default_channel_layout(audio_dec_ctx->channels);
if (out_channels == 1)
{
dst_ch_layout = AV_CH_LAYOUT_MONO;
//printf("dst_ch_layout: AV_CH_LAYOUT_MONO\n");
}
else if (out_channels == 2)
{
dst_ch_layout = AV_CH_LAYOUT_STEREO;
//printf("dst_ch_layout: AV_CH_LAYOUT_STEREO\n");
}
else
{
dst_ch_layout = AV_CH_LAYOUT_SURROUND;
//printf("dst_ch_layout: AV_CH_LAYOUT_SURROUND\n");
}
if (src_ch_layout <= 0)
{
printf("src_ch_layout error \n");
return -1;
}
src_nb_samples = pAudioDecodeFrame->nb_samples;
if (src_nb_samples <= 0)
{
printf("src_nb_samples error \n");
return -1;
}
av_opt_set_int(swr_ctx, "in_channel_layout", src_ch_layout, 0);
av_opt_set_int(swr_ctx, "in_sample_rate", audio_dec_ctx->sample_rate, 0);
av_opt_set_sample_fmt(swr_ctx, "in_sample_fmt", audio_dec_ctx->sample_fmt, 0);
av_opt_set_int(swr_ctx, "out_channel_layout", dst_ch_layout, 0);
av_opt_set_int(swr_ctx, "out_sample_rate", out_sample_rate, 0);
av_opt_set_sample_fmt(swr_ctx, "out_sample_fmt", (AVSampleFormat)out_sample_fmt, 0);
if ((ret = swr_init(swr_ctx)) < 0) {
printf("Failed to initialize the resampling context\n");
return -1;
}
max_dst_nb_samples = dst_nb_samples = av_rescale_rnd(src_nb_samples,
out_sample_rate, audio_dec_ctx->sample_rate, AV_ROUND_UP);
if (max_dst_nb_samples <= 0)
{
printf("av_rescale_rnd error \n");
return -1;
}
dst_nb_channels = av_get_channel_layout_nb_channels(dst_ch_layout);
ret = av_samples_alloc_array_and_samples(&dst_data, &dst_linesize, dst_nb_channels,
dst_nb_samples, (AVSampleFormat)out_sample_fmt, 0);
if (ret < 0)
{
printf("av_samples_alloc_array_and_samples error \n");
return -1;
}
dst_nb_samples = av_rescale_rnd(swr_get_delay(swr_ctx, audio_dec_ctx->sample_rate) +
src_nb_samples, out_sample_rate, audio_dec_ctx->sample_rate, AV_ROUND_UP);
if (dst_nb_samples <= 0)
{
printf("av_rescale_rnd error \n");
return -1;
}
if (dst_nb_samples > max_dst_nb_samples)
{
av_free(dst_data[0]);
ret = av_samples_alloc(dst_data, &dst_linesize, dst_nb_channels,
dst_nb_samples, (AVSampleFormat)out_sample_fmt, 1);
max_dst_nb_samples = dst_nb_samples;
}
if (swr_ctx)
{
ret = swr_convert(swr_ctx, dst_data, dst_nb_samples,
(const uint8_t **)pAudioDecodeFrame->data, pAudioDecodeFrame->nb_samples);
if (ret < 0)
{
printf("swr_convert error \n");
return -1;
}
resampled_data_size = av_samples_get_buffer_size(&dst_linesize, dst_nb_channels,
ret, (AVSampleFormat)out_sample_fmt, 1);
if (resampled_data_size < 0)
{
printf("av_samples_get_buffer_size error \n");
return -1;
}
}
else
{
printf("swr_ctx null error \n");
return -1;
}
memcpy(out_buf, dst_data[0], resampled_data_size);
if (dst_data)
{
av_freep(&dst_data[0]);
}
av_freep(&dst_data);
dst_data = NULL;
if (swr_ctx)
{
swr_free(&swr_ctx);
}
return resampled_data_size;
}
//创建一个全局的结构体变量以便于我们从文件中得到的声音包有地方存
//放同时也保证SDL中的声音回调函数audio_callback 能从这个地方得到声音数据
typedef struct PacketQueue{
AVPacketList *first_pkt, *last_pkt;
int nb_packets;
int size;
SDL_mutex *mutex;//因为SDL 是在一个独立的线程中来进行音频处理的。如果我们没有正确的锁定这个队列,我们有 可能把数据搞乱。
SDL_cond *cond;
}PacketQueue;
PacketQueue audioq;
void packet_queue_init(PacketQueue *pq){
memset(pq, 0, sizeof(PacketQueue));
pq->mutex = SDL_CreateMutex();
pq->cond = SDL_CreateCond();
}
int packet_queue_put(PacketQueue *q, AVPacket *pkt){
AVPacketList *pkt1;
if (av_dup_packet(pkt) < 0){
printf("error");
return -1;
}
pkt1 = (AVPacketList*)av_malloc(sizeof(AVPacketList));
if (!pkt1){
printf("error");
return -1;
}
pkt1->pkt = *pkt;
pkt1->next = NULL;
//函数SDL_LockMutex()锁定队列的互斥量以便于我们向队列中添加东西,然后函
//数SDL_CondSignal()通过我们的条件变量为一个接 收函数(如果它在等待)发
//出一个信号来告诉它现在已经有数据了,接着就会解锁互斥量并让队列可以自由
//访问。
SDL_LockMutex(q->mutex);
if (!q->last_pkt)//队列为空
q->first_pkt = pkt1;
else//队列不为空
q->last_pkt->next = pkt1;
q->last_pkt = pkt1;
q->nb_packets++;
q->size += pkt1->pkt.size;
SDL_CondSignal(q->cond);
SDL_UnlockMutex(q->mutex);
return 0;
}
int quit = 0;
int decode_interrupt_cb(void){
return quit;
}
static int packet_queue_get(PacketQueue *q, AVPacket *pkt, int block){
AVPacketList *pkt1;
int ret;
SDL_LockMutex(q->mutex);
for (;;){
if (quit){
ret = -1;
break;
}
pkt1 = q->first_pkt;
if (pkt1){
q->first_pkt = pkt1->next;
if (!q->first_pkt)
q->last_pkt = NULL;
q->nb_packets--;
q->size -= pkt1->pkt.size;
*pkt = pkt1->pkt;
av_free(pkt1);
ret = 1;
break;
}
else if (!block){
ret = 0;
break;
}
else{
SDL_CondWait(q->cond, q->mutex);
}
}
SDL_UnlockMutex(q->mutex);
return ret;
}
int audio_decode_frame(AVCodecContext *aCodecCtx, uint8_t *audio_buf, int buf_size){
static AVPacket pkt;
static uint8_t *audio_pkt_data = NULL;
static int audio_pkt_size = 0;
int len1, data_size, ret = 0;
static AVFrame *pFrame;
pFrame = av_frame_alloc();
/*if (packet_queue_get(&audioq, &pkt, 1) < 0){//从这里开始,取得main线程放入队列的包
printf("error, can't get packet from the queue");
return -1;
}
len1 = avcodec_decode_audio4(aCodecCtx, pFrame, &ret, &pkt);
if (len1 < 0)
return -1;
return AudioResampling(aCodecCtx, pFrame, AV_SAMPLE_FMT_S16, 2, 44100, audio_buf);*/
for (;;){
while (audio_pkt_size > 0){
data_size = buf_size;
len1 = avcodec_decode_audio4(aCodecCtx, pFrame, &ret, &pkt);
//len1 = avcodec_decode_audio3(aCodecCtx, (int16_t *)audio_buf,
// &data_size, &pkt);
if (len1 < 0){//if error, skip frame
printf("error\n");
audio_pkt_size = 0;
break;
}
data_size = AudioResampling(aCodecCtx, pFrame, AV_SAMPLE_FMT_S16, 2, 44100, audio_buf);
audio_pkt_data += len1;
audio_pkt_size -= len1;
if (data_size <= 0)//No data yet, get more frames
continue;
return data_size;
}
if (pkt.data)
av_free_packet(&pkt);
if (quit)
return -1;
if (packet_queue_get(&audioq, &pkt, 1) < 0){//从这里开始,取得main线程放入队列的包
printf("error, can't get packet from the queue");
return -1;
}
audio_pkt_data = pkt.data;
audio_pkt_size = pkt.size;
}
}
//声音回调函数
//userdata是输入,stream是输出,len是输入,len的值一般为4096(调试中发现的),
//audio_callback函数的功能是调用audio_decode_frame函数,把解码后数据块audio_buf追加在stream的后面,
//通过SDL库对audio_callback的不断调用,不断解码数据,然后放到stream的末尾,
//SDL库认为stream中数据够播放一帧音频了,就播放它,
//第三个参数len是向stream中写数据的内存分配尺度,是分配给audio_callback函数写入缓存大小。
void audio_callback(void *userdata, Uint8 *stream, int len){
//SDL_memset(stream, 0, len);
AVCodecContext *aCodecCtx = (AVCodecContext*)userdata;
int len1, audio_size;
//audio_buf 的大小为 1.5 倍的声音帧的大 小以便于有一个比较好的缓冲
static uint8_t audio_buf[(AVCODEC_MAX_AUDIO_FRAME_SIZE * 3) / 2];
static unsigned int audio_buf_size = 0;
static unsigned int audio_buf_index = 0;
while (len > 0){
if (audio_buf_index >= audio_buf_size){//already send all our data, get more
audio_size = audio_decode_frame(aCodecCtx, audio_buf, sizeof(audio_buf));
if (audio_size < 0){//error, output silence
printf("error, output silence\n");
audio_buf_size = SDL_AUDIO_BUFFER_SIZE;
memset(audio_buf, 0, audio_buf_size);
}
else
audio_buf_size = audio_size;
audio_buf_index = 0;
}
len1 = audio_buf_size - audio_buf_index;
if (len1>len){
len1 = len;
}
memcpy(stream, (uint8_t *)audio_buf + audio_buf_index, len1);
len -= len1;
stream += len1;
audio_buf_index += len1;
}
}
int _tmain(int argc, char *agrv[]){
av_register_all(); //注册了所有的文件格式和编解码的库,它们将被自动的使用在被打开的合适格式的文件上
AVFormatContext *pFormatCtx;
pFormatCtx = avformat_alloc_context();
char filepath[] = "../../resource/test.mp3";
//Open an input stream and read the header
if (avformat_open_input(&pFormatCtx, filepath, NULL, NULL) != 0){
printf("Can't open the file\n");
return -1;
}
//Retrieve stream information
if (avformat_find_stream_info(pFormatCtx, NULL) < 0){
printf("Couldn't find stream information.\n");
return -1;
}
//output file information
cout << "文件信息----------------------------------" << endl;
av_dump_format(pFormatCtx, 0, filepath, 0);
cout << "--------------------------------------------" << endl;
int i, videoIndex, audioIndex;
//Find the first video stream
videoIndex = -1;
audioIndex = -1;
for (i = 0; i < pFormatCtx->nb_streams; i++){//视音频流的个数
if (pFormatCtx->streams[i]->codec->codec_type == AVMEDIA_TYPE_VIDEO
&& videoIndex < 0){
videoIndex = i;
}
if (pFormatCtx->streams[i]->codec->codec_type == AVMEDIA_TYPE_AUDIO
&& audioIndex < 0)
audioIndex = i;
}
if (videoIndex == -1)
return -1;
if (audioIndex == -1)
return -1;
AVCodecContext *pCodecCtx, *paCodecCtx;
AVCodec *pCodec, *paCodec;
//Get a pointer to the codec context for the video stream
//流中关于编解码器的信息就是被我们叫做"codec context"(编解码器上下文)
//的东西。这里面包含了流中所使用的关于编解码器的所有信
pCodecCtx = pFormatCtx->streams[videoIndex]->codec;
paCodecCtx = pFormatCtx->streams[audioIndex]->codec;
//Find the decoder for the video stream
pCodec = avcodec_find_decoder(pCodecCtx->codec_id);
paCodec = avcodec_find_decoder(paCodecCtx->codec_id);
if (pCodec == NULL || paCodecCtx == NULL){
printf("Unsupported codec!\n");
return -1;
}
//Open codec
if (avcodec_open2(pCodecCtx, pCodec, NULL) < 0){
printf("Could not open video codec.\n");
return -1;
}
if (avcodec_open2(paCodecCtx, paCodec, NULL) < 0){
printf("Could not open audio codec.\n");
return -1;
}
//--------------------------------------------------------//
printf("比特率 %3d\n", pFormatCtx->bit_rate);
printf("解码器名称 %s\n", paCodecCtx->codec->long_name);
printf("time_base %d \n", paCodecCtx->time_base);
printf("声道数 %d \n", paCodecCtx->channels);
printf("sample per second %d \n", paCodecCtx->sample_rate);
//--------------------------------------------------------//
//allocate video frame and set its fileds to default value
AVFrame *pFrame, *pFrameYUV;
pFrame = av_frame_alloc();
pFrameYUV = av_frame_alloc();
//即使我们申请了一帧的内存,当转换的时候,我们仍然需要一个地方来放置原始
//的数据。我们使用avpicture_get_size 来获得我们需要的大小, 然后手工申请
//内存空间:
uint8_t *out_buffer;
int numBytes;
numBytes = avpicture_get_size(PIX_FMT_YUV420P, pCodecCtx->width, pCodecCtx->height);
//av_malloc 是ffmpeg 的malloc,用来实现一个简单的malloc 的包装,这样来保
//证内存地址是对齐的(4 字节对齐或者2 字节对齐)。它并不能保 护你不被内
//存泄漏,重复释放或者其它malloc 的问题所困扰。
out_buffer = (uint8_t *)av_malloc(numBytes*sizeof(uint8_t));
//Assign appropriate parts of buffer to image planes in pFrameYUV
//Note that pFrameYUV is an AVFrame, but AVFrame is a superset of AVPicture
avpicture_fill((AVPicture*)pFrameYUV, out_buffer, PIX_FMT_YUV420P, pCodecCtx->width, pCodecCtx->height);
//----------------SDL--------------------------------------//
if (SDL_Init(SDL_INIT_VIDEO | SDL_INIT_AUDIO | SDL_INIT_TIMER)){
printf("Could not initialize SDL -%s\n", SDL_GetError());
exit(1);
}
//先设置声音的选项:采样率,声音通道数和其它的参 数,然后我们
//设置一个回调函数和一些用户数据userdata。当开始播放音频的时候,SDL 将不
//断地调用这个回调函数并且要求它来向声音缓冲填入一个特定的数量的字节。
//当我们把这些信息放到SDL_AudioSpec 结构体中后,我们调用函数
//SDL_OpenAudio()就会打开声音设备并且给我们送 回另外一个AudioSpec 结构
//体。这个结构体是我们实际上用到的--因为我们不能保证得到我们所要求的。
SDL_AudioSpec wanted_spec;
wanted_spec.freq = paCodecCtx->sample_rate;
wanted_spec.format = AUDIO_S16SYS;
wanted_spec.channels = paCodecCtx->channels; //声音的通道数
wanted_spec.silence = 0; //用来表示静音的值
wanted_spec.samples = SDL_AUDIO_BUFFER_SIZE; //声音缓冲区的大小
wanted_spec.callback = audio_callback;
wanted_spec.userdata = paCodecCtx;
if (SDL_OpenAudio(&wanted_spec, NULL) < 0){
printf("SDL_OpenAudio error: %s\n", SDL_GetError());
return -1;
}
packet_queue_init(&audioq);
SDL_PauseAudio(0);
SDL_Window *window = nullptr;
window = SDL_CreateWindow("MyPlayer", SDL_WINDOWPOS_CENTERED, SDL_WINDOWPOS_CENTERED,
pCodecCtx->width, pCodecCtx->height, SDL_WINDOW_SHOWN);
if (!window){
cout << SDL_GetError() << endl;
return 1;
}
SDL_Renderer *ren = nullptr;
ren = SDL_CreateRenderer(window, -1, SDL_RENDERER_ACCELERATED | SDL_RENDERER_PRESENTVSYNC);
if (ren == nullptr){
cout << SDL_GetError() << endl;
return -1;
}
SDL_Texture *texture = nullptr;
texture = SDL_CreateTexture(ren, SDL_PIXELFORMAT_YV12,
SDL_TEXTUREACCESS_STREAMING, pCodecCtx->width, pCodecCtx->height);
SDL_Rect rect;
rect.x = 0, rect.y = 0;
rect.w = pCodecCtx->width;
rect.h = pCodecCtx->height;
//*************************************************************//
//通过读取包来读取整个视频流,然后把它解码成帧,最后转换格式并且保存
int frameFinished;
//int psize = pCodecCtx->width * pCodecCtx->height;
AVPacket packet;
av_new_packet(&packet, numBytes);
i = 0;
int ret;
static struct SwsContext *img_convert_ctx;
img_convert_ctx = sws_getContext(pCodecCtx->width, pCodecCtx->height,
pCodecCtx->pix_fmt, pCodecCtx->width, pCodecCtx->height, PIX_FMT_YUV420P,
SWS_BICUBIC, NULL, NULL, NULL);
//Read the next frame of a stream
while (av_read_frame(pFormatCtx, &packet) >= 0){
//Is this a packet from the video stream?
if (packet.stream_index == videoIndex){
//decode video frame of size packet.size from packet.data into picture
ret = avcodec_decode_video2(pCodecCtx, pFrame, &frameFinished, &packet);
//Did we get a video frame?
if (ret >= 0){
//Convert the image from its native format to YUV
if (frameFinished){
sws_scale(img_convert_ctx, (const uint8_t* const*)pFrame->data,
pFrame->linesize, 0, pCodecCtx->height, pFrameYUV->data, pFrameYUV->linesize);
SDL_UpdateYUVTexture(texture, &rect, pFrameYUV->data[0], pFrameYUV->linesize[0],
pFrameYUV->data[1], pFrameYUV->linesize[1], pFrameYUV->data[2], pFrameYUV->linesize[2]);
SDL_RenderClear(ren);
SDL_RenderCopy(ren, texture, &rect, &rect);
SDL_RenderPresent(ren);
}
SDL_Delay(50);
}
else{
av_free_packet(&packet);
cout << "decode error" << endl;
return -1;
}
}
else if (packet.stream_index == audioIndex){
//packet_queue_put(&audioq, &packet);
/*ret = avcodec_decode_audio4(paCodecCtx, pFrame, &frameFinished, &packet);
cout << pFrame->format << endl;
if (ret < 0){
printf("Error in decoding audio frame\n");
exit(0);
}
if (frameFinished){
printf("pts %5d\n", packet.pts);
printf("dts %5d\n", packet.dts);
printf("packet_size %5d\n", packet.size);
}
audio_chunk = (Uint8*)pFrame->data[0];
audio_len = pFrame->linesize[0];
audio_pos = audio_chunk;
//SDL_PauseAudio(0);
while (audio_len>0)
SDL_Delay(1);*/
packet_queue_put(&audioq, &packet);
}
}
SDL_Event event;
while (true){
SDL_PollEvent(&event);
switch (event.type){
case SDL_QUIT:
SDL_Quit();
exit(0);
break;
case SDL_KEYDOWN:
default:
break;
}
}
SDL_DestroyTexture(texture);
av_frame_free(&pFrame);
av_frame_free(&pFrameYUV);
avcodec_close(pCodecCtx);
avformat_close_input(&pFormatCtx);
return 0;
}