Work on aac encoding

This commit is contained in:
Isaac Connor 2016-09-16 15:54:13 -04:00
parent d4645cd94a
commit 61df6e9d75
6 changed files with 268 additions and 80 deletions

View File

@ -412,13 +412,13 @@ int hacked_up_context2_for_older_ffmpeg(AVFormatContext **avctx, AVOutputFormat
static void zm_log_fps(double d, const char *postfix) { static void zm_log_fps(double d, const char *postfix) {
uint64_t v = lrintf(d * 100); uint64_t v = lrintf(d * 100);
if (!v) { if (!v) {
Debug(3, "%1.4f %s", d, postfix); Debug(1, "%1.4f %s", d, postfix);
} else if (v % 100) { } else if (v % 100) {
Debug(3, "%3.2f %s", d, postfix); Debug(1, "%3.2f %s", d, postfix);
} else if (v % (100 * 1000)) { } else if (v % (100 * 1000)) {
Debug(3, "%1.0f %s", d, postfix); Debug(1, "%1.0f %s", d, postfix);
} else } else
Debug(3, "%1.0fk %s", d / 1000, postfix); Debug(1, "%1.0fk %s", d / 1000, postfix);
} }
/* "user interface" functions */ /* "user interface" functions */
@ -429,17 +429,17 @@ void zm_dump_stream_format(AVFormatContext *ic, int i, int index, int is_output)
AVDictionaryEntry *lang = av_dict_get(st->metadata, "language", NULL, 0); AVDictionaryEntry *lang = av_dict_get(st->metadata, "language", NULL, 0);
avcodec_string(buf, sizeof(buf), st->codec, is_output); avcodec_string(buf, sizeof(buf), st->codec, is_output);
Debug(3, " Stream #%d:%d", index, i); Debug(1, " Stream #%d:%d", index, i);
/* the pid is an important information, so we display it */ /* the pid is an important information, so we display it */
/* XXX: add a generic system */ /* XXX: add a generic system */
if (flags & AVFMT_SHOW_IDS) if (flags & AVFMT_SHOW_IDS)
Debug(3, "[0x%x]", st->id); Debug(1, "[0x%x]", st->id);
if (lang) if (lang)
Debug(3, "(%s)", lang->value); Debug(1, "(%s)", lang->value);
av_log(NULL, AV_LOG_DEBUG, ", %d, %d/%d", st->codec_info_nb_frames, av_log(NULL, AV_LOG_DEBUG, ", %d, %d/%d", st->codec_info_nb_frames,
st->time_base.num, st->time_base.den); st->time_base.num, st->time_base.den);
Debug(3, ": %s", buf); Debug(1, ": %s", buf);
if (st->sample_aspect_ratio.num && // default if (st->sample_aspect_ratio.num && // default
av_cmp_q(st->sample_aspect_ratio, st->codec->sample_aspect_ratio)) { av_cmp_q(st->sample_aspect_ratio, st->codec->sample_aspect_ratio)) {
@ -448,7 +448,7 @@ void zm_dump_stream_format(AVFormatContext *ic, int i, int index, int is_output)
st->codec->width * (int64_t)st->sample_aspect_ratio.num, st->codec->width * (int64_t)st->sample_aspect_ratio.num,
st->codec->height * (int64_t)st->sample_aspect_ratio.den, st->codec->height * (int64_t)st->sample_aspect_ratio.den,
1024 * 1024); 1024 * 1024);
Debug(3, ", SAR %d:%d DAR %d:%d", Debug(1, ", SAR %d:%d DAR %d:%d",
st->sample_aspect_ratio.num, st->sample_aspect_ratio.den, st->sample_aspect_ratio.num, st->sample_aspect_ratio.den,
display_aspect_ratio.num, display_aspect_ratio.den); display_aspect_ratio.num, display_aspect_ratio.den);
} }
@ -470,28 +470,40 @@ void zm_dump_stream_format(AVFormatContext *ic, int i, int index, int is_output)
} }
if (st->disposition & AV_DISPOSITION_DEFAULT) if (st->disposition & AV_DISPOSITION_DEFAULT)
Debug(3, " (default)"); Debug(1, " (default)");
if (st->disposition & AV_DISPOSITION_DUB) if (st->disposition & AV_DISPOSITION_DUB)
Debug(3, " (dub)"); Debug(1, " (dub)");
if (st->disposition & AV_DISPOSITION_ORIGINAL) if (st->disposition & AV_DISPOSITION_ORIGINAL)
Debug(3, " (original)"); Debug(1, " (original)");
if (st->disposition & AV_DISPOSITION_COMMENT) if (st->disposition & AV_DISPOSITION_COMMENT)
Debug(3, " (comment)"); Debug(1, " (comment)");
if (st->disposition & AV_DISPOSITION_LYRICS) if (st->disposition & AV_DISPOSITION_LYRICS)
Debug(3, " (lyrics)"); Debug(1, " (lyrics)");
if (st->disposition & AV_DISPOSITION_KARAOKE) if (st->disposition & AV_DISPOSITION_KARAOKE)
Debug(3, " (karaoke)"); Debug(1, " (karaoke)");
if (st->disposition & AV_DISPOSITION_FORCED) if (st->disposition & AV_DISPOSITION_FORCED)
Debug(3, " (forced)"); Debug(1, " (forced)");
if (st->disposition & AV_DISPOSITION_HEARING_IMPAIRED) if (st->disposition & AV_DISPOSITION_HEARING_IMPAIRED)
Debug(3, " (hearing impaired)"); Debug(1, " (hearing impaired)");
if (st->disposition & AV_DISPOSITION_VISUAL_IMPAIRED) if (st->disposition & AV_DISPOSITION_VISUAL_IMPAIRED)
Debug(3, " (visual impaired)"); Debug(1, " (visual impaired)");
if (st->disposition & AV_DISPOSITION_CLEAN_EFFECTS) if (st->disposition & AV_DISPOSITION_CLEAN_EFFECTS)
Debug(3, " (clean effects)"); Debug(1, " (clean effects)");
Debug(3, "\n"); Debug(1, "\n");
//dump_metadata(NULL, st->metadata, " "); //dump_metadata(NULL, st->metadata, " ");
//dump_sidedata(NULL, st, " "); //dump_sidedata(NULL, st, " ");
} }
int check_sample_fmt(AVCodec *codec, enum AVSampleFormat sample_fmt) {
const enum AVSampleFormat *p = codec->sample_fmts;
while (*p != AV_SAMPLE_FMT_NONE) {
if (*p == sample_fmt)
return 1;
else Debug(2, "Not %s", av_get_sample_fmt_name( *p ) );
p++;
}
return 0;
}

View File

@ -334,4 +334,6 @@ void zm_dump_stream_format(AVFormatContext *ic, int i, int index, int is_output)
#define zm_avcodec_decode_video(context, rawFrame, frameComplete, packet ) avcodec_decode_video( context, rawFrame, frameComplete, packet->data, packet->size) #define zm_avcodec_decode_video(context, rawFrame, frameComplete, packet ) avcodec_decode_video( context, rawFrame, frameComplete, packet->data, packet->size)
#endif #endif
int check_sample_fmt(AVCodec *codec, enum AVSampleFormat sample_fmt);
#endif // ZM_FFMPEG_H #endif // ZM_FFMPEG_H

View File

@ -290,7 +290,6 @@ int FfmpegCamera::OpenFfmpeg() {
Debug ( 1, "Opened input" ); Debug ( 1, "Opened input" );
Info( "Stream open %s", mPath.c_str() ); Info( "Stream open %s", mPath.c_str() );
startTime=av_gettime();//FIXME here or after find_Stream_info
//FIXME can speed up initial analysis but need sensible parameters... //FIXME can speed up initial analysis but need sensible parameters...
//mFormatContext->probesize = 32; //mFormatContext->probesize = 32;
@ -356,20 +355,6 @@ int FfmpegCamera::OpenFfmpeg() {
} else { } else {
Debug(1, "Video Found decoder"); Debug(1, "Video Found decoder");
zm_dump_stream_format(mFormatContext, mVideoStreamId, 0, 0); zm_dump_stream_format(mFormatContext, mVideoStreamId, 0, 0);
}
if (mAudioStreamId >= 0) {
mAudioCodecContext = mFormatContext->streams[mAudioStreamId]->codec;
if ((mAudioCodec = avcodec_find_decoder(mAudioCodecContext->codec_id)) == NULL) {
Debug(1, "Can't find codec for audio stream from %s", mPath.c_str());
} else {
Debug(1, "Audio Found decoder");
zm_dump_stream_format(mFormatContext, mAudioStreamId, 0, 0);
}
}
//
// Open the codec // Open the codec
#if !LIBAVFORMAT_VERSION_CHECK(53, 8, 0, 8, 0) #if !LIBAVFORMAT_VERSION_CHECK(53, 8, 0, 8, 0)
Debug ( 1, "Calling avcodec_open" ); Debug ( 1, "Calling avcodec_open" );
@ -379,6 +364,29 @@ int FfmpegCamera::OpenFfmpeg() {
if (avcodec_open2(mVideoCodecContext, mVideoCodec, 0) < 0) if (avcodec_open2(mVideoCodecContext, mVideoCodec, 0) < 0)
#endif #endif
Fatal( "Unable to open codec for video stream from %s", mPath.c_str() ); Fatal( "Unable to open codec for video stream from %s", mPath.c_str() );
}
if (mAudioStreamId >= 0) {
mAudioCodecContext = mFormatContext->streams[mAudioStreamId]->codec;
if ((mAudioCodec = avcodec_find_decoder(mAudioCodecContext->codec_id)) == NULL) {
Debug(1, "Can't find codec for audio stream from %s", mPath.c_str());
} else {
Debug(1, "Audio Found decoder");
zm_dump_stream_format(mFormatContext, mAudioStreamId, 0, 0);
// Open the codec
#if !LIBAVFORMAT_VERSION_CHECK(53, 8, 0, 8, 0)
Debug ( 1, "Calling avcodec_open" );
if (avcodec_open(mAudioCodecContext, mAudioCodec) < 0)
#else
Debug ( 1, "Calling avcodec_open2" );
if (avcodec_open2(mAudioCodecContext, mAudioCodec, 0) < 0)
#endif
Fatal( "Unable to open codec for video stream from %s", mPath.c_str() );
}
}
//
Debug ( 1, "Opened codec" ); Debug ( 1, "Opened codec" );
@ -556,7 +564,7 @@ int FfmpegCamera::CaptureAndRecord( Image &image, bool recording, char* event_fi
while ( !frameComplete ) { while ( !frameComplete ) {
// We are now allocating dynamically because we need to queue these and may go out of scope. // We are now allocating dynamically because we need to queue these and may go out of scope.
AVPacket *packet = (AVPacket *)av_malloc(sizeof(AVPacket)); AVPacket *packet = (AVPacket *)av_malloc(sizeof(AVPacket));
av_init_packet( packet); av_init_packet( packet );
Debug(5, "Before av_read_frame"); Debug(5, "Before av_read_frame");
ret = av_read_frame( mFormatContext, packet ); ret = av_read_frame( mFormatContext, packet );
Debug(5, "After av_read_frame (%d)", ret ); Debug(5, "After av_read_frame (%d)", ret );
@ -593,6 +601,7 @@ Debug(5, "After av_read_frame (%d)", ret );
if ( ! videoStore ) { if ( ! videoStore ) {
//Instantiate the video storage module //Instantiate the video storage module
startTime=av_gettime();//FIXME here or after find_Stream_info
if (record_audio) { if (record_audio) {
if (mAudioStreamId == -1) { if (mAudioStreamId == -1) {
@ -627,10 +636,11 @@ Debug(5, "After av_read_frame (%d)", ret );
while ( ( queued_packet = packetqueue.popPacket() ) ) { while ( ( queued_packet = packetqueue.popPacket() ) ) {
packet_count += 1; packet_count += 1;
//Write the packet to our video store //Write the packet to our video store
Debug(2, "Writing queued packet stream: %d KEY %d", queued_packet->stream_index, packet->flags & AV_PKT_FLAG_KEY );
if ( queued_packet->stream_index == mVideoStreamId ) { if ( queued_packet->stream_index == mVideoStreamId ) {
ret = videoStore->writeVideoFramePacket( queued_packet, mFormatContext->streams[mVideoStreamId]); ret = videoStore->writeVideoFramePacket( queued_packet, mFormatContext->streams[mVideoStreamId]);
} else if ( queued_packet->stream_index == mAudioStreamId ) { } else if ( queued_packet->stream_index == mAudioStreamId ) {
//ret = videoStore->writeAudioFramePacket(&queued_packet, mFormatContext->streams[mAudioStreamId]); ret = videoStore->writeAudioFramePacket( queued_packet, mFormatContext->streams[mAudioStreamId]);
} else { } else {
Warning("Unknown stream id in queued packet (%d)", queued_packet->stream_index ); Warning("Unknown stream id in queued packet (%d)", queued_packet->stream_index );
ret = -1; ret = -1;
@ -643,12 +653,6 @@ Debug(5, "After av_read_frame (%d)", ret );
Debug(2, "Wrote %d queued packets", packet_count ); Debug(2, "Wrote %d queued packets", packet_count );
} // end if ! wasRecording } // end if ! wasRecording
//Write the packet to our video store
int ret = videoStore->writeVideoFramePacket( packet, mFormatContext->streams[mVideoStreamId] );
if ( ret < 0 ) { //Less than zero and we skipped a frame
zm_av_unref_packet( packet );
return 0;
}
} else { } else {
// Not recording // Not recording
if ( videoStore ) { if ( videoStore ) {
@ -656,15 +660,23 @@ Debug(5, "After av_read_frame (%d)", ret );
delete videoStore; delete videoStore;
videoStore = NULL; videoStore = NULL;
} }
//Buffer video packets, since we are not recording
if ( (packet->stream_index == mVideoStreamId) && ( packet->flags & AV_PKT_FLAG_KEY ) ) {
packetqueue.clearQueue();
}
packetqueue.queuePacket(packet);
} // end if } // end if
//Buffer video packets
if ( packet->flags & AV_PKT_FLAG_KEY ) {
packetqueue.clearQueue();
}
packetqueue.queuePacket(packet);
if ( packet->stream_index == mVideoStreamId ) { if ( packet->stream_index == mVideoStreamId ) {
if ( videoStore ) {
//Write the packet to our video store
int ret = videoStore->writeVideoFramePacket( packet, mFormatContext->streams[mVideoStreamId] );
if ( ret < 0 ) { //Less than zero and we skipped a frame
zm_av_unref_packet( packet );
return 0;
}
}
ret = zm_avcodec_decode_video( mVideoCodecContext, mRawFrame, &frameComplete, packet ); ret = zm_avcodec_decode_video( mVideoCodecContext, mRawFrame, &frameComplete, packet );
if ( ret < 0 ) { if ( ret < 0 ) {
av_strerror( ret, errbuf, AV_ERROR_MAX_STRING_SIZE ); av_strerror( ret, errbuf, AV_ERROR_MAX_STRING_SIZE );

View File

@ -20,9 +20,9 @@
#ifndef ZM_PACKETQUEUE_H #ifndef ZM_PACKETQUEUE_H
#define ZM_PACKETQUEUE_H #define ZM_PACKETQUEUE_H
#include <boost/interprocess/managed_shared_memory.hpp> //#include <boost/interprocess/managed_shared_memory.hpp>
#include <boost/interprocess/containers/map.hpp> //#include <boost/interprocess/containers/map.hpp>
#include <boost/interprocess/allocators/allocator.hpp> //#include <boost/interprocess/allocators/allocator.hpp>
#include <queue> #include <queue>
extern "C" { extern "C" {

View File

@ -49,8 +49,9 @@ VideoStore::VideoStore(const char *filename_in, const char *format_in,
Info("Opening video storage stream %s format: %s\n", filename, format); Info("Opening video storage stream %s format: %s\n", filename, format);
int ret; int ret;
//Init everything we need static char error_buffer[255];
av_register_all(); //Init everything we need, shouldn't have to do this, ffmpeg_camera or something else will call it.
//av_register_all();
ret = avformat_alloc_output_context2(&oc, NULL, NULL, filename); ret = avformat_alloc_output_context2(&oc, NULL, NULL, filename);
if ( ret < 0 ) { if ( ret < 0 ) {
@ -153,12 +154,68 @@ VideoStore::VideoStore(const char *filename_in, const char *format_in,
Warning( "Unsupported Orientation(%d)", orientation ); Warning( "Unsupported Orientation(%d)", orientation );
} }
} }
audio_output_codec = NULL;
if (input_audio_stream) { if (input_audio_stream) {
if ( input_audio_stream->codec->codec_id != AV_CODEC_ID_AAC ) { if ( input_audio_stream->codec->codec_id != AV_CODEC_ID_AAC ) {
Warning("Can't transcode to AAC at this time"); Warning("Can't transcode to AAC at this time");
audio_stream = NULL; audio_stream = NULL;
audio_output_codec = avcodec_find_encoder(AV_CODEC_ID_AAC);
if ( audio_output_codec ) {
audio_stream = avformat_new_stream(oc, audio_output_codec );
audio_output_context = audio_stream->codec;
//audio_output_context = avcodec_alloc_context3( audio_output_codec );
if ( audio_output_context ) {
Debug(2, "Have audio_output_context");
AVDictionary *opts = NULL;
av_dict_set(&opts, "strict", "experimental", 0);
/* put sample parameters */
audio_output_context->bit_rate = input_audio_stream->codec->bit_rate;
audio_output_context->sample_rate = input_audio_stream->codec->sample_rate;
audio_output_context->channels = input_audio_stream->codec->channels;
audio_output_context->channel_layout = input_audio_stream->codec->channel_layout;
audio_output_context->sample_fmt = input_audio_stream->codec->sample_fmt;
/* check that the encoder supports s16 pcm input */
if (!check_sample_fmt( audio_output_codec, audio_output_context->sample_fmt)) {
Error( "Encoder does not support sample format %s, setting to FLTP",
av_get_sample_fmt_name( audio_output_context->sample_fmt));
audio_output_context->sample_fmt = AV_SAMPLE_FMT_FLTP;
}
Debug(1, "Audio output bit_rate (%d) sample_rate(%d) channels(%d) fmt(%d) layout(%d)",
audio_output_context->bit_rate,
audio_output_context->sample_rate,
audio_output_context->channels,
audio_output_context->sample_fmt,
audio_output_context->channel_layout
);
/** Set the sample rate for the container. */
audio_stream->time_base.den = input_audio_stream->codec->sample_rate;
audio_stream->time_base.num = 1;
ret = avcodec_open2(audio_output_context, audio_output_codec, &opts );
if ( ret < 0 ) {
av_strerror(ret, error_buffer, sizeof(error_buffer));
Fatal( "could not open codec (%d) (%s)\n", ret, error_buffer );
} else {
Debug(2, "Success opening AAC codec");
}
av_dict_free(&opts);
} else {
Error( "could not allocate codec context for AAC\n");
}
} else {
Error( "could not find codec for AAC\n");
}
} else { } else {
Debug(3, "Got something other than AAC (%d)", input_audio_stream->codec->codec_id ); Debug(3, "Got something other than AAC (%d)", input_audio_stream->codec->codec_id );
@ -205,7 +262,9 @@ VideoStore::VideoStore(const char *filename_in, const char *format_in,
ret = avformat_write_header(oc, NULL); ret = avformat_write_header(oc, NULL);
if (ret < 0) { if (ret < 0) {
zm_dump_stream_format( oc, 0, 0, 1 ); zm_dump_stream_format( oc, 0, 0, 1 );
Fatal("Error occurred when writing output file header to %s: %s\n", if ( audio_stream )
zm_dump_stream_format( oc, 1, 0, 1 );
Error("Error occurred when writing output file header to %s: %s\n",
filename, filename,
av_make_error_string(ret).c_str()); av_make_error_string(ret).c_str());
} }
@ -271,13 +330,11 @@ void VideoStore::dumpPacket( AVPacket *pkt ){
int VideoStore::writeVideoFramePacket(AVPacket *ipkt, AVStream *input_video_stream){ int VideoStore::writeVideoFramePacket(AVPacket *ipkt, AVStream *input_video_stream){
Debug(2, "writeVideoFrame"); Debug(4, "writeVideoFrame");
Debug(3, "before ost_tbcket starttime %d, timebase%d", startTime, video_stream->time_base ); Debug(3, "before ost_tbcket starttime %d, timebase%d", startTime, video_stream->time_base );
//zm_dump_stream_format( oc, ipkt->stream_index, 0, 1 ); //zm_dump_stream_format( oc, ipkt->stream_index, 0, 1 );
Debug(2, "writeVideoFrame %x", video_stream);
int64_t ost_tb_start_time = av_rescale_q(startTime, AV_TIME_BASE_Q, video_stream->time_base); int64_t ost_tb_start_time = av_rescale_q(startTime, AV_TIME_BASE_Q, video_stream->time_base);
Debug(3, "before ost_tbcket starttime %d, ost_tbcket %d", startTime, ost_tb_start_time ); Debug(2, "before ost_tbcket starttime %d, ost_tbcket %d", startTime, ost_tb_start_time );
Debug(2, "writeVideoFrame");
AVPacket opkt; AVPacket opkt;
AVPicture pict; AVPicture pict;
@ -314,17 +371,21 @@ int VideoStore::writeVideoFramePacket(AVPacket *ipkt, AVStream *input_video_stre
/*opkt.flags |= AV_PKT_FLAG_KEY;*/ /*opkt.flags |= AV_PKT_FLAG_KEY;*/
if (video_stream->codec->codec_type == AVMEDIA_TYPE_VIDEO && (output_format->flags & AVFMT_RAWPICTURE)) { if (video_stream->codec->codec_type == AVMEDIA_TYPE_VIDEO && (output_format->flags & AVFMT_RAWPICTURE)) {
Debug(3, "video and RAWPICTURE");
/* store AVPicture in AVPacket, as expected by the output format */ /* store AVPicture in AVPacket, as expected by the output format */
avpicture_fill(&pict, opkt.data, video_stream->codec->pix_fmt, video_stream->codec->width, video_stream->codec->height); avpicture_fill(&pict, opkt.data, video_stream->codec->pix_fmt, video_stream->codec->width, video_stream->codec->height);
opkt.data = (uint8_t *)&pict; opkt.data = (uint8_t *)&pict;
opkt.size = sizeof(AVPicture); opkt.size = sizeof(AVPicture);
opkt.flags |= AV_PKT_FLAG_KEY; opkt.flags |= AV_PKT_FLAG_KEY;
} else {
Debug(3, "Not video and RAWPICTURE");
} }
//memcpy(&safepkt, &opkt, sizeof(AVPacket)); //memcpy(&safepkt, &opkt, sizeof(AVPacket));
if ((opkt.data == NULL)||(opkt.size < 1)) { if ((opkt.data == NULL)||(opkt.size < 1)) {
Warning("%s:%d: Mangled AVPacket: discarding frame", __FILE__, __LINE__ ); Warning("%s:%d: Mangled AVPacket: discarding frame", __FILE__, __LINE__ );
dumpPacket( ipkt);
dumpPacket(&opkt); dumpPacket(&opkt);
} else if ((prevDts > 0) && (prevDts > opkt.dts)) { } else if ((prevDts > 0) && (prevDts > opkt.dts)) {
@ -350,7 +411,7 @@ int VideoStore::writeVideoFramePacket(AVPacket *ipkt, AVStream *input_video_stre
} }
int VideoStore::writeAudioFramePacket(AVPacket *ipkt, AVStream *input_video_stream){ int VideoStore::writeAudioFramePacket(AVPacket *ipkt, AVStream *input_audio_stream){
Debug(2, "writeAudioFrame"); Debug(2, "writeAudioFrame");
if(!audio_stream) { if(!audio_stream) {
@ -361,6 +422,7 @@ int VideoStore::writeAudioFramePacket(AVPacket *ipkt, AVStream *input_video_stre
return -1;*/ return -1;*/
//zm_dump_stream_format( oc, ipkt->stream_index, 0, 1 ); //zm_dump_stream_format( oc, ipkt->stream_index, 0, 1 );
int ret;
// What is this doing? Getting the time of the start of this video chunk? Does that actually make sense? // What is this doing? Getting the time of the start of this video chunk? Does that actually make sense?
int64_t ost_tb_start_time = av_rescale_q(startTime, AV_TIME_BASE_Q, audio_stream->time_base); int64_t ost_tb_start_time = av_rescale_q(startTime, AV_TIME_BASE_Q, audio_stream->time_base);
@ -371,51 +433,146 @@ int VideoStore::writeAudioFramePacket(AVPacket *ipkt, AVStream *input_video_stre
//Scale the PTS of the outgoing packet to be the correct time base //Scale the PTS of the outgoing packet to be the correct time base
if (ipkt->pts != AV_NOPTS_VALUE) { if (ipkt->pts != AV_NOPTS_VALUE) {
Debug(3, "Rescaling output pts"); Debug(2, "Rescaling output pts");
opkt.pts = av_rescale_q(ipkt->pts-startPts, input_video_stream->time_base, audio_stream->time_base) - ost_tb_start_time; opkt.pts = av_rescale_q(ipkt->pts-startPts, input_audio_stream->time_base, audio_stream->time_base) - ost_tb_start_time;
} else { } else {
Debug(3, "Setting output pts to AV_NOPTS_VALUE"); Debug(2, "Setting output pts to AV_NOPTS_VALUE");
opkt.pts = AV_NOPTS_VALUE; opkt.pts = AV_NOPTS_VALUE;
} }
//Scale the DTS of the outgoing packet to be the correct time base //Scale the DTS of the outgoing packet to be the correct time base
if(ipkt->dts == AV_NOPTS_VALUE) { if(ipkt->dts == AV_NOPTS_VALUE) {
Debug(4, "ipkt->dts == AV_NOPTS_VALUE %d to %d", AV_NOPTS_VALUE, opkt.dts ); Debug(2, "ipkt->dts == AV_NOPTS_VALUE %d to %d", AV_NOPTS_VALUE, opkt.dts );
opkt.dts = av_rescale_q(input_video_stream->cur_dts-startDts, AV_TIME_BASE_Q, audio_stream->time_base); opkt.dts = av_rescale_q(input_audio_stream->cur_dts-startDts, AV_TIME_BASE_Q, audio_stream->time_base);
Debug(4, "ipkt->dts == AV_NOPTS_VALUE %d to %d", AV_NOPTS_VALUE, opkt.dts ); Debug(2, "ipkt->dts == AV_NOPTS_VALUE %d to %d", AV_NOPTS_VALUE, opkt.dts );
} else { } else {
Debug(4, "ipkt->dts != AV_NOPTS_VALUE %d to %d", AV_NOPTS_VALUE, opkt.dts ); Debug(2, "ipkt->dts != AV_NOPTS_VALUE %d to %d", AV_NOPTS_VALUE, opkt.dts );
opkt.dts = av_rescale_q(ipkt->dts-startDts, input_video_stream->time_base, audio_stream->time_base); opkt.dts = av_rescale_q(ipkt->dts-startDts, input_audio_stream->time_base, audio_stream->time_base);
Debug(4, "ipkt->dts != AV_NOPTS_VALUE %d to %d", AV_NOPTS_VALUE, opkt.dts ); Debug(2, "ipkt->dts != AV_NOPTS_VALUE %d to %d", AV_NOPTS_VALUE, opkt.dts );
} }
Debug(2, "Not sure what ost_tb_start_time is (%d) - (%d)", opkt.dts, ost_tb_start_time );
opkt.dts -= ost_tb_start_time; opkt.dts -= ost_tb_start_time;
// Seems like it would be really weird for the codec type to NOT be audiu // Seems like it would be really weird for the codec type to NOT be audiu
if (audio_stream->codec->codec_type == AVMEDIA_TYPE_AUDIO && ipkt->dts != AV_NOPTS_VALUE) { if (audio_stream->codec->codec_type == AVMEDIA_TYPE_AUDIO && ipkt->dts != AV_NOPTS_VALUE) {
int duration = av_get_audio_frame_duration(input_video_stream->codec, ipkt->size); int duration = av_get_audio_frame_duration(input_audio_stream->codec, ipkt->size);
Debug( 4, "code is audio, dts != AV_NOPTS_VALUE got duration(%d)", duration ); Debug( 1, "code is audio, dts != AV_NOPTS_VALUE got duration(%d)", duration );
if ( ! duration ) { if ( ! duration ) {
duration = input_video_stream->codec->frame_size; duration = input_audio_stream->codec->frame_size;
Warning( "got no duration from av_get_audio_frame_duration. Using frame size(%d)", duration ); Warning( "got no duration from av_get_audio_frame_duration. Using frame size(%d)", duration );
} }
//FIXME where to get filter_in_rescale_delta_last //FIXME where to get filter_in_rescale_delta_last
//FIXME av_rescale_delta doesn't exist in ubuntu vivid libavtools //FIXME av_rescale_delta doesn't exist in ubuntu vivid libavtools
opkt.dts = opkt.pts = av_rescale_delta(input_video_stream->time_base, ipkt->dts, opkt.dts = opkt.pts = av_rescale_delta(input_audio_stream->time_base, ipkt->dts,
(AVRational){1, input_video_stream->codec->sample_rate}, duration, &filter_in_rescale_delta_last, (AVRational){1, input_audio_stream->codec->sample_rate}, duration, &filter_in_rescale_delta_last,
audio_stream->time_base) - ost_tb_start_time; audio_stream->time_base) - ost_tb_start_time;
Debug(4, "rescaled dts is: (%d)", opkt.dts ); Debug(2, "rescaled dts is: (%d)", opkt.dts );
} }
opkt.duration = av_rescale_q(ipkt->duration, input_video_stream->time_base, audio_stream->time_base); opkt.duration = av_rescale_q(ipkt->duration, input_audio_stream->time_base, audio_stream->time_base);
opkt.pos=-1; opkt.pos=-1;
opkt.flags = ipkt->flags; opkt.flags = ipkt->flags;
opkt.data = ipkt->data;
opkt.size = ipkt->size;
opkt.stream_index = ipkt->stream_index; opkt.stream_index = ipkt->stream_index;
int ret; if ( audio_output_codec ) {
AVFrame *input_frame;
AVFrame *output_frame;
// Need to re-encode
if ( 0 ) {
//avcodec_send_packet( input_audio_stream->codec, ipkt);
//avcodec_receive_frame( input_audio_stream->codec, input_frame );
//avcodec_send_frame( audio_stream->codec, input_frame );
//
////avcodec_receive_packet( audio_stream->codec, &opkt );
} else {
/** Create a new frame to store the audio samples. */
if (!(input_frame = av_frame_alloc())) {
Error("Could not allocate input frame");
zm_av_unref_packet(&opkt);
return 0;
} else {
Debug(2, "Got input frame alloc");
}
/**
* Decode the audio frame stored in the packet.
* The input audio stream decoder is used to do this.
* If we are at the end of the file, pass an empty packet to the decoder
* to flush it.
*/
if ((ret = avcodec_decode_audio4(input_audio_stream->codec, input_frame,
&data_present, ipkt)) < 0) {
Error( "Could not decode frame (error '%s')\n",
av_make_error_string(ret).c_str());
dumpPacket( ipkt);
av_frame_free(&input_frame);
zm_av_unref_packet(&opkt);
return 0;
}
/** Create a new frame to store the audio samples. */
if (!(output_frame = av_frame_alloc())) {
Error("Could not allocate output frame");
av_frame_free(&input_frame);
zm_av_unref_packet(&opkt);
return 0;
} else {
Debug(2, "Got output frame alloc");
}
/**
* Set the frame's parameters, especially its size and format.
* av_frame_get_buffer needs this to allocate memory for the
* audio samples of the frame.
* Default channel layouts based on the number of channels
* are assumed for simplicity.
*/
output_frame->nb_samples = audio_stream->codec->frame_size;
output_frame->channel_layout = audio_output_context->channel_layout;
output_frame->channels = audio_output_context->channels;
output_frame->format = audio_output_context->sample_fmt;
output_frame->sample_rate = audio_output_context->sample_rate;
/**
* Allocate the samples of the created frame. This call will make
* sure that the audio frame can hold as many samples as specified.
*/
Debug(2, "getting buffer");
if (( ret = av_frame_get_buffer( output_frame, 0)) < 0) {
Error( "Couldnt allocate output frame buffer samples (error '%s')",
av_make_error_string(ret).c_str() );
Error("Frame: samples(%d) layout (%d) format(%d) rate(%d)", output_frame->nb_samples,
output_frame->channel_layout, output_frame->format , output_frame->sample_rate
);
av_frame_free(&input_frame);
av_frame_free(&output_frame);
zm_av_unref_packet(&opkt);
return 0;
}
/** Set a timestamp based on the sample rate for the container. */
if (output_frame) {
output_frame->pts = opkt.pts;
}
/**
* Encode the audio frame and store it in the temporary packet.
* The output audio stream encoder is used to do this.
*/
if (( ret = avcodec_encode_audio2( audio_output_context, &opkt,
input_frame, &data_present )) < 0) {
Error( "Could not encode frame (error '%s')",
av_make_error_string(ret).c_str());
zm_av_unref_packet(&opkt);
return 0;
}
}
} else {
opkt.data = ipkt->data;
opkt.size = ipkt->size;
}
ret = av_interleaved_write_frame(oc, &opkt); ret = av_interleaved_write_frame(oc, &opkt);
if(ret!=0){ if(ret!=0){
Fatal("Error encoding audio frame packet: %s\n", av_make_error_string(ret).c_str()); Fatal("Error encoding audio frame packet: %s\n", av_make_error_string(ret).c_str());

View File

@ -14,6 +14,11 @@ private:
AVFormatContext *oc; AVFormatContext *oc;
AVStream *video_stream; AVStream *video_stream;
AVStream *audio_stream; AVStream *audio_stream;
// The following are used when encoding the audio stream to AAC
AVCodec *audio_output_codec;
AVCodecContext *audio_output_context;
int data_present;
const char *filename; const char *filename;
const char *format; const char *format;