Work on aac encoding
This commit is contained in:
parent
d4645cd94a
commit
61df6e9d75
|
@ -412,13 +412,13 @@ int hacked_up_context2_for_older_ffmpeg(AVFormatContext **avctx, AVOutputFormat
|
||||||
static void zm_log_fps(double d, const char *postfix) {
|
static void zm_log_fps(double d, const char *postfix) {
|
||||||
uint64_t v = lrintf(d * 100);
|
uint64_t v = lrintf(d * 100);
|
||||||
if (!v) {
|
if (!v) {
|
||||||
Debug(3, "%1.4f %s", d, postfix);
|
Debug(1, "%1.4f %s", d, postfix);
|
||||||
} else if (v % 100) {
|
} else if (v % 100) {
|
||||||
Debug(3, "%3.2f %s", d, postfix);
|
Debug(1, "%3.2f %s", d, postfix);
|
||||||
} else if (v % (100 * 1000)) {
|
} else if (v % (100 * 1000)) {
|
||||||
Debug(3, "%1.0f %s", d, postfix);
|
Debug(1, "%1.0f %s", d, postfix);
|
||||||
} else
|
} else
|
||||||
Debug(3, "%1.0fk %s", d / 1000, postfix);
|
Debug(1, "%1.0fk %s", d / 1000, postfix);
|
||||||
}
|
}
|
||||||
|
|
||||||
/* "user interface" functions */
|
/* "user interface" functions */
|
||||||
|
@ -429,17 +429,17 @@ void zm_dump_stream_format(AVFormatContext *ic, int i, int index, int is_output)
|
||||||
AVDictionaryEntry *lang = av_dict_get(st->metadata, "language", NULL, 0);
|
AVDictionaryEntry *lang = av_dict_get(st->metadata, "language", NULL, 0);
|
||||||
|
|
||||||
avcodec_string(buf, sizeof(buf), st->codec, is_output);
|
avcodec_string(buf, sizeof(buf), st->codec, is_output);
|
||||||
Debug(3, " Stream #%d:%d", index, i);
|
Debug(1, " Stream #%d:%d", index, i);
|
||||||
|
|
||||||
/* the pid is an important information, so we display it */
|
/* the pid is an important information, so we display it */
|
||||||
/* XXX: add a generic system */
|
/* XXX: add a generic system */
|
||||||
if (flags & AVFMT_SHOW_IDS)
|
if (flags & AVFMT_SHOW_IDS)
|
||||||
Debug(3, "[0x%x]", st->id);
|
Debug(1, "[0x%x]", st->id);
|
||||||
if (lang)
|
if (lang)
|
||||||
Debug(3, "(%s)", lang->value);
|
Debug(1, "(%s)", lang->value);
|
||||||
av_log(NULL, AV_LOG_DEBUG, ", %d, %d/%d", st->codec_info_nb_frames,
|
av_log(NULL, AV_LOG_DEBUG, ", %d, %d/%d", st->codec_info_nb_frames,
|
||||||
st->time_base.num, st->time_base.den);
|
st->time_base.num, st->time_base.den);
|
||||||
Debug(3, ": %s", buf);
|
Debug(1, ": %s", buf);
|
||||||
|
|
||||||
if (st->sample_aspect_ratio.num && // default
|
if (st->sample_aspect_ratio.num && // default
|
||||||
av_cmp_q(st->sample_aspect_ratio, st->codec->sample_aspect_ratio)) {
|
av_cmp_q(st->sample_aspect_ratio, st->codec->sample_aspect_ratio)) {
|
||||||
|
@ -448,7 +448,7 @@ void zm_dump_stream_format(AVFormatContext *ic, int i, int index, int is_output)
|
||||||
st->codec->width * (int64_t)st->sample_aspect_ratio.num,
|
st->codec->width * (int64_t)st->sample_aspect_ratio.num,
|
||||||
st->codec->height * (int64_t)st->sample_aspect_ratio.den,
|
st->codec->height * (int64_t)st->sample_aspect_ratio.den,
|
||||||
1024 * 1024);
|
1024 * 1024);
|
||||||
Debug(3, ", SAR %d:%d DAR %d:%d",
|
Debug(1, ", SAR %d:%d DAR %d:%d",
|
||||||
st->sample_aspect_ratio.num, st->sample_aspect_ratio.den,
|
st->sample_aspect_ratio.num, st->sample_aspect_ratio.den,
|
||||||
display_aspect_ratio.num, display_aspect_ratio.den);
|
display_aspect_ratio.num, display_aspect_ratio.den);
|
||||||
}
|
}
|
||||||
|
@ -470,28 +470,40 @@ void zm_dump_stream_format(AVFormatContext *ic, int i, int index, int is_output)
|
||||||
}
|
}
|
||||||
|
|
||||||
if (st->disposition & AV_DISPOSITION_DEFAULT)
|
if (st->disposition & AV_DISPOSITION_DEFAULT)
|
||||||
Debug(3, " (default)");
|
Debug(1, " (default)");
|
||||||
if (st->disposition & AV_DISPOSITION_DUB)
|
if (st->disposition & AV_DISPOSITION_DUB)
|
||||||
Debug(3, " (dub)");
|
Debug(1, " (dub)");
|
||||||
if (st->disposition & AV_DISPOSITION_ORIGINAL)
|
if (st->disposition & AV_DISPOSITION_ORIGINAL)
|
||||||
Debug(3, " (original)");
|
Debug(1, " (original)");
|
||||||
if (st->disposition & AV_DISPOSITION_COMMENT)
|
if (st->disposition & AV_DISPOSITION_COMMENT)
|
||||||
Debug(3, " (comment)");
|
Debug(1, " (comment)");
|
||||||
if (st->disposition & AV_DISPOSITION_LYRICS)
|
if (st->disposition & AV_DISPOSITION_LYRICS)
|
||||||
Debug(3, " (lyrics)");
|
Debug(1, " (lyrics)");
|
||||||
if (st->disposition & AV_DISPOSITION_KARAOKE)
|
if (st->disposition & AV_DISPOSITION_KARAOKE)
|
||||||
Debug(3, " (karaoke)");
|
Debug(1, " (karaoke)");
|
||||||
if (st->disposition & AV_DISPOSITION_FORCED)
|
if (st->disposition & AV_DISPOSITION_FORCED)
|
||||||
Debug(3, " (forced)");
|
Debug(1, " (forced)");
|
||||||
if (st->disposition & AV_DISPOSITION_HEARING_IMPAIRED)
|
if (st->disposition & AV_DISPOSITION_HEARING_IMPAIRED)
|
||||||
Debug(3, " (hearing impaired)");
|
Debug(1, " (hearing impaired)");
|
||||||
if (st->disposition & AV_DISPOSITION_VISUAL_IMPAIRED)
|
if (st->disposition & AV_DISPOSITION_VISUAL_IMPAIRED)
|
||||||
Debug(3, " (visual impaired)");
|
Debug(1, " (visual impaired)");
|
||||||
if (st->disposition & AV_DISPOSITION_CLEAN_EFFECTS)
|
if (st->disposition & AV_DISPOSITION_CLEAN_EFFECTS)
|
||||||
Debug(3, " (clean effects)");
|
Debug(1, " (clean effects)");
|
||||||
Debug(3, "\n");
|
Debug(1, "\n");
|
||||||
|
|
||||||
//dump_metadata(NULL, st->metadata, " ");
|
//dump_metadata(NULL, st->metadata, " ");
|
||||||
|
|
||||||
//dump_sidedata(NULL, st, " ");
|
//dump_sidedata(NULL, st, " ");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
int check_sample_fmt(AVCodec *codec, enum AVSampleFormat sample_fmt) {
|
||||||
|
const enum AVSampleFormat *p = codec->sample_fmts;
|
||||||
|
|
||||||
|
while (*p != AV_SAMPLE_FMT_NONE) {
|
||||||
|
if (*p == sample_fmt)
|
||||||
|
return 1;
|
||||||
|
else Debug(2, "Not %s", av_get_sample_fmt_name( *p ) );
|
||||||
|
p++;
|
||||||
|
}
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
|
@ -334,4 +334,6 @@ void zm_dump_stream_format(AVFormatContext *ic, int i, int index, int is_output)
|
||||||
#define zm_avcodec_decode_video(context, rawFrame, frameComplete, packet ) avcodec_decode_video( context, rawFrame, frameComplete, packet->data, packet->size)
|
#define zm_avcodec_decode_video(context, rawFrame, frameComplete, packet ) avcodec_decode_video( context, rawFrame, frameComplete, packet->data, packet->size)
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
int check_sample_fmt(AVCodec *codec, enum AVSampleFormat sample_fmt);
|
||||||
|
|
||||||
#endif // ZM_FFMPEG_H
|
#endif // ZM_FFMPEG_H
|
||||||
|
|
|
@ -290,7 +290,6 @@ int FfmpegCamera::OpenFfmpeg() {
|
||||||
Debug ( 1, "Opened input" );
|
Debug ( 1, "Opened input" );
|
||||||
|
|
||||||
Info( "Stream open %s", mPath.c_str() );
|
Info( "Stream open %s", mPath.c_str() );
|
||||||
startTime=av_gettime();//FIXME here or after find_Stream_info
|
|
||||||
|
|
||||||
//FIXME can speed up initial analysis but need sensible parameters...
|
//FIXME can speed up initial analysis but need sensible parameters...
|
||||||
//mFormatContext->probesize = 32;
|
//mFormatContext->probesize = 32;
|
||||||
|
@ -356,20 +355,6 @@ int FfmpegCamera::OpenFfmpeg() {
|
||||||
} else {
|
} else {
|
||||||
Debug(1, "Video Found decoder");
|
Debug(1, "Video Found decoder");
|
||||||
zm_dump_stream_format(mFormatContext, mVideoStreamId, 0, 0);
|
zm_dump_stream_format(mFormatContext, mVideoStreamId, 0, 0);
|
||||||
}
|
|
||||||
|
|
||||||
if (mAudioStreamId >= 0) {
|
|
||||||
mAudioCodecContext = mFormatContext->streams[mAudioStreamId]->codec;
|
|
||||||
if ((mAudioCodec = avcodec_find_decoder(mAudioCodecContext->codec_id)) == NULL) {
|
|
||||||
Debug(1, "Can't find codec for audio stream from %s", mPath.c_str());
|
|
||||||
} else {
|
|
||||||
Debug(1, "Audio Found decoder");
|
|
||||||
zm_dump_stream_format(mFormatContext, mAudioStreamId, 0, 0);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
//
|
|
||||||
|
|
||||||
// Open the codec
|
// Open the codec
|
||||||
#if !LIBAVFORMAT_VERSION_CHECK(53, 8, 0, 8, 0)
|
#if !LIBAVFORMAT_VERSION_CHECK(53, 8, 0, 8, 0)
|
||||||
Debug ( 1, "Calling avcodec_open" );
|
Debug ( 1, "Calling avcodec_open" );
|
||||||
|
@ -379,6 +364,29 @@ int FfmpegCamera::OpenFfmpeg() {
|
||||||
if (avcodec_open2(mVideoCodecContext, mVideoCodec, 0) < 0)
|
if (avcodec_open2(mVideoCodecContext, mVideoCodec, 0) < 0)
|
||||||
#endif
|
#endif
|
||||||
Fatal( "Unable to open codec for video stream from %s", mPath.c_str() );
|
Fatal( "Unable to open codec for video stream from %s", mPath.c_str() );
|
||||||
|
}
|
||||||
|
|
||||||
|
if (mAudioStreamId >= 0) {
|
||||||
|
mAudioCodecContext = mFormatContext->streams[mAudioStreamId]->codec;
|
||||||
|
if ((mAudioCodec = avcodec_find_decoder(mAudioCodecContext->codec_id)) == NULL) {
|
||||||
|
Debug(1, "Can't find codec for audio stream from %s", mPath.c_str());
|
||||||
|
} else {
|
||||||
|
Debug(1, "Audio Found decoder");
|
||||||
|
zm_dump_stream_format(mFormatContext, mAudioStreamId, 0, 0);
|
||||||
|
// Open the codec
|
||||||
|
#if !LIBAVFORMAT_VERSION_CHECK(53, 8, 0, 8, 0)
|
||||||
|
Debug ( 1, "Calling avcodec_open" );
|
||||||
|
if (avcodec_open(mAudioCodecContext, mAudioCodec) < 0)
|
||||||
|
#else
|
||||||
|
Debug ( 1, "Calling avcodec_open2" );
|
||||||
|
if (avcodec_open2(mAudioCodecContext, mAudioCodec, 0) < 0)
|
||||||
|
#endif
|
||||||
|
Fatal( "Unable to open codec for video stream from %s", mPath.c_str() );
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
//
|
||||||
|
|
||||||
|
|
||||||
Debug ( 1, "Opened codec" );
|
Debug ( 1, "Opened codec" );
|
||||||
|
|
||||||
|
@ -556,7 +564,7 @@ int FfmpegCamera::CaptureAndRecord( Image &image, bool recording, char* event_fi
|
||||||
while ( !frameComplete ) {
|
while ( !frameComplete ) {
|
||||||
// We are now allocating dynamically because we need to queue these and may go out of scope.
|
// We are now allocating dynamically because we need to queue these and may go out of scope.
|
||||||
AVPacket *packet = (AVPacket *)av_malloc(sizeof(AVPacket));
|
AVPacket *packet = (AVPacket *)av_malloc(sizeof(AVPacket));
|
||||||
av_init_packet( packet);
|
av_init_packet( packet );
|
||||||
Debug(5, "Before av_read_frame");
|
Debug(5, "Before av_read_frame");
|
||||||
ret = av_read_frame( mFormatContext, packet );
|
ret = av_read_frame( mFormatContext, packet );
|
||||||
Debug(5, "After av_read_frame (%d)", ret );
|
Debug(5, "After av_read_frame (%d)", ret );
|
||||||
|
@ -593,6 +601,7 @@ Debug(5, "After av_read_frame (%d)", ret );
|
||||||
|
|
||||||
if ( ! videoStore ) {
|
if ( ! videoStore ) {
|
||||||
//Instantiate the video storage module
|
//Instantiate the video storage module
|
||||||
|
startTime=av_gettime();//FIXME here or after find_Stream_info
|
||||||
|
|
||||||
if (record_audio) {
|
if (record_audio) {
|
||||||
if (mAudioStreamId == -1) {
|
if (mAudioStreamId == -1) {
|
||||||
|
@ -627,10 +636,11 @@ Debug(5, "After av_read_frame (%d)", ret );
|
||||||
while ( ( queued_packet = packetqueue.popPacket() ) ) {
|
while ( ( queued_packet = packetqueue.popPacket() ) ) {
|
||||||
packet_count += 1;
|
packet_count += 1;
|
||||||
//Write the packet to our video store
|
//Write the packet to our video store
|
||||||
|
Debug(2, "Writing queued packet stream: %d KEY %d", queued_packet->stream_index, packet->flags & AV_PKT_FLAG_KEY );
|
||||||
if ( queued_packet->stream_index == mVideoStreamId ) {
|
if ( queued_packet->stream_index == mVideoStreamId ) {
|
||||||
ret = videoStore->writeVideoFramePacket( queued_packet, mFormatContext->streams[mVideoStreamId]);
|
ret = videoStore->writeVideoFramePacket( queued_packet, mFormatContext->streams[mVideoStreamId]);
|
||||||
} else if ( queued_packet->stream_index == mAudioStreamId ) {
|
} else if ( queued_packet->stream_index == mAudioStreamId ) {
|
||||||
//ret = videoStore->writeAudioFramePacket(&queued_packet, mFormatContext->streams[mAudioStreamId]);
|
ret = videoStore->writeAudioFramePacket( queued_packet, mFormatContext->streams[mAudioStreamId]);
|
||||||
} else {
|
} else {
|
||||||
Warning("Unknown stream id in queued packet (%d)", queued_packet->stream_index );
|
Warning("Unknown stream id in queued packet (%d)", queued_packet->stream_index );
|
||||||
ret = -1;
|
ret = -1;
|
||||||
|
@ -643,12 +653,6 @@ Debug(5, "After av_read_frame (%d)", ret );
|
||||||
Debug(2, "Wrote %d queued packets", packet_count );
|
Debug(2, "Wrote %d queued packets", packet_count );
|
||||||
} // end if ! wasRecording
|
} // end if ! wasRecording
|
||||||
|
|
||||||
//Write the packet to our video store
|
|
||||||
int ret = videoStore->writeVideoFramePacket( packet, mFormatContext->streams[mVideoStreamId] );
|
|
||||||
if ( ret < 0 ) { //Less than zero and we skipped a frame
|
|
||||||
zm_av_unref_packet( packet );
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
} else {
|
} else {
|
||||||
// Not recording
|
// Not recording
|
||||||
if ( videoStore ) {
|
if ( videoStore ) {
|
||||||
|
@ -656,15 +660,23 @@ Debug(5, "After av_read_frame (%d)", ret );
|
||||||
delete videoStore;
|
delete videoStore;
|
||||||
videoStore = NULL;
|
videoStore = NULL;
|
||||||
}
|
}
|
||||||
} // end if
|
|
||||||
|
|
||||||
//Buffer video packets
|
//Buffer video packets, since we are not recording
|
||||||
if ( packet->flags & AV_PKT_FLAG_KEY ) {
|
if ( (packet->stream_index == mVideoStreamId) && ( packet->flags & AV_PKT_FLAG_KEY ) ) {
|
||||||
packetqueue.clearQueue();
|
packetqueue.clearQueue();
|
||||||
}
|
}
|
||||||
packetqueue.queuePacket(packet);
|
packetqueue.queuePacket(packet);
|
||||||
|
} // end if
|
||||||
|
|
||||||
if ( packet->stream_index == mVideoStreamId ) {
|
if ( packet->stream_index == mVideoStreamId ) {
|
||||||
|
if ( videoStore ) {
|
||||||
|
//Write the packet to our video store
|
||||||
|
int ret = videoStore->writeVideoFramePacket( packet, mFormatContext->streams[mVideoStreamId] );
|
||||||
|
if ( ret < 0 ) { //Less than zero and we skipped a frame
|
||||||
|
zm_av_unref_packet( packet );
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
}
|
||||||
ret = zm_avcodec_decode_video( mVideoCodecContext, mRawFrame, &frameComplete, packet );
|
ret = zm_avcodec_decode_video( mVideoCodecContext, mRawFrame, &frameComplete, packet );
|
||||||
if ( ret < 0 ) {
|
if ( ret < 0 ) {
|
||||||
av_strerror( ret, errbuf, AV_ERROR_MAX_STRING_SIZE );
|
av_strerror( ret, errbuf, AV_ERROR_MAX_STRING_SIZE );
|
||||||
|
|
|
@ -20,9 +20,9 @@
|
||||||
#ifndef ZM_PACKETQUEUE_H
|
#ifndef ZM_PACKETQUEUE_H
|
||||||
#define ZM_PACKETQUEUE_H
|
#define ZM_PACKETQUEUE_H
|
||||||
|
|
||||||
#include <boost/interprocess/managed_shared_memory.hpp>
|
//#include <boost/interprocess/managed_shared_memory.hpp>
|
||||||
#include <boost/interprocess/containers/map.hpp>
|
//#include <boost/interprocess/containers/map.hpp>
|
||||||
#include <boost/interprocess/allocators/allocator.hpp>
|
//#include <boost/interprocess/allocators/allocator.hpp>
|
||||||
#include <queue>
|
#include <queue>
|
||||||
|
|
||||||
extern "C" {
|
extern "C" {
|
||||||
|
|
|
@ -49,8 +49,9 @@ VideoStore::VideoStore(const char *filename_in, const char *format_in,
|
||||||
Info("Opening video storage stream %s format: %s\n", filename, format);
|
Info("Opening video storage stream %s format: %s\n", filename, format);
|
||||||
|
|
||||||
int ret;
|
int ret;
|
||||||
//Init everything we need
|
static char error_buffer[255];
|
||||||
av_register_all();
|
//Init everything we need, shouldn't have to do this, ffmpeg_camera or something else will call it.
|
||||||
|
//av_register_all();
|
||||||
|
|
||||||
ret = avformat_alloc_output_context2(&oc, NULL, NULL, filename);
|
ret = avformat_alloc_output_context2(&oc, NULL, NULL, filename);
|
||||||
if ( ret < 0 ) {
|
if ( ret < 0 ) {
|
||||||
|
@ -153,12 +154,68 @@ VideoStore::VideoStore(const char *filename_in, const char *format_in,
|
||||||
Warning( "Unsupported Orientation(%d)", orientation );
|
Warning( "Unsupported Orientation(%d)", orientation );
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
audio_output_codec = NULL;
|
||||||
|
|
||||||
if (input_audio_stream) {
|
if (input_audio_stream) {
|
||||||
|
|
||||||
if ( input_audio_stream->codec->codec_id != AV_CODEC_ID_AAC ) {
|
if ( input_audio_stream->codec->codec_id != AV_CODEC_ID_AAC ) {
|
||||||
Warning("Can't transcode to AAC at this time");
|
Warning("Can't transcode to AAC at this time");
|
||||||
audio_stream = NULL;
|
audio_stream = NULL;
|
||||||
|
|
||||||
|
audio_output_codec = avcodec_find_encoder(AV_CODEC_ID_AAC);
|
||||||
|
if ( audio_output_codec ) {
|
||||||
|
audio_stream = avformat_new_stream(oc, audio_output_codec );
|
||||||
|
|
||||||
|
audio_output_context = audio_stream->codec;
|
||||||
|
|
||||||
|
//audio_output_context = avcodec_alloc_context3( audio_output_codec );
|
||||||
|
if ( audio_output_context ) {
|
||||||
|
|
||||||
|
Debug(2, "Have audio_output_context");
|
||||||
|
AVDictionary *opts = NULL;
|
||||||
|
av_dict_set(&opts, "strict", "experimental", 0);
|
||||||
|
|
||||||
|
/* put sample parameters */
|
||||||
|
audio_output_context->bit_rate = input_audio_stream->codec->bit_rate;
|
||||||
|
audio_output_context->sample_rate = input_audio_stream->codec->sample_rate;
|
||||||
|
audio_output_context->channels = input_audio_stream->codec->channels;
|
||||||
|
audio_output_context->channel_layout = input_audio_stream->codec->channel_layout;
|
||||||
|
audio_output_context->sample_fmt = input_audio_stream->codec->sample_fmt;
|
||||||
|
|
||||||
|
/* check that the encoder supports s16 pcm input */
|
||||||
|
if (!check_sample_fmt( audio_output_codec, audio_output_context->sample_fmt)) {
|
||||||
|
Error( "Encoder does not support sample format %s, setting to FLTP",
|
||||||
|
av_get_sample_fmt_name( audio_output_context->sample_fmt));
|
||||||
|
audio_output_context->sample_fmt = AV_SAMPLE_FMT_FLTP;
|
||||||
|
}
|
||||||
|
|
||||||
|
Debug(1, "Audio output bit_rate (%d) sample_rate(%d) channels(%d) fmt(%d) layout(%d)",
|
||||||
|
audio_output_context->bit_rate,
|
||||||
|
audio_output_context->sample_rate,
|
||||||
|
audio_output_context->channels,
|
||||||
|
audio_output_context->sample_fmt,
|
||||||
|
audio_output_context->channel_layout
|
||||||
|
);
|
||||||
|
|
||||||
|
/** Set the sample rate for the container. */
|
||||||
|
audio_stream->time_base.den = input_audio_stream->codec->sample_rate;
|
||||||
|
audio_stream->time_base.num = 1;
|
||||||
|
|
||||||
|
ret = avcodec_open2(audio_output_context, audio_output_codec, &opts );
|
||||||
|
if ( ret < 0 ) {
|
||||||
|
av_strerror(ret, error_buffer, sizeof(error_buffer));
|
||||||
|
Fatal( "could not open codec (%d) (%s)\n", ret, error_buffer );
|
||||||
|
} else {
|
||||||
|
Debug(2, "Success opening AAC codec");
|
||||||
|
}
|
||||||
|
av_dict_free(&opts);
|
||||||
|
} else {
|
||||||
|
Error( "could not allocate codec context for AAC\n");
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
Error( "could not find codec for AAC\n");
|
||||||
|
}
|
||||||
|
|
||||||
} else {
|
} else {
|
||||||
Debug(3, "Got something other than AAC (%d)", input_audio_stream->codec->codec_id );
|
Debug(3, "Got something other than AAC (%d)", input_audio_stream->codec->codec_id );
|
||||||
|
|
||||||
|
@ -205,7 +262,9 @@ VideoStore::VideoStore(const char *filename_in, const char *format_in,
|
||||||
ret = avformat_write_header(oc, NULL);
|
ret = avformat_write_header(oc, NULL);
|
||||||
if (ret < 0) {
|
if (ret < 0) {
|
||||||
zm_dump_stream_format( oc, 0, 0, 1 );
|
zm_dump_stream_format( oc, 0, 0, 1 );
|
||||||
Fatal("Error occurred when writing output file header to %s: %s\n",
|
if ( audio_stream )
|
||||||
|
zm_dump_stream_format( oc, 1, 0, 1 );
|
||||||
|
Error("Error occurred when writing output file header to %s: %s\n",
|
||||||
filename,
|
filename,
|
||||||
av_make_error_string(ret).c_str());
|
av_make_error_string(ret).c_str());
|
||||||
}
|
}
|
||||||
|
@ -271,13 +330,11 @@ void VideoStore::dumpPacket( AVPacket *pkt ){
|
||||||
|
|
||||||
int VideoStore::writeVideoFramePacket(AVPacket *ipkt, AVStream *input_video_stream){
|
int VideoStore::writeVideoFramePacket(AVPacket *ipkt, AVStream *input_video_stream){
|
||||||
|
|
||||||
Debug(2, "writeVideoFrame");
|
Debug(4, "writeVideoFrame");
|
||||||
Debug(3, "before ost_tbcket starttime %d, timebase%d", startTime, video_stream->time_base );
|
Debug(3, "before ost_tbcket starttime %d, timebase%d", startTime, video_stream->time_base );
|
||||||
//zm_dump_stream_format( oc, ipkt->stream_index, 0, 1 );
|
//zm_dump_stream_format( oc, ipkt->stream_index, 0, 1 );
|
||||||
Debug(2, "writeVideoFrame %x", video_stream);
|
|
||||||
int64_t ost_tb_start_time = av_rescale_q(startTime, AV_TIME_BASE_Q, video_stream->time_base);
|
int64_t ost_tb_start_time = av_rescale_q(startTime, AV_TIME_BASE_Q, video_stream->time_base);
|
||||||
Debug(3, "before ost_tbcket starttime %d, ost_tbcket %d", startTime, ost_tb_start_time );
|
Debug(2, "before ost_tbcket starttime %d, ost_tbcket %d", startTime, ost_tb_start_time );
|
||||||
Debug(2, "writeVideoFrame");
|
|
||||||
|
|
||||||
AVPacket opkt;
|
AVPacket opkt;
|
||||||
AVPicture pict;
|
AVPicture pict;
|
||||||
|
@ -314,17 +371,21 @@ int VideoStore::writeVideoFramePacket(AVPacket *ipkt, AVStream *input_video_stre
|
||||||
/*opkt.flags |= AV_PKT_FLAG_KEY;*/
|
/*opkt.flags |= AV_PKT_FLAG_KEY;*/
|
||||||
|
|
||||||
if (video_stream->codec->codec_type == AVMEDIA_TYPE_VIDEO && (output_format->flags & AVFMT_RAWPICTURE)) {
|
if (video_stream->codec->codec_type == AVMEDIA_TYPE_VIDEO && (output_format->flags & AVFMT_RAWPICTURE)) {
|
||||||
|
Debug(3, "video and RAWPICTURE");
|
||||||
/* store AVPicture in AVPacket, as expected by the output format */
|
/* store AVPicture in AVPacket, as expected by the output format */
|
||||||
avpicture_fill(&pict, opkt.data, video_stream->codec->pix_fmt, video_stream->codec->width, video_stream->codec->height);
|
avpicture_fill(&pict, opkt.data, video_stream->codec->pix_fmt, video_stream->codec->width, video_stream->codec->height);
|
||||||
opkt.data = (uint8_t *)&pict;
|
opkt.data = (uint8_t *)&pict;
|
||||||
opkt.size = sizeof(AVPicture);
|
opkt.size = sizeof(AVPicture);
|
||||||
opkt.flags |= AV_PKT_FLAG_KEY;
|
opkt.flags |= AV_PKT_FLAG_KEY;
|
||||||
|
} else {
|
||||||
|
Debug(3, "Not video and RAWPICTURE");
|
||||||
}
|
}
|
||||||
|
|
||||||
//memcpy(&safepkt, &opkt, sizeof(AVPacket));
|
//memcpy(&safepkt, &opkt, sizeof(AVPacket));
|
||||||
|
|
||||||
if ((opkt.data == NULL)||(opkt.size < 1)) {
|
if ((opkt.data == NULL)||(opkt.size < 1)) {
|
||||||
Warning("%s:%d: Mangled AVPacket: discarding frame", __FILE__, __LINE__ );
|
Warning("%s:%d: Mangled AVPacket: discarding frame", __FILE__, __LINE__ );
|
||||||
|
dumpPacket( ipkt);
|
||||||
dumpPacket(&opkt);
|
dumpPacket(&opkt);
|
||||||
|
|
||||||
} else if ((prevDts > 0) && (prevDts > opkt.dts)) {
|
} else if ((prevDts > 0) && (prevDts > opkt.dts)) {
|
||||||
|
@ -350,7 +411,7 @@ int VideoStore::writeVideoFramePacket(AVPacket *ipkt, AVStream *input_video_stre
|
||||||
|
|
||||||
}
|
}
|
||||||
|
|
||||||
int VideoStore::writeAudioFramePacket(AVPacket *ipkt, AVStream *input_video_stream){
|
int VideoStore::writeAudioFramePacket(AVPacket *ipkt, AVStream *input_audio_stream){
|
||||||
Debug(2, "writeAudioFrame");
|
Debug(2, "writeAudioFrame");
|
||||||
|
|
||||||
if(!audio_stream) {
|
if(!audio_stream) {
|
||||||
|
@ -361,6 +422,7 @@ int VideoStore::writeAudioFramePacket(AVPacket *ipkt, AVStream *input_video_stre
|
||||||
return -1;*/
|
return -1;*/
|
||||||
//zm_dump_stream_format( oc, ipkt->stream_index, 0, 1 );
|
//zm_dump_stream_format( oc, ipkt->stream_index, 0, 1 );
|
||||||
|
|
||||||
|
int ret;
|
||||||
// What is this doing? Getting the time of the start of this video chunk? Does that actually make sense?
|
// What is this doing? Getting the time of the start of this video chunk? Does that actually make sense?
|
||||||
int64_t ost_tb_start_time = av_rescale_q(startTime, AV_TIME_BASE_Q, audio_stream->time_base);
|
int64_t ost_tb_start_time = av_rescale_q(startTime, AV_TIME_BASE_Q, audio_stream->time_base);
|
||||||
|
|
||||||
|
@ -371,51 +433,146 @@ int VideoStore::writeAudioFramePacket(AVPacket *ipkt, AVStream *input_video_stre
|
||||||
|
|
||||||
//Scale the PTS of the outgoing packet to be the correct time base
|
//Scale the PTS of the outgoing packet to be the correct time base
|
||||||
if (ipkt->pts != AV_NOPTS_VALUE) {
|
if (ipkt->pts != AV_NOPTS_VALUE) {
|
||||||
Debug(3, "Rescaling output pts");
|
Debug(2, "Rescaling output pts");
|
||||||
opkt.pts = av_rescale_q(ipkt->pts-startPts, input_video_stream->time_base, audio_stream->time_base) - ost_tb_start_time;
|
opkt.pts = av_rescale_q(ipkt->pts-startPts, input_audio_stream->time_base, audio_stream->time_base) - ost_tb_start_time;
|
||||||
} else {
|
} else {
|
||||||
Debug(3, "Setting output pts to AV_NOPTS_VALUE");
|
Debug(2, "Setting output pts to AV_NOPTS_VALUE");
|
||||||
opkt.pts = AV_NOPTS_VALUE;
|
opkt.pts = AV_NOPTS_VALUE;
|
||||||
}
|
}
|
||||||
|
|
||||||
//Scale the DTS of the outgoing packet to be the correct time base
|
//Scale the DTS of the outgoing packet to be the correct time base
|
||||||
if(ipkt->dts == AV_NOPTS_VALUE) {
|
if(ipkt->dts == AV_NOPTS_VALUE) {
|
||||||
Debug(4, "ipkt->dts == AV_NOPTS_VALUE %d to %d", AV_NOPTS_VALUE, opkt.dts );
|
Debug(2, "ipkt->dts == AV_NOPTS_VALUE %d to %d", AV_NOPTS_VALUE, opkt.dts );
|
||||||
opkt.dts = av_rescale_q(input_video_stream->cur_dts-startDts, AV_TIME_BASE_Q, audio_stream->time_base);
|
opkt.dts = av_rescale_q(input_audio_stream->cur_dts-startDts, AV_TIME_BASE_Q, audio_stream->time_base);
|
||||||
Debug(4, "ipkt->dts == AV_NOPTS_VALUE %d to %d", AV_NOPTS_VALUE, opkt.dts );
|
Debug(2, "ipkt->dts == AV_NOPTS_VALUE %d to %d", AV_NOPTS_VALUE, opkt.dts );
|
||||||
} else {
|
} else {
|
||||||
Debug(4, "ipkt->dts != AV_NOPTS_VALUE %d to %d", AV_NOPTS_VALUE, opkt.dts );
|
Debug(2, "ipkt->dts != AV_NOPTS_VALUE %d to %d", AV_NOPTS_VALUE, opkt.dts );
|
||||||
opkt.dts = av_rescale_q(ipkt->dts-startDts, input_video_stream->time_base, audio_stream->time_base);
|
opkt.dts = av_rescale_q(ipkt->dts-startDts, input_audio_stream->time_base, audio_stream->time_base);
|
||||||
Debug(4, "ipkt->dts != AV_NOPTS_VALUE %d to %d", AV_NOPTS_VALUE, opkt.dts );
|
Debug(2, "ipkt->dts != AV_NOPTS_VALUE %d to %d", AV_NOPTS_VALUE, opkt.dts );
|
||||||
}
|
}
|
||||||
|
Debug(2, "Not sure what ost_tb_start_time is (%d) - (%d)", opkt.dts, ost_tb_start_time );
|
||||||
opkt.dts -= ost_tb_start_time;
|
opkt.dts -= ost_tb_start_time;
|
||||||
|
|
||||||
// Seems like it would be really weird for the codec type to NOT be audiu
|
// Seems like it would be really weird for the codec type to NOT be audiu
|
||||||
if (audio_stream->codec->codec_type == AVMEDIA_TYPE_AUDIO && ipkt->dts != AV_NOPTS_VALUE) {
|
if (audio_stream->codec->codec_type == AVMEDIA_TYPE_AUDIO && ipkt->dts != AV_NOPTS_VALUE) {
|
||||||
int duration = av_get_audio_frame_duration(input_video_stream->codec, ipkt->size);
|
int duration = av_get_audio_frame_duration(input_audio_stream->codec, ipkt->size);
|
||||||
Debug( 4, "code is audio, dts != AV_NOPTS_VALUE got duration(%d)", duration );
|
Debug( 1, "code is audio, dts != AV_NOPTS_VALUE got duration(%d)", duration );
|
||||||
if ( ! duration ) {
|
if ( ! duration ) {
|
||||||
duration = input_video_stream->codec->frame_size;
|
duration = input_audio_stream->codec->frame_size;
|
||||||
Warning( "got no duration from av_get_audio_frame_duration. Using frame size(%d)", duration );
|
Warning( "got no duration from av_get_audio_frame_duration. Using frame size(%d)", duration );
|
||||||
}
|
}
|
||||||
|
|
||||||
//FIXME where to get filter_in_rescale_delta_last
|
//FIXME where to get filter_in_rescale_delta_last
|
||||||
//FIXME av_rescale_delta doesn't exist in ubuntu vivid libavtools
|
//FIXME av_rescale_delta doesn't exist in ubuntu vivid libavtools
|
||||||
opkt.dts = opkt.pts = av_rescale_delta(input_video_stream->time_base, ipkt->dts,
|
opkt.dts = opkt.pts = av_rescale_delta(input_audio_stream->time_base, ipkt->dts,
|
||||||
(AVRational){1, input_video_stream->codec->sample_rate}, duration, &filter_in_rescale_delta_last,
|
(AVRational){1, input_audio_stream->codec->sample_rate}, duration, &filter_in_rescale_delta_last,
|
||||||
audio_stream->time_base) - ost_tb_start_time;
|
audio_stream->time_base) - ost_tb_start_time;
|
||||||
Debug(4, "rescaled dts is: (%d)", opkt.dts );
|
Debug(2, "rescaled dts is: (%d)", opkt.dts );
|
||||||
}
|
}
|
||||||
|
|
||||||
opkt.duration = av_rescale_q(ipkt->duration, input_video_stream->time_base, audio_stream->time_base);
|
opkt.duration = av_rescale_q(ipkt->duration, input_audio_stream->time_base, audio_stream->time_base);
|
||||||
opkt.pos=-1;
|
opkt.pos=-1;
|
||||||
opkt.flags = ipkt->flags;
|
opkt.flags = ipkt->flags;
|
||||||
|
|
||||||
opkt.data = ipkt->data;
|
|
||||||
opkt.size = ipkt->size;
|
|
||||||
opkt.stream_index = ipkt->stream_index;
|
opkt.stream_index = ipkt->stream_index;
|
||||||
|
|
||||||
int ret;
|
if ( audio_output_codec ) {
|
||||||
|
|
||||||
|
|
||||||
|
|
||||||
|
AVFrame *input_frame;
|
||||||
|
AVFrame *output_frame;
|
||||||
|
// Need to re-encode
|
||||||
|
if ( 0 ) {
|
||||||
|
//avcodec_send_packet( input_audio_stream->codec, ipkt);
|
||||||
|
//avcodec_receive_frame( input_audio_stream->codec, input_frame );
|
||||||
|
//avcodec_send_frame( audio_stream->codec, input_frame );
|
||||||
|
//
|
||||||
|
////avcodec_receive_packet( audio_stream->codec, &opkt );
|
||||||
|
} else {
|
||||||
|
|
||||||
|
/** Create a new frame to store the audio samples. */
|
||||||
|
if (!(input_frame = av_frame_alloc())) {
|
||||||
|
Error("Could not allocate input frame");
|
||||||
|
zm_av_unref_packet(&opkt);
|
||||||
|
return 0;
|
||||||
|
} else {
|
||||||
|
Debug(2, "Got input frame alloc");
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Decode the audio frame stored in the packet.
|
||||||
|
* The input audio stream decoder is used to do this.
|
||||||
|
* If we are at the end of the file, pass an empty packet to the decoder
|
||||||
|
* to flush it.
|
||||||
|
*/
|
||||||
|
if ((ret = avcodec_decode_audio4(input_audio_stream->codec, input_frame,
|
||||||
|
&data_present, ipkt)) < 0) {
|
||||||
|
Error( "Could not decode frame (error '%s')\n",
|
||||||
|
av_make_error_string(ret).c_str());
|
||||||
|
dumpPacket( ipkt);
|
||||||
|
av_frame_free(&input_frame);
|
||||||
|
zm_av_unref_packet(&opkt);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
/** Create a new frame to store the audio samples. */
|
||||||
|
if (!(output_frame = av_frame_alloc())) {
|
||||||
|
Error("Could not allocate output frame");
|
||||||
|
av_frame_free(&input_frame);
|
||||||
|
zm_av_unref_packet(&opkt);
|
||||||
|
return 0;
|
||||||
|
} else {
|
||||||
|
Debug(2, "Got output frame alloc");
|
||||||
|
}
|
||||||
|
/**
|
||||||
|
* Set the frame's parameters, especially its size and format.
|
||||||
|
* av_frame_get_buffer needs this to allocate memory for the
|
||||||
|
* audio samples of the frame.
|
||||||
|
* Default channel layouts based on the number of channels
|
||||||
|
* are assumed for simplicity.
|
||||||
|
*/
|
||||||
|
output_frame->nb_samples = audio_stream->codec->frame_size;
|
||||||
|
output_frame->channel_layout = audio_output_context->channel_layout;
|
||||||
|
output_frame->channels = audio_output_context->channels;
|
||||||
|
output_frame->format = audio_output_context->sample_fmt;
|
||||||
|
output_frame->sample_rate = audio_output_context->sample_rate;
|
||||||
|
/**
|
||||||
|
* Allocate the samples of the created frame. This call will make
|
||||||
|
* sure that the audio frame can hold as many samples as specified.
|
||||||
|
*/
|
||||||
|
Debug(2, "getting buffer");
|
||||||
|
if (( ret = av_frame_get_buffer( output_frame, 0)) < 0) {
|
||||||
|
Error( "Couldnt allocate output frame buffer samples (error '%s')",
|
||||||
|
av_make_error_string(ret).c_str() );
|
||||||
|
Error("Frame: samples(%d) layout (%d) format(%d) rate(%d)", output_frame->nb_samples,
|
||||||
|
output_frame->channel_layout, output_frame->format , output_frame->sample_rate
|
||||||
|
);
|
||||||
|
av_frame_free(&input_frame);
|
||||||
|
av_frame_free(&output_frame);
|
||||||
|
zm_av_unref_packet(&opkt);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
/** Set a timestamp based on the sample rate for the container. */
|
||||||
|
if (output_frame) {
|
||||||
|
output_frame->pts = opkt.pts;
|
||||||
|
}
|
||||||
|
/**
|
||||||
|
* Encode the audio frame and store it in the temporary packet.
|
||||||
|
* The output audio stream encoder is used to do this.
|
||||||
|
*/
|
||||||
|
if (( ret = avcodec_encode_audio2( audio_output_context, &opkt,
|
||||||
|
input_frame, &data_present )) < 0) {
|
||||||
|
Error( "Could not encode frame (error '%s')",
|
||||||
|
av_make_error_string(ret).c_str());
|
||||||
|
zm_av_unref_packet(&opkt);
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
opkt.data = ipkt->data;
|
||||||
|
opkt.size = ipkt->size;
|
||||||
|
}
|
||||||
ret = av_interleaved_write_frame(oc, &opkt);
|
ret = av_interleaved_write_frame(oc, &opkt);
|
||||||
if(ret!=0){
|
if(ret!=0){
|
||||||
Fatal("Error encoding audio frame packet: %s\n", av_make_error_string(ret).c_str());
|
Fatal("Error encoding audio frame packet: %s\n", av_make_error_string(ret).c_str());
|
||||||
|
|
|
@ -15,6 +15,11 @@ private:
|
||||||
AVStream *video_stream;
|
AVStream *video_stream;
|
||||||
AVStream *audio_stream;
|
AVStream *audio_stream;
|
||||||
|
|
||||||
|
// The following are used when encoding the audio stream to AAC
|
||||||
|
AVCodec *audio_output_codec;
|
||||||
|
AVCodecContext *audio_output_context;
|
||||||
|
int data_present;
|
||||||
|
|
||||||
const char *filename;
|
const char *filename;
|
||||||
const char *format;
|
const char *format;
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue