From 966fbb1a10682f1e285d5f50c9a3666d65367e41 Mon Sep 17 00:00:00 2001 From: SteveGilvarry Date: Mon, 4 Apr 2016 20:04:14 +1000 Subject: [PATCH] Add Packet Queue --- src/CMakeLists.txt | 2 +- src/zm_ffmpeg_camera.cpp | 915 +++++++++++++++++++-------------------- src/zm_ffmpeg_camera.h | 1 + src/zm_packetqueue.cpp | 92 ++++ src/zm_packetqueue.h | 53 +++ 5 files changed, 593 insertions(+), 470 deletions(-) create mode 100644 src/zm_packetqueue.cpp create mode 100644 src/zm_packetqueue.h diff --git a/src/CMakeLists.txt b/src/CMakeLists.txt index e29717642..21844e12f 100644 --- a/src/CMakeLists.txt +++ b/src/CMakeLists.txt @@ -4,7 +4,7 @@ configure_file(zm_config.h.in "${CMAKE_CURRENT_BINARY_DIR}/zm_config.h" @ONLY) # Group together all the source files that are used by all the binaries (zmc, zma, zmu, zms etc) -set(ZM_BIN_SRC_FILES zm_box.cpp zm_buffer.cpp zm_camera.cpp zm_comms.cpp zm_config.cpp zm_coord.cpp zm_curl_camera.cpp zm.cpp zm_db.cpp zm_logger.cpp zm_event.cpp zm_exception.cpp zm_file_camera.cpp zm_ffmpeg_camera.cpp zm_image.cpp zm_jpeg.cpp zm_libvlc_camera.cpp zm_local_camera.cpp zm_monitor.cpp zm_ffmpeg.cpp zm_mpeg.cpp zm_poly.cpp zm_regexp.cpp zm_remote_camera.cpp zm_remote_camera_http.cpp zm_remote_camera_rtsp.cpp zm_rtp.cpp zm_rtp_ctrl.cpp zm_rtp_data.cpp zm_rtp_source.cpp zm_rtsp.cpp zm_rtsp_auth.cpp zm_sdp.cpp zm_signal.cpp zm_stream.cpp zm_thread.cpp zm_time.cpp zm_timer.cpp zm_user.cpp zm_utils.cpp zm_video.cpp zm_videostore.cpp zm_zone.cpp) +set(ZM_BIN_SRC_FILES zm_box.cpp zm_buffer.cpp zm_camera.cpp zm_comms.cpp zm_config.cpp zm_coord.cpp zm_curl_camera.cpp zm.cpp zm_db.cpp zm_logger.cpp zm_event.cpp zm_exception.cpp zm_file_camera.cpp zm_ffmpeg_camera.cpp zm_image.cpp zm_jpeg.cpp zm_libvlc_camera.cpp zm_local_camera.cpp zm_monitor.cpp zm_ffmpeg.cpp zm_mpeg.cpp zm_packetqueue.cpp zm_poly.cpp zm_regexp.cpp zm_remote_camera.cpp zm_remote_camera_http.cpp zm_remote_camera_rtsp.cpp zm_rtp.cpp zm_rtp_ctrl.cpp zm_rtp_data.cpp zm_rtp_source.cpp zm_rtsp.cpp zm_rtsp_auth.cpp zm_sdp.cpp zm_signal.cpp zm_stream.cpp zm_thread.cpp zm_time.cpp zm_timer.cpp zm_user.cpp zm_utils.cpp zm_video.cpp zm_videostore.cpp zm_zone.cpp) # A fix for cmake recompiling the source files for every target. add_library(zm STATIC ${ZM_BIN_SRC_FILES}) diff --git a/src/zm_ffmpeg_camera.cpp b/src/zm_ffmpeg_camera.cpp index 36bafaedf..19df58227 100755 --- a/src/zm_ffmpeg_camera.cpp +++ b/src/zm_ffmpeg_camera.cpp @@ -23,7 +23,8 @@ #include "zm_ffmpeg_camera.h" -extern "C"{ + +extern "C" { #include "libavutil/time.h" } #ifndef AV_ERROR_MAX_STRING_SIZE @@ -31,600 +32,576 @@ extern "C"{ #endif #ifdef SOLARIS -#include // for ESRCH +#include // for ESRCH #include #include #endif -FfmpegCamera::FfmpegCamera( int p_id, const std::string &p_path, const std::string &p_method, const std::string &p_options, int p_width, int p_height, int p_colours, int p_brightness, int p_contrast, int p_hue, int p_colour, bool p_capture ) : - Camera( p_id, FFMPEG_SRC, p_width, p_height, p_colours, ZM_SUBPIX_ORDER_DEFAULT_FOR_COLOUR(p_colours), p_brightness, p_contrast, p_hue, p_colour, p_capture ), - mPath( p_path ), - mMethod( p_method ), - mOptions( p_options ) -{ - if ( capture ) - { - Initialise(); - } - - mFormatContext = NULL; - mVideoStreamId = -1; - mAudioStreamId = -1; - mCodecContext = NULL; - mCodec = NULL; - mRawFrame = NULL; - mFrame = NULL; - frameCount = 0; - startTime=0; - mIsOpening = false; - mCanCapture = false; - mOpenStart = 0; - mReopenThread = 0; - wasRecording = false; - +FfmpegCamera::FfmpegCamera(int p_id, const std::string &p_path, const std::string &p_method, const std::string &p_options, int p_width, int p_height, int p_colours, int p_brightness, int p_contrast, int p_hue, int p_colour, bool p_capture) : + Camera(p_id, FFMPEG_SRC, p_width, p_height, p_colours, ZM_SUBPIX_ORDER_DEFAULT_FOR_COLOUR(p_colours), p_brightness, p_contrast, p_hue, p_colour, p_capture), + mPath(p_path), + mMethod(p_method), + mOptions(p_options) { + if (capture) { + Initialise(); + } + + mFormatContext = NULL; + mVideoStreamId = -1; + mAudioStreamId = -1; + mCodecContext = NULL; + mCodec = NULL; + mRawFrame = NULL; + mFrame = NULL; + frameCount = 0; + startTime = 0; + mIsOpening = false; + mCanCapture = false; + mOpenStart = 0; + mReopenThread = 0; + wasRecording = false; + #if HAVE_LIBSWSCALE - mConvertContext = NULL; + mConvertContext = NULL; #endif - /* Has to be located inside the constructor so other components such as zma will receive correct colours and subpixel order */ - if(colours == ZM_COLOUR_RGB32) { - subpixelorder = ZM_SUBPIX_ORDER_RGBA; - imagePixFormat = AV_PIX_FMT_RGBA; - } else if(colours == ZM_COLOUR_RGB24) { - subpixelorder = ZM_SUBPIX_ORDER_RGB; - imagePixFormat = AV_PIX_FMT_RGB24; - } else if(colours == ZM_COLOUR_GRAY8) { - subpixelorder = ZM_SUBPIX_ORDER_NONE; - imagePixFormat = AV_PIX_FMT_GRAY8; - } else { - Panic("Unexpected colours: %d",colours); - } - + /* Has to be located inside the constructor so other components such as zma will receive correct colours and subpixel order */ + if (colours == ZM_COLOUR_RGB32) { + subpixelorder = ZM_SUBPIX_ORDER_RGBA; + imagePixFormat = AV_PIX_FMT_RGBA; + } else if (colours == ZM_COLOUR_RGB24) { + subpixelorder = ZM_SUBPIX_ORDER_RGB; + imagePixFormat = AV_PIX_FMT_RGB24; + } else if (colours == ZM_COLOUR_GRAY8) { + subpixelorder = ZM_SUBPIX_ORDER_NONE; + imagePixFormat = AV_PIX_FMT_GRAY8; + } else { + Panic("Unexpected colours: %d", colours); + } + } -FfmpegCamera::~FfmpegCamera() -{ - CloseFfmpeg(); +FfmpegCamera::~FfmpegCamera() { + CloseFfmpeg(); - if ( capture ) - { - Terminate(); - } + if (capture) { + Terminate(); + } } -void FfmpegCamera::Initialise() -{ - if ( logDebugging() ) - av_log_set_level( AV_LOG_DEBUG ); - else - av_log_set_level( AV_LOG_QUIET ); +void FfmpegCamera::Initialise() { + if (logDebugging()) + av_log_set_level(AV_LOG_DEBUG); + else + av_log_set_level(AV_LOG_QUIET); - av_register_all(); + av_register_all(); } -void FfmpegCamera::Terminate() -{ +void FfmpegCamera::Terminate() { } -int FfmpegCamera::PrimeCapture() -{ - mVideoStreamId = -1; - mAudioStreamId = -1; - Info( "Priming capture from %s", mPath.c_str() ); +int FfmpegCamera::PrimeCapture() { + mVideoStreamId = -1; + mAudioStreamId = -1; + Info("Priming capture from %s", mPath.c_str()); - if (OpenFfmpeg() != 0){ + if (OpenFfmpeg() != 0) { + ReopenFfmpeg(); + } + return 0; +} + +int FfmpegCamera::PreCapture() { + // Nothing to do here + return ( 0); +} + +int FfmpegCamera::Capture(Image &image) { + if (!mCanCapture) { + return -1; + } + + // If the reopen thread has a value, but mCanCapture != 0, then we have just reopened the connection to the ffmpeg device, and we can clean up the thread. + if (mReopenThread != 0) { + void *retval = 0; + int ret; + + ret = pthread_join(mReopenThread, &retval); + if (ret != 0) { + Error("Could not join reopen thread."); + } + + Info("Successfully reopened stream."); + mReopenThread = 0; + } + + AVPacket packet; + uint8_t* directbuffer; + + /* Request a writeable buffer of the target image */ + directbuffer = image.WriteBuffer(width, height, colours, subpixelorder); + if (directbuffer == NULL) { + Error("Failed requesting writeable buffer for the captured image."); + return (-1); + } + + int frameComplete = false; + while (!frameComplete) { + int avResult = av_read_frame(mFormatContext, &packet); + if (avResult < 0) { + char errbuf[AV_ERROR_MAX_STRING_SIZE]; + av_strerror(avResult, errbuf, AV_ERROR_MAX_STRING_SIZE); + if ( + // Check if EOF. + (avResult == AVERROR_EOF || (mFormatContext->pb && mFormatContext->pb->eof_reached)) || + // Check for Connection failure. + (avResult == -110) + ) { + Info("av_read_frame returned \"%s\". Reopening stream.", errbuf); ReopenFfmpeg(); + } + + Error("Unable to read packet from stream %d: error %d \"%s\".", packet.stream_index, avResult, errbuf); + return ( -1); } - return 0; -} - -int FfmpegCamera::PreCapture() -{ - // Nothing to do here - return( 0 ); -} - -int FfmpegCamera::Capture( Image &image ) -{ - if (!mCanCapture){ - return -1; - } - - // If the reopen thread has a value, but mCanCapture != 0, then we have just reopened the connection to the ffmpeg device, and we can clean up the thread. - if (mReopenThread != 0) { - void *retval = 0; - int ret; - - ret = pthread_join(mReopenThread, &retval); - if (ret != 0){ - Error("Could not join reopen thread."); - } - - Info( "Successfully reopened stream." ); - mReopenThread = 0; - } - - AVPacket packet; - uint8_t* directbuffer; - - /* Request a writeable buffer of the target image */ - directbuffer = image.WriteBuffer(width, height, colours, subpixelorder); - if(directbuffer == NULL) { - Error("Failed requesting writeable buffer for the captured image."); - return (-1); - } - - int frameComplete = false; - while ( !frameComplete ) - { - int avResult = av_read_frame( mFormatContext, &packet ); - if ( avResult < 0 ) - { - char errbuf[AV_ERROR_MAX_STRING_SIZE]; - av_strerror(avResult, errbuf, AV_ERROR_MAX_STRING_SIZE); - if ( - // Check if EOF. - (avResult == AVERROR_EOF || (mFormatContext->pb && mFormatContext->pb->eof_reached)) || - // Check for Connection failure. - (avResult == -110) - ) - { - Info( "av_read_frame returned \"%s\". Reopening stream.", errbuf); - ReopenFfmpeg(); - } - - Error( "Unable to read packet from stream %d: error %d \"%s\".", packet.stream_index, avResult, errbuf ); - return( -1 ); - } - Debug( 5, "Got packet from stream %d", packet.stream_index ); - if ( packet.stream_index == mVideoStreamId ) - { + Debug(5, "Got packet from stream %d", packet.stream_index); + if (packet.stream_index == mVideoStreamId) { #if LIBAVCODEC_VERSION_CHECK(52, 23, 0, 23, 0) - if ( avcodec_decode_video2( mCodecContext, mRawFrame, &frameComplete, &packet ) < 0 ) + if (avcodec_decode_video2(mCodecContext, mRawFrame, &frameComplete, &packet) < 0) #else - if ( avcodec_decode_video( mCodecContext, mRawFrame, &frameComplete, packet.data, packet.size ) < 0 ) + if (avcodec_decode_video(mCodecContext, mRawFrame, &frameComplete, packet.data, packet.size) < 0) #endif - Fatal( "Unable to decode frame at frame %d", frameCount ); + Fatal("Unable to decode frame at frame %d", frameCount); - Debug( 4, "Decoded video packet at frame %d", frameCount ); + Debug(4, "Decoded video packet at frame %d", frameCount); - if ( frameComplete ) - { - Debug( 3, "Got frame %d", frameCount ); + if (frameComplete) { + Debug(3, "Got frame %d", frameCount); + + avpicture_fill((AVPicture *) mFrame, directbuffer, imagePixFormat, width, height); - avpicture_fill( (AVPicture *)mFrame, directbuffer, imagePixFormat, width, height); - #if HAVE_LIBSWSCALE - if(mConvertContext == NULL) { - mConvertContext = sws_getContext( mCodecContext->width, mCodecContext->height, mCodecContext->pix_fmt, width, height, imagePixFormat, SWS_BICUBIC, NULL, NULL, NULL ); + if (mConvertContext == NULL) { + mConvertContext = sws_getContext(mCodecContext->width, mCodecContext->height, mCodecContext->pix_fmt, width, height, imagePixFormat, SWS_BICUBIC, NULL, NULL, NULL); - if(mConvertContext == NULL) - Fatal( "Unable to create conversion context for %s", mPath.c_str() ); - } - - if ( sws_scale( mConvertContext, mRawFrame->data, mRawFrame->linesize, 0, mCodecContext->height, mFrame->data, mFrame->linesize ) < 0 ) - Fatal( "Unable to convert raw format %u to target format %u at frame %d", mCodecContext->pix_fmt, imagePixFormat, frameCount ); -#else // HAVE_LIBSWSCALE - Fatal( "You must compile ffmpeg with the --enable-swscale option to use ffmpeg cameras" ); -#endif // HAVE_LIBSWSCALE - - frameCount++; - } + if (mConvertContext == NULL) + Fatal("Unable to create conversion context for %s", mPath.c_str()); } -#if LIBAVCODEC_VERSION_CHECK(57, 8, 0, 12, 100) - av_packet_unref( &packet); -#else - av_free_packet( &packet ); -#endif + + if (sws_scale(mConvertContext, mRawFrame->data, mRawFrame->linesize, 0, mCodecContext->height, mFrame->data, mFrame->linesize) < 0) + Fatal("Unable to convert raw format %u to target format %u at frame %d", mCodecContext->pix_fmt, imagePixFormat, frameCount); +#else // HAVE_LIBSWSCALE + Fatal("You must compile ffmpeg with the --enable-swscale option to use ffmpeg cameras"); +#endif // HAVE_LIBSWSCALE + + frameCount++; + } } - return (0); +#if LIBAVCODEC_VERSION_CHECK(57, 8, 0, 12, 100) + av_packet_unref(&packet); +#else + av_free_packet(&packet); +#endif + } + return (0); } -int FfmpegCamera::PostCapture() -{ - // Nothing to do here - return( 0 ); +int FfmpegCamera::PostCapture() { + // Nothing to do here + return ( 0); } int FfmpegCamera::OpenFfmpeg() { - Debug ( 2, "OpenFfmpeg called." ); + Debug(2, "OpenFfmpeg called."); - mOpenStart = time(NULL); - mIsOpening = true; + mOpenStart = time(NULL); + mIsOpening = true; - // Open the input, not necessarily a file + // Open the input, not necessarily a file #if !LIBAVFORMAT_VERSION_CHECK(53, 2, 0, 4, 0) - Debug ( 1, "Calling av_open_input_file" ); - if ( av_open_input_file( &mFormatContext, mPath.c_str(), NULL, 0, NULL ) !=0 ) + Debug(1, "Calling av_open_input_file"); + if (av_open_input_file(&mFormatContext, mPath.c_str(), NULL, 0, NULL) != 0) #else - // Handle options - AVDictionary *opts = 0; - StringVector opVect = split(Options(), ","); - - // Set transport method as specified by method field, rtpUni is default - if ( Method() == "rtpMulti" ) - opVect.push_back("rtsp_transport=udp_multicast"); - else if ( Method() == "rtpRtsp" ) - opVect.push_back("rtsp_transport=tcp"); - else if ( Method() == "rtpRtspHttp" ) - opVect.push_back("rtsp_transport=http"); - - Debug(2, "Number of Options: %d",opVect.size()); - for (size_t i=0; i 1) { - parts[0] = trimSpaces(parts[0]); - parts[1] = trimSpaces(parts[1]); - if ( av_dict_set(&opts, parts[0].c_str(), parts[1].c_str(), 0) == 0 ) { - Debug(2, "set option %d '%s' to '%s'", i, parts[0].c_str(), parts[1].c_str()); - } - else - { - Warning( "Error trying to set option %d '%s' to '%s'", i, parts[0].c_str(), parts[1].c_str() ); - } - - } - } - Debug ( 1, "Calling avformat_open_input" ); + // Handle options + AVDictionary *opts = 0; + StringVector opVect = split(Options(), ","); - mFormatContext = avformat_alloc_context( ); - mFormatContext->interrupt_callback.callback = FfmpegInterruptCallback; - mFormatContext->interrupt_callback.opaque = this; + // Set transport method as specified by method field, rtpUni is default + if (Method() == "rtpMulti") + opVect.push_back("rtsp_transport=udp_multicast"); + else if (Method() == "rtpRtsp") + opVect.push_back("rtsp_transport=tcp"); + else if (Method() == "rtpRtspHttp") + opVect.push_back("rtsp_transport=http"); + + Debug(2, "Number of Options: %d", opVect.size()); + for (size_t i = 0; i < opVect.size(); i++) { + StringVector parts = split(opVect[i], "="); + if (parts.size() > 1) { + parts[0] = trimSpaces(parts[0]); + parts[1] = trimSpaces(parts[1]); + if (av_dict_set(&opts, parts[0].c_str(), parts[1].c_str(), 0) == 0) { + Debug(2, "set option %d '%s' to '%s'", i, parts[0].c_str(), parts[1].c_str()); + } else { + Warning("Error trying to set option %d '%s' to '%s'", i, parts[0].c_str(), parts[1].c_str()); + } - if ( avformat_open_input( &mFormatContext, mPath.c_str(), NULL, &opts ) !=0 ) -#endif - { - mIsOpening = false; - Error( "Unable to open input %s due to: %s", mPath.c_str(), strerror(errno) ); - return -1; } + } + Debug(1, "Calling avformat_open_input"); + mFormatContext = avformat_alloc_context(); + mFormatContext->interrupt_callback.callback = FfmpegInterruptCallback; + mFormatContext->interrupt_callback.opaque = this; + + if (avformat_open_input(&mFormatContext, mPath.c_str(), NULL, &opts) != 0) +#endif + { mIsOpening = false; - Debug ( 1, "Opened input" ); + Error("Unable to open input %s due to: %s", mPath.c_str(), strerror(errno)); + return -1; + } - Info( "Stream open %s", mPath.c_str() ); - startTime=av_gettime();//FIXME here or after find_Stream_info - - //FIXME can speed up initial analysis but need sensible parameters... - //mFormatContext->probesize = 32; - //mFormatContext->max_analyze_duration = 32; - // Locate stream info from avformat_open_input + mIsOpening = false; + Debug(1, "Opened input"); + + Info("Stream open %s", mPath.c_str()); + startTime = av_gettime(); //FIXME here or after find_Stream_info + + //FIXME can speed up initial analysis but need sensible parameters... + //mFormatContext->probesize = 32; + //mFormatContext->max_analyze_duration = 32; + // Locate stream info from avformat_open_input #if !LIBAVFORMAT_VERSION_CHECK(53, 6, 0, 6, 0) - Debug ( 1, "Calling av_find_stream_info" ); - if ( av_find_stream_info( mFormatContext ) < 0 ) + Debug(1, "Calling av_find_stream_info"); + if (av_find_stream_info(mFormatContext) < 0) #else - Debug ( 1, "Calling avformat_find_stream_info" ); - if ( avformat_find_stream_info( mFormatContext, 0 ) < 0 ) + Debug(1, "Calling avformat_find_stream_info"); + if (avformat_find_stream_info(mFormatContext, 0) < 0) #endif - Fatal( "Unable to find stream info from %s due to: %s", mPath.c_str(), strerror(errno) ); + Fatal("Unable to find stream info from %s due to: %s", mPath.c_str(), strerror(errno)); - Debug ( 1, "Got stream info" ); + Debug(1, "Got stream info"); - // Find first video stream present - mVideoStreamId = -1; - for (unsigned int i=0; i < mFormatContext->nb_streams; i++ ) + // Find first video stream present + mVideoStreamId = -1; + for (unsigned int i = 0; i < mFormatContext->nb_streams; i++) { +#if (LIBAVCODEC_VERSION_CHECK(52, 64, 0, 64, 0) || LIBAVUTIL_VERSION_CHECK(50, 14, 0, 14, 0)) + if (mFormatContext->streams[i]->codec->codec_type == AVMEDIA_TYPE_VIDEO) +#else + if (mFormatContext->streams[i]->codec->codec_type == CODEC_TYPE_VIDEO) +#endif + { + mVideoStreamId = i; + break; + } + if (mAudioStreamId == -1) //FIXME best way to copy all other streams? { #if (LIBAVCODEC_VERSION_CHECK(52, 64, 0, 64, 0) || LIBAVUTIL_VERSION_CHECK(50, 14, 0, 14, 0)) - if ( mFormatContext->streams[i]->codec->codec_type == AVMEDIA_TYPE_VIDEO ) + if (mFormatContext->streams[i]->codec->codec_type == AVMEDIA_TYPE_AUDIO) #else - if ( mFormatContext->streams[i]->codec->codec_type == CODEC_TYPE_VIDEO ) + if (mFormatContext->streams[i]->codec->codec_type == CODEC_TYPE_AUDIO) #endif - { - mVideoStreamId = i; - break; - } - if(mAudioStreamId == -1) //FIXME best way to copy all other streams? - { -#if (LIBAVCODEC_VERSION_CHECK(52, 64, 0, 64, 0) || LIBAVUTIL_VERSION_CHECK(50, 14, 0, 14, 0)) - if ( mFormatContext->streams[i]->codec->codec_type == AVMEDIA_TYPE_AUDIO ) -#else - if ( mFormatContext->streams[i]->codec->codec_type == CODEC_TYPE_AUDIO ) -#endif - { - mAudioStreamId = i; - } - } + { + mAudioStreamId = i; + } } - if ( mVideoStreamId == -1 ) - Fatal( "Unable to locate video stream in %s", mPath.c_str() ); + } + if (mVideoStreamId == -1) + Fatal("Unable to locate video stream in %s", mPath.c_str()); - Debug ( 1, "Found video stream" ); + Debug(1, "Found video stream"); - mCodecContext = mFormatContext->streams[mVideoStreamId]->codec; + mCodecContext = mFormatContext->streams[mVideoStreamId]->codec; - // Try and get the codec from the codec context - if ( (mCodec = avcodec_find_decoder( mCodecContext->codec_id )) == NULL ) - Fatal( "Can't find codec for video stream from %s", mPath.c_str() ); + // Try and get the codec from the codec context + if ((mCodec = avcodec_find_decoder(mCodecContext->codec_id)) == NULL) + Fatal("Can't find codec for video stream from %s", mPath.c_str()); - Debug ( 1, "Found decoder" ); + Debug(1, "Found decoder"); - // Open the codec + // Open the codec #if !LIBAVFORMAT_VERSION_CHECK(53, 8, 0, 8, 0) - Debug ( 1, "Calling avcodec_open" ); - if ( avcodec_open( mCodecContext, mCodec ) < 0 ) + Debug(1, "Calling avcodec_open"); + if (avcodec_open(mCodecContext, mCodec) < 0) #else - Debug ( 1, "Calling avcodec_open2" ); - if ( avcodec_open2( mCodecContext, mCodec, 0 ) < 0 ) + Debug(1, "Calling avcodec_open2"); + if (avcodec_open2(mCodecContext, mCodec, 0) < 0) #endif - Fatal( "Unable to open codec for video stream from %s", mPath.c_str() ); + Fatal("Unable to open codec for video stream from %s", mPath.c_str()); - Debug ( 1, "Opened codec" ); + Debug(1, "Opened codec"); - // Allocate space for the native video frame + // Allocate space for the native video frame #if LIBAVCODEC_VERSION_CHECK(55, 28, 1, 45, 101) - mRawFrame = av_frame_alloc(); + mRawFrame = av_frame_alloc(); #else - mRawFrame = avcodec_alloc_frame(); + mRawFrame = avcodec_alloc_frame(); #endif - // Allocate space for the converted video frame + // Allocate space for the converted video frame #if LIBAVCODEC_VERSION_CHECK(55, 28, 1, 45, 101) - mFrame = av_frame_alloc(); + mFrame = av_frame_alloc(); #else - mFrame = avcodec_alloc_frame(); + mFrame = avcodec_alloc_frame(); #endif - if(mRawFrame == NULL || mFrame == NULL) - Fatal( "Unable to allocate frame for %s", mPath.c_str() ); + if (mRawFrame == NULL || mFrame == NULL) + Fatal("Unable to allocate frame for %s", mPath.c_str()); - Debug ( 1, "Allocated frames" ); - - int pSize = avpicture_get_size( imagePixFormat, width, height ); - if( (unsigned int)pSize != imagesize) { - Fatal("Image size mismatch. Required: %d Available: %d",pSize,imagesize); - } + Debug(1, "Allocated frames"); + + int pSize = avpicture_get_size(imagePixFormat, width, height); + if ((unsigned int) pSize != imagesize) { + Fatal("Image size mismatch. Required: %d Available: %d", pSize, imagesize); + } + + Debug(1, "Validated imagesize"); - Debug ( 1, "Validated imagesize" ); - #if HAVE_LIBSWSCALE - Debug ( 1, "Calling sws_isSupportedInput" ); - if(!sws_isSupportedInput(mCodecContext->pix_fmt)) { - Fatal("swscale does not support the codec format: %c%c%c%c",(mCodecContext->pix_fmt)&0xff,((mCodecContext->pix_fmt>>8)&0xff),((mCodecContext->pix_fmt>>16)&0xff),((mCodecContext->pix_fmt>>24)&0xff)); - } - - if(!sws_isSupportedOutput(imagePixFormat)) { - Fatal("swscale does not support the target format: %c%c%c%c",(imagePixFormat)&0xff,((imagePixFormat>>8)&0xff),((imagePixFormat>>16)&0xff),((imagePixFormat>>24)&0xff)); - } - + Debug(1, "Calling sws_isSupportedInput"); + if (!sws_isSupportedInput(mCodecContext->pix_fmt)) { + Fatal("swscale does not support the codec format: %c%c%c%c", (mCodecContext->pix_fmt)&0xff, ((mCodecContext->pix_fmt >> 8)&0xff), ((mCodecContext->pix_fmt >> 16)&0xff), ((mCodecContext->pix_fmt >> 24)&0xff)); + } + + if (!sws_isSupportedOutput(imagePixFormat)) { + Fatal("swscale does not support the target format: %c%c%c%c", (imagePixFormat)&0xff, ((imagePixFormat >> 8)&0xff), ((imagePixFormat >> 16)&0xff), ((imagePixFormat >> 24)&0xff)); + } + #else // HAVE_LIBSWSCALE - Fatal( "You must compile ffmpeg with the --enable-swscale option to use ffmpeg cameras" ); + Fatal("You must compile ffmpeg with the --enable-swscale option to use ffmpeg cameras"); #endif // HAVE_LIBSWSCALE - mCanCapture = true; + mCanCapture = true; - return 0; + return 0; } int FfmpegCamera::ReopenFfmpeg() { - Debug(2, "ReopenFfmpeg called."); + Debug(2, "ReopenFfmpeg called."); - mCanCapture = false; - if (pthread_create( &mReopenThread, NULL, ReopenFfmpegThreadCallback, (void*) this) != 0){ - // Log a fatal error and exit the process. - Fatal( "ReopenFfmpeg failed to create worker thread." ); - } + mCanCapture = false; + if (pthread_create(&mReopenThread, NULL, ReopenFfmpegThreadCallback, (void*) this) != 0) { + // Log a fatal error and exit the process. + Fatal("ReopenFfmpeg failed to create worker thread."); + } - return 0; + return 0; } -int FfmpegCamera::CloseFfmpeg(){ +int FfmpegCamera::CloseFfmpeg() { - Debug(2, "CloseFfmpeg called."); + Debug(2, "CloseFfmpeg called."); - mCanCapture = false; + mCanCapture = false; #if LIBAVCODEC_VERSION_CHECK(55, 28, 1, 45, 101) - av_frame_free( &mFrame ); - av_frame_free( &mRawFrame ); + av_frame_free(&mFrame); + av_frame_free(&mRawFrame); #else - av_freep( &mFrame ); - av_freep( &mRawFrame ); + av_freep(&mFrame); + av_freep(&mRawFrame); #endif - + #if HAVE_LIBSWSCALE - if ( mConvertContext ) - { - sws_freeContext( mConvertContext ); - mConvertContext = NULL; - } + if (mConvertContext) { + sws_freeContext(mConvertContext); + mConvertContext = NULL; + } #endif - if ( mCodecContext ) - { - avcodec_close( mCodecContext ); - mCodecContext = NULL; // Freed by av_close_input_file - } - if ( mFormatContext ) - { + if (mCodecContext) { + avcodec_close(mCodecContext); + mCodecContext = NULL; // Freed by av_close_input_file + } + if (mFormatContext) { #if !LIBAVFORMAT_VERSION_CHECK(53, 17, 0, 25, 0) - av_close_input_file( mFormatContext ); + av_close_input_file(mFormatContext); #else - avformat_close_input( &mFormatContext ); + avformat_close_input(&mFormatContext); #endif - mFormatContext = NULL; - } + mFormatContext = NULL; + } - return 0; + return 0; } -int FfmpegCamera::FfmpegInterruptCallback(void *ctx) -{ - FfmpegCamera* camera = reinterpret_cast(ctx); - if (camera->mIsOpening){ - int now = time(NULL); - if ((now - camera->mOpenStart) > config.ffmpeg_open_timeout) { - Error ( "Open video took more than %d seconds.", config.ffmpeg_open_timeout ); - return 1; - } +int FfmpegCamera::FfmpegInterruptCallback(void *ctx) { + FfmpegCamera* camera = reinterpret_cast (ctx); + if (camera->mIsOpening) { + int now = time(NULL); + if ((now - camera->mOpenStart) > config.ffmpeg_open_timeout) { + Error("Open video took more than %d seconds.", config.ffmpeg_open_timeout); + return 1; } + } - return 0; + return 0; } -void *FfmpegCamera::ReopenFfmpegThreadCallback(void *ctx){ - if (ctx == NULL) return NULL; +void *FfmpegCamera::ReopenFfmpegThreadCallback(void *ctx) { + if (ctx == NULL) return NULL; - FfmpegCamera* camera = reinterpret_cast(ctx); + FfmpegCamera* camera = reinterpret_cast (ctx); - while (1){ - // Close current stream. - camera->CloseFfmpeg(); + while (1) { + // Close current stream. + camera->CloseFfmpeg(); - // Sleep if necessary to not reconnect too fast. - int wait = config.ffmpeg_open_timeout - (time(NULL) - camera->mOpenStart); - wait = wait < 0 ? 0 : wait; - if (wait > 0){ - Debug( 1, "Sleeping %d seconds before reopening stream.", wait ); - sleep(wait); - } - - if (camera->OpenFfmpeg() == 0){ - return NULL; - } + // Sleep if necessary to not reconnect too fast. + int wait = config.ffmpeg_open_timeout - (time(NULL) - camera->mOpenStart); + wait = wait < 0 ? 0 : wait; + if (wait > 0) { + Debug(1, "Sleeping %d seconds before reopening stream.", wait); + sleep(wait); } + + if (camera->OpenFfmpeg() == 0) { + return NULL; + } + } } //Function to handle capture and store -int FfmpegCamera::CaptureAndRecord( Image &image, bool recording, char* event_file ) -{ - if (!mCanCapture){ - return -1; - } - - // If the reopen thread has a value, but mCanCapture != 0, then we have just reopened the connection to the ffmpeg device, and we can clean up the thread. - if (mReopenThread != 0) { - void *retval = 0; - int ret; - - ret = pthread_join(mReopenThread, &retval); - if (ret != 0){ - Error("Could not join reopen thread."); - } - - Info( "Successfully reopened stream." ); - mReopenThread = 0; + +int FfmpegCamera::CaptureAndRecord(Image &image, bool recording, char* event_file) { + if (!mCanCapture) { + return -1; + } + + // If the reopen thread has a value, but mCanCapture != 0, then we have just reopened the connection to the ffmpeg device, and we can clean up the thread. + if (mReopenThread != 0) { + void *retval = 0; + int ret; + + ret = pthread_join(mReopenThread, &retval); + if (ret != 0) { + Error("Could not join reopen thread."); } - AVPacket packet; - uint8_t* directbuffer; - - /* Request a writeable buffer of the target image */ - directbuffer = image.WriteBuffer(width, height, colours, subpixelorder); - if(directbuffer == NULL) { - Error("Failed requesting writeable buffer for the captured image."); - return (-1); - } - - int frameComplete = false; - while ( !frameComplete ) - { - int avResult = av_read_frame( mFormatContext, &packet ); - if ( avResult < 0 ) - { - char errbuf[AV_ERROR_MAX_STRING_SIZE]; - av_strerror(avResult, errbuf, AV_ERROR_MAX_STRING_SIZE); - if ( - // Check if EOF. - (avResult == AVERROR_EOF || (mFormatContext->pb && mFormatContext->pb->eof_reached)) || - // Check for Connection failure. - (avResult == -110) - ) - { - Info( "av_read_frame returned \"%s\". Reopening stream.", errbuf); - ReopenFfmpeg(); - } + Info("Successfully reopened stream."); + mReopenThread = 0; + } - Error( "Unable to read packet from stream %d: error %d \"%s\".", packet.stream_index, avResult, errbuf ); - return( -1 ); - } - Debug( 5, "Got packet from stream %d", packet.stream_index ); - if ( packet.stream_index == mVideoStreamId ) - { + AVPacket packet; + uint8_t* directbuffer; + zm_packetqueue packetqueue; + + /* Request a writeable buffer of the target image */ + directbuffer = image.WriteBuffer(width, height, colours, subpixelorder); + if (directbuffer == NULL) { + Error("Failed requesting writeable buffer for the captured image."); + return (-1); + } + + int frameComplete = false; + while (!frameComplete) { + int avResult = av_read_frame(mFormatContext, &packet); + if (avResult < 0) { + char errbuf[AV_ERROR_MAX_STRING_SIZE]; + av_strerror(avResult, errbuf, AV_ERROR_MAX_STRING_SIZE); + if ( + // Check if EOF. + (avResult == AVERROR_EOF || (mFormatContext->pb && mFormatContext->pb->eof_reached)) || + // Check for Connection failure. + (avResult == -110) + ) { + Info("av_read_frame returned \"%s\". Reopening stream.", errbuf); + ReopenFfmpeg(); + } + + Error("Unable to read packet from stream %d: error %d \"%s\".", packet.stream_index, avResult, errbuf); + return ( -1); + } + Debug(5, "Got packet from stream %d", packet.stream_index); + if (packet.stream_index == mVideoStreamId) { #if LIBAVCODEC_VERSION_CHECK(52, 23, 0, 23, 0) - if ( avcodec_decode_video2( mCodecContext, mRawFrame, &frameComplete, &packet ) < 0 ) + if (avcodec_decode_video2(mCodecContext, mRawFrame, &frameComplete, &packet) < 0) #else - if ( avcodec_decode_video( mCodecContext, mRawFrame, &frameComplete, packet.data, packet.size ) < 0 ) + if (avcodec_decode_video(mCodecContext, mRawFrame, &frameComplete, packet.data, packet.size) < 0) #endif - Fatal( "Unable to decode frame at frame %d", frameCount ); + Fatal("Unable to decode frame at frame %d", frameCount); - Debug( 4, "Decoded video packet at frame %d", frameCount ); + Debug(4, "Decoded video packet at frame %d", frameCount); - if ( frameComplete ) - { - Debug( 3, "Got frame %d", frameCount ); - - avpicture_fill( (AVPicture *)mFrame, directbuffer, imagePixFormat, width, height); + if (frameComplete) { + Debug(3, "Got frame %d", frameCount); - //Keep the last keyframe so we can establish immediate video - /*if(packet.flags & AV_PKT_FLAG_KEY) - av_copy_packet(&lastKeyframePkt, &packet);*/ - //TODO I think we need to store the key frame location for seeking as part of the event - - //Video recording - if(recording && !wasRecording){ - //Instantiate the video storage module + avpicture_fill((AVPicture *) mFrame, directbuffer, imagePixFormat, width, height); - videoStore = new VideoStore((const char *)event_file, "mp4", mFormatContext->streams[mVideoStreamId],mAudioStreamId==-1?NULL:mFormatContext->streams[mAudioStreamId],startTime); - wasRecording = true; - strcpy(oldDirectory, event_file); - - }else if(!recording && wasRecording && videoStore){ - Info("Deleting videoStore instance"); - delete videoStore; - videoStore = NULL; - } - - //The directory we are recording to is no longer tied to the current event. Need to re-init the videostore with the correct directory and start recording again - if(recording && wasRecording && (strcmp(oldDirectory, event_file)!=0) && (packet.flags & AV_PKT_FLAG_KEY) ){ //don't open new videostore until we're on a key frame..would this require an offset adjustment for the event as a result?...if we store our key frame location with the event will that be enough? - Info("Re-starting video storage module"); - if(videoStore){ - delete videoStore; - videoStore = NULL; - } - - videoStore = new VideoStore((const char *)event_file, "mp4", mFormatContext->streams[mVideoStreamId],mAudioStreamId==-1?NULL:mFormatContext->streams[mAudioStreamId],startTime); - strcpy(oldDirectory, event_file); - } - - if(videoStore && recording){ - //Write the packet to our video store - int ret = videoStore->writeVideoFramePacket(&packet, mFormatContext->streams[mVideoStreamId]);//, &lastKeyframePkt); - if(ret<0){//Less than zero and we skipped a frame - av_free_packet( &packet ); - return 0; - } - } - -#if HAVE_LIBSWSCALE - if(mConvertContext == NULL) { - mConvertContext = sws_getContext( mCodecContext->width, mCodecContext->height, mCodecContext->pix_fmt, width, height, imagePixFormat, SWS_BICUBIC, NULL, NULL, NULL ); - if(mConvertContext == NULL) - Fatal( "Unable to create conversion context for %s", mPath.c_str() ); - } - - if ( sws_scale( mConvertContext, mRawFrame->data, mRawFrame->linesize, 0, mCodecContext->height, mFrame->data, mFrame->linesize ) < 0 ) - Fatal( "Unable to convert raw format %u to target format %u at frame %d", mCodecContext->pix_fmt, imagePixFormat, frameCount ); -#else // HAVE_LIBSWSCALE - Fatal( "You must compile ffmpeg with the --enable-swscale option to use ffmpeg cameras" ); -#endif // HAVE_LIBSWSCALE - - frameCount++; - } - }else if(packet.stream_index == mAudioStreamId){//FIXME best way to copy all other streams - if(videoStore && recording){ - //Write the packet to our video store - int ret = videoStore->writeAudioFramePacket(&packet, mFormatContext->streams[packet.stream_index]); //FIXME no relevance of last key frame - if(ret<0){//Less than zero and we skipped a frame - av_free_packet( &packet ); - return 0; - } - } + //Buffer video packets + if (!recording) { + if(packet.flags & AV_PKT_FLAG_KEY){ + packetqueue.clearQueues(); + } + packetqueue.queueVideoPacket(&packet); } - av_free_packet( &packet ); + + //Video recording + if (recording && !wasRecording) { + //Instantiate the video storage module + + videoStore = new VideoStore((const char *) event_file, "mp4", mFormatContext->streams[mVideoStreamId], mAudioStreamId == -1 ? NULL : mFormatContext->streams[mAudioStreamId], startTime); + wasRecording = true; + strcpy(oldDirectory, event_file); + + } else if (!recording && wasRecording && videoStore) { + Info("Deleting videoStore instance"); + delete videoStore; + videoStore = NULL; + } + + //The directory we are recording to is no longer tied to the current event. Need to re-init the videostore with the correct directory and start recording again + if (recording && wasRecording && (strcmp(oldDirectory, event_file) != 0) && (packet.flags & AV_PKT_FLAG_KEY)) { //don't open new videostore until we're on a key frame..would this require an offset adjustment for the event as a result?...if we store our key frame location with the event will that be enough? + Info("Re-starting video storage module"); + if (videoStore) { + delete videoStore; + videoStore = NULL; + } + + videoStore = new VideoStore((const char *) event_file, "mp4", mFormatContext->streams[mVideoStreamId], mAudioStreamId == -1 ? NULL : mFormatContext->streams[mAudioStreamId], startTime); + strcpy(oldDirectory, event_file); + } + + if (videoStore && recording) { + //Write the packet to our video store + int ret = videoStore->writeVideoFramePacket(&packet, mFormatContext->streams[mVideoStreamId]); //, &lastKeyframePkt); + if (ret < 0) {//Less than zero and we skipped a frame + av_free_packet(&packet); + return 0; + } + } + +#if HAVE_LIBSWSCALE + if (mConvertContext == NULL) { + mConvertContext = sws_getContext(mCodecContext->width, mCodecContext->height, mCodecContext->pix_fmt, width, height, imagePixFormat, SWS_BICUBIC, NULL, NULL, NULL); + if (mConvertContext == NULL) + Fatal("Unable to create conversion context for %s", mPath.c_str()); + } + + if (sws_scale(mConvertContext, mRawFrame->data, mRawFrame->linesize, 0, mCodecContext->height, mFrame->data, mFrame->linesize) < 0) + Fatal("Unable to convert raw format %u to target format %u at frame %d", mCodecContext->pix_fmt, imagePixFormat, frameCount); +#else // HAVE_LIBSWSCALE + Fatal("You must compile ffmpeg with the --enable-swscale option to use ffmpeg cameras"); +#endif // HAVE_LIBSWSCALE + + frameCount++; + } + } else if (packet.stream_index == mAudioStreamId) {//FIXME best way to copy all other streams + if (videoStore && recording) { + //Write the packet to our video store + int ret = videoStore->writeAudioFramePacket(&packet, mFormatContext->streams[packet.stream_index]); //FIXME no relevance of last key frame + if (ret < 0) {//Less than zero and we skipped a frame + av_free_packet(&packet); + return 0; + } + } } - return (frameCount); + av_free_packet(&packet); + } + return (frameCount); } diff --git a/src/zm_ffmpeg_camera.h b/src/zm_ffmpeg_camera.h index 46aa40210..4f38b4f6b 100755 --- a/src/zm_ffmpeg_camera.h +++ b/src/zm_ffmpeg_camera.h @@ -26,6 +26,7 @@ //#include "zm_utils.h" #include "zm_ffmpeg.h" #include "zm_videostore.h" +#include "zm_packetqueue.h" // // Class representing 'ffmpeg' cameras, i.e. those which are diff --git a/src/zm_packetqueue.cpp b/src/zm_packetqueue.cpp new file mode 100644 index 000000000..ec37e7139 --- /dev/null +++ b/src/zm_packetqueue.cpp @@ -0,0 +1,92 @@ +//ZoneMinder Packet Queue Implementation Class +//Copyright 2016 Steve Gilvarry +// +//This file is part of ZoneMinder. +// +//ZoneMinder is free software: you can redistribute it and/or modify +//it under the terms of the GNU General Public License as published by +//the Free Software Foundation, either version 3 of the License, or +//(at your option) any later version. +// +//ZoneMinder is distributed in the hope that it will be useful, +//but WITHOUT ANY WARRANTY; without even the implied warranty of +//MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +//GNU General Public License for more details. +// +//You should have received a copy of the GNU General Public License +//along with ZoneMinder. If not, see . + + +#include "zm_packetqueue.h" + +#define VIDEO_QUEUESIZE 200 +#define AUDIO_QUEUESIZE 50 + +using namespace std; + +zm_packetqueue::zm_packetqueue() { +} + +zm_packetqueue::zm_packetqueue(const zm_packetqueue& orig) { +} + +zm_packetqueue::~zm_packetqueue() { +} + +bool zm_packetqueue::queueVideoPacket(AVPacket* packet){ + return queuePacket(m_VideoQueue, packet); +} + +bool zm_packetqueue::queueAudioPacket(AVPacket* packet) +{ + return queuePacket(m_AudioQueue, packet); +} + +bool zm_packetqueue::queuePacket(queue& packetQueue, AVPacket* packet){ + + AVPacket input_ref = { 0 }; + if (av_packet_ref(&input_ref, packet) < 0){ + return false; + } + packetQueue.push(*packet); + + return true; +} + +bool zm_packetqueue::popPacket(queue& packetQueue, AVPacket* packet) +{ + if (packetQueue.empty()) + { + return false; + } + + *packet = packetQueue.front(); + packetQueue.pop(); + + return true; +} + +void zm_packetqueue::clearQueue(std::queue& packetQueue) +{ + while(!packetQueue.empty()) + { + packetQueue.pop(); + } +} + +void zm_packetqueue::clearQueues() +{ + clearQueue(m_VideoQueue); + clearQueue(m_AudioQueue); +} + +bool zm_packetqueue::popAudioPacket(AVPacket* packet) +{ + return popPacket(m_AudioQueue, packet); +} + +bool zm_packetqueue::popVideoPacket(AVPacket* packet) +{ + return popPacket(m_VideoQueue, packet); +} + diff --git a/src/zm_packetqueue.h b/src/zm_packetqueue.h new file mode 100644 index 000000000..17fe0e9e5 --- /dev/null +++ b/src/zm_packetqueue.h @@ -0,0 +1,53 @@ +//ZoneMinder Packet Queue Interface Class +//Copyright 2016 Steve Gilvarry +// +//This file is part of ZoneMinder. +// +//ZoneMinder is free software: you can redistribute it and/or modify +//it under the terms of the GNU General Public License as published by +//the Free Software Foundation, either version 3 of the License, or +//(at your option) any later version. +// +//ZoneMinder is distributed in the hope that it will be useful, +//but WITHOUT ANY WARRANTY; without even the implied warranty of +//MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +//GNU General Public License for more details. +// +//You should have received a copy of the GNU General Public License +//along with ZoneMinder. If not, see . + + +#ifndef ZM_PACKETQUEUE_H +#define ZM_PACKETQUEUE_H + +#include + +extern "C" { +#include +} + +class zm_packetqueue { +public: + zm_packetqueue(); + zm_packetqueue(const zm_packetqueue& orig); + virtual ~zm_packetqueue(); + bool queuePacket(std::queue& packetQueue, AVPacket* packet); + bool queueVideoPacket(AVPacket* packet); + bool queueAudioPacket(AVPacket* packet); + bool popPacket(std::queue& packetQueue, AVPacket* packet); + bool popVideoPacket(AVPacket* packet); + bool popAudioPacket(AVPacket* packet); + void clearQueues(); + void clearQueue(std::queue& packetQueue); +private: + int m_MaxVideoQueueSize; + int m_MaxAudioQueueSize; + std::queue m_VideoQueue; + std::queue m_AudioQueue; + +}; + + + +#endif /* ZM_PACKETQUEUE_H */ +