Merge branch 'master' of github.com:ZoneMinder/zoneminder
This commit is contained in:
commit
b88e699d7e
|
@ -293,18 +293,16 @@ int FfmpegCamera::OpenFfmpeg() {
|
||||||
mFormatContext->interrupt_callback.opaque = this;
|
mFormatContext->interrupt_callback.opaque = this;
|
||||||
|
|
||||||
ret = avformat_open_input(&mFormatContext, mPath.c_str(), nullptr, &opts);
|
ret = avformat_open_input(&mFormatContext, mPath.c_str(), nullptr, &opts);
|
||||||
if ( ret != 0 )
|
if (ret != 0) {
|
||||||
{
|
|
||||||
logPrintf(Logger::ERROR + monitor->Importance(),
|
logPrintf(Logger::ERROR + monitor->Importance(),
|
||||||
"Unable to open input %s due to: %s", mPath.c_str(),
|
"Unable to open input %s due to: %s", mPath.c_str(),
|
||||||
av_make_error_string(ret).c_str());
|
av_make_error_string(ret).c_str());
|
||||||
|
|
||||||
if ( mFormatContext ) {
|
if (mFormatContext) {
|
||||||
avformat_close_input(&mFormatContext);
|
avformat_close_input(&mFormatContext);
|
||||||
mFormatContext = nullptr;
|
mFormatContext = nullptr;
|
||||||
}
|
}
|
||||||
av_dict_free(&opts);
|
av_dict_free(&opts);
|
||||||
|
|
||||||
return -1;
|
return -1;
|
||||||
}
|
}
|
||||||
AVDictionaryEntry *e = nullptr;
|
AVDictionaryEntry *e = nullptr;
|
||||||
|
|
|
@ -3306,9 +3306,9 @@ void neon32_armv7_fastblend(const uint8_t* col1, const uint8_t* col2, uint8_t* r
|
||||||
__attribute__((noinline)) void neon64_armv8_fastblend(const uint8_t* col1, const uint8_t* col2, uint8_t* result, unsigned long count, double blendpercent) {
|
__attribute__((noinline)) void neon64_armv8_fastblend(const uint8_t* col1, const uint8_t* col2, uint8_t* result, unsigned long count, double blendpercent) {
|
||||||
#if (defined(__aarch64__) && !defined(ZM_STRIP_NEON))
|
#if (defined(__aarch64__) && !defined(ZM_STRIP_NEON))
|
||||||
static double current_blendpercent = 0.0;
|
static double current_blendpercent = 0.0;
|
||||||
|
static int8_t divider = 0;
|
||||||
|
|
||||||
if (current_blendpercent != blendpercent) {
|
if (current_blendpercent != blendpercent) {
|
||||||
static int8_t divider = 0;
|
|
||||||
/* Attempt to match the blending percent to one of the possible values */
|
/* Attempt to match the blending percent to one of the possible values */
|
||||||
if(blendpercent < 2.34375) {
|
if(blendpercent < 2.34375) {
|
||||||
// 1.5625% blending
|
// 1.5625% blending
|
||||||
|
|
|
@ -487,7 +487,7 @@ void Monitor::Load(MYSQL_ROW dbrow, bool load_zones=true, Purpose p = QUERY) {
|
||||||
} else if ( ! strcmp(dbrow[col], "Libvlc") ) {
|
} else if ( ! strcmp(dbrow[col], "Libvlc") ) {
|
||||||
type = LIBVLC;
|
type = LIBVLC;
|
||||||
} else if ( ! strcmp(dbrow[col], "cURL") ) {
|
} else if ( ! strcmp(dbrow[col], "cURL") ) {
|
||||||
type = CURL;
|
type = LIBCURL;
|
||||||
} else if ( ! strcmp(dbrow[col], "VNC") ) {
|
} else if ( ! strcmp(dbrow[col], "VNC") ) {
|
||||||
type = VNC;
|
type = VNC;
|
||||||
} else {
|
} else {
|
||||||
|
@ -859,7 +859,7 @@ void Monitor::LoadCamera() {
|
||||||
#endif // HAVE_LIBVLC
|
#endif // HAVE_LIBVLC
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
case CURL: {
|
case LIBCURL: {
|
||||||
#if HAVE_LIBCURL
|
#if HAVE_LIBCURL
|
||||||
camera = zm::make_unique<cURLCamera>(this,
|
camera = zm::make_unique<cURLCamera>(this,
|
||||||
path.c_str(),
|
path.c_str(),
|
||||||
|
|
|
@ -76,7 +76,7 @@ public:
|
||||||
FILE,
|
FILE,
|
||||||
FFMPEG,
|
FFMPEG,
|
||||||
LIBVLC,
|
LIBVLC,
|
||||||
CURL,
|
LIBCURL,
|
||||||
NVSOCKET,
|
NVSOCKET,
|
||||||
VNC,
|
VNC,
|
||||||
} CameraType;
|
} CameraType;
|
||||||
|
|
182
src/zm_mpeg.cpp
182
src/zm_mpeg.cpp
|
@ -1,17 +1,17 @@
|
||||||
/*
|
/*
|
||||||
* ZoneMinder MPEG class implementation, $Date$, $Revision$
|
* ZoneMinder MPEG class implementation, $Date$, $Revision$
|
||||||
* Copyright (C) 2001-2008 Philip Coombes
|
* Copyright (C) 2001-2008 Philip Coombes
|
||||||
*
|
*
|
||||||
* This program is free software; you can redistribute it and/or
|
* This program is free software; you can redistribute it and/or
|
||||||
* modify it under the terms of the GNU General Public License
|
* modify it under the terms of the GNU General Public License
|
||||||
* as published by the Free Software Foundation; either version 2
|
* as published by the Free Software Foundation; either version 2
|
||||||
* of the License, or (at your option) any later version.
|
* of the License, or (at your option) any later version.
|
||||||
*
|
*
|
||||||
* This program is distributed in the hope that it will be useful,
|
* This program is distributed in the hope that it will be useful,
|
||||||
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
||||||
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
||||||
* GNU General Public License for more details.
|
* GNU General Public License for more details.
|
||||||
*
|
*
|
||||||
* You should have received a copy of the GNU General Public License
|
* You should have received a copy of the GNU General Public License
|
||||||
* along with this program; if not, write to the Free Software
|
* along with this program; if not, write to the Free Software
|
||||||
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
|
* Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA.
|
||||||
|
@ -42,19 +42,26 @@ void VideoStream::SetupFormat( ) {
|
||||||
ofc = nullptr;
|
ofc = nullptr;
|
||||||
avformat_alloc_output_context2(&ofc, nullptr, format, filename);
|
avformat_alloc_output_context2(&ofc, nullptr, format, filename);
|
||||||
|
|
||||||
if ( !ofc ) {
|
if (!ofc) {
|
||||||
Fatal("avformat_alloc_..._context failed");
|
Fatal("avformat_alloc_..._context failed");
|
||||||
}
|
}
|
||||||
|
|
||||||
of = ofc->oformat;
|
of = ofc->oformat;
|
||||||
Debug(1, "Using output format: %s (%s)", of->name, of->long_name);
|
Debug(1, "Using output format: %s (%s)", of->name, of->long_name);
|
||||||
}
|
}
|
||||||
|
|
||||||
void VideoStream::SetupCodec( int colours, int subpixelorder, int width, int height, int bitrate, double frame_rate ) {
|
int VideoStream::SetupCodec(
|
||||||
|
int colours,
|
||||||
|
int subpixelorder,
|
||||||
|
int width,
|
||||||
|
int height,
|
||||||
|
int bitrate,
|
||||||
|
double frame_rate
|
||||||
|
) {
|
||||||
/* ffmpeg format matching */
|
/* ffmpeg format matching */
|
||||||
switch ( colours ) {
|
switch (colours) {
|
||||||
case ZM_COLOUR_RGB24:
|
case ZM_COLOUR_RGB24:
|
||||||
if ( subpixelorder == ZM_SUBPIX_ORDER_BGR ) {
|
if (subpixelorder == ZM_SUBPIX_ORDER_BGR) {
|
||||||
/* BGR subpixel order */
|
/* BGR subpixel order */
|
||||||
pf = AV_PIX_FMT_BGR24;
|
pf = AV_PIX_FMT_BGR24;
|
||||||
} else {
|
} else {
|
||||||
|
@ -63,13 +70,13 @@ void VideoStream::SetupCodec( int colours, int subpixelorder, int width, int hei
|
||||||
}
|
}
|
||||||
break;
|
break;
|
||||||
case ZM_COLOUR_RGB32:
|
case ZM_COLOUR_RGB32:
|
||||||
if ( subpixelorder == ZM_SUBPIX_ORDER_ARGB ) {
|
if (subpixelorder == ZM_SUBPIX_ORDER_ARGB) {
|
||||||
/* ARGB subpixel order */
|
/* ARGB subpixel order */
|
||||||
pf = AV_PIX_FMT_ARGB;
|
pf = AV_PIX_FMT_ARGB;
|
||||||
} else if ( subpixelorder == ZM_SUBPIX_ORDER_ABGR ) {
|
} else if (subpixelorder == ZM_SUBPIX_ORDER_ABGR) {
|
||||||
/* ABGR subpixel order */
|
/* ABGR subpixel order */
|
||||||
pf = AV_PIX_FMT_ABGR;
|
pf = AV_PIX_FMT_ABGR;
|
||||||
} else if ( subpixelorder == ZM_SUBPIX_ORDER_BGRA ) {
|
} else if (subpixelorder == ZM_SUBPIX_ORDER_BGRA) {
|
||||||
/* BGRA subpixel order */
|
/* BGRA subpixel order */
|
||||||
pf = AV_PIX_FMT_BGRA;
|
pf = AV_PIX_FMT_BGRA;
|
||||||
} else {
|
} else {
|
||||||
|
@ -85,22 +92,22 @@ void VideoStream::SetupCodec( int colours, int subpixelorder, int width, int hei
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
|
||||||
if ( strcmp("rtp", of->name) == 0 ) {
|
if (strcmp("rtp", of->name) == 0) {
|
||||||
// RTP must have a packet_size.
|
// RTP must have a packet_size.
|
||||||
// Not sure what this value should be really...
|
// Not sure what this value should be really...
|
||||||
ofc->packet_size = width*height;
|
ofc->packet_size = width*height;
|
||||||
Debug(1,"Setting packet_size to %d", ofc->packet_size);
|
Debug(1,"Setting packet_size to %d", ofc->packet_size);
|
||||||
|
|
||||||
if ( of->video_codec == AV_CODEC_ID_NONE ) {
|
if (of->video_codec == AV_CODEC_ID_NONE) {
|
||||||
// RTP does not have a default codec in ffmpeg <= 0.8.
|
// RTP does not have a default codec in ffmpeg <= 0.8.
|
||||||
of->video_codec = AV_CODEC_ID_MPEG4;
|
of->video_codec = AV_CODEC_ID_MPEG4;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
_AVCODECID codec_id = of->video_codec;
|
_AVCODECID codec_id = of->video_codec;
|
||||||
if ( codec_name ) {
|
if (codec_name) {
|
||||||
AVCodec *a = avcodec_find_encoder_by_name(codec_name);
|
AVCodec *a = avcodec_find_encoder_by_name(codec_name);
|
||||||
if ( a ) {
|
if (a) {
|
||||||
codec_id = a->id;
|
codec_id = a->id;
|
||||||
Debug(1, "Using codec \"%s\"", codec_name);
|
Debug(1, "Using codec \"%s\"", codec_name);
|
||||||
} else {
|
} else {
|
||||||
|
@ -111,31 +118,29 @@ void VideoStream::SetupCodec( int colours, int subpixelorder, int width, int hei
|
||||||
/* add the video streams using the default format codecs
|
/* add the video streams using the default format codecs
|
||||||
and initialize the codecs */
|
and initialize the codecs */
|
||||||
ost = nullptr;
|
ost = nullptr;
|
||||||
if ( codec_id != AV_CODEC_ID_NONE ) {
|
if (codec_id != AV_CODEC_ID_NONE) {
|
||||||
codec = avcodec_find_encoder(codec_id);
|
codec = avcodec_find_encoder(codec_id);
|
||||||
if ( !codec ) {
|
if (!codec) {
|
||||||
Fatal("Could not find encoder for '%s'", avcodec_get_name(codec_id));
|
Error("Could not find encoder for '%s'", avcodec_get_name(codec_id));
|
||||||
|
return -1;
|
||||||
}
|
}
|
||||||
|
|
||||||
Debug(1, "Found encoder for '%s'", avcodec_get_name(codec_id));
|
Debug(1, "Found encoder for '%s'", avcodec_get_name(codec_id));
|
||||||
|
ost = avformat_new_stream(ofc, codec);
|
||||||
ost = avformat_new_stream( ofc, codec );
|
if (!ost) {
|
||||||
|
Error("Could not alloc stream");
|
||||||
if ( !ost ) {
|
return -1;
|
||||||
Fatal( "Could not alloc stream" );
|
|
||||||
return;
|
|
||||||
}
|
}
|
||||||
Debug( 1, "Allocated stream (%d) !=? (%d)", ost->id , ofc->nb_streams - 1 );
|
Debug(1, "Allocated stream (%d) !=? (%d)", ost->id , ofc->nb_streams - 1);
|
||||||
ost->id = ofc->nb_streams - 1;
|
ost->id = ofc->nb_streams - 1;
|
||||||
|
|
||||||
codec_context = avcodec_alloc_context3(nullptr);
|
codec_context = avcodec_alloc_context3(nullptr);
|
||||||
//avcodec_parameters_to_context(codec_context, ost->codecpar);
|
//avcodec_parameters_to_context(codec_context, ost->codecpar);
|
||||||
|
|
||||||
codec_context->codec_id = codec->id;
|
codec_context->codec_id = codec->id;
|
||||||
codec_context->codec_type = codec->type;
|
codec_context->codec_type = codec->type;
|
||||||
|
|
||||||
codec_context->pix_fmt = strcmp("mjpeg", ofc->oformat->name) == 0 ? AV_PIX_FMT_YUVJ422P : AV_PIX_FMT_YUV420P;
|
codec_context->pix_fmt = strcmp("mjpeg", ofc->oformat->name) == 0 ? AV_PIX_FMT_YUVJ422P : AV_PIX_FMT_YUV420P;
|
||||||
if ( bitrate <= 100 ) {
|
if (bitrate <= 100) {
|
||||||
// Quality based bitrate control (VBR). Scale is 1..31 where 1 is best.
|
// Quality based bitrate control (VBR). Scale is 1..31 where 1 is best.
|
||||||
// This gets rid of artifacts in the beginning of the movie; and well, even quality.
|
// This gets rid of artifacts in the beginning of the movie; and well, even quality.
|
||||||
codec_context->flags |= AV_CODEC_FLAG_QSCALE;
|
codec_context->flags |= AV_CODEC_FLAG_QSCALE;
|
||||||
|
@ -155,22 +160,22 @@ void VideoStream::SetupCodec( int colours, int subpixelorder, int width, int hei
|
||||||
codec_context->time_base.num = 1;
|
codec_context->time_base.num = 1;
|
||||||
ost->time_base.den = frame_rate;
|
ost->time_base.den = frame_rate;
|
||||||
ost->time_base.num = 1;
|
ost->time_base.num = 1;
|
||||||
|
|
||||||
|
|
||||||
Debug( 1, "Will encode in %d fps. %dx%d", codec_context->time_base.den, width, height );
|
Debug( 1, "Will encode in %d fps. %dx%d", codec_context->time_base.den, width, height );
|
||||||
|
|
||||||
/* emit one intra frame every second */
|
/* emit one intra frame every second */
|
||||||
codec_context->gop_size = frame_rate;
|
codec_context->gop_size = frame_rate;
|
||||||
|
|
||||||
// some formats want stream headers to be separate
|
// some formats want stream headers to be separate
|
||||||
if ( of->flags & AVFMT_GLOBALHEADER )
|
if (of->flags & AVFMT_GLOBALHEADER)
|
||||||
codec_context->flags |= AV_CODEC_FLAG_GLOBAL_HEADER;
|
codec_context->flags |= AV_CODEC_FLAG_GLOBAL_HEADER;
|
||||||
|
|
||||||
avcodec_parameters_from_context(ost->codecpar, codec_context);
|
avcodec_parameters_from_context(ost->codecpar, codec_context);
|
||||||
zm_dump_codecpar(ost->codecpar);
|
zm_dump_codecpar(ost->codecpar);
|
||||||
} else {
|
} else {
|
||||||
Fatal( "of->video_codec == AV_CODEC_ID_NONE" );
|
Error("of->video_codec == AV_CODEC_ID_NONE");
|
||||||
}
|
return -1;
|
||||||
|
}
|
||||||
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
void VideoStream::SetParameters( ) {
|
void VideoStream::SetParameters( ) {
|
||||||
|
@ -198,11 +203,11 @@ const char *VideoStream::MimeType() const {
|
||||||
bool VideoStream::OpenStream( ) {
|
bool VideoStream::OpenStream( ) {
|
||||||
int ret;
|
int ret;
|
||||||
|
|
||||||
/* now that all the parameters are set, we can open the
|
/* now that all the parameters are set, we can open the
|
||||||
video codecs and allocate the necessary encode buffers */
|
video codecs and allocate the necessary encode buffers */
|
||||||
if ( ost ) {
|
if ( ost ) {
|
||||||
Debug(1,"Opening codec");
|
Debug(1,"Opening codec");
|
||||||
|
|
||||||
/* open the codec */
|
/* open the codec */
|
||||||
|
|
||||||
if ((ret = avcodec_open2(codec_context, codec, nullptr)) < 0) {
|
if ((ret = avcodec_open2(codec_context, codec, nullptr)) < 0) {
|
||||||
|
@ -319,7 +324,7 @@ VideoStream::VideoStream( const char *in_filename, const char *in_format, int bi
|
||||||
if ( !initialised ) {
|
if ( !initialised ) {
|
||||||
Initialise( );
|
Initialise( );
|
||||||
}
|
}
|
||||||
|
|
||||||
if ( format ) {
|
if ( format ) {
|
||||||
int length = strlen(format);
|
int length = strlen(format);
|
||||||
codec_and_format = new char[length+1];;
|
codec_and_format = new char[length+1];;
|
||||||
|
@ -337,13 +342,13 @@ VideoStream::VideoStream( const char *in_filename, const char *in_format, int bi
|
||||||
SetupFormat( );
|
SetupFormat( );
|
||||||
SetupCodec( colours, subpixelorder, width, height, bitrate, frame_rate );
|
SetupCodec( colours, subpixelorder, width, height, bitrate, frame_rate );
|
||||||
SetParameters( );
|
SetParameters( );
|
||||||
|
|
||||||
// Allocate buffered packets.
|
// Allocate buffered packets.
|
||||||
packet_buffers = new AVPacket*[2];
|
packet_buffers = new AVPacket*[2];
|
||||||
packet_buffers[0] = new AVPacket();
|
packet_buffers[0] = new AVPacket();
|
||||||
packet_buffers[1] = new AVPacket();
|
packet_buffers[1] = new AVPacket();
|
||||||
packet_index = 0;
|
packet_index = 0;
|
||||||
|
|
||||||
// Initialize mutex used by streaming thread.
|
// Initialize mutex used by streaming thread.
|
||||||
if ( pthread_mutex_init( buffer_copy_lock, nullptr ) != 0 ) {
|
if ( pthread_mutex_init( buffer_copy_lock, nullptr ) != 0 ) {
|
||||||
Fatal("pthread_mutex_init failed");
|
Fatal("pthread_mutex_init failed");
|
||||||
|
@ -353,35 +358,35 @@ VideoStream::VideoStream( const char *in_filename, const char *in_format, int bi
|
||||||
|
|
||||||
VideoStream::~VideoStream( ) {
|
VideoStream::~VideoStream( ) {
|
||||||
Debug( 1, "VideoStream destructor." );
|
Debug( 1, "VideoStream destructor." );
|
||||||
|
|
||||||
// Stop streaming thread.
|
// Stop streaming thread.
|
||||||
if ( streaming_thread ) {
|
if ( streaming_thread ) {
|
||||||
do_streaming = false;
|
do_streaming = false;
|
||||||
void* thread_exit_code;
|
void* thread_exit_code;
|
||||||
|
|
||||||
Debug( 1, "Asking streaming thread to exit." );
|
Debug( 1, "Asking streaming thread to exit." );
|
||||||
|
|
||||||
// Wait for thread to exit.
|
// Wait for thread to exit.
|
||||||
pthread_join(streaming_thread, &thread_exit_code);
|
pthread_join(streaming_thread, &thread_exit_code);
|
||||||
}
|
}
|
||||||
|
|
||||||
if ( buffer_copy != nullptr ) {
|
if ( buffer_copy != nullptr ) {
|
||||||
av_free( buffer_copy );
|
av_free( buffer_copy );
|
||||||
}
|
}
|
||||||
|
|
||||||
if ( buffer_copy_lock ) {
|
if ( buffer_copy_lock ) {
|
||||||
if ( pthread_mutex_destroy( buffer_copy_lock ) != 0 ) {
|
if ( pthread_mutex_destroy( buffer_copy_lock ) != 0 ) {
|
||||||
Error( "pthread_mutex_destroy failed" );
|
Error( "pthread_mutex_destroy failed" );
|
||||||
}
|
}
|
||||||
delete buffer_copy_lock;
|
delete buffer_copy_lock;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (packet_buffers) {
|
if (packet_buffers) {
|
||||||
delete packet_buffers[0];
|
delete packet_buffers[0];
|
||||||
delete packet_buffers[1];
|
delete packet_buffers[1];
|
||||||
delete[] packet_buffers;
|
delete[] packet_buffers;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* close each codec */
|
/* close each codec */
|
||||||
if ( ost ) {
|
if ( ost ) {
|
||||||
avcodec_close( codec_context );
|
avcodec_close( codec_context );
|
||||||
|
@ -409,7 +414,7 @@ VideoStream::~VideoStream( ) {
|
||||||
|
|
||||||
/* free the stream */
|
/* free the stream */
|
||||||
av_free( ofc );
|
av_free( ofc );
|
||||||
|
|
||||||
/* free format and codec_name data. */
|
/* free format and codec_name data. */
|
||||||
if ( codec_and_format ) {
|
if ( codec_and_format ) {
|
||||||
delete codec_and_format;
|
delete codec_and_format;
|
||||||
|
@ -420,12 +425,12 @@ double VideoStream::EncodeFrame( const uint8_t *buffer, int buffer_size, bool _a
|
||||||
if ( pthread_mutex_lock(buffer_copy_lock) != 0 ) {
|
if ( pthread_mutex_lock(buffer_copy_lock) != 0 ) {
|
||||||
Fatal( "EncodeFrame: pthread_mutex_lock failed." );
|
Fatal( "EncodeFrame: pthread_mutex_lock failed." );
|
||||||
}
|
}
|
||||||
|
|
||||||
if (buffer_copy_size < buffer_size) {
|
if (buffer_copy_size < buffer_size) {
|
||||||
if ( buffer_copy ) {
|
if ( buffer_copy ) {
|
||||||
av_free(buffer_copy);
|
av_free(buffer_copy);
|
||||||
}
|
}
|
||||||
|
|
||||||
// Allocate a buffer to store source images for the streaming thread to encode.
|
// Allocate a buffer to store source images for the streaming thread to encode.
|
||||||
buffer_copy = (uint8_t *)av_malloc(buffer_size);
|
buffer_copy = (uint8_t *)av_malloc(buffer_size);
|
||||||
if ( !buffer_copy ) {
|
if ( !buffer_copy ) {
|
||||||
|
@ -435,35 +440,34 @@ double VideoStream::EncodeFrame( const uint8_t *buffer, int buffer_size, bool _a
|
||||||
}
|
}
|
||||||
buffer_copy_size = buffer_size;
|
buffer_copy_size = buffer_size;
|
||||||
}
|
}
|
||||||
|
|
||||||
add_timestamp = _add_timestamp;
|
add_timestamp = _add_timestamp;
|
||||||
timestamp = _timestamp;
|
timestamp = _timestamp;
|
||||||
buffer_copy_used = buffer_size;
|
buffer_copy_used = buffer_size;
|
||||||
memcpy(buffer_copy, buffer, buffer_size);
|
memcpy(buffer_copy, buffer, buffer_size);
|
||||||
|
|
||||||
if ( pthread_mutex_unlock(buffer_copy_lock) != 0 ) {
|
if ( pthread_mutex_unlock(buffer_copy_lock) != 0 ) {
|
||||||
Fatal( "EncodeFrame: pthread_mutex_unlock failed." );
|
Fatal( "EncodeFrame: pthread_mutex_unlock failed." );
|
||||||
}
|
}
|
||||||
|
|
||||||
if ( streaming_thread == 0 ) {
|
if ( streaming_thread == 0 ) {
|
||||||
Debug( 1, "Starting streaming thread" );
|
Debug( 1, "Starting streaming thread" );
|
||||||
|
|
||||||
// Start a thread for streaming encoded video.
|
// Start a thread for streaming encoded video.
|
||||||
if (pthread_create( &streaming_thread, nullptr, StreamingThreadCallback, (void*) this) != 0){
|
if (pthread_create( &streaming_thread, nullptr, StreamingThreadCallback, (void*) this) != 0){
|
||||||
// Log a fatal error and exit the process.
|
// Log a fatal error and exit the process.
|
||||||
Fatal( "VideoStream failed to create streaming thread." );
|
Fatal( "VideoStream failed to create streaming thread." );
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
//return ActuallyEncodeFrame( buffer, buffer_size, add_timestamp, timestamp);
|
//return ActuallyEncodeFrame( buffer, buffer_size, add_timestamp, timestamp);
|
||||||
|
|
||||||
return _timestamp;
|
return _timestamp;
|
||||||
}
|
}
|
||||||
|
|
||||||
double VideoStream::ActuallyEncodeFrame( const uint8_t *buffer, int buffer_size, bool add_timestamp, unsigned int timestamp ) {
|
double VideoStream::ActuallyEncodeFrame( const uint8_t *buffer, int buffer_size, bool add_timestamp, unsigned int timestamp ) {
|
||||||
|
|
||||||
if ( codec_context->pix_fmt != pf ) {
|
if ( codec_context->pix_fmt != pf ) {
|
||||||
static struct SwsContext *img_convert_ctx = nullptr;
|
static struct SwsContext *img_convert_ctx = nullptr;
|
||||||
memcpy( tmp_opicture->data[0], buffer, buffer_size );
|
memcpy( tmp_opicture->data[0], buffer, buffer_size );
|
||||||
if ( !img_convert_ctx ) {
|
if ( !img_convert_ctx ) {
|
||||||
img_convert_ctx = sws_getCachedContext( nullptr, codec_context->width, codec_context->height, pf, codec_context->width, codec_context->height, codec_context->pix_fmt, SWS_BICUBIC, nullptr, nullptr, nullptr );
|
img_convert_ctx = sws_getCachedContext( nullptr, codec_context->width, codec_context->height, pf, codec_context->width, codec_context->height, codec_context->pix_fmt, SWS_BICUBIC, nullptr, nullptr, nullptr );
|
||||||
|
@ -475,37 +479,36 @@ double VideoStream::ActuallyEncodeFrame( const uint8_t *buffer, int buffer_size,
|
||||||
memcpy( opicture->data[0], buffer, buffer_size );
|
memcpy( opicture->data[0], buffer, buffer_size );
|
||||||
}
|
}
|
||||||
AVFrame *opicture_ptr = opicture;
|
AVFrame *opicture_ptr = opicture;
|
||||||
|
|
||||||
AVPacket *pkt = packet_buffers[packet_index];
|
AVPacket *pkt = packet_buffers[packet_index];
|
||||||
av_init_packet( pkt );
|
av_init_packet( pkt );
|
||||||
int got_packet = 0;
|
int got_packet = 0;
|
||||||
if (codec_context->codec_type == AVMEDIA_TYPE_VIDEO &&
|
if (codec_context->codec_type == AVMEDIA_TYPE_VIDEO &&
|
||||||
codec_context->codec_id == AV_CODEC_ID_RAWVIDEO) {
|
codec_context->codec_id == AV_CODEC_ID_RAWVIDEO) {
|
||||||
pkt->flags |= AV_PKT_FLAG_KEY;
|
pkt->flags |= AV_PKT_FLAG_KEY;
|
||||||
pkt->stream_index = ost->index;
|
pkt->stream_index = ost->index;
|
||||||
pkt->data = (uint8_t *)opicture_ptr;
|
pkt->data = (uint8_t *)opicture_ptr;
|
||||||
pkt->size = sizeof (AVPicture);
|
pkt->size = sizeof (AVPicture);
|
||||||
got_packet = 1;
|
got_packet = 1;
|
||||||
} else {
|
} else {
|
||||||
opicture_ptr->pts = codec_context->frame_number;
|
opicture_ptr->pts = codec_context->frame_number;
|
||||||
opicture_ptr->quality = codec_context->global_quality;
|
opicture_ptr->quality = codec_context->global_quality;
|
||||||
|
|
||||||
avcodec_send_frame(codec_context, opicture_ptr);
|
avcodec_send_frame(codec_context, opicture_ptr);
|
||||||
int ret = avcodec_receive_packet(codec_context, pkt);
|
int ret = avcodec_receive_packet(codec_context, pkt);
|
||||||
if ( ret < 0 ) {
|
if (ret < 0) {
|
||||||
if ( AVERROR_EOF != ret ) {
|
if (AVERROR_EOF != ret) {
|
||||||
Error("ERror encoding video (%d) (%s)", ret,
|
Error("ERror encoding video (%d) (%s)", ret, av_err2str(ret));
|
||||||
av_err2str(ret));
|
|
||||||
}
|
|
||||||
} else {
|
|
||||||
got_packet = 1;
|
|
||||||
}
|
}
|
||||||
|
} else {
|
||||||
|
got_packet = 1;
|
||||||
|
}
|
||||||
|
|
||||||
if ( got_packet ) {
|
if (got_packet) {
|
||||||
// if ( c->coded_frame->key_frame )
|
// if ( c->coded_frame->key_frame )
|
||||||
// {
|
// {
|
||||||
// pkt->flags |= AV_PKT_FLAG_KEY;
|
// pkt->flags |= AV_PKT_FLAG_KEY;
|
||||||
// }
|
// }
|
||||||
|
|
||||||
if ( pkt->pts != (int64_t)AV_NOPTS_VALUE ) {
|
if ( pkt->pts != (int64_t)AV_NOPTS_VALUE ) {
|
||||||
pkt->pts = av_rescale_q( pkt->pts, codec_context->time_base, ost->time_base );
|
pkt->pts = av_rescale_q( pkt->pts, codec_context->time_base, ost->time_base );
|
||||||
|
@ -517,18 +520,17 @@ double VideoStream::ActuallyEncodeFrame( const uint8_t *buffer, int buffer_size,
|
||||||
pkt->stream_index = ost->index;
|
pkt->stream_index = ost->index;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
return ( opicture_ptr->pts);
|
return opicture_ptr->pts;
|
||||||
}
|
}
|
||||||
|
|
||||||
int VideoStream::SendPacket(AVPacket *packet) {
|
int VideoStream::SendPacket(AVPacket *packet) {
|
||||||
|
int ret = av_write_frame(ofc, packet);
|
||||||
int ret = av_write_frame( ofc, packet );
|
if (ret < 0) {
|
||||||
if ( ret != 0 ) {
|
Error("Error %d while writing video frame: %s", ret, av_err2str(errno));
|
||||||
Fatal( "Error %d while writing video frame: %s", ret, av_err2str( errno ) );
|
}
|
||||||
}
|
av_packet_unref(packet);
|
||||||
av_packet_unref( packet );
|
return ret;
|
||||||
return ret;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
void *VideoStream::StreamingThreadCallback(void *ctx) {
|
void *VideoStream::StreamingThreadCallback(void *ctx) {
|
||||||
|
|
|
@ -68,7 +68,7 @@ protected:
|
||||||
static void Initialise();
|
static void Initialise();
|
||||||
|
|
||||||
void SetupFormat( );
|
void SetupFormat( );
|
||||||
void SetupCodec( int colours, int subpixelorder, int width, int height, int bitrate, double frame_rate );
|
int SetupCodec( int colours, int subpixelorder, int width, int height, int bitrate, double frame_rate );
|
||||||
void SetParameters();
|
void SetParameters();
|
||||||
void ActuallyOpenStream();
|
void ActuallyOpenStream();
|
||||||
double ActuallyEncodeFrame( const uint8_t *buffer, int buffer_size, bool add_timestamp=false, unsigned int timestamp=0 );
|
double ActuallyEncodeFrame( const uint8_t *buffer, int buffer_size, bool add_timestamp=false, unsigned int timestamp=0 );
|
||||||
|
|
|
@ -21,7 +21,7 @@ class ZoneMinderDeviceSource;
|
||||||
|
|
||||||
class BaseServerMediaSubsession {
|
class BaseServerMediaSubsession {
|
||||||
public:
|
public:
|
||||||
BaseServerMediaSubsession(StreamReplicator* replicator):
|
explicit BaseServerMediaSubsession(StreamReplicator* replicator):
|
||||||
m_replicator(replicator) {};
|
m_replicator(replicator) {};
|
||||||
|
|
||||||
FramedSource* createSource(
|
FramedSource* createSource(
|
||||||
|
|
|
@ -157,7 +157,6 @@ Image *StreamBase::prepareImage(Image *image) {
|
||||||
int disp_image_width = (image->Width() * scale) / ZM_SCALE_BASE, disp_image_height = (image->Height() * scale) / ZM_SCALE_BASE;
|
int disp_image_width = (image->Width() * scale) / ZM_SCALE_BASE, disp_image_height = (image->Height() * scale) / ZM_SCALE_BASE;
|
||||||
int last_disp_image_width = (image->Width() * last_scale) / ZM_SCALE_BASE, last_disp_image_height = (image->Height() * last_scale) / ZM_SCALE_BASE;
|
int last_disp_image_width = (image->Width() * last_scale) / ZM_SCALE_BASE, last_disp_image_height = (image->Height() * last_scale) / ZM_SCALE_BASE;
|
||||||
int send_image_width = (disp_image_width * act_mag ) / mag, send_image_height = (disp_image_height * act_mag ) / mag;
|
int send_image_width = (disp_image_width * act_mag ) / mag, send_image_height = (disp_image_height * act_mag ) / mag;
|
||||||
int last_send_image_width = (last_disp_image_width * last_act_mag ) / last_mag, last_send_image_height = (last_disp_image_height * last_act_mag ) / last_mag;
|
|
||||||
|
|
||||||
Debug(3,
|
Debug(3,
|
||||||
"Scaling by %d, zooming by %d = magnifying by %d(%d)\n"
|
"Scaling by %d, zooming by %d = magnifying by %d(%d)\n"
|
||||||
|
@ -169,8 +168,7 @@ Image *StreamBase::prepareImage(Image *image) {
|
||||||
"Last actual image width = %d, height = %d\n"
|
"Last actual image width = %d, height = %d\n"
|
||||||
"Display image width = %d, height = %d\n"
|
"Display image width = %d, height = %d\n"
|
||||||
"Last display image width = %d, height = %d\n"
|
"Last display image width = %d, height = %d\n"
|
||||||
"Send image width = %d, height = %d\n"
|
"Send image width = %d, height = %d\n",
|
||||||
"Last send image width = %d, height = %d\n",
|
|
||||||
scale, zoom, mag, act_mag,
|
scale, zoom, mag, act_mag,
|
||||||
last_scale, last_zoom, last_mag, last_act_mag,
|
last_scale, last_zoom, last_mag, last_act_mag,
|
||||||
base_image_width, base_image_height,
|
base_image_width, base_image_height,
|
||||||
|
@ -180,8 +178,7 @@ Image *StreamBase::prepareImage(Image *image) {
|
||||||
last_act_image_width, last_act_image_height,
|
last_act_image_width, last_act_image_height,
|
||||||
disp_image_width, disp_image_height,
|
disp_image_width, disp_image_height,
|
||||||
last_disp_image_width, last_disp_image_height,
|
last_disp_image_width, last_disp_image_height,
|
||||||
send_image_width, send_image_height,
|
send_image_width, send_image_height
|
||||||
last_send_image_width, last_send_image_height
|
|
||||||
);
|
);
|
||||||
|
|
||||||
if ( ( mag != ZM_SCALE_BASE ) && (act_mag != ZM_SCALE_BASE) ) {
|
if ( ( mag != ZM_SCALE_BASE ) && (act_mag != ZM_SCALE_BASE) ) {
|
||||||
|
|
|
@ -22,7 +22,14 @@
|
||||||
#include "zm_image.h"
|
#include "zm_image.h"
|
||||||
#include "zm_logger.h"
|
#include "zm_logger.h"
|
||||||
|
|
||||||
SWScale::SWScale() : gotdefaults(false), swscale_ctx(nullptr), input_avframe(nullptr), output_avframe(nullptr) {
|
SWScale::SWScale() :
|
||||||
|
gotdefaults(false),
|
||||||
|
swscale_ctx(nullptr),
|
||||||
|
input_avframe(nullptr),
|
||||||
|
output_avframe(nullptr),
|
||||||
|
default_width(0),
|
||||||
|
default_height(0)
|
||||||
|
{
|
||||||
Debug(4, "SWScale object created");
|
Debug(4, "SWScale object created");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -122,13 +122,14 @@ std::string stringtf(const char* format, ...) {
|
||||||
va_start(args, format);
|
va_start(args, format);
|
||||||
va_list args2;
|
va_list args2;
|
||||||
va_copy(args2, args);
|
va_copy(args2, args);
|
||||||
|
int size = vsnprintf(nullptr, 0, format, args);
|
||||||
int size = vsnprintf(nullptr, 0, format, args) + 1; // Extra space for '\0'
|
|
||||||
va_end(args);
|
va_end(args);
|
||||||
|
|
||||||
if (size <= 0) {
|
if (size < 0) {
|
||||||
|
va_end(args2);
|
||||||
throw std::runtime_error("Error during formatting.");
|
throw std::runtime_error("Error during formatting.");
|
||||||
}
|
}
|
||||||
|
size += 1; // Extra space for '\0'
|
||||||
|
|
||||||
std::unique_ptr<char[]> buf(new char[size]);
|
std::unique_ptr<char[]> buf(new char[size]);
|
||||||
vsnprintf(buf.get(), size, format, args2);
|
vsnprintf(buf.get(), size, format, args2);
|
||||||
|
@ -259,6 +260,8 @@ void HwCapsDetect() {
|
||||||
unsigned long auxval = 0;
|
unsigned long auxval = 0;
|
||||||
elf_aux_info(AT_HWCAP, &auxval, sizeof(auxval));
|
elf_aux_info(AT_HWCAP, &auxval, sizeof(auxval));
|
||||||
if (auxval & HWCAP_NEON) {
|
if (auxval & HWCAP_NEON) {
|
||||||
|
#else
|
||||||
|
{
|
||||||
#error Unsupported OS.
|
#error Unsupported OS.
|
||||||
#endif
|
#endif
|
||||||
Debug(1,"Detected ARM (AArch32) processor with Neon");
|
Debug(1,"Detected ARM (AArch32) processor with Neon");
|
||||||
|
|
|
@ -401,6 +401,10 @@ bool VideoStore::open() {
|
||||||
} else {
|
} else {
|
||||||
audio_in_ctx = avcodec_alloc_context3(audio_out_codec);
|
audio_in_ctx = avcodec_alloc_context3(audio_out_codec);
|
||||||
ret = avcodec_parameters_to_context(audio_in_ctx, audio_in_stream->codecpar);
|
ret = avcodec_parameters_to_context(audio_in_ctx, audio_in_stream->codecpar);
|
||||||
|
if (ret < 0)
|
||||||
|
Error("Failure from avcodec_parameters_to_context %s",
|
||||||
|
av_make_error_string(ret).c_str());
|
||||||
|
|
||||||
audio_in_ctx->time_base = audio_in_stream->time_base;
|
audio_in_ctx->time_base = audio_in_stream->time_base;
|
||||||
|
|
||||||
audio_out_ctx = avcodec_alloc_context3(audio_out_codec);
|
audio_out_ctx = avcodec_alloc_context3(audio_out_codec);
|
||||||
|
@ -735,7 +739,6 @@ bool VideoStore::setup_resampler() {
|
||||||
audio_out_ctx->sample_fmt = audio_in_ctx->sample_fmt;
|
audio_out_ctx->sample_fmt = audio_in_ctx->sample_fmt;
|
||||||
audio_out_ctx->channels = audio_in_ctx->channels;
|
audio_out_ctx->channels = audio_in_ctx->channels;
|
||||||
audio_out_ctx->channel_layout = audio_in_ctx->channel_layout;
|
audio_out_ctx->channel_layout = audio_in_ctx->channel_layout;
|
||||||
audio_out_ctx->sample_fmt = audio_in_ctx->sample_fmt;
|
|
||||||
if (!audio_out_ctx->channel_layout) {
|
if (!audio_out_ctx->channel_layout) {
|
||||||
Debug(3, "Correcting channel layout from (%" PRIi64 ") to (%" PRIi64 ")",
|
Debug(3, "Correcting channel layout from (%" PRIi64 ") to (%" PRIi64 ")",
|
||||||
audio_out_ctx->channel_layout,
|
audio_out_ctx->channel_layout,
|
||||||
|
@ -858,7 +861,7 @@ bool VideoStore::setup_resampler() {
|
||||||
return false;
|
return false;
|
||||||
}
|
}
|
||||||
if ((ret = swr_init(resample_ctx)) < 0) {
|
if ((ret = swr_init(resample_ctx)) < 0) {
|
||||||
Error("Could not open resampler");
|
Error("Could not open resampler %d", ret);
|
||||||
av_frame_free(&in_frame);
|
av_frame_free(&in_frame);
|
||||||
av_frame_free(&out_frame);
|
av_frame_free(&out_frame);
|
||||||
swr_free(&resample_ctx);
|
swr_free(&resample_ctx);
|
||||||
|
|
|
@ -219,7 +219,7 @@ bool Zone::CheckAlarms(const Image *delta_image) {
|
||||||
int alarm_mid_x = -1;
|
int alarm_mid_x = -1;
|
||||||
int alarm_mid_y = -1;
|
int alarm_mid_y = -1;
|
||||||
|
|
||||||
unsigned int lo_x = polygon.Extent().Lo().x_;
|
//unsigned int lo_x = polygon.Extent().Lo().x_;
|
||||||
unsigned int lo_y = polygon.Extent().Lo().y_;
|
unsigned int lo_y = polygon.Extent().Lo().y_;
|
||||||
unsigned int hi_x = polygon.Extent().Hi().x_;
|
unsigned int hi_x = polygon.Extent().Hi().x_;
|
||||||
unsigned int hi_y = polygon.Extent().Hi().y_;
|
unsigned int hi_y = polygon.Extent().Hi().y_;
|
||||||
|
@ -699,6 +699,7 @@ bool Zone::CheckAlarms(const Image *delta_image) {
|
||||||
|
|
||||||
if ((type < PRECLUSIVE) && (check_method >= BLOBS) && (monitor->GetOptSaveJPEGs() > 1)) {
|
if ((type < PRECLUSIVE) && (check_method >= BLOBS) && (monitor->GetOptSaveJPEGs() > 1)) {
|
||||||
|
|
||||||
|
unsigned int lo_x = polygon.Extent().Lo().x_;
|
||||||
// First mask out anything we don't want
|
// First mask out anything we don't want
|
||||||
for (unsigned int y = lo_y; y <= hi_y; y++) {
|
for (unsigned int y = lo_y; y <= hi_y; y++) {
|
||||||
pdiff = diff_buff + ((diff_width * y) + lo_x);
|
pdiff = diff_buff + ((diff_width * y) + lo_x);
|
||||||
|
|
|
@ -88,9 +88,15 @@ function initialAlarmCues(eventId) {
|
||||||
}
|
}
|
||||||
|
|
||||||
function setAlarmCues(data) {
|
function setAlarmCues(data) {
|
||||||
cueFrames = data.frames;
|
if (!data) {
|
||||||
alarmSpans = renderAlarmCues(vid ? $j("#videoobj") : $j("#evtStream"));//use videojs width or zms width
|
Error('No data in setAlarmCues for event ' + eventData.Id);
|
||||||
$j(".alarmCue").html(alarmSpans);
|
} else if (!data.frames) {
|
||||||
|
Error('No data.frames in setAlarmCues for event ' + eventData.Id);
|
||||||
|
} else {
|
||||||
|
cueFrames = data.frames;
|
||||||
|
alarmSpans = renderAlarmCues(vid ? $j("#videoobj") : $j("#evtStream"));//use videojs width or zms width
|
||||||
|
$j(".alarmCue").html(alarmSpans);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
function renderAlarmCues(containerEl) {
|
function renderAlarmCues(containerEl) {
|
||||||
|
|
Loading…
Reference in New Issue