// // ZoneMinder Local Camera Class Implementation, $Date$, $Revision$ // Copyright (C) 2001-2008 Philip Coombes // // This program is free software; you can redistribute it and/or // modify it under the terms of the GNU General Public License // as published by the Free Software Foundation; either version 2 // of the License, or (at your option) any later version. // // This program is distributed in the hope that it will be useful, // but WITHOUT ANY WARRANTY; without even the implied warranty of // MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the // GNU General Public License for more details. // // You should have received a copy of the GNU General Public License // along with this program; if not, write to the Free Software // Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. // #include "zm_local_camera.h" #include #include #include #include #include static int vidioctl( int fd, int request, void *arg ) { int result = -1; do { result = ioctl( fd, request, arg ); } while ( result == -1 && errno == EINTR ); return( result ); } int LocalCamera::camera_count = 0; int LocalCamera::channel_count = 0; int LocalCamera::last_channel = -1; int LocalCamera::vid_fd; #ifdef ZM_V4L2 LocalCamera::V4L2Data LocalCamera::v4l2_data; #endif // ZM_V4L2 LocalCamera::V4L1Data LocalCamera::v4l1_data; unsigned char *LocalCamera::y_table; signed char *LocalCamera::uv_table; short *LocalCamera::r_v_table; short *LocalCamera::g_v_table; short *LocalCamera::g_u_table; short *LocalCamera::b_u_table; LocalCamera::LocalCamera( int p_id, const std::string &p_device, int p_channel, int p_format, const std::string &p_method, int p_width, int p_height, int p_palette, int p_brightness, int p_contrast, int p_hue, int p_colour, bool p_capture ) : #ifdef ZM_V4L2 Camera( p_id, LOCAL_SRC, p_width, p_height, ((palette==VIDEO_PALETTE_GREY||palette==V4L2_PIX_FMT_GREY)?1:3), p_brightness, p_contrast, p_hue, p_colour, p_capture ), #else // ZM_V4L2 Camera( p_id, LOCAL_SRC, p_width, p_height, (palette==VIDEO_PALETTE_GREY?1:3), p_brightness, p_contrast, p_hue, p_colour, p_capture ), #endif // ZM_V4L2 device( p_device ), channel( p_channel ), format( p_format ), palette( p_palette ), v4l2( p_method == "v4l2" ) { // If we are the first, or only, input on this device then // do the initial opening etc device_prime = (camera_count++ == 0); if ( device_prime && capture ) Initialise(); if ( channel != last_channel ) { // We are the first, or only, input that uses this channel channel_prime = true; channel_count++; last_channel = channel; } else { // We are the second, or subsequent, input using this channel channel_prime = false; } } LocalCamera::~LocalCamera() { if ( device_prime && capture ) Terminate(); } void LocalCamera::Initialise() { struct stat st; if ( stat( device.c_str(), &st ) < 0 ) Fatal( "Failed to stat video device %s: %s", device.c_str(), strerror(errno) ); if ( !S_ISCHR(st.st_mode) ) Fatal( "File %s is not device file: %s", device.c_str(), strerror(errno) ); Debug( 3, "Opening video device %s", device.c_str() ); //if ( (vid_fd = open( device.c_str(), O_RDWR|O_NONBLOCK, 0 )) < 0 ) if ( (vid_fd = open( device.c_str(), O_RDWR, 0 )) < 0 ) Fatal( "Failed to open video device %s: %s", device.c_str(), strerror(errno) ); #ifdef ZM_V4L2 Debug( 2, "V4L2 support enabled, using %s api", v4l2?"V4L2":"V4L1" ); if ( v4l2 ) { #if HAVE_LIBSWSCALE ffPixFormat = PIX_FMT_NONE; switch( palette ) { case V4L2_PIX_FMT_RGB444 : ffPixFormat = PIX_FMT_RGB32; break; case V4L2_PIX_FMT_RGB555 : ffPixFormat = PIX_FMT_RGB555; break; case V4L2_PIX_FMT_RGB565 : ffPixFormat = PIX_FMT_RGB565; break; case V4L2_PIX_FMT_BGR24 : ffPixFormat = PIX_FMT_BGR24; break; case V4L2_PIX_FMT_RGB24 : ffPixFormat = PIX_FMT_RGB24; break; case V4L2_PIX_FMT_BGR32 : ffPixFormat = PIX_FMT_BGR32; break; case V4L2_PIX_FMT_RGB32 : ffPixFormat = PIX_FMT_RGB32; break; case V4L2_PIX_FMT_GREY : ffPixFormat = PIX_FMT_GRAY8; break; case V4L2_PIX_FMT_YUYV : ffPixFormat = PIX_FMT_YUYV422; break; case V4L2_PIX_FMT_YUV422P : ffPixFormat = PIX_FMT_YUV422P; break; case V4L2_PIX_FMT_YUV411P : ffPixFormat = PIX_FMT_YUV411P; break; case V4L2_PIX_FMT_YUV444 : ffPixFormat = PIX_FMT_YUV444P; break; case V4L2_PIX_FMT_YUV410 : ffPixFormat = PIX_FMT_YUV410P; break; case V4L2_PIX_FMT_YUV420 : ffPixFormat = PIX_FMT_YUV420P; break; case V4L2_PIX_FMT_JPEG : ffPixFormat = PIX_FMT_YUVJ444P; break; // These don't seem to have ffmpeg equivalents // See if you can match any of the ones in the default clause below!? case V4L2_PIX_FMT_UYVY : case V4L2_PIX_FMT_RGB332 : case V4L2_PIX_FMT_RGB555X : case V4L2_PIX_FMT_RGB565X : case V4L2_PIX_FMT_Y16 : case V4L2_PIX_FMT_PAL8 : case V4L2_PIX_FMT_YVU410 : case V4L2_PIX_FMT_YVU420 : case V4L2_PIX_FMT_Y41P : case V4L2_PIX_FMT_YUV555 : case V4L2_PIX_FMT_YUV565 : case V4L2_PIX_FMT_YUV32 : case V4L2_PIX_FMT_NV12 : case V4L2_PIX_FMT_NV21 : case V4L2_PIX_FMT_YYUV : case V4L2_PIX_FMT_HI240 : case V4L2_PIX_FMT_HM12 : case V4L2_PIX_FMT_SBGGR8 : case V4L2_PIX_FMT_SGBRG8 : case V4L2_PIX_FMT_SBGGR16 : case V4L2_PIX_FMT_MJPEG : case V4L2_PIX_FMT_DV : case V4L2_PIX_FMT_MPEG : case V4L2_PIX_FMT_WNVA : case V4L2_PIX_FMT_SN9C10X : case V4L2_PIX_FMT_PWC1 : case V4L2_PIX_FMT_PWC2 : case V4L2_PIX_FMT_ET61X251 : case V4L2_PIX_FMT_SPCA501 : case V4L2_PIX_FMT_SPCA505 : case V4L2_PIX_FMT_SPCA508 : case V4L2_PIX_FMT_SPCA561 : case V4L2_PIX_FMT_PAC207 : case V4L2_PIX_FMT_PJPG : case V4L2_PIX_FMT_YVYU : default : { Fatal( "Can't find swscale format for palette %d", palette ); break; // These are all spare and may match some of the above ffPixFormat = PIX_FMT_YUVJ420P; ffPixFormat = PIX_FMT_YUVJ422P; ffPixFormat = PIX_FMT_XVMC_MPEG2_MC; ffPixFormat = PIX_FMT_XVMC_MPEG2_IDCT; ffPixFormat = PIX_FMT_UYVY422; ffPixFormat = PIX_FMT_UYYVYY411; ffPixFormat = PIX_FMT_BGR565; ffPixFormat = PIX_FMT_BGR555; ffPixFormat = PIX_FMT_BGR8; ffPixFormat = PIX_FMT_BGR4; ffPixFormat = PIX_FMT_BGR4_BYTE; ffPixFormat = PIX_FMT_RGB8; ffPixFormat = PIX_FMT_RGB4; ffPixFormat = PIX_FMT_RGB4_BYTE; ffPixFormat = PIX_FMT_NV12; ffPixFormat = PIX_FMT_NV21; ffPixFormat = PIX_FMT_RGB32_1; ffPixFormat = PIX_FMT_BGR32_1; ffPixFormat = PIX_FMT_GRAY16BE; ffPixFormat = PIX_FMT_GRAY16LE; ffPixFormat = PIX_FMT_YUV440P; ffPixFormat = PIX_FMT_YUVJ440P; ffPixFormat = PIX_FMT_YUVA420P; ffPixFormat = PIX_FMT_VDPAU_H264; ffPixFormat = PIX_FMT_VDPAU_MPEG1; ffPixFormat = PIX_FMT_VDPAU_MPEG2; ffPixFormat = PIX_FMT_NB; } } #endif // HAVE_LIBSWSCALE struct v4l2_capability vid_cap; Debug( 3, "Checking video device capabilities" ); if ( vidioctl( vid_fd, VIDIOC_QUERYCAP, &vid_cap ) < 0 ) Fatal( "Failed to query video device: %s", strerror(errno) ); if ( !(vid_cap.capabilities & V4L2_CAP_VIDEO_CAPTURE) ) Fatal( "Video device is not video capture device" ); if ( !(vid_cap.capabilities & V4L2_CAP_STREAMING) ) Fatal( "Video device does not support streaming i/o" ); Debug( 3, "Setting up video format" ); memset( &v4l2_data.fmt, 0, sizeof(v4l2_data.fmt) ); v4l2_data.fmt.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; if ( vidioctl( vid_fd, VIDIOC_G_FMT, &v4l2_data.fmt ) < 0 ) Fatal( "Failed to get video format: %s", strerror(errno) ); v4l2_data.fmt.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; v4l2_data.fmt.fmt.pix.width = width; v4l2_data.fmt.fmt.pix.height = height; v4l2_data.fmt.fmt.pix.pixelformat = palette; //v4l2_data.fmt.fmt.pix.field = V4L2_FIELD_INTERLACED; v4l2_data.fmt.fmt.pix.field = V4L2_FIELD_ANY; if ( vidioctl( vid_fd, VIDIOC_S_FMT, &v4l2_data.fmt ) < 0 ) Fatal( "Failed to set video format: %s", strerror(errno) ); /* Note VIDIOC_S_FMT may change width and height. */ /* Buggy driver paranoia. */ unsigned int min; min = v4l2_data.fmt.fmt.pix.width * 2; if (v4l2_data.fmt.fmt.pix.bytesperline < min) v4l2_data.fmt.fmt.pix.bytesperline = min; min = v4l2_data.fmt.fmt.pix.bytesperline * v4l2_data.fmt.fmt.pix.height; if (v4l2_data.fmt.fmt.pix.sizeimage < min) v4l2_data.fmt.fmt.pix.sizeimage = min; Debug( 3, "Setting up request buffers" ); memset( &v4l2_data.reqbufs, 0, sizeof(v4l2_data.reqbufs) ); v4l2_data.reqbufs.count = 8; //v4l2_data.reqbufs.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; v4l2_data.reqbufs.type = v4l2_data.fmt.type; v4l2_data.reqbufs.memory = V4L2_MEMORY_MMAP; if ( vidioctl( vid_fd, VIDIOC_REQBUFS, &v4l2_data.reqbufs ) < 0 ) { if ( errno == EINVAL ) { Fatal( "Unable to initialise memory mapping, unsupported in device" ); } else { Fatal( "Unable to initialise memory mapping: %s", strerror(errno) ); } } if ( v4l2_data.reqbufs.count < 2 ) Fatal( "Insufficient buffer memory %d on video device", v4l2_data.reqbufs.count ); Debug( 3, "Setting up %d data buffers", v4l2_data.reqbufs.count ); v4l2_data.buffers = new V4L2MappedBuffer[v4l2_data.reqbufs.count]; #if HAVE_LIBSWSCALE ffPictures = new AVFrame *[v4l2_data.reqbufs.count]; #endif // HAVE_LIBSWSCALE for ( int i = 0; i < v4l2_data.reqbufs.count; i++ ) { struct v4l2_buffer vid_buf; memset( &vid_buf, 0, sizeof(vid_buf) ); //vid_buf.type = V4L2_BUF_TYPE_VIDEO_CAPTURE; vid_buf.type = v4l2_data.fmt.type; //vid_buf.memory = V4L2_MEMORY_MMAP; vid_buf.memory = v4l2_data.reqbufs.memory; vid_buf.index = i; if ( vidioctl( vid_fd, VIDIOC_QUERYBUF, &vid_buf ) < 0 ) Fatal( "Unable to query video buffer: %s", strerror(errno) ); v4l2_data.buffers[i].length = vid_buf.length; v4l2_data.buffers[i].start = mmap( NULL, vid_buf.length, PROT_READ|PROT_WRITE, MAP_SHARED, vid_fd, vid_buf.m.offset ); if ( v4l2_data.buffers[i].start == MAP_FAILED ) Fatal( "Can't map video buffer %d (%d bytes) to memory: %s(%d)", i, vid_buf.length, strerror(errno), errno ); #if HAVE_LIBSWSCALE ffPictures[i] = avcodec_alloc_frame(); if ( !ffPictures[i] ) Fatal( "Could not allocate picture" ); avpicture_fill( (AVPicture *)ffPictures[i], (unsigned char *)v4l2_data.buffers[i].start, ffPixFormat, width, height ); #endif // HAVE_LIBSWSCALE } Debug( 3, "Configuring video source" ); if ( vidioctl( vid_fd, VIDIOC_S_INPUT, &channel ) < 0 ) { Fatal( "Failed to set camera source %d: %s", channel, strerror(errno) ); } struct v4l2_input input; v4l2_std_id stdId; memset( &input, 0, sizeof(input) ); if ( vidioctl( vid_fd, VIDIOC_ENUMINPUT, &input ) < 0 ) { Fatal( "Failed to enumerate input %d: %s", channel, strerror(errno) ); } if ( (input.std != V4L2_STD_UNKNOWN) && ((input.std & format) == V4L2_STD_UNKNOWN) ) { Fatal( "Device does not support video standard %d", format ); } stdId = format; if ( vidioctl( vid_fd, VIDIOC_S_STD, &stdId ) < 0 ) { Fatal( "Failed to set video standard %d: %s", format, strerror(errno) ); } } else #endif // ZM_V4L2 { #if HAVE_LIBSWSCALE ffPixFormat = PIX_FMT_NONE; switch( palette ) { case VIDEO_PALETTE_RGB555 : ffPixFormat = PIX_FMT_RGB555; break; case VIDEO_PALETTE_RGB565 : ffPixFormat = PIX_FMT_RGB565; break; case VIDEO_PALETTE_RGB24 : if ( config.local_bgr_invert ) ffPixFormat = PIX_FMT_BGR24; else ffPixFormat = PIX_FMT_RGB24; break; case VIDEO_PALETTE_GREY : ffPixFormat = PIX_FMT_GRAY8; break; case VIDEO_PALETTE_YUYV : case VIDEO_PALETTE_YUV422 : ffPixFormat = PIX_FMT_YUYV422; break; case VIDEO_PALETTE_YUV422P : ffPixFormat = PIX_FMT_YUV422P; break; case VIDEO_PALETTE_YUV420P : ffPixFormat = PIX_FMT_YUV420P; break; default : { Fatal( "Can't find swscale format for palette %d", palette ); break; // These are all spare and may match some of the above ffPixFormat = PIX_FMT_YUVJ420P; ffPixFormat = PIX_FMT_YUVJ422P; ffPixFormat = PIX_FMT_YUVJ444P; ffPixFormat = PIX_FMT_XVMC_MPEG2_MC; ffPixFormat = PIX_FMT_XVMC_MPEG2_IDCT; ffPixFormat = PIX_FMT_UYVY422; ffPixFormat = PIX_FMT_UYYVYY411; ffPixFormat = PIX_FMT_BGR565; ffPixFormat = PIX_FMT_BGR555; ffPixFormat = PIX_FMT_BGR8; ffPixFormat = PIX_FMT_BGR4; ffPixFormat = PIX_FMT_BGR4_BYTE; ffPixFormat = PIX_FMT_RGB8; ffPixFormat = PIX_FMT_RGB4; ffPixFormat = PIX_FMT_RGB4_BYTE; ffPixFormat = PIX_FMT_NV12; ffPixFormat = PIX_FMT_NV21; ffPixFormat = PIX_FMT_RGB32_1; ffPixFormat = PIX_FMT_BGR32_1; ffPixFormat = PIX_FMT_GRAY16BE; ffPixFormat = PIX_FMT_GRAY16LE; ffPixFormat = PIX_FMT_YUV440P; ffPixFormat = PIX_FMT_YUVJ440P; ffPixFormat = PIX_FMT_YUVA420P; ffPixFormat = PIX_FMT_VDPAU_H264; ffPixFormat = PIX_FMT_VDPAU_MPEG1; ffPixFormat = PIX_FMT_VDPAU_MPEG2; ffPixFormat = PIX_FMT_NB; } } #endif // HAVE_LIBSWSCALE Debug( 3, "Configuring picture attributes" ); struct video_picture vid_pic; memset( &vid_pic, 0, sizeof(vid_pic) ); if ( ioctl( vid_fd, VIDIOCGPICT, &vid_pic) < 0 ) Fatal( "Failed to get picture attributes: %s", strerror(errno) ); Debug( 4, "Old P:%d", vid_pic.palette ); Debug( 4, "Old D:%d", vid_pic.depth ); Debug( 4, "Old B:%d", vid_pic.brightness ); Debug( 4, "Old h:%d", vid_pic.hue ); Debug( 4, "Old Cl:%d", vid_pic.colour ); Debug( 4, "Old Cn:%d", vid_pic.contrast ); switch (vid_pic.palette = palette) { case VIDEO_PALETTE_GREY : { vid_pic.depth = 8; break; } case VIDEO_PALETTE_RGB565 : case VIDEO_PALETTE_YUYV : case VIDEO_PALETTE_YUV422 : case VIDEO_PALETTE_YUV420P : case VIDEO_PALETTE_YUV422P : { vid_pic.depth = 16; break; } case VIDEO_PALETTE_RGB24 : { vid_pic.depth = 24; break; } } if ( brightness >= 0 ) vid_pic.brightness = brightness; if ( hue >= 0 ) vid_pic.hue = hue; if ( colour >= 0 ) vid_pic.colour = colour; if ( contrast >= 0 ) vid_pic.contrast = contrast; if ( ioctl( vid_fd, VIDIOCSPICT, &vid_pic ) < 0 ) { Error( "Failed to set picture attributes: %s", strerror(errno) ); if ( config.strict_video_config ) exit(-1); } Debug( 3, "Configuring window attributes" ); struct video_window vid_win; memset( &vid_win, 0, sizeof(vid_win) ); if ( ioctl( vid_fd, VIDIOCGWIN, &vid_win) < 0 ) { Error( "Failed to get window attributes: %s", strerror(errno) ); exit(-1); } Debug( 4, "Old X:%d", vid_win.x ); Debug( 4, "Old Y:%d", vid_win.y ); Debug( 4, "Old W:%d", vid_win.width ); Debug( 4, "Old H:%d", vid_win.height ); vid_win.x = 0; vid_win.y = 0; vid_win.width = width; vid_win.height = height; vid_win.flags &= ~VIDEO_WINDOW_INTERLACE; if ( ioctl( vid_fd, VIDIOCSWIN, &vid_win ) < 0 ) { Error( "Failed to set window attributes: %s", strerror(errno) ); if ( config.strict_video_config ) exit(-1); } Debug( 3, "Setting up request buffers" ); if ( ioctl( vid_fd, VIDIOCGMBUF, &v4l1_data.frames ) < 0 ) Fatal( "Failed to setup memory: %s", strerror(errno) ); v4l1_data.buffers = new video_mmap[v4l1_data.frames.frames]; Debug( 4, "vmb.frames = %d", v4l1_data.frames.frames ); Debug( 4, "vmb.size = %d", v4l1_data.frames.size ); Debug( 3, "Setting up %d frame buffers", v4l1_data.frames.frames ); v4l1_data.buffer = (unsigned char *)mmap( 0, v4l1_data.frames.size, PROT_READ|PROT_WRITE, MAP_SHARED, vid_fd, 0 ); if ( v4l1_data.buffer == MAP_FAILED ) Fatal( "Could not mmap video: %s", strerror(errno) ); #if HAVE_LIBSWSCALE ffPictures = new AVFrame *[v4l1_data.frames.frames]; for ( int i = 0; i < v4l1_data.frames.frames; i++ ) { v4l1_data.buffers[i].frame = i; v4l1_data.buffers[i].width = width; v4l1_data.buffers[i].height = height; v4l1_data.buffers[i].format = palette; ffPictures[i] = avcodec_alloc_frame(); if ( !ffPictures[i] ) Fatal( "Could not allocate picture" ); avpicture_fill( (AVPicture *)ffPictures[i], (unsigned char *)v4l1_data.buffer+(i*v4l1_data.frames.size/v4l1_data.frames.frames), ffPixFormat, width, height ); } #endif // HAVE_LIBSWSCALE Debug( 3, "Configuring video source" ); struct video_channel vid_src; memset( &vid_src, 0, sizeof(vid_src) ); vid_src.channel = channel; if ( ioctl( vid_fd, VIDIOCGCHAN, &vid_src) < 0 ) Fatal( "Failed to get camera source: %s", strerror(errno) ); Debug( 4, "Old C:%d", vid_src.channel ); Debug( 4, "Old F:%d", vid_src.norm ); Debug( 4, "Old Fl:%x", vid_src.flags ); Debug( 4, "Old T:%d", vid_src.type ); vid_src.norm = format; vid_src.flags = 0; vid_src.type = VIDEO_TYPE_CAMERA; if ( ioctl( vid_fd, VIDIOCSCHAN, &vid_src ) < 0 ) { Error( "Failed to set camera source %d: %s", channel, strerror(errno) ); if ( config.strict_video_config ) exit(-1); } if ( ioctl( vid_fd, VIDIOCGWIN, &vid_win) < 0 ) Fatal( "Failed to get window data: %s", strerror(errno) ); Debug( 4, "New X:%d", vid_win.x ); Debug( 4, "New Y:%d", vid_win.y ); Debug( 4, "New W:%d", vid_win.width ); Debug( 4, "New H:%d", vid_win.height ); if ( ioctl( vid_fd, VIDIOCGPICT, &vid_pic) < 0 ) Fatal( "Failed to get window data: %s", strerror(errno) ); Debug( 4, "New P:%d", vid_pic.palette ); Debug( 4, "New D:%d", vid_pic.depth ); Debug( 4, "New B:%d", vid_pic.brightness ); Debug( 4, "New h:%d", vid_pic.hue ); Debug( 4, "New Cl:%d", vid_pic.colour ); Debug( 4, "New Cn:%d", vid_pic.contrast ); } Debug( 3, "Setting up static colour tables" ); y_table = new unsigned char[256]; for ( int i = 0; i <= 255; i++ ) { unsigned char c = i; if ( c <= 16 ) y_table[c] = 0; else if ( c >= 235 ) y_table[c] = 255; else y_table[c] = (255*(c-16))/219; } uv_table = new signed char[256]; for ( int i = 0; i <= 255; i++ ) { unsigned char c = i; if ( c <= 16 ) uv_table[c] = -127; else if ( c >= 240 ) uv_table[c] = 127; else uv_table[c] = (127*(c-128))/112; } r_v_table = new short[255]; g_v_table = new short[255]; g_u_table = new short[255]; b_u_table = new short[255]; r_v_table += 127; g_v_table += 127; g_u_table += 127; b_u_table += 127; for ( int i = -127; i <= 127; i++ ) { signed char c = i; r_v_table[c] = (1402*c)/1000; g_u_table[c] = (344*c)/1000; g_v_table[c] = (714*c)/1000; b_u_table[c] = (1772*c)/1000; } } void LocalCamera::Terminate() { #ifdef ZM_V4L2 if ( v4l2 ) { Debug( 3, "Terminating video stream" ); //enum v4l2_buf_type type = V4L2_BUF_TYPE_VIDEO_CAPTURE; enum v4l2_buf_type type = v4l2_data.fmt.type; if ( vidioctl( vid_fd, VIDIOC_STREAMOFF, &type ) < 0 ) Error( "Failed to stop capture stream: %s", strerror(errno) ); Debug( 3, "Unmapping video buffers" ); for ( int i = 0; i < v4l2_data.reqbufs.count; i++ ) if ( munmap( v4l2_data.buffers[i].start, v4l2_data.buffers[i].length ) < 0 ) Error( "Failed to munmap buffer %d: %s", i, strerror(errno) ); } else #endif // ZM_V4L2 { Debug( 3, "Unmapping video buffers" ); if ( munmap((char*)v4l1_data.buffer, v4l1_data.frames.size) < 0 ) Error( "Failed to munmap buffers: %s", strerror(errno) ); delete[] v4l1_data.buffers; } close( vid_fd ); } bool LocalCamera::GetCurrentSettings( const char *device, char *output, bool verbose ) { output[0] = 0; if ( verbose ) sprintf( output, output+strlen(output), "Checking Video Device: %s\n", device ); if ( (vid_fd=open(device, O_RDWR)) <=0 ) { Error( "Failed to open video device %s: %s", device, strerror(errno) ); if ( verbose ) sprintf( output+strlen(output), "Error, failed to open video device %s: %s\n", device, strerror(errno) ); else sprintf( output+strlen(output), "error%d\n", errno ); return( false ); } struct video_capability vid_cap; if ( ioctl( vid_fd, VIDIOCGCAP, &vid_cap ) < 0 ) { Error( "Failed to get video capabilities: %s", strerror(errno) ); if ( verbose ) sprintf( output, "Error, failed to get video capabilities %s: %s\n", device, strerror(errno) ); else sprintf( output, "error%d\n", errno ); return( false ); } if ( verbose ) { sprintf( output+strlen(output), "Video Capabilities\n" ); sprintf( output+strlen(output), " Name: %s\n", vid_cap.name ); sprintf( output+strlen(output), " Type: %d\n%s%s%s%s%s%s%s%s%s%s%s%s%s%s", vid_cap.type, vid_cap.type&VID_TYPE_CAPTURE?" Can capture\n":"", vid_cap.type&VID_TYPE_TUNER?" Can tune\n":"", vid_cap.type&VID_TYPE_TELETEXT?" Does teletext\n":"", vid_cap.type&VID_TYPE_OVERLAY?" Overlay onto frame buffer\n":"", vid_cap.type&VID_TYPE_CHROMAKEY?" Overlay by chromakey\n":"", vid_cap.type&VID_TYPE_CLIPPING?" Can clip\n":"", vid_cap.type&VID_TYPE_FRAMERAM?" Uses the frame buffer memory\n":"", vid_cap.type&VID_TYPE_SCALES?" Scalable\n":"", vid_cap.type&VID_TYPE_MONOCHROME?" Monochrome only\n":"", vid_cap.type&VID_TYPE_SUBCAPTURE?" Can capture subareas of the image\n":"", vid_cap.type&VID_TYPE_MPEG_DECODER?" Can decode MPEG streams\n":"", vid_cap.type&VID_TYPE_MPEG_ENCODER?" Can encode MPEG streams\n":"", vid_cap.type&VID_TYPE_MJPEG_DECODER?" Can decode MJPEG streams\n":"", vid_cap.type&VID_TYPE_MJPEG_ENCODER?" Can encode MJPEG streams\n":"" ); sprintf( output+strlen(output), " Video Channels: %d\n", vid_cap.channels ); sprintf( output+strlen(output), " Audio Channels: %d\n", vid_cap.audios ); sprintf( output+strlen(output), " Maximum Width: %d\n", vid_cap.maxwidth ); sprintf( output+strlen(output), " Maximum Height: %d\n", vid_cap.maxheight ); sprintf( output+strlen(output), " Minimum Width: %d\n", vid_cap.minwidth ); sprintf( output+strlen(output), " Minimum Height: %d\n", vid_cap.minheight ); } else { sprintf( output+strlen(output), "N:%s,", vid_cap.name ); sprintf( output+strlen(output), "T:%d,", vid_cap.type ); sprintf( output+strlen(output), "nC:%d,", vid_cap.channels ); sprintf( output+strlen(output), "nA:%d,", vid_cap.audios ); sprintf( output+strlen(output), "mxW:%d,", vid_cap.maxwidth ); sprintf( output+strlen(output), "mxH:%d,", vid_cap.maxheight ); sprintf( output+strlen(output), "mnW:%d,", vid_cap.minwidth ); sprintf( output+strlen(output), "mnH:%d,", vid_cap.minheight ); } struct video_window vid_win; if ( ioctl( vid_fd, VIDIOCGWIN, &vid_win ) < 0 ) { Error( "Failed to get window attributes: %s", strerror(errno) ); if ( verbose ) sprintf( output, "Error, failed to get window attributes: %s\n", strerror(errno) ); else sprintf( output, "error%d\n", errno ); return( false ); } if ( verbose ) { sprintf( output+strlen(output), "Window Attributes\n" ); sprintf( output+strlen(output), " X Offset: %d\n", vid_win.x ); sprintf( output+strlen(output), " Y Offset: %d\n", vid_win.y ); sprintf( output+strlen(output), " Width: %d\n", vid_win.width ); sprintf( output+strlen(output), " Height: %d\n", vid_win.height ); } else { sprintf( output+strlen(output), "X:%d,", vid_win.x ); sprintf( output+strlen(output), "Y:%d,", vid_win.y ); sprintf( output+strlen(output), "W:%d,", vid_win.width ); sprintf( output+strlen(output), "H:%d,", vid_win.height ); } struct video_picture vid_pic; if ( ioctl( vid_fd, VIDIOCGPICT, &vid_pic ) < 0 ) { Error( "Failed to get picture attributes: %s", strerror(errno) ); if ( verbose ) sprintf( output, "Error, failed to get picture attributes: %s\n", strerror(errno) ); else sprintf( output, "error%d\n", errno ); return( false ); } if ( verbose ) { sprintf( output+strlen(output), "Picture Attributes\n" ); sprintf( output+strlen(output), " Palette: %d - %s\n", vid_pic.palette, vid_pic.palette==VIDEO_PALETTE_GREY?"Linear greyscale":( vid_pic.palette==VIDEO_PALETTE_HI240?"High 240 cube (BT848)":( vid_pic.palette==VIDEO_PALETTE_RGB565?"565 16 bit RGB":( vid_pic.palette==VIDEO_PALETTE_RGB24?"24bit RGB":( vid_pic.palette==VIDEO_PALETTE_RGB32?"32bit RGB":( vid_pic.palette==VIDEO_PALETTE_RGB555?"555 15bit RGB":( vid_pic.palette==VIDEO_PALETTE_YUV422?"YUV422 capture":( vid_pic.palette==VIDEO_PALETTE_YUYV?"YUYV":( vid_pic.palette==VIDEO_PALETTE_UYVY?"UVYV":( vid_pic.palette==VIDEO_PALETTE_YUV420?"YUV420":( vid_pic.palette==VIDEO_PALETTE_YUV411?"YUV411 capture":( vid_pic.palette==VIDEO_PALETTE_RAW?"RAW capture (BT848)":( vid_pic.palette==VIDEO_PALETTE_YUYV?"YUYV":( vid_pic.palette==VIDEO_PALETTE_YUV422?"YUV422":( vid_pic.palette==VIDEO_PALETTE_YUV422P?"YUV 4:2:2 Planar":( vid_pic.palette==VIDEO_PALETTE_YUV411P?"YUV 4:1:1 Planar":( vid_pic.palette==VIDEO_PALETTE_YUV420P?"YUV 4:2:0 Planar":( vid_pic.palette==VIDEO_PALETTE_YUV410P?"YUV 4:1:0 Planar":"Unknown" )))))))))))))))))); sprintf( output+strlen(output), " Colour Depth: %d\n", vid_pic.depth ); sprintf( output+strlen(output), " Brightness: %d\n", vid_pic.brightness ); sprintf( output+strlen(output), " Hue: %d\n", vid_pic.hue ); sprintf( output+strlen(output), " Colour :%d\n", vid_pic.colour ); sprintf( output+strlen(output), " Contrast: %d\n", vid_pic.contrast ); sprintf( output+strlen(output), " Whiteness: %d\n", vid_pic.whiteness ); } else { sprintf( output+strlen(output), "P:%d,", vid_pic.palette ); sprintf( output+strlen(output), "D:%d,", vid_pic.depth ); sprintf( output+strlen(output), "B:%d,", vid_pic.brightness ); sprintf( output+strlen(output), "h:%d,", vid_pic.hue ); sprintf( output+strlen(output), "Cl:%d,", vid_pic.colour ); sprintf( output+strlen(output), "Cn:%d,", vid_pic.contrast ); sprintf( output+strlen(output), "w:%d,", vid_pic.whiteness ); } for ( int chan = 0; chan < vid_cap.channels; chan++ ) { struct video_channel vid_src; vid_src.channel = chan; if ( ioctl( vid_fd, VIDIOCGCHAN, &vid_src ) < 0 ) { Error( "Failed to get channel %d attributes: %s", chan, strerror(errno) ); if ( verbose ) sprintf( output, "Error, failed to get channel %d attributes: %s\n", chan, strerror(errno) ); else sprintf( output, "error%d\n", errno ); return( false ); } if ( verbose ) { sprintf( output+strlen(output), "Channel %d Attributes\n", chan ); sprintf( output+strlen(output), " Name: %s\n", vid_src.name ); sprintf( output+strlen(output), " Channel: %d\n", vid_src.channel ); sprintf( output+strlen(output), " Flags: %d\n%s%s", vid_src.flags, vid_src.flags&VIDEO_VC_TUNER?" Channel has a tuner\n":"", vid_src.flags&VIDEO_VC_AUDIO?" Channel has audio\n":"" ); sprintf( output+strlen(output), " Type: %d - %s\n", vid_src.type, vid_src.type==VIDEO_TYPE_TV?"TV":( vid_src.type==VIDEO_TYPE_CAMERA?"Camera":"Unknown" )); sprintf( output+strlen(output), " Format: %d - %s\n", vid_src.norm, vid_src.norm==VIDEO_MODE_PAL?"PAL":( vid_src.norm==VIDEO_MODE_NTSC?"NTSC":( vid_src.norm==VIDEO_MODE_SECAM?"SECAM":( vid_src.norm==VIDEO_MODE_AUTO?"AUTO":"Unknown" )))); } else { sprintf( output+strlen(output), "n%d:%s,", chan, vid_src.name ); sprintf( output+strlen(output), "C%d:%d,", chan, vid_src.channel ); sprintf( output+strlen(output), "Fl%d:%x,", chan, vid_src.flags ); sprintf( output+strlen(output), "T%d:%d", chan, vid_src.type ); sprintf( output+strlen(output), "F%d:%d%s,", chan, vid_src.norm, chan==(vid_cap.channels-1)?"":"," ); } } return( true ); } int LocalCamera::Brightness( int p_brightness ) { #ifdef ZM_V4L2 if ( v4l2 ) { struct v4l2_control vid_control; memset( &vid_control, 0, sizeof(vid_control) ); vid_control.id = V4L2_CID_BRIGHTNESS; if ( vidioctl( vid_fd, VIDIOC_G_CTRL, &vid_control ) < 0 ) { if ( errno != EINVAL ) Error( "Unable to query brightness: %s", strerror(errno) ) else Warning( "Brightness control is not suppported" ) Info( "Brightness 1 %d", vid_control.value ); } else if ( p_brightness >= 0 ) { vid_control.value = p_brightness; Info( "Brightness 2 %d", vid_control.value ); /* The driver may clamp the value or return ERANGE, ignored here */ if ( vidioctl ( vid_fd, VIDIOC_S_CTRL, &vid_control ) ) { if ( errno != ERANGE ) Error( "Unable to set brightness: %s", strerror(errno) ) else Warning( "Given brightness value (%d) may be out-of-range", p_brightness ) } Info( "Brightness 3 %d", vid_control.value ); } return( vid_control.value ); } else #endif // ZM_V4L2 { struct video_picture vid_pic; if ( ioctl( vid_fd, VIDIOCGPICT, &vid_pic) < 0 ) { Error( "Failed to get picture attributes: %s", strerror(errno) ); return( -1 ); } if ( p_brightness >= 0 ) { vid_pic.brightness = p_brightness; if ( ioctl( vid_fd, VIDIOCSPICT, &vid_pic ) < 0 ) { Error( "Failed to set picture attributes: %s", strerror(errno) ); return( -1 ); } } return( vid_pic.brightness ); } } int LocalCamera::Hue( int p_hue ) { #ifdef ZM_V4L2 if ( v4l2 ) { struct v4l2_control vid_control; memset( &vid_control, 0, sizeof(vid_control) ); vid_control.id = V4L2_CID_HUE; if ( vidioctl( vid_fd, VIDIOC_G_CTRL, &vid_control ) < 0 ) { if ( errno != EINVAL ) Error( "Unable to query hue: %s", strerror(errno) ) else Warning( "Hue control is not suppported" ) } else if ( p_hue >= 0 ) { vid_control.value = p_hue; /* The driver may clamp the value or return ERANGE, ignored here */ if ( vidioctl ( vid_fd, VIDIOC_S_CTRL, &vid_control ) < 0 ) { if ( errno != ERANGE ) Error( "Unable to set hue: %s", strerror(errno) ) else Warning( "Given hue value (%d) may be out-of-range", p_hue ) } } return( vid_control.value ); } else #endif // ZM_V4L2 { struct video_picture vid_pic; if ( ioctl( vid_fd, VIDIOCGPICT, &vid_pic) < 0 ) { Error( "Failed to get picture attributes: %s", strerror(errno) ); return( -1 ); } if ( p_hue >= 0 ) { vid_pic.hue = p_hue; if ( ioctl( vid_fd, VIDIOCSPICT, &vid_pic ) < 0 ) { Error( "Failed to set picture attributes: %s", strerror(errno) ); return( -1 ); } } return( vid_pic.hue ); } } int LocalCamera::Colour( int p_colour ) { #ifdef ZM_V4L2 if ( v4l2 ) { struct v4l2_control vid_control; memset( &vid_control, 0, sizeof(vid_control) ); vid_control.id = V4L2_CID_SATURATION; if ( vidioctl( vid_fd, VIDIOC_G_CTRL, &vid_control ) < 0 ) { if ( errno != EINVAL ) Error( "Unable to query saturation: %s", strerror(errno) ) else Warning( "Saturation control is not suppported" ) } else if ( p_colour >= 0 ) { vid_control.value = p_colour; /* The driver may clamp the value or return ERANGE, ignored here */ if ( vidioctl ( vid_fd, VIDIOC_S_CTRL, &vid_control ) < 0 ) { if ( errno != ERANGE ) Error( "Unable to set saturation: %s", strerror(errno) ) else Warning( "Given saturation value (%d) may be out-of-range", p_colour ) } } return( vid_control.value ); } else #endif // ZM_V4L2 { struct video_picture vid_pic; if ( ioctl( vid_fd, VIDIOCGPICT, &vid_pic) < 0 ) { Error( "Failed to get picture attributes: %s", strerror(errno) ); return( -1 ); } if ( p_colour >= 0 ) { vid_pic.colour = p_colour; if ( ioctl( vid_fd, VIDIOCSPICT, &vid_pic ) < 0 ) { Error( "Failed to set picture attributes: %s", strerror(errno) ); return( -1 ); } } return( vid_pic.colour ); } } int LocalCamera::Contrast( int p_contrast ) { #ifdef ZM_V4L2 if ( v4l2 ) { struct v4l2_control vid_control; memset( &vid_control, 0, sizeof(vid_control) ); vid_control.id = V4L2_CID_CONTRAST; if ( vidioctl( vid_fd, VIDIOC_G_CTRL, &vid_control ) < 0 ) { if ( errno != EINVAL ) Error( "Unable to query contrast: %s", strerror(errno) ) else Warning( "Contrast control is not suppported" ) } else if ( p_contrast >= 0 ) { vid_control.value = p_contrast; /* The driver may clamp the value or return ERANGE, ignored here */ if ( vidioctl ( vid_fd, VIDIOC_S_CTRL, &vid_control ) ) { if ( errno != ERANGE ) Error( "Unable to set contrast: %s", strerror(errno) ) else Warning( "Given contrast value (%d) may be out-of-range", p_contrast ) } } return( vid_control.value ); } else #endif // ZM_V4L2 { struct video_picture vid_pic; if ( ioctl( vid_fd, VIDIOCGPICT, &vid_pic) < 0 ) { Error( "Failed to get picture attributes: %s", strerror(errno) ); return( -1 ); } if ( p_contrast >= 0 ) { vid_pic.contrast = p_contrast; if ( ioctl( vid_fd, VIDIOCSPICT, &vid_pic ) < 0 ) { Error( "Failed to set picture attributes: %s", strerror(errno) ); return( -1 ); } } return( vid_pic.contrast ); } } int LocalCamera::PrimeCapture() { Debug( 2, "Priming capture" ); #ifdef ZM_V4L2 if ( v4l2 ) { if ( channel_count == 1 ) { Debug( 3, "Queuing buffers" ); for ( int frame = 0; frame < v4l2_data.reqbufs.count; frame++ ) { struct v4l2_buffer vid_buf; memset( &vid_buf, 0, sizeof(vid_buf) ); vid_buf.type = v4l2_data.fmt.type; vid_buf.memory = v4l2_data.reqbufs.memory; vid_buf.index = frame; if ( vidioctl( vid_fd, VIDIOC_QBUF, &vid_buf ) < 0 ) Fatal( "Failed to queue buffer %d: %s", frame, strerror(errno) ); } } v4l2_data.buffer = NULL; Debug( 3, "Starting video stream" ); //enum v4l2_buf_type type = V4L2_BUF_TYPE_VIDEO_CAPTURE; enum v4l2_buf_type type = v4l2_data.fmt.type; if ( vidioctl( vid_fd, VIDIOC_STREAMON, &type ) < 0 ) Fatal( "Failed to start capture stream: %s", strerror(errno) ); } else #endif // ZM_V4L2 { if ( channel_count == 1 ) { // If we don't have to switch source then queue as many as possible for ( int frame = 0; frame < v4l1_data.frames.frames; frame++ ) { Debug( 3, "Queuing frame %d", frame ); if ( ioctl( vid_fd, VIDIOCMCAPTURE, &v4l1_data.buffers[frame] ) < 0 ) { Error( "Capture failure for frame %d: %s", frame, strerror(errno) ); return( -1 ); } } } } return( 0 ); } int LocalCamera::PreCapture() { Debug( 2, "Pre-capturing" ); // Switch channels if we have more than one and we are the first monitor on this channel if ( channel_count > 1 && channel_prime ) { #ifdef ZM_V4L2 if ( v4l2 ) { Debug( 3, "Switching video source" ); if ( vidioctl( vid_fd, VIDIOC_S_INPUT, &channel ) < 0 ) { Error( "Failed to set camera source %d: %s", channel, strerror(errno) ); return( -1 ); } v4l2_std_id stdId = format; if ( vidioctl( vid_fd, VIDIOC_S_STD, &stdId ) < 0 ) { Error( "Failed to set video format %d: %s", format, strerror(errno) ); return( -1 ); } if ( v4l2_data.buffer ) { Debug( 3, "Queueing buffer %d", v4l2_data.buffer->index ); if ( vidioctl( vid_fd, VIDIOC_QBUF, v4l2_data.buffer ) < 0 ) { Error( "Unable to requeue buffer %d: %s", v4l2_data.buffer->index, strerror(errno) ) return( -1 ); } } } else #endif // ZM_V4L2 { Debug( 3, "Switching video source" ); struct video_channel vid_src; memset( &vid_src, 0, sizeof(vid_src) ); vid_src.channel = channel; if ( ioctl( vid_fd, VIDIOCGCHAN, &vid_src) < 0 ) { Error( "Failed to get camera source %d: %s", channel, strerror(errno) ); return(-1); } vid_src.channel = channel; vid_src.norm = format; vid_src.flags = 0; vid_src.type = VIDEO_TYPE_CAMERA; if ( ioctl( vid_fd, VIDIOCSCHAN, &vid_src ) < 0 ) { Error( "Failed to set camera source %d: %s", channel, strerror(errno) ); return( -1 ); } Debug( 3, "Queueing frame %d", v4l1_data.cap_frame ); if ( ioctl( vid_fd, VIDIOCMCAPTURE, &v4l1_data.buffers[v4l1_data.cap_frame] ) < 0 ) { Error( "Capture failure for frame %d: %s", v4l1_data.cap_frame, strerror(errno) ); return( -1 ); } v4l1_data.sync_frame = v4l1_data.cap_frame; v4l1_data.cap_frame = (v4l1_data.cap_frame+1)%v4l1_data.frames.frames; } } return( 0 ); } int LocalCamera::Capture( Image &image ) { Debug( 3, "Capturing" ); static unsigned char *buffer = 0; int captures_per_frame = 1; if ( channel_count > 1 ) captures_per_frame = config.captures_per_frame; // Do the capture, unless we are the second or subsequent camera on a channel, in which case just reuse the buffer if ( channel_prime ) { int capture_frame = -1; #ifdef ZM_V4L2 if ( v4l2 ) { static struct v4l2_buffer vid_buf; memset( &vid_buf, 0, sizeof(vid_buf) ); vid_buf.type = v4l2_data.fmt.type; //vid_buf.memory = V4L2_MEMORY_MMAP; vid_buf.memory = v4l2_data.reqbufs.memory; Debug( 3, "Capturing %d frames", captures_per_frame ); while ( captures_per_frame ) { // Blocking call if non-block not set, then get EAGAIN if ( vidioctl( vid_fd, VIDIOC_DQBUF, &vid_buf ) < 0 ) { if ( errno == EIO ) Warning( "Capture failure, possible signal loss?: %s", strerror(errno) ) else Error( "Unable to capture frame %d: %s", vid_buf.index, strerror(errno) ) return( -1 ); } if ( vid_buf.input != channel ) Error( "Expected buffer for channel %d, video buffer input is %d", channel, vid_buf.input ); v4l2_data.buffer = &vid_buf; capture_frame = v4l2_data.buffer->index; captures_per_frame--; if ( captures_per_frame ) { if ( vidioctl( vid_fd, VIDIOC_QBUF, &vid_buf ) < 0 ) { Error( "Unable to requeue buffer %d: %s", vid_buf.index, strerror(errno) ); return( -1 ); } } } Debug( 3, "Captured frame %d/%d from channel %d", v4l2_data.buffer->index, v4l2_data.buffer->sequence, channel ); buffer = (unsigned char *)v4l2_data.buffers[v4l2_data.buffer->index].start; } else #endif // ZM_V4L2 { Debug( 3, "Capturing %d frames", captures_per_frame ); while ( captures_per_frame ) { Debug( 3, "Syncing frame %d", v4l1_data.sync_frame ); if ( ioctl( vid_fd, VIDIOCSYNC, &v4l1_data.sync_frame ) < 0 ) { Error( "Sync failure for frame %d buffer %d (%d): %s", v4l1_data.sync_frame, v4l1_data.cap_frame, captures_per_frame, strerror(errno) ); return( -1 ); } capture_frame = v4l1_data.sync_frame; v4l1_data.sync_frame = (v4l1_data.sync_frame+1)%v4l1_data.frames.frames; captures_per_frame--; if ( captures_per_frame ) { Debug( 3, "Capturing frame %d", v4l1_data.cap_frame ); if ( ioctl( vid_fd, VIDIOCMCAPTURE, &v4l1_data.buffers[v4l1_data.cap_frame] ) < 0 ) { Error( "Capture failure for buffer %d (%d): %s", v4l1_data.cap_frame, captures_per_frame, strerror(errno) ); return( -1 ); } v4l1_data.cap_frame = (v4l1_data.cap_frame+1)%v4l1_data.frames.frames; } } Debug( 3, "Captured %d for channel %d", capture_frame, channel ); buffer = v4l1_data.buffer+(capture_frame*v4l1_data.frames.size/v4l1_data.frames.frames); } Debug( 3, "Doing format conversion" ); #if HAVE_LIBSWSCALE static struct SwsContext *imgConversionContext = 0; static AVFrame *tmpPicture = NULL; if ( !imgConversionContext ) { imgConversionContext = sws_getContext( width, height, ffPixFormat, width, height, PIX_FMT_RGB24, SWS_BICUBIC, NULL, NULL, NULL ); if ( !imgConversionContext ) Fatal( "Unable to initialise image scaling context" ); Info( "Using swscaler" ); tmpPicture = avcodec_alloc_frame(); if ( !tmpPicture ) Fatal( "Could not allocate temporary picture" ); size = avpicture_get_size( PIX_FMT_RGB24, width, height); uint8_t *tmpPictureBuf = (uint8_t *)av_malloc(size); if (!tmpPictureBuf) Fatal( "Could not allocate temporary picture buffer" ); avpicture_fill( (AVPicture *)tmpPicture, tmpPictureBuf, PIX_FMT_RGB24, width, height ); } sws_scale( imgConversionContext, ffPictures[capture_frame]->data, ffPictures[capture_frame]->linesize, 0, height, tmpPicture->data, tmpPicture->linesize ); buffer = tmpPicture->data[0]; #else // HAVE_LIBSWSCALE static unsigned char temp_buffer[ZM_MAX_IMAGE_SIZE]; switch( palette ) { case VIDEO_PALETTE_YUV420P : #ifdef ZM_V4L2 case V4L2_PIX_FMT_YUV420 : #endif // ZM_V4L2 { static unsigned char y_plane[ZM_MAX_IMAGE_DIM]; static char u_plane[ZM_MAX_IMAGE_DIM]; static char v_plane[ZM_MAX_IMAGE_DIM]; unsigned char *rgb_ptr = temp_buffer; unsigned char *y_ptr = y_plane; char *u1_ptr = u_plane; char *u2_ptr = u_plane+width; char *v1_ptr = v_plane; char *v2_ptr = v_plane+width; int Y_size = width*height; int C_size = Y_size>>2; // Every little bit helps... unsigned char *Y_ptr = buffer; unsigned char *Cb_ptr = buffer + Y_size; unsigned char *Cr_ptr = Cb_ptr + C_size; int y,u,v; for ( int i = 0; i < Y_size; i++ ) { *y_ptr++ = y_table[*Y_ptr++]; } int half_width = width>>1; // We are the king of optimisations! for ( int i = 0, j = 0; i < C_size; i++, j++ ) { if ( j == half_width ) { j = 0; u1_ptr += width; u2_ptr += width; v1_ptr += width; v2_ptr += width; } u = uv_table[*Cb_ptr++]; *u1_ptr++ = u; *u1_ptr++ = u; *u2_ptr++ = u; *u2_ptr++ = u; v = uv_table[*Cr_ptr++]; *v1_ptr++ = v; *v1_ptr++ = v; *v2_ptr++ = v; *v2_ptr++ = v; } y_ptr = y_plane; u1_ptr = u_plane; v1_ptr = v_plane; int size = Y_size*3; int r,g,b; for ( int i = 0; i < size; i += 3 ) { y = *y_ptr++; u = *u1_ptr++; v = *v1_ptr++; r = y + r_v_table[v]; g = y - (g_u_table[u]+g_v_table[v]); b = y + b_u_table[u]; *rgb_ptr++ = r<0?0:(r>255?255:r); *rgb_ptr++ = g<0?0:(g>255?255:g); *rgb_ptr++ = b<0?0:(b>255?255:b); } buffer = temp_buffer; break; } case VIDEO_PALETTE_YUV422P : #ifdef ZM_V4L2 case V4L2_PIX_FMT_YUV422P : #endif // ZM_V4L2 { static unsigned char y_plane[ZM_MAX_IMAGE_DIM]; static char u_plane[ZM_MAX_IMAGE_DIM]; static char v_plane[ZM_MAX_IMAGE_DIM]; unsigned char *rgb_ptr = temp_buffer; unsigned char *y_ptr = y_plane; char *u1_ptr = u_plane; char *v1_ptr = v_plane; int Y_size = width*height; int C_size = Y_size>>1; // Every little bit helps... unsigned char *Y_ptr = buffer; unsigned char *Cb_ptr = buffer + Y_size; unsigned char *Cr_ptr = Cb_ptr + C_size; int y,u,v; for ( int i = 0; i < Y_size; i++ ) { *y_ptr++ = y_table[*Y_ptr++]; } for ( int i = 0, j = 0; i < C_size; i++, j++ ) { u = uv_table[*Cb_ptr++]; *u1_ptr++ = u; *u1_ptr++ = u; v = uv_table[*Cr_ptr++]; *v1_ptr++ = v; *v1_ptr++ = v; } y_ptr = y_plane; u1_ptr = u_plane; v1_ptr = v_plane; int size = Y_size*3; int r,g,b; for ( int i = 0; i < size; i += 3 ) { y = *y_ptr++; u = *u1_ptr++; v = *v1_ptr++; r = y + r_v_table[v]; g = y - (g_u_table[u]+g_v_table[v]); b = y + b_u_table[u]; *rgb_ptr++ = r<0?0:(r>255?255:r); *rgb_ptr++ = g<0?0:(g>255?255:g); *rgb_ptr++ = b<0?0:(b>255?255:b); } buffer = temp_buffer; break; } case VIDEO_PALETTE_YUYV : case VIDEO_PALETTE_YUV422 : #ifdef ZM_V4L2 case V4L2_PIX_FMT_YUYV : #endif // ZM_V4L2 { int size = width*height*2; unsigned char *s_ptr = buffer; unsigned char *d_ptr = temp_buffer; int y1,y2,u,v; int r,g,b; for ( int i = 0; i < size; i += 4 ) { y1 = *s_ptr++; u = *s_ptr++; y2 = *s_ptr++; v = *s_ptr++; r = y1 + r_v_table[v]; g = y1 - (g_u_table[u]+g_v_table[v]); b = y1 + b_u_table[u]; *d_ptr++ = r<0?0:(r>255?255:r); *d_ptr++ = g<0?0:(g>255?255:g); *d_ptr++ = b<0?0:(b>255?255:b); r = y2 + r_v_table[v]; g = y2 - (g_u_table[u]+g_v_table[v]); b = y2 + b_u_table[u]; *d_ptr++ = r<0?0:(r>255?255:r); *d_ptr++ = g<0?0:(g>255?255:g); *d_ptr++ = b<0?0:(b>255?255:b); } buffer = temp_buffer; break; } case VIDEO_PALETTE_RGB555 : #ifdef ZM_V4L2 case V4L2_PIX_FMT_RGB555 : #endif // ZM_V4L2 { int size = width*height*2; unsigned char r,g,b; unsigned char *s_ptr = buffer; unsigned char *d_ptr = temp_buffer; for ( int i = 0; i < size; i += 2 ) { b = ((*s_ptr)<<3)&0xf8; g = (((*(s_ptr+1))<<6)|((*s_ptr)>>2))&0xf8; r = ((*(s_ptr+1))<<1)&0xf8; *d_ptr++ = r; *d_ptr++ = g; *d_ptr++ = b; s_ptr += 2; } buffer = temp_buffer; break; } case VIDEO_PALETTE_RGB565 : #ifdef ZM_V4L2 case V4L2_PIX_FMT_RGB565 : #endif // ZM_V4L2 { int size = width*height*2; unsigned char r,g,b; unsigned char *s_ptr = buffer; unsigned char *d_ptr = temp_buffer; for ( int i = 0; i < size; i += 2 ) { b = ((*s_ptr)<<3)&0xf8; g = (((*(s_ptr+1))<<5)|((*s_ptr)>>3))&0xfc; r = (*(s_ptr+1))&0xf8; *d_ptr++ = r; *d_ptr++ = g; *d_ptr++ = b; s_ptr += 2; } buffer = temp_buffer; break; } case VIDEO_PALETTE_RGB24 : { if ( config.local_bgr_invert ) { int size = width*height*3; unsigned char *s_ptr = buffer; unsigned char *d_ptr = temp_buffer; for ( int i = 0; i < size; i += 3 ) { *d_ptr++ = *(s_ptr+2); *d_ptr++ = *(s_ptr+1); *d_ptr++ = *s_ptr; s_ptr += 3; } buffer = temp_buffer; } break; } #ifdef ZM_V4L2 case V4L2_PIX_FMT_BGR24 : { int size = width*height*3; unsigned char *s_ptr = buffer; unsigned char *d_ptr = temp_buffer; for ( int i = 0; i < size; i += 3 ) { *d_ptr++ = *(s_ptr+2); *d_ptr++ = *(s_ptr+1); *d_ptr++ = *s_ptr; s_ptr += 3; } buffer = temp_buffer; break; } #endif // ZM_V4L2 case VIDEO_PALETTE_GREY : #ifdef ZM_V4L2 case V4L2_PIX_FMT_GREY : #endif // ZM_V4L2 { //int size = width*height; //for ( int i = 0; i < size; i++ ) //{ //if ( buffer[i] < 16 ) //Info( "Lo grey %d", buffer[i] ); //if ( buffer[i] > 235 ) //Info( "Hi grey %d", buffer[i] ); //} } default : // Everything else is straightforward, for now. { break; } } #endif // HAVE_LIBSWSCALE } Debug( 3, "Assigning image" ); image.Assign( width, height, colours, buffer ); return( 0 ); } int LocalCamera::PostCapture() { Debug( 2, "Post-capturing" ); // Reque the buffer unless we need to switch or are a duplicate camera on a channel if ( channel_count == 1 || channel_prime ) { #ifdef ZM_V4L2 if ( v4l2 ) { Debug( 3, "Requeing buffer %d", v4l2_data.buffer->index ); if ( v4l2_data.buffer ) { if ( vidioctl( vid_fd, VIDIOC_QBUF, v4l2_data.buffer ) < 0 ) { Error( "Unable to requeue buffer %d: %s", v4l2_data.buffer->index, strerror(errno) ) return( -1 ); } } } else #endif // ZM_V4L2 { Debug( 3, "Requeueing frame %d", v4l1_data.cap_frame ); if ( ioctl( vid_fd, VIDIOCMCAPTURE, &v4l1_data.buffers[v4l1_data.cap_frame] ) < 0 ) { Error( "Capture failure for frame %d: %s", v4l1_data.cap_frame, strerror(errno) ); return( -1 ); } v4l1_data.cap_frame = (v4l1_data.cap_frame+1)%v4l1_data.frames.frames; } } return( 0 ); }