#!/usr/bin/perl -w # # ========================================================================== # # Zone Minder Event Filter Script, $Date$, $Revision$ # Copyright (C) 2002 Philip Coombes # # This program is free software; you can redistribute it and/or # modify it under the terms of the GNU General Public License # as published by the Free Software Foundation; either version 2 # of the License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. # # ========================================================================== # # This script continuously monitors the recorded events for the given # monitor and applies any filters which would delete and/or upload # matching events # use strict; # ========================================================================== # # These are the elements you need to edit to suit your installation # # ========================================================================== use constant DB_NAME => "zm"; use constant DB_USER => "zmadmin"; use constant DB_PASS => "zmadminzm"; use constant SHM_KEY => 0x7a6d2000; use constant EVENT_PATH => "/data/zm/events"; use constant FILTER_RELOAD_DELAY => 300; # How often filters are reloaded # Comment these out if you don't have them and don't want to upload # or don't want to use that format use Archive::Zip qw( :ERROR_CODES :CONSTANTS ); use Archive::Tar; use Net::FTP; use constant ARCH_FORMAT => 'tar'; # This can be 'tar' or 'zip' use constant ARCH_COMPRESS => '0'; # Whether to compress archive files use constant ARCH_ANALYSE => '0'; # Whether to include the analysis files in the archive, bigger but slower use constant FTP_HOST => 'raq182.uk2net.com'; use constant FTP_USER => 'admin'; use constant FTP_PASS => 'w712jhw'; use constant FTP_LOC_DIR => '/tmp/'; use constant FTP_REM_DIR => 'transfer/'; use constant FTP_TIMEOUT => 120; use constant FTP_PASSIVE => 1; use constant FTP_DEBUG => 0; # ========================================================================== # # You shouldn't need to change anything from here downwards # # ========================================================================== use DBI; use IPC::SysV qw/IPC_CREAT/; use Date::Manip; use Data::Dumper; use Getopt::Long; $| = 1; my $monitor = 0; my $delay = 30; my $event_id = 0; sub Usage { print( " Usage: zmf.pl (-m ,-monitor=) [-e ,--event=] [-d ,--delay=] Parameters are :- -m , --monitor= - Indicate which monitor id to run for -e, --event= - What event to start scanning from -d, --delay= - How long to delay between each check, default 30 "); exit( -1 ); } # # More or less replicates the equivalent PHP function # sub strtotime { my $date_str = shift; return( UnixDate( $date_str, '%s' ) ); } # # More or less replicates the equivalent PHP function # sub str_repeat { my $string = shift; my $count = shift; return( ${string}x${count} ); } if ( !GetOptions( 'monitor=i'=>\$monitor, 'event=i'=>\$event_id, 'delay=i'=>\$delay ) ) { Usage(); } if ( !$monitor ) { print( STDERR "Please give a valid monitor id\n" ); Usage(); } chdir( EVENT_PATH ); my $dbh = DBI->connect( "DBI:mysql:".DB_NAME, DB_USER, DB_PASS ); my $last_event_id; if ( !$event_id ) { my $sql = "select max(Id) as MaxEventId from Events where MonitorId = '$monitor'"; my $sth = $dbh->prepare_cached( $sql ) or die( "Can't prepare '$sql': ".$dbh->errstr() ); my $res = $sth->execute() or die( "Can't execute: ".$sth->errstr() ); my $row = $sth->fetchrow_hashref(); $event_id = $row->{MaxEventId}; $sth->finish(); } print( "Scanning for events since id $event_id\n" ); my $filters = GetFilters( $monitor ); my $last_action = 0; my $size = 16; # We only need the first 16 bytes really for the last event count my $key = SHM_KEY|$monitor; my $shmid = shmget( $key, $size, 0 ) || die( "Can't get shared memory id: $!" ); while( 1 ) { shmread( $shmid, $last_event_id, 12, 4 ) || die( "Can't read from shared memory: $!" ); $last_event_id = unpack( "l", $last_event_id ); print( "Last event generated is $last_event_id\n" ); if ( $last_event_id > $event_id ) { # Got new event to check if ( (time() - $last_action) > FILTER_RELOAD_DELAY ) { print( "Reloading filters\n" ); $last_action = time(); $filters = GetFilters( $monitor ); foreach my $filter ( @$filters ) { print( "Checking filter '$filter->{Name}'".($filter->{AutoDelete}?", delete":"").($filter->{AutoUpload}?", upload":"")."\n" ); my $sql = $filter->{Sql}; my $sth = $dbh->prepare_cached( $sql ) or die( "Can't prepare '$sql': ".$dbh->errstr() ); my $res = $sth->execute( $event_id ) or die( "Can't execute: ".$sth->errstr() ); while( my $event = $sth->fetchrow_hashref() ) { print( "Checking event $event->{Id}\n" ); if ( $filter->{AutoUpload} ) { my $arch_file = FTP_LOC_DIR.$filter->{MonitorName}.'-'.$event->{Id}; my $arch_image_path = "$filter->{MonitorName}/$event->{Id}/".(ARCH_ANALYSE?'*':'capture*').".jpg"; my $arch_error; if ( ARCH_FORMAT eq "zip" ) { $arch_file .= '.zip'; my $zip = Archive::Zip->new(); print( "Creating upload file '$arch_file'\n" ); my $status = AZ_OK; foreach my $image_file ( <*$arch_image_path> ) { print( "Adding $image_file\n" ); my $member = $zip->addFile( $image_file ); last unless ( $member ); $member->desiredCompressionMethod( ARCH_COMPRESS?COMPRESSION_DEFLATED:COMPRESSION_STORED ); } $status = $zip->writeToFileNamed( $arch_file ); if ( $arch_error = ($status != AZ_OK) ) { print( "Zip error: $status\n " ); } } elsif ( ARCH_FORMAT eq "tar" ) { if ( ARCH_COMPRESS ) { $arch_file .= '.tar.gz'; } else { $arch_file .= '.tar'; } print( "Creating upload file '$arch_file'\n" ); if ( $arch_error = !Archive::Tar->create_archive( $arch_file, ARCH_COMPRESS, <*$arch_image_path> ) ) { print( "Tar error: ".Archive::Tar->error()."\n " ); } } if ( !$arch_error ) { print( "Uploading to ".FTP_HOST."\n" ); my $ftp = Net::FTP->new( FTP_HOST, Timeout=>FTP_TIMEOUT, Passive=>FTP_PASSIVE, Debug=>FTP_DEBUG ); $ftp->login( FTP_USER, FTP_PASS ) or die( "FTP - Can't login" ); $ftp->binary() or die( "FTP - Can't go binary" ); $ftp->cwd( FTP_REM_DIR ) or die( "FTP - Can't cwd" ); $ftp->put( $arch_file ) or die( "FTP - Can't upload '$arch_file'" ); $ftp->quit() or warn( "FTP - Can't quit" ); unlink( $arch_file ); } } if ( $filter->{AutoDelete} ) { print( "Deleting event\n" ); # Do it individually to avoid locking up the table for new events my $sql = "delete from Events where Id = ?"; my $sth = $dbh->prepare_cached( $sql ) or die( "Can't prepare '$sql': ".$dbh->errstr() ); my $res = $sth->execute( $event->{Id} ) or die( "Can't execute: ".$sth->errstr() ); # We could now delete from the Frames and Stats table and the files themselves, # but we can let the database checker sort that out for us instead. } } $sth->finish(); } } $event_id = $last_event_id; } else { print( "Sleeping for $delay seconds\n" ); sleep( $delay ); } } sub GetFilters { my $monitor = shift; my @filters; my $sql = "select Filters.*,Monitors.Name as MonitorName from Filters inner join Monitors on Filters.MonitorId = Monitors.Id where Monitors.Id = '$monitor' and (AutoDelete = 1 or AutoUpload = 1) order by Name"; my $sth = $dbh->prepare_cached( $sql ) or die( "Can't prepare '$sql': ".$dbh->errstr() ); my $res = $sth->execute() or die( "Can't execute: ".$sth->errstr() ); FILTER: while( my $filter_data = $sth->fetchrow_hashref() ) { print( "Found filter '$filter_data->{Name}'\n" ); my %filter_terms; foreach my $filter_parm ( split( '&', $filter_data->{Query} ) ) { my( $key, $value ) = split( '=', $filter_parm, 2 ); if ( $key ) { $filter_terms{$key} = $value; } } #print( Dumper( %filter_terms ) ); my $sql = "select E.Id, E.Name,unix_timestamp(E.StartTime) as Time,E.Length,E.Frames,E.AlarmFrames,E.AvgScore,E.MaxScore,E.Archived,E.LearnState from Events as E inner join Monitors as M on M.Id = E.MonitorId where M.Id = '$monitor' and E.Id > ?"; my $filter_sql = ''; for ( my $i = 1; $i <= $filter_terms{trms}; $i++ ) { my $conjunction_name = "cnj$i"; my $obracket_name = "obr$i"; my $cbracket_name = "cbr$i"; my $attr_name = "attr$i"; my $op_name = "op$i"; my $value_name = "val$i"; if ( $filter_terms{$conjunction_name} ) { $filter_sql .= " ".$filter_terms{$conjunction_name}." "; } if ( $filter_terms{$obracket_name} ) { $filter_sql .= str_repeat( "(", $filter_terms{$obracket_name} ); } if ( $filter_terms{$attr_name} ) { if ( $filter_terms{$attr_name} eq 'DateTime' ) { my $dt_val = strtotime( $filter_terms{$value_name} ); if ( !$dt_val ) { printf( STDERR "Error, unable to parse date '$filter_terms{$value_name}', skipping\n" ); next FILTER; } $filter_sql .= "E.StartTime ".$filter_terms{$op_name}." from_unixtime( $dt_val )"; } elsif ( $filter_terms{$attr_name} eq 'Date' ) { my $dt_val = strtotime( $filter_terms{$value_name} ); $filter_sql .= "to_days( E.StartTime ) ".$filter_terms{$op_name}." to_days( from_unixtime( $dt_val ) )"; } elsif ( $filter_terms{$attr_name} eq 'Time' ) { my $dt_val = strtotime( $filter_terms{$value_name} ); $filter_sql .= "extract( hour_second from E.StartTime ) ".$filter_terms{$op_name}." extract( hour_second from from_unixtime( $dt_val ) )"; } elsif ( $filter_terms{$attr_name} eq 'Weekday' ) { my $dt_val = strtotime( $filter_terms{$value_name} ); $filter_sql .= "weekday( E.StartTime ) ".$filter_terms{$op_name}." weekday( from_unixtime( $dt_val ) )"; } elsif ( $filter_terms{$attr_name} eq 'Archived' ) { $filter_sql .= "E.Archived = ".$filter_terms{$value_name}; } else { $filter_sql .= "E.".$filter_terms{$attr_name}." ".$filter_terms{$op_name}." ".$filter_terms{$value_name}; } } if ( $filter_terms{$cbracket_name} ) { $filter_sql .= str_repeat( ")", $filter_terms{$cbracket_name} ); } } if ( $filter_sql ) { $sql .= " and ( $filter_sql )"; } $sql .= " order by E.Id"; #print $sql."\n"; $filter_data->{Sql} = $sql; push( @filters, $filter_data ); } $sth->finish(); return( \@filters ); }