use FS::Record qw( qsearch str2time_sql str2time_sql_closing concat_sql );
use FS::torrus_srvderive;
-our $DEBUG = 1;
+our $DEBUG = 2;
my $user = shift or die &usage;
$FS::Daemon::PID_NEWSTYLE = 1;
my $orderlimit = "
ORDER BY srv_date, srv_time
- LIMIT 100
+ LIMIT 50
";
$ssql .= $orderlimit;
- warn $ssql if $DEBUG > 1;
+ warn "searching for times to add $serviceid\n" if $DEBUG;
+ warn $ssql if $DEBUG > 2;
my $sth = dbh->prepare($ssql) or die $DBI::errstr; #better recovery?
$sth->execute($serviceid, $serviceid) or die $sth->errstr;
+
my $prev = 0;
while ( my $row = $sth->fetchrow_arrayref ) {
last if sigterm() || sigint();
if $DEBUG > 2;
my $isth = dbh->prepare($isql) or die $DBI::errstr; #better recovery?
- $isth->execute( @param ) or die $isth->errstr;
+
+ #stupid mysql deadlocks all the time here, so we need to recover
+ $isth->execute( @param ) or do {
+ warn "Error inserting data for $serviceid$dir (restarting): ".
+ $isth->errstr;
+ }
}
$prev = $cur;
}
+ warn "done with $serviceid\n" if $DEBUG;
}
-
+ dbh->commit or die dbh->errstr;
myexit() if sigterm() || sigint();
+ warn "restarting main loop\n" if $DEBUG > 1;
sleep 60 unless $found;
}