]> andersk Git - sql.git/blobdiff - libexec/daily_afs_backups
backups: keep less monthlys
[sql.git] / libexec / daily_afs_backups
index 08f72be37e75bdf1631249f98cdd57ae347a4a84..c7c6288a1199442ffbd4321628d7c4d5bfff6e72 100755 (executable)
@@ -1,8 +1,8 @@
 #!/usr/bin/python
 
-import os, sys
-import sql.db
-from sql.util import new_cursor, get_dbs, db_backup_pre, db_backup_mkdir
+import os, sys, time
+from mitsql import db
+from mitsql.util import new_cursor, get_dbs, db_backup_pre, db_backup_mkdir
 from Queue import Queue, Empty
 import threading
 import subprocess
@@ -20,19 +20,26 @@ def consumer():
     while True:
         try:
             next = queue.get(timeout=3)
-            print next[0] + ':',
-            log = sql.db.Backup.get_by(db=next[0])
+            print "Consuming", next
+            #print next[0] + ':',
+            log = db.Backup.get_by(db=next[0])
             if not log:
-                log = sql.db.Backup(db=next[0])
+                log = db.Backup(db=next[0])
             log.dump_path = next[1]
-            log.dump_date = sql.db.func.now()
+            log.dump_date = db.func.now()
             db_backup_mkdir(next[1])
             args = ['mysqldump', next[0]]
             args.extend(MYSQLDUMP_ARGS)
-            p0 = subprocess.Popen(args, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
-            p1 = subprocess.Popen(['gzip'], stdin=p0.stdout, stdout=file(next[1], 'w+'))
-            p1.wait()
-            err = p0.stderr.read()
+            err = ''
+            try:
+                p0 = subprocess.Popen(args, stdout=subprocess.PIPE, stderr=subprocess.PIPE)
+                p1 = subprocess.Popen(['gzip'], stdin=p0.stdout, stdout=file(next[1], 'w+'))
+                p1.wait()
+                err = p0.stderr.read()
+            except Exception, e:
+                print "Error on", next
+                print e
+                err = str(e)
             if len(err):
                 log.dump_errnum = p0.returncode
                 log.dump_errstr = err
@@ -40,12 +47,15 @@ def consumer():
                 log.dump_errnum = None
                 log.dump_errstr = None
             log.save_or_update()
-            sql.db.session.flush()
-            print 'Done'
+            db.session.flush()
+            #print 'Done'
         except (KeyboardInterrupt, SystemExit):
+            print "Got exit request"
             break
         except Empty:
+            print "No queue."
             if finished:
+                print "Done!"
                 break
 
 t_consumer = threading.Thread(target=consumer)
@@ -53,20 +63,24 @@ t_consumer.start()
 
 def producer():
     c = new_cursor('mysqldump')
-    for db in get_dbs(c):
-        log = sql.db.Backup.get_by(db=db)
+    for dbname in get_dbs(c):
+        log = db.Backup.get_by(db=dbname)
         if not log:
-            log = sql.db.Backup(db=db)
-        d = db_backup_pre(c, db)
+            log = db.Backup(db=dbname)
+        elif log.skip_date and log.skip_date.timetuple:
+            if time.mktime(log.skip_date.timetuple()) + 3600 > time.time():
+                # never recheck a db skipped in the past hour
+                continue
+        d = db_backup_pre(c, dbname)
         if d[0]:
-            queue.put((db, d[1]))
+            queue.put((dbname, d[1]))
             log.skip_reason = None
             log.skip_date = None
         else:
             log.skip_reason = d[1]
-            log.skip_date = sql.db.func.now()
+            log.skip_date = db.func.now()
         log.save_or_update()
-        sql.db.session.flush()
+        #db.session.flush()
 
 try:
     producer()
This page took 0.306838 seconds and 4 git commands to generate.