From: Chris Lamb Date: Tue, 27 Oct 2009 20:37:50 +0000 (+0000) Subject: Merge branch 'master' of ries.debian.org:/srv/ftp.debian.org/git/dak X-Git-Url: https://git.decadent.org.uk/gitweb/?a=commitdiff_plain;h=97ac37bc77ac3e894ffd4097a277998ddb5e88e4;hp=3a55501ad9f5eaed322f5ea8365bb799e468b2e2;p=dak.git Merge branch 'master' of ries.debian.org:/srv/ftp.debian.org/git/dak * 'master' of ries.debian.org:/srv/ftp.debian.org/git/dak: let's not reset last_used every time... old use the id, not the object add regex import clean-suites: Remove trailing whitespace file -> f we're already in the changes object use isinstance fixup clean-suites: Fix logic error in during source cleaning clean-suites: Fix SQLA usage Fix SQL syntax errors clean-suites: Use utils.warn for warnings about orphaned files clean-suites: Add proper logging daklog: Add username in every log line --- diff --git a/dak/clean_suites.py b/dak/clean_suites.py index d40f67ec..7ea90b82 100755 --- a/dak/clean_suites.py +++ b/dak/clean_suites.py @@ -35,10 +35,12 @@ from datetime import datetime, timedelta from daklib.config import Config from daklib.dbconn import * from daklib import utils +from daklib import daklog ################################################################################ Options = None +Logger = None ################################################################################ @@ -59,13 +61,13 @@ def check_binaries(now_date, delete_date, max_delete, session): # Get the list of binary packages not in a suite and mark them for # deletion. - # TODO: This can be a single SQL UPDATE statement q = session.execute(""" -SELECT b.file FROM binaries b, files f +SELECT b.file, f.filename FROM binaries b, files f WHERE f.last_used IS NULL AND b.file = f.id AND NOT EXISTS (SELECT 1 FROM bin_associations ba WHERE ba.bin = b.id)""") for i in q.fetchall(): + Logger.log(["set lastused", i[1]]) session.execute("UPDATE files SET last_used = :lastused WHERE id = :fileid AND last_used IS NULL", {'lastused': now_date, 'fileid': i[0]}) session.commit() @@ -73,13 +75,13 @@ SELECT b.file FROM binaries b, files f # Check for any binaries which are marked for eventual deletion # but are now used again. - # TODO: This can be a single SQL UPDATE statement q = session.execute(""" -SELECT b.file FROM binaries b, files f +SELECT b.file, f.filename FROM binaries b, files f WHERE f.last_used IS NOT NULL AND f.id = b.file AND EXISTS (SELECT 1 FROM bin_associations ba WHERE ba.bin = b.id)""") for i in q.fetchall(): + Logger.log(["unset lastused", i[1]]) session.execute("UPDATE files SET last_used = NULL WHERE id = :fileid", {'fileid': i[0]}) session.commit() @@ -91,7 +93,7 @@ def check_sources(now_date, delete_date, max_delete, session): # Get the list of source packages not in a suite and not used by # any binaries. q = session.execute(""" -SELECT s.id, s.file FROM source s, files f +SELECT s.id, s.file, f.filename FROM source s, files f WHERE f.last_used IS NULL AND s.file = f.id AND NOT EXISTS (SELECT 1 FROM src_associations sa WHERE sa.source = s.id) AND NOT EXISTS (SELECT 1 FROM binaries b WHERE b.source = s.id)""") @@ -103,20 +105,24 @@ SELECT s.id, s.file FROM source s, files f for i in q.fetchall(): source_id = i[0] dsc_file_id = i[1] + dsc_fname = i[2] # Mark the .dsc file for deletion + Logger.log(["set lastused", dsc_fname]) session.execute("""UPDATE files SET last_used = :last_used WHERE id = :dscfileid AND last_used IS NULL""", {'last_used': now_date, 'dscfileid': dsc_file_id}) # Mark all other files references by .dsc too if they're not used by anyone else - x = session.execute("""SELECT f.id FROM files f, dsc_files d + x = session.execute("""SELECT f.id, f.filename FROM files f, dsc_files d WHERE d.source = :sourceid AND d.file = f.id""", {'sourceid': source_id}) for j in x.fetchall(): file_id = j[0] + file_name = j[1] y = session.execute("SELECT id FROM dsc_files d WHERE d.file = :fileid", {'fileid': file_id}) if len(y.fetchall()) == 1: + Logger.log(["set lastused", file_name]) session.execute("""UPDATE files SET last_used = :lastused WHERE id = :fileid AND last_used IS NULL""", {'lastused': now_date, 'fileid': file_id}) @@ -127,7 +133,7 @@ SELECT s.id, s.file FROM source s, files f # are now used again. q = session.execute(""" -SELECT f.id FROM source s, files f, dsc_files df +SELECT f.id, f.filename FROM source s, files f, dsc_files df WHERE f.last_used IS NOT NULL AND s.id = df.source AND df.file = f.id AND ((EXISTS (SELECT 1 FROM src_associations sa WHERE sa.source = s.id)) OR (EXISTS (SELECT 1 FROM binaries b WHERE b.source = s.id)))""") @@ -135,9 +141,8 @@ SELECT f.id FROM source s, files f, dsc_files df #### XXX: this should also handle deleted binaries specially (ie, not #### reinstate sources because of them - # Could be done in SQL; but left this way for hysterical raisins - # [and freedom to innovate don'cha know?] for i in q.fetchall(): + Logger.log(["unset lastused", i[1]]) session.execute("UPDATE files SET last_used = NULL WHERE id = :fileid", {'fileid': i[0]}) @@ -158,13 +163,15 @@ def check_files(now_date, delete_date, max_delete, session): SELECT id, filename FROM files f WHERE NOT EXISTS (SELECT 1 FROM binaries b WHERE b.file = f.id) AND NOT EXISTS (SELECT 1 FROM dsc_files df WHERE df.file = f.id) + AND last_used IS NULL ORDER BY filename""") ql = q.fetchall() if len(ql) > 0: - print "WARNING: check_files found something it shouldn't" + utils.warn("check_files found something it shouldn't") for x in ql: - print x + utils.warn("orphaned file: %s" % x) + Logger.log(["set lastused", x[1], "ORPHANED FILE"]) session.execute("UPDATE files SET last_used = :lastused WHERE id = :fileid", {'lastused': now_date, 'fileid': x[0]}) @@ -177,12 +184,13 @@ def clean_binaries(now_date, delete_date, max_delete, session): # XXX: why doesn't this remove the files here as well? I don't think it # buys anything keeping this separate print "Cleaning binaries from the DB..." + print "Deleting from binaries table... " + for bin in session.query(DBBinary).join(DBBinary.poolfile).filter(PoolFile.last_used <= delete_date): + Logger.log(["delete binary", bin.poolfile.filename]) + if not Options["No-Action"]: + session.delete(bin) if not Options["No-Action"]: - print "Deleting from binaries table... " - session.execute("""DELETE FROM binaries WHERE EXISTS - (SELECT 1 FROM files WHERE binaries.file = files.id - AND files.last_used <= :deldate)""", - {'deldate': delete_date}) + session.commit() ######################################## @@ -200,41 +208,37 @@ def clean(now_date, delete_date, max_delete, session): os.mkdir(dest) # Delete from source - if not Options["No-Action"]: - print "Deleting from source table... " - session.execute("""DELETE FROM dsc_files - WHERE EXISTS - (SELECT 1 FROM source s, files f, dsc_files df - WHERE f.last_used <= :deletedate - AND s.file = f.id AND s.id = df.source - AND df.id = dsc_files.id)""", {'deletedate': delete_date}) - session.execute("""DELETE FROM source - WHERE EXISTS - (SELECT 1 FROM files - WHERE source.file = files.id - AND files.last_used <= :deletedate)""", {'deletedate': delete_date}) + print "Deleting from source table... " + q = session.execute(""" +SELECT s.id, f.filename FROM source s, files f + WHERE f.last_used <= :deletedate + AND s.file = f.id""", {'deletedate': delete_date}) + for s in q.fetchall(): + Logger.log(["delete source", s[1], s[0]]) + if not Options["No-Action"]: + session.execute("DELETE FROM dsc_files WHERE source = :s_id", {"s_id":s[0]}) + session.execute("DELETE FROM source WHERE id = :s_id", {"s_id":s[0]}) + if not Options["No-Action"]: session.commit() # Delete files from the pool - query = """SELECT l.path, f.filename FROM location l, files f - WHERE f.last_used <= :deletedate AND l.id = f.location""" + old_files = session.query(PoolFile).filter(PoolFile.last_used <= delete_date) if max_delete is not None: - query += " LIMIT %d" % max_delete + old_files = old_files.limit(max_delete) print "Limiting removals to %d" % max_delete - q = session.execute(query, {'deletedate': delete_date}) - for i in q.fetchall(): - filename = i[0] + i[1] + for pf in old_files: + filename = os.path.join(pf.location.path, pf.filename) if not os.path.exists(filename): utils.warn("can not find '%s'." % (filename)) continue + Logger.log(["delete pool file", filename]) if os.path.isfile(filename): if os.path.islink(filename): count += 1 - if Options["No-Action"]: - print "Removing symlink %s..." % (filename) - else: + Logger.log(["delete symlink", filename]) + if not Options["No-Action"]: os.unlink(filename) else: size += os.stat(filename)[stat.ST_SIZE] @@ -245,20 +249,17 @@ def clean(now_date, delete_date, max_delete, session): if os.path.exists(dest_filename): dest_filename = utils.find_next_free(dest_filename) - if Options["No-Action"]: - print "Cleaning %s -> %s ..." % (filename, dest_filename) - else: + Logger.log(["move to morgue", filename, dest_filename]) + if not Options["No-Action"]: utils.move(filename, dest_filename) + + if not Options["No-Action"]: + session.delete(pf) + else: utils.fubar("%s is neither symlink nor file?!" % (filename)) - # Delete from the 'files' table - # XXX: I've a horrible feeling that the max_delete stuff breaks here - mhy - # TODO: Change it so we do the DELETEs as we go; it'll be slower but - # more reliable if not Options["No-Action"]: - print "Deleting from files table... " - session.execute("DELETE FROM files WHERE last_used <= :deletedate", {'deletedate': delete_date}) session.commit() if count > 0: @@ -271,7 +272,7 @@ def clean_maintainers(now_date, delete_date, max_delete, session): # TODO Replace this whole thing with one SQL statement q = session.execute(""" -SELECT m.id FROM maintainer m +SELECT m.id, m.name FROM maintainer m WHERE NOT EXISTS (SELECT 1 FROM binaries b WHERE b.maintainer = m.id) AND NOT EXISTS (SELECT 1 FROM source s WHERE s.maintainer = m.id OR s.changedby = m.id) AND NOT EXISTS (SELECT 1 FROM src_uploaders u WHERE u.maintainer = m.id)""") @@ -280,9 +281,10 @@ SELECT m.id FROM maintainer m for i in q.fetchall(): maintainer_id = i[0] + Logger.log(["delete maintainer", i[1]]) if not Options["No-Action"]: session.execute("DELETE FROM maintainer WHERE id = :maint", {'maint': maintainer_id}) - count += 1 + count += 1 if not Options["No-Action"]: session.commit() @@ -297,7 +299,7 @@ def clean_fingerprints(now_date, delete_date, max_delete, session): # TODO Replace this whole thing with one SQL statement q = session.execute(""" -SELECT f.id FROM fingerprint f +SELECT f.id, f.fingerprint FROM fingerprint f WHERE f.keyring IS NULL AND NOT EXISTS (SELECT 1 FROM binaries b WHERE b.sig_fpr = f.id) AND NOT EXISTS (SELECT 1 FROM source s WHERE s.sig_fpr = f.id)""") @@ -306,9 +308,10 @@ SELECT f.id FROM fingerprint f for i in q.fetchall(): fingerprint_id = i[0] + Logger.log(["delete fingerprint", i[1]]) if not Options["No-Action"]: session.execute("DELETE FROM fingerprint WHERE id = :fpr", {'fpr': fingerprint_id}) - count += 1 + count += 1 if not Options["No-Action"]: session.commit() @@ -330,24 +333,22 @@ def clean_queue_build(now_date, delete_date, max_delete, session): our_delete_date = now_date - timedelta(seconds = int(cnf["Clean-Suites::QueueBuildStayOfExecution"])) count = 0 - q = session.execute("SELECT filename FROM queue_build WHERE last_used <= :deletedate", - {'deletedate': our_delete_date}) - for i in q.fetchall(): - filename = i[0] - if not os.path.exists(filename): - utils.warn("%s (from queue_build) doesn't exist." % (filename)) + for qf in session.query(QueueBuild).filter(QueueBuild.last_used <= our_delete_date): + if not os.path.exists(qf.filename): + utils.warn("%s (from queue_build) doesn't exist." % (qf.filename)) continue - if not cnf.FindB("Dinstall::SecurityQueueBuild") and not os.path.islink(filename): - utils.fubar("%s (from queue_build) should be a symlink but isn't." % (filename)) + if not cnf.FindB("Dinstall::SecurityQueueBuild") and not os.path.islink(qf.filename): + utils.fubar("%s (from queue_build) should be a symlink but isn't." % (qf.filename)) - os.unlink(filename) + Logger.log(["delete queue build", qf.filename]) + if not Options["No-Action"]: + os.unlink(qf.filename) + session.delete(qf) count += 1 - session.execute("DELETE FROM queue_build WHERE last_used <= :deletedate", - {'deletedate': our_delete_date}) - - session.commit() + if not Options["No-Action"]: + session.commit() if count: print "Cleaned %d queue_build files." % (count) @@ -355,7 +356,7 @@ def clean_queue_build(now_date, delete_date, max_delete, session): ################################################################################ def main(): - global Options + global Options, Logger cnf = Config() @@ -384,6 +385,8 @@ def main(): if Options["Help"]: usage() + Logger = daklog.Logger(cnf, "clean-suites", debug=Options["No-Action"]) + session = DBConn().session() now_date = datetime.now() @@ -398,6 +401,8 @@ def main(): clean_fingerprints(now_date, delete_date, max_delete, session) clean_queue_build(now_date, delete_date, max_delete, session) + Logger.close() + ################################################################################ if __name__ == '__main__': diff --git a/dak/process_new.py b/dak/process_new.py index f01f4b91..1423d705 100755 --- a/dak/process_new.py +++ b/dak/process_new.py @@ -695,7 +695,7 @@ def do_new(upload, session): try: check_daily_lock() done = add_overrides (new, upload, session) - Logger.log([utils.getusername(), "NEW ACCEPT: %s" % (upload.pkg.changes_file)]) + Logger.log(["NEW ACCEPT: %s" % (upload.pkg.changes_file)]) except CantGetLockError: print "Hello? Operator! Give me the number for 911!" print "Dinstall in the locked area, cant process packages, come back later" @@ -708,7 +708,7 @@ def do_new(upload, session): reject_message=Options["Manual-Reject"], note=get_new_comments(changes.get("source", ""), session=session)) if not aborted: - Logger.log([utils.getusername(), "NEW REJECT: %s" % (upload.pkg.changes_file)]) + Logger.log(["NEW REJECT: %s" % (upload.pkg.changes_file)]) os.unlink(upload.pkg.changes_file[:-8]+".dak") done = 1 elif answer == 'N': @@ -717,7 +717,7 @@ def do_new(upload, session): elif answer == 'P' and not Options["Trainee"]: prod_maintainer(get_new_comments(changes.get("source", ""), session=session), upload) - Logger.log([utils.getusername(), "NEW PROD: %s" % (upload.pkg.changes_file)]) + Logger.log(["NEW PROD: %s" % (upload.pkg.changes_file)]) elif answer == 'R' and not Options["Trainee"]: confirm = utils.our_raw_input("Really clear note (y/N)? ").lower() if confirm == "y": @@ -792,12 +792,12 @@ def do_byhand(upload, session): done = 1 for f in byhand: del files[f] - Logger.log([utils.getusername(), "BYHAND ACCEPT: %s" % (upload.pkg.changes_file)]) + Logger.log(["BYHAND ACCEPT: %s" % (upload.pkg.changes_file)]) except CantGetLockError: print "Hello? Operator! Give me the number for 911!" print "Dinstall in the locked area, cant process packages, come back later" elif answer == 'M': - Logger.log([utils.getusername(), "BYHAND REJECT: %s" % (upload.pkg.changes_file)]) + Logger.log(["BYHAND REJECT: %s" % (upload.pkg.changes_file)]) upload.do_reject(manual=1, reject_message=Options["Manual-Reject"]) os.unlink(upload.pkg.changes_file[:-8]+".dak") done = 1 @@ -1017,7 +1017,7 @@ def end(): if accept_count > 1: sets = "sets" sys.stderr.write("Accepted %d package %s, %s.\n" % (accept_count, sets, utils.size_type(int(accept_bytes)))) - Logger.log([utils.getusername(), "total",accept_count,accept_bytes]) + Logger.log(["total",accept_count,accept_bytes]) if not Options["No-Action"] and not Options["Trainee"]: Logger.close() diff --git a/dak/rm.py b/dak/rm.py index fe763605..51400194 100755 --- a/dak/rm.py +++ b/dak/rm.py @@ -505,8 +505,8 @@ def main (): logfile.write("----------------------------------------------\n") logfile.flush() - dsc_type_id = get_override_type('dsc', session) - deb_type_id = get_override_type('deb', session) + dsc_type_id = get_override_type('dsc', session).overridetype_id + deb_type_id = get_override_type('deb', session).overridetype_id # Do the actual deletion print "Deleting...", diff --git a/daklib/changes.py b/daklib/changes.py index 59c7da17..ff232224 100755 --- a/daklib/changes.py +++ b/daklib/changes.py @@ -191,8 +191,8 @@ class Changes(object): self.dsc_files.update(p.load()) next_obj = p.load() - if type(next_obj) is DictType: - self.pkg.orig_files.update(next_obj) + if isinstance(next_obj, dict): + self.orig_files.update(next_obj) else: # Auto-convert old dak files to new format supporting # multiple tarballs diff --git a/daklib/daklog.py b/daklib/daklog.py index 0cca205e..dfcae368 100755 --- a/daklib/daklog.py +++ b/daklib/daklog.py @@ -58,13 +58,12 @@ class Logger: logfile = utils.open_file(logfilename, 'a') os.umask(umask) self.logfile = logfile - # Log the start of the program - user = pwd.getpwuid(os.getuid())[0] - self.log(["program start", user]) + self.log(["program start"]) def log (self, details): "Log an event" - # Prepend the timestamp and program name + # Prepend timestamp, program name, and user name + details.insert(0, utils.getusername()) details.insert(0, self.program) timestamp = time.strftime("%Y%m%d%H%M%S") details.insert(0, timestamp) diff --git a/daklib/dbconn.py b/daklib/dbconn.py index 224935dd..56ff65aa 100755 --- a/daklib/dbconn.py +++ b/daklib/dbconn.py @@ -84,6 +84,7 @@ def session_wrapper(fn): # Session is last argument in args session = args[-1] if session is None: + args = list(args) session = args[-1] = DBConn().session() private_transaction = True @@ -1417,6 +1418,7 @@ class Queue(object): # them (if one doesn't already exist) for dsc_file in changes.dsc_files.keys(): # Skip all files except orig tarballs + from daklib.regexes import re_is_orig_source if not re_is_orig_source.match(dsc_file): continue # Skip orig files not identified in the pool diff --git a/daklib/queue.py b/daklib/queue.py index 1ece0d1d..8bb06307 100755 --- a/daklib/queue.py +++ b/daklib/queue.py @@ -71,7 +71,7 @@ def get_type(f, session): """ # Determine the type if f.has_key("dbtype"): - file_type = file["dbtype"] + file_type = f["dbtype"] elif re_source_ext.match(f["type"]): file_type = "dsc" else: @@ -998,11 +998,6 @@ class Upload(object): for field_name in [ "build-depends", "build-depends-indep" ]: field = self.pkg.dsc.get(field_name) if field: - # Check for broken dpkg-dev lossage... - if field.startswith("ARRAY"): - self.rejects.append("%s: invalid %s field produced by a broken version of dpkg-dev (1.10.11)" % \ - (dsc_filename, field_name.title())) - # Have apt try to parse them... try: apt_pkg.ParseSrcDepends(field)