X-Git-Url: https://git.decadent.org.uk/gitweb/?a=blobdiff_plain;f=dak%2Fprocess_unchecked.py;h=ec649f0bb825258500cec14b20fc0062c46f7123;hb=909c2bf3b4a10db63f1694b5878faf4576b1e949;hp=1d87447a7d61280e17375e5f96b4d3db3e783451;hpb=30413cf0ff7bc21b8d2b8b4346406357fe55dc19;p=dak.git diff --git a/dak/process_unchecked.py b/dak/process_unchecked.py index 1d87447a..ec649f0b 100755 --- a/dak/process_unchecked.py +++ b/dak/process_unchecked.py @@ -1,8 +1,7 @@ #!/usr/bin/env python # Checks Debian packages from Incoming -# Copyright (C) 2000, 2001, 2002, 2003, 2004, 2005 James Troup -# $Id: jennifer,v 1.65 2005-12-05 05:35:47 ajt Exp $ +# Copyright (C) 2000, 2001, 2002, 2003, 2004, 2005, 2006 James Troup # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by @@ -31,7 +30,10 @@ import commands, errno, fcntl, os, re, shutil, stat, sys, time, tempfile, traceback import apt_inst, apt_pkg -import db_access, katie, logging, utils +import daklib.database +import daklib.logging +import daklib.queue +import daklib.utils from types import * @@ -45,17 +47,15 @@ re_strip_revision = re.compile(r"-([^-]+)$") ################################################################################ # Globals -jennifer_version = "$Revision: 1.65 $" - Cnf = None Options = None Logger = None -Katie = None +Upload = None reprocess = 0 in_holding = {} -# Aliases to the real vars in the Katie class; hysterical raisins. +# Aliases to the real vars in the Upload class; hysterical raisins. reject_message = "" changes = {} dsc = {} @@ -66,19 +66,18 @@ pkg = {} ############################################################################### def init(): - global Cnf, Options, Katie, changes, dsc, dsc_files, files, pkg + global Cnf, Options, Upload, changes, dsc, dsc_files, files, pkg apt_pkg.init() Cnf = apt_pkg.newConfiguration() - apt_pkg.ReadConfigFileISC(Cnf,utils.which_conf_file()) + apt_pkg.ReadConfigFileISC(Cnf,daklib.utils.which_conf_file()) Arguments = [('a',"automatic","Dinstall::Options::Automatic"), ('h',"help","Dinstall::Options::Help"), ('n',"no-action","Dinstall::Options::No-Action"), ('p',"no-lock", "Dinstall::Options::No-Lock"), - ('s',"no-mail", "Dinstall::Options::No-Mail"), - ('V',"version","Dinstall::Options::Version")] + ('s',"no-mail", "Dinstall::Options::No-Mail")] for i in ["automatic", "help", "no-action", "no-lock", "no-mail", "override-distribution", "version"]: @@ -89,17 +88,14 @@ def init(): if Options["Help"]: usage() - elif Options["Version"]: - print "jennifer %s" % (jennifer_version) - sys.exit(0) - Katie = katie.Katie(Cnf) + Upload = daklib.queue.Upload(Cnf) - changes = Katie.pkg.changes - dsc = Katie.pkg.dsc - dsc_files = Katie.pkg.dsc_files - files = Katie.pkg.files - pkg = Katie.pkg + changes = Upload.pkg.changes + dsc = Upload.pkg.dsc + dsc_files = Upload.pkg.dsc_files + files = Upload.pkg.files + pkg = Upload.pkg return changes_files @@ -169,7 +165,7 @@ def clean_holding(): for file in in_holding.keys(): if os.path.exists(file): if file.find('/') != -1: - utils.fubar("WTF? clean_holding() got a file ('%s') with / in it!" % (file)) + daklib.utils.fubar("WTF? clean_holding() got a file ('%s') with / in it!" % (file)) else: os.unlink(file) in_holding = {} @@ -182,20 +178,20 @@ def check_changes(): # Parse the .changes field into a dictionary try: - changes.update(utils.parse_changes(filename)) - except utils.cant_open_exc: + changes.update(daklib.utils.parse_changes(filename)) + except daklib.utils.cant_open_exc: reject("%s: can't read file." % (filename)) return 0 - except utils.changes_parse_error_exc, line: + except daklib.utils.changes_parse_error_exc, line: reject("%s: parse error, can't grok: %s." % (filename, line)) return 0 # Parse the Files field from the .changes into another dictionary try: - files.update(utils.build_file_list(changes)) - except utils.changes_parse_error_exc, line: + files.update(daklib.utils.build_file_list(changes)) + except daklib.utils.changes_parse_error_exc, line: reject("%s: parse error, can't grok: %s." % (filename, line)) - except utils.nk_format_exc, format: + except daklib.utils.nk_format_exc, format: reject("%s: unknown format '%s'." % (filename, format)) return 0 @@ -219,8 +215,8 @@ def check_changes(): try: (changes["maintainer822"], changes["maintainer2047"], changes["maintainername"], changes["maintaineremail"]) = \ - utils.fix_maintainer (changes["maintainer"]) - except utils.ParseMaintError, msg: + daklib.utils.fix_maintainer (changes["maintainer"]) + except daklib.utils.ParseMaintError, msg: reject("%s: Maintainer field ('%s') failed to parse: %s" \ % (filename, changes["maintainer"], msg)) @@ -228,8 +224,8 @@ def check_changes(): try: (changes["changedby822"], changes["changedby2047"], changes["changedbyname"], changes["changedbyemail"]) = \ - utils.fix_maintainer (changes.get("changed-by", "")) - except utils.ParseMaintError, msg: + daklib.utils.fix_maintainer (changes.get("changed-by", "")) + except daklib.utils.ParseMaintError, msg: (changes["changedby822"], changes["changedby2047"], changes["changedbyname"], changes["changedbyemail"]) = \ ("", "", "", "") @@ -239,18 +235,18 @@ def check_changes(): # Ensure all the values in Closes: are numbers if changes.has_key("closes"): for i in changes["closes"].keys(): - if katie.re_isanum.match (i) == None: + if daklib.queue.re_isanum.match (i) == None: reject("%s: `%s' from Closes field isn't a number." % (filename, i)) # chopversion = no epoch; chopversion2 = no epoch and no revision (e.g. for .orig.tar.gz comparison) - changes["chopversion"] = utils.re_no_epoch.sub('', changes["version"]) - changes["chopversion2"] = utils.re_no_revision.sub('', changes["chopversion"]) + changes["chopversion"] = daklib.utils.re_no_epoch.sub('', changes["version"]) + changes["chopversion2"] = daklib.utils.re_no_revision.sub('', changes["chopversion"]) # Check there isn't already a changes file of the same name in one # of the queue directories. base_filename = os.path.basename(filename) - for dir in [ "Accepted", "Byhand", "Done", "New" ]: + for dir in [ "Accepted", "Byhand", "Done", "New", "ProposedUpdates" ]: if os.path.exists(Cnf["Dir::Queue::%s" % (dir) ]+'/'+base_filename): reject("%s: a file with this name already exists in the %s directory." % (base_filename, dir)) @@ -331,7 +327,7 @@ Pre-Depends on dpkg (>= 1.10.24).""" (result, output) = commands.getstatusoutput(cmd) if result != 0: reject("%s: 'ar t' invocation failed." % (filename)) - reject(utils.prefix_multi_line_string(output, " [ar output:] "), "") + reject(daklib.utils.prefix_multi_line_string(output, " [ar output:] "), "") chunks = output.split('\n') if len(chunks) != 3: reject("%s: found %d chunks, expected 3." % (filename, len(chunks))) @@ -362,7 +358,7 @@ Pre-Depends on dpkg (>= 1.10.24).""" def check_files(): global reprocess - archive = utils.where_am_i() + archive = daklib.utils.where_am_i() file_keys = files.keys() # if reprocess is 2 we've already done this and we're checking @@ -375,11 +371,11 @@ def check_files(): copy_to_holding(file) os.chdir(cwd) - # Check there isn't already a .changes or .katie file of the same name in - # the proposed-updates "CopyChanges" or "CopyKatie" storage directories. + # Check there isn't already a .changes or .dak file of the same name in + # the proposed-updates "CopyChanges" or "CopyDotDak" storage directories. # [NB: this check must be done post-suite mapping] base_filename = os.path.basename(pkg.changes_file) - katie_filename = base_filename[:-8]+".katie" + dot_dak_filename = base_filename[:-8]+".dak" for suite in changes["distribution"].keys(): copychanges = "Suite::%s::CopyChanges" % (suite) if Cnf.has_key(copychanges) and \ @@ -387,11 +383,11 @@ def check_files(): reject("%s: a file with this name already exists in %s" \ % (base_filename, Cnf[copychanges])) - copykatie = "Suite::%s::CopyKatie" % (suite) - if Cnf.has_key(copykatie) and \ - os.path.exists(Cnf[copykatie]+"/"+katie_filename): + copy_dot_dak = "Suite::%s::CopyDotDak" % (suite) + if Cnf.has_key(copy_dot_dak) and \ + os.path.exists(Cnf[copy_dot_dak]+"/"+dot_dak_filename): reject("%s: a file with this name already exists in %s" \ - % (katie_filename, Cnf[copykatie])) + % (dot_dak_filename, Cnf[copy_dot_dak])) reprocess = 0 has_binaries = 0 @@ -399,10 +395,10 @@ def check_files(): for file in file_keys: # Ensure the file does not already exist in one of the accepted directories - for dir in [ "Accepted", "Byhand", "New" ]: + for dir in [ "Accepted", "Byhand", "New", "ProposedUpdates" ]: if os.path.exists(Cnf["Dir::Queue::%s" % (dir) ]+'/'+file): reject("%s file already exists in the %s directory." % (file, dir)) - if not utils.re_taint_free.match(file): + if not daklib.utils.re_taint_free.match(file): reject("!!WARNING!! tainted filename: '%s'." % (file)) # Check the file is readable if os.access(file,os.R_OK) == 0: @@ -420,12 +416,12 @@ def check_files(): files[file]["byhand"] = 1 files[file]["type"] = "byhand" # Checks for a binary package... - elif utils.re_isadeb.match(file): + elif daklib.utils.re_isadeb.match(file): has_binaries = 1 files[file]["type"] = "deb" # Extract package control information - deb_file = utils.open_file(file) + deb_file = daklib.utils.open_file(file) try: control = apt_pkg.ParseSection(apt_inst.debExtractControl(deb_file)) except: @@ -493,7 +489,7 @@ def check_files(): source = files[file]["source"] source_version = "" if source.find("(") != -1: - m = utils.re_extract_src_version.match(source) + m = daklib.utils.re_extract_src_version.match(source) source = m.group(1) source_version = m.group(2) if not source_version: @@ -502,12 +498,12 @@ def check_files(): files[file]["source version"] = source_version # Ensure the filename matches the contents of the .deb - m = utils.re_isadeb.match(file) + m = daklib.utils.re_isadeb.match(file) # package name file_package = m.group(1) if files[file]["package"] != file_package: reject("%s: package part of filename (%s) does not match package name in the %s (%s)." % (file, file_package, files[file]["dbtype"], files[file]["package"])) - epochless_version = utils.re_no_epoch.sub('', control.Find("Version")) + epochless_version = daklib.utils.re_no_epoch.sub('', control.Find("Version")) # version file_version = m.group(2) if epochless_version != file_version: @@ -525,9 +521,9 @@ def check_files(): reject("source version (%s) for %s doesn't match changes version %s." % (source_version, file, changes["version"])) else: # Check in the SQL database - if not Katie.source_exists(source_package, source_version, changes["distribution"].keys()): + if not Upload.source_exists(source_package, source_version, changes["distribution"].keys()): # Check in one of the other directories - source_epochless_version = utils.re_no_epoch.sub('', source_version) + source_epochless_version = daklib.utils.re_no_epoch.sub('', source_version) dsc_filename = "%s_%s.dsc" % (source_package, source_epochless_version) if os.path.exists(Cnf["Dir::Queue::Byhand"] + '/' + dsc_filename): files[file]["byhand"] = 1 @@ -536,13 +532,13 @@ def check_files(): elif not os.path.exists(Cnf["Dir::Queue::Accepted"] + '/' + dsc_filename): reject("no source found for %s %s (%s)." % (source_package, source_version, file)) # Check the version and for file overwrites - reject(Katie.check_binary_against_db(file),"") + reject(Upload.check_binary_against_db(file),"") check_deb_ar(file, control) # Checks for a source package... else: - m = utils.re_issource.match(file) + m = daklib.utils.re_issource.match(file) if m: has_source = 1 files[file]["package"] = m.group(1) @@ -567,7 +563,7 @@ def check_files(): # Check the signature of a .dsc file if files[file]["type"] == "dsc": - dsc["fingerprint"] = utils.check_signature(file, reject) + dsc["fingerprint"] = daklib.utils.check_signature(file, reject) files[file]["architecture"] = "source" @@ -598,13 +594,13 @@ def check_files(): # Validate the component component = files[file]["component"] - component_id = db_access.get_component_id(component) + component_id = daklib.database.get_component_id(component) if component_id == -1: reject("file '%s' has unknown component '%s'." % (file, component)) continue # See if the package is NEW - if not Katie.in_override_p(files[file]["package"], files[file]["component"], suite, files[file].get("dbtype",""), file): + if not Upload.in_override_p(files[file]["package"], files[file]["component"], suite, files[file].get("dbtype",""), file): files[file]["new"] = 1 # Validate the priority @@ -613,14 +609,14 @@ def check_files(): # Determine the location location = Cnf["Dir::Pool"] - location_id = db_access.get_location_id (location, component, archive) + location_id = daklib.database.get_location_id (location, component, archive) if location_id == -1: reject("[INTERNAL ERROR] couldn't determine location (Component: %s, Archive: %s)" % (component, archive)) files[file]["location id"] = location_id # Check the md5sum & size against existing files (if any) - files[file]["pool name"] = utils.poolify (changes["source"], files[file]["component"]) - files_id = db_access.get_files_id(files[file]["pool name"] + file, files[file]["size"], files[file]["md5sum"], files[file]["location id"]) + files[file]["pool name"] = daklib.utils.poolify (changes["source"], files[file]["component"]) + files_id = daklib.database.get_files_id(files[file]["pool name"] + file, files[file]["size"], files[file]["md5sum"], files[file]["location id"]) if files_id == -1: reject("INTERNAL ERROR, get_files_id() returned multiple matches for %s." % (file)) elif files_id == -2: @@ -628,7 +624,7 @@ def check_files(): files[file]["files id"] = files_id # Check for packages that have moved from one component to another - q = Katie.projectB.query(""" + q = Upload.projectB.query(""" SELECT c.name FROM binaries b, bin_associations ba, suite s, location l, component c, architecture a, files f WHERE b.package = '%s' AND s.suite_name = '%s' @@ -675,22 +671,22 @@ def check_dsc(): # Parse the .dsc file try: - dsc.update(utils.parse_changes(dsc_filename, signing_rules=1)) - except utils.cant_open_exc: + dsc.update(daklib.utils.parse_changes(dsc_filename, signing_rules=1)) + except daklib.utils.cant_open_exc: # if not -n copy_to_holding() will have done this for us... if Options["No-Action"]: reject("%s: can't read file." % (dsc_filename)) - except utils.changes_parse_error_exc, line: + except daklib.utils.changes_parse_error_exc, line: reject("%s: parse error, can't grok: %s." % (dsc_filename, line)) - except utils.invalid_dsc_format_exc, line: + except daklib.utils.invalid_dsc_format_exc, line: reject("%s: syntax error on line %s." % (dsc_filename, line)) # Build up the file list of files mentioned by the .dsc try: - dsc_files.update(utils.build_file_list(dsc, is_a_dsc=1)) - except utils.no_files_exc: + dsc_files.update(daklib.utils.build_file_list(dsc, is_a_dsc=1)) + except daklib.utils.no_files_exc: reject("%s: no Files: field." % (dsc_filename)) return 0 - except utils.changes_parse_error_exc, line: + except daklib.utils.changes_parse_error_exc, line: reject("%s: parse error, can't grok: %s." % (dsc_filename, line)) return 0 @@ -713,8 +709,8 @@ def check_dsc(): # Validate the Maintainer field try: - utils.fix_maintainer (dsc["maintainer"]) - except utils.ParseMaintError, msg: + daklib.utils.fix_maintainer (dsc["maintainer"]) + except daklib.utils.ParseMaintError, msg: reject("%s: Maintainer field ('%s') failed to parse: %s" \ % (dsc_filename, dsc["maintainer"], msg)) @@ -734,7 +730,7 @@ def check_dsc(): pass # Ensure the version number in the .dsc matches the version number in the .changes - epochless_dsc_version = utils.re_no_epoch.sub('', dsc["version"]) + epochless_dsc_version = daklib.utils.re_no_epoch.sub('', dsc["version"]) changes_version = files[dsc_filename]["version"] if epochless_dsc_version != files[dsc_filename]["version"]: reject("version ('%s') in .dsc does not match version ('%s') in .changes." % (epochless_dsc_version, changes_version)) @@ -742,7 +738,7 @@ def check_dsc(): # Ensure there is a .tar.gz in the .dsc file has_tar = 0 for f in dsc_files.keys(): - m = utils.re_issource.match(f) + m = daklib.utils.re_issource.match(f) if not m: reject("%s: %s in Files field not recognised as source." % (dsc_filename, f)) type = m.group(3) @@ -752,9 +748,9 @@ def check_dsc(): reject("%s: no .tar.gz or .orig.tar.gz in 'Files' field." % (dsc_filename)) # Ensure source is newer than existing source in target suites - reject(Katie.check_source_against_db(dsc_filename),"") + reject(Upload.check_source_against_db(dsc_filename),"") - (reject_msg, is_in_incoming) = Katie.check_dsc_against_db(dsc_filename) + (reject_msg, is_in_incoming) = Upload.check_dsc_against_db(dsc_filename) reject(reject_msg, "") if is_in_incoming: if not Options["No-Action"]: @@ -789,7 +785,7 @@ def get_changelog_versions(source_dir): # Create a symlink mirror of the source files in our temporary directory for f in files.keys(): - m = utils.re_issource.match(f) + m = daklib.utils.re_issource.match(f) if m: src = os.path.join(source_dir, f) # If a file is missing for whatever reason, give up. @@ -812,14 +808,14 @@ def get_changelog_versions(source_dir): (result, output) = commands.getstatusoutput(cmd) if (result != 0): reject("'dpkg-source -x' failed for %s [return code: %s]." % (dsc_filename, result)) - reject(utils.prefix_multi_line_string(output, " [dpkg-source output:] "), "") + reject(daklib.utils.prefix_multi_line_string(output, " [dpkg-source output:] "), "") return if not Cnf.Find("Dir::Queue::BTSVersionTrack"): return # Get the upstream version - upstr_version = utils.re_no_epoch.sub('', dsc["version"]) + upstr_version = daklib.utils.re_no_epoch.sub('', dsc["version"]) if re_strip_revision.search(upstr_version): upstr_version = re_strip_revision.sub('', upstr_version) @@ -831,7 +827,7 @@ def get_changelog_versions(source_dir): # Parse the changelog dsc["bts changelog"] = "" - changelog_file = utils.open_file(changelog_filename) + changelog_file = daklib.utils.open_file(changelog_filename) for line in changelog_file.readlines(): m = re_changelog_versions.match(line) if m: @@ -874,7 +870,7 @@ def check_source(): shutil.rmtree(tmpdir) except OSError, e: if errno.errorcode[e.errno] != 'EACCES': - utils.fubar("%s: couldn't remove tmp dir for source tree." % (dsc["source"])) + daklib.utils.fubar("%s: couldn't remove tmp dir for source tree." % (dsc["source"])) reject("%s: source tree could not be cleanly removed." % (dsc["source"])) # We probably have u-r or u-w directories so chmod everything @@ -882,10 +878,10 @@ def check_source(): cmd = "chmod -R u+rwx %s" % (tmpdir) result = os.system(cmd) if result != 0: - utils.fubar("'%s' failed with result %s." % (cmd, result)) + daklib.utils.fubar("'%s' failed with result %s." % (cmd, result)) shutil.rmtree(tmpdir) except: - utils.fubar("%s: couldn't remove tmp dir for source tree." % (dsc["source"])) + daklib.utils.fubar("%s: couldn't remove tmp dir for source tree." % (dsc["source"])) ################################################################################ @@ -905,8 +901,8 @@ def check_urgency (): def check_md5sums (): for file in files.keys(): try: - file_handle = utils.open_file(file) - except utils.cant_open_exc: + file_handle = daklib.utils.open_file(file) + except daklib.utils.cant_open_exc: continue # Check md5sum @@ -922,8 +918,8 @@ def check_md5sums (): for file in dsc_files.keys(): try: - file_handle = utils.open_file(file) - except utils.cant_open_exc: + file_handle = daklib.utils.open_file(file) + except daklib.utils.cant_open_exc: continue # Check md5sum @@ -968,7 +964,7 @@ def check_timestamps(): if files[filename]["type"] == "deb": tar.reset() try: - deb_file = utils.open_file(filename) + deb_file = daklib.utils.open_file(filename) apt_inst.debExtract(deb_file,tar.callback,"control.tar.gz") deb_file.seek(0) try: @@ -1034,18 +1030,21 @@ def action (): if not changes.has_key("distribution") or not isinstance(changes["distribution"], DictType): changes["distribution"] = {} - (summary, short_summary) = Katie.build_summaries() + (summary, short_summary) = Upload.build_summaries() # q-unapproved hax0ring queue_info = { "New": { "is": is_new, "process": acknowledge_new }, "Byhand" : { "is": is_byhand, "process": do_byhand }, + "StableUpdate" : { "is": is_stableupdate, "process": do_stableupdate }, "Unembargo" : { "is": is_unembargo, "process": queue_unembargo }, "Embargo" : { "is": is_embargo, "process": queue_embargo }, } queues = [ "New", "Byhand" ] if Cnf.FindB("Dinstall::SecurityQueueHandling"): queues += [ "Unembargo", "Embargo" ] + else: + queues += [ "StableUpdate" ] (prompt, answer) = ("", "XXX") if Options["No-Action"] or Options["Automatic"]: @@ -1087,15 +1086,15 @@ def action (): answer = 'A' while prompt.find(answer) == -1: - answer = utils.our_raw_input(prompt) - m = katie.re_default_answer.match(prompt) + answer = daklib.utils.our_raw_input(prompt) + m = daklib.queue.re_default_answer.match(prompt) if answer == "": answer = m.group(1) answer = answer[:1].upper() if answer == 'R': os.chdir (pkg.directory) - Katie.do_reject(0, reject_message) + Upload.do_reject(0, reject_message) elif answer == 'A': accept(summary, short_summary) remove_from_unchecked() @@ -1114,21 +1113,21 @@ def remove_from_unchecked(): ################################################################################ def accept (summary, short_summary): - Katie.accept(summary, short_summary) - Katie.check_override() + Upload.accept(summary, short_summary) + Upload.check_override() ################################################################################ def move_to_dir (dest, perms=0660, changesperms=0664): - utils.move (pkg.changes_file, dest, perms=changesperms) + daklib.utils.move (pkg.changes_file, dest, perms=changesperms) file_keys = files.keys() for file in file_keys: - utils.move (file, dest, perms=perms) + daklib.utils.move (file, dest, perms=perms) ################################################################################ def is_unembargo (): - q = Katie.projectB.query( + q = Upload.projectB.query( "SELECT package FROM disembargo WHERE package = '%s' AND version = '%s'" % (changes["source"], changes["version"])) ql = q.getresult() @@ -1139,7 +1138,7 @@ def is_unembargo (): if changes["architecture"].has_key("source"): if Options["No-Action"]: return 1 - Katie.projectB.query( + Upload.projectB.query( "INSERT INTO disembargo (package, version) VALUES ('%s', '%s')" % (changes["source"], changes["version"])) return 1 @@ -1150,13 +1149,13 @@ def queue_unembargo (summary): print "Moving to UNEMBARGOED holding area." Logger.log(["Moving to unembargoed", pkg.changes_file]) - Katie.dump_vars(Cnf["Dir::Queue::Unembargoed"]) + Upload.dump_vars(Cnf["Dir::Queue::Unembargoed"]) move_to_dir(Cnf["Dir::Queue::Unembargoed"]) - Katie.queue_build("unembargoed", Cnf["Dir::Queue::Unembargoed"]) + Upload.queue_build("unembargoed", Cnf["Dir::Queue::Unembargoed"]) # Check for override disparities - Katie.Subst["__SUMMARY__"] = summary - Katie.check_override() + Upload.Subst["__SUMMARY__"] = summary + Upload.check_override() ################################################################################ @@ -1167,13 +1166,31 @@ def queue_embargo (summary): print "Moving to EMBARGOED holding area." Logger.log(["Moving to embargoed", pkg.changes_file]) - Katie.dump_vars(Cnf["Dir::Queue::Embargoed"]) + Upload.dump_vars(Cnf["Dir::Queue::Embargoed"]) move_to_dir(Cnf["Dir::Queue::Embargoed"]) - Katie.queue_build("embargoed", Cnf["Dir::Queue::Embargoed"]) + Upload.queue_build("embargoed", Cnf["Dir::Queue::Embargoed"]) + + # Check for override disparities + Upload.Subst["__SUMMARY__"] = summary + Upload.check_override() + +################################################################################ + +def is_stableupdate (): + if changes["distribution"].has_key("proposed-updates"): + return 1 + return 0 + +def do_stableupdate (summary): + print "Moving to PROPOSED-UPDATES holding area." + Logger.log(["Moving to proposed-updates", pkg.changes_file]); + + Upload.dump_vars(Cnf["Dir::Queue::ProposedUpdates"]); + move_to_dir(Cnf["Dir::Queue::ProposedUpdates"]) # Check for override disparities - Katie.Subst["__SUMMARY__"] = summary - Katie.check_override() + Upload.Subst["__SUMMARY__"] = summary; + Upload.check_override(); ################################################################################ @@ -1187,12 +1204,12 @@ def do_byhand (summary): print "Moving to BYHAND holding area." Logger.log(["Moving to byhand", pkg.changes_file]) - Katie.dump_vars(Cnf["Dir::Queue::Byhand"]) + Upload.dump_vars(Cnf["Dir::Queue::Byhand"]) move_to_dir(Cnf["Dir::Queue::Byhand"]) # Check for override disparities - Katie.Subst["__SUMMARY__"] = summary - Katie.check_override() + Upload.Subst["__SUMMARY__"] = summary + Upload.check_override() ################################################################################ @@ -1203,37 +1220,37 @@ def is_new (): return 0 def acknowledge_new (summary): - Subst = Katie.Subst + Subst = Upload.Subst print "Moving to NEW holding area." Logger.log(["Moving to new", pkg.changes_file]) - Katie.dump_vars(Cnf["Dir::Queue::New"]) + Upload.dump_vars(Cnf["Dir::Queue::New"]) move_to_dir(Cnf["Dir::Queue::New"]) if not Options["No-Mail"]: print "Sending new ack." Subst["__SUMMARY__"] = summary - new_ack_message = utils.TemplateSubst(Subst,Cnf["Dir::Templates"]+"/jennifer.new") - utils.send_mail(new_ack_message) + new_ack_message = daklib.utils.TemplateSubst(Subst,Cnf["Dir::Templates"]+"/process-unchecked.new") + daklib.utils.send_mail(new_ack_message) ################################################################################ # reprocess is necessary for the case of foo_1.2-1 and foo_1.2-2 in # Incoming. -1 will reference the .orig.tar.gz, but -2 will not. -# Katie.check_dsc_against_db() can find the .orig.tar.gz but it will +# Upload.check_dsc_against_db() can find the .orig.tar.gz but it will # not have processed it during it's checks of -2. If -1 has been -# deleted or otherwise not checked by jennifer, the .orig.tar.gz will -# not have been checked at all. To get round this, we force the -# .orig.tar.gz into the .changes structure and reprocess the .changes -# file. +# deleted or otherwise not checked by 'dak process-unchecked', the +# .orig.tar.gz will not have been checked at all. To get round this, +# we force the .orig.tar.gz into the .changes structure and reprocess +# the .changes file. def process_it (changes_file): global reprocess, reject_message # Reset some globals reprocess = 1 - Katie.init_vars() + Upload.init_vars() # Some defaults in case we can't fully process the .changes file changes["maintainer2047"] = Cnf["Dinstall::MyEmailAddress"] changes["changedby2047"] = Cnf["Dinstall::MyEmailAddress"] @@ -1256,7 +1273,7 @@ def process_it (changes_file): # Relativize the filename so we use the copy in holding # rather than the original... pkg.changes_file = os.path.basename(pkg.changes_file) - changes["fingerprint"] = utils.check_signature(pkg.changes_file, reject) + changes["fingerprint"] = daklib.utils.check_signature(pkg.changes_file, reject) if changes["fingerprint"]: valid_changes_p = check_changes() else: @@ -1271,7 +1288,7 @@ def process_it (changes_file): check_md5sums() check_urgency() check_timestamps() - Katie.update_subst(reject_message) + Upload.update_subst(reject_message) action() except SystemExit: raise @@ -1297,16 +1314,16 @@ def main(): # Ensure all the arguments we were given are .changes files for file in changes_files: if not file.endswith(".changes"): - utils.warn("Ignoring '%s' because it's not a .changes file." % (file)) + daklib.utils.warn("Ignoring '%s' because it's not a .changes file." % (file)) changes_files.remove(file) if changes_files == []: - utils.fubar("Need at least one .changes file as an argument.") + daklib.utils.fubar("Need at least one .changes file as an argument.") # Check that we aren't going to clash with the daily cron job if not Options["No-Action"] and os.path.exists("%s/daily.lock" % (Cnf["Dir::Lock"])) and not Options["No-Lock"]: - utils.fubar("Archive maintenance in progress. Try again later.") + daklib.utils.fubar("Archive maintenance in progress. Try again later.") # Obtain lock if not in no-action mode and initialize the log @@ -1316,21 +1333,21 @@ def main(): fcntl.lockf(lock_fd, fcntl.LOCK_EX | fcntl.LOCK_NB) except IOError, e: if errno.errorcode[e.errno] == 'EACCES' or errno.errorcode[e.errno] == 'EAGAIN': - utils.fubar("Couldn't obtain lock; assuming another jennifer is already running.") + daklib.utils.fubar("Couldn't obtain lock; assuming another 'dak process-unchecked' is already running.") else: raise - Logger = Katie.Logger = logging.Logger(Cnf, "jennifer") + Logger = Upload.Logger = daklib.logging.Logger(Cnf, "process-unchecked") # debian-{devel-,}-changes@lists.debian.org toggles writes access based on this header - bcc = "X-Katie: %s" % (jennifer_version) + bcc = "X-DAK: dak process-unchecked\nX-Katie: this header is obsolete" if Cnf.has_key("Dinstall::Bcc"): - Katie.Subst["__BCC__"] = bcc + "\nBcc: %s" % (Cnf["Dinstall::Bcc"]) + Upload.Subst["__BCC__"] = bcc + "\nBcc: %s" % (Cnf["Dinstall::Bcc"]) else: - Katie.Subst["__BCC__"] = bcc + Upload.Subst["__BCC__"] = bcc # Sort the .changes files so that we process sourceful ones first - changes_files.sort(utils.changes_compare) + changes_files.sort(daklib.utils.changes_compare) # Process the changes files for changes_file in changes_files: @@ -1341,13 +1358,13 @@ def main(): if not Options["No-Action"]: clean_holding() - accept_count = Katie.accept_count - accept_bytes = Katie.accept_bytes + accept_count = Upload.accept_count + accept_bytes = Upload.accept_bytes if accept_count: sets = "set" if accept_count > 1: sets = "sets" - print "Accepted %d package %s, %s." % (accept_count, sets, utils.size_type(int(accept_bytes))) + print "Accepted %d package %s, %s." % (accept_count, sets, daklib.utils.size_type(int(accept_bytes))) Logger.log(["total",accept_count,accept_bytes]) if not Options["No-Action"]: