X-Git-Url: https://git.decadent.org.uk/gitweb/?a=blobdiff_plain;ds=sidebyside;f=dak%2Fqueue_report.py;h=0ab0b732d48324c6fef38b3f834a1aa8bdf72a06;hb=fb5b24b09968fc7351f2dc307b8cdc76463bbb0a;hp=a929a4b6b8e5eb87ede949284cd368b2dd3005e3;hpb=7b3c2a52a559c459ddf5da96cfeceba8b19cba9d;p=dak.git diff --git a/dak/queue_report.py b/dak/queue_report.py index a929a4b6..0ab0b732 100755 --- a/dak/queue_report.py +++ b/dak/queue_report.py @@ -43,8 +43,9 @@ except ImportError: pass from daklib import utils -from daklib.queue import Upload -from daklib.dbconn import DBConn, has_new_comment, DBChange, DBSource, get_uid_from_fingerprint +from daklib.dbconn import DBConn, DBSource, has_new_comment, PolicyQueue, \ + get_uid_from_fingerprint +from daklib.policy import PolicyQueueUploadHandler from daklib.textutils import fix_maintainer from daklib.dak_exceptions import * @@ -56,7 +57,7 @@ row_number = 0 def usage(exit_code=0): print """Usage: dak queue-report -Prints a report of packages in queue directories (usually new and byhand). +Prints a report of packages in queues (usually new and byhand). -h, --help show this help and exit. -8, --822 writes 822 formated output to the location set in dak.conf @@ -178,6 +179,7 @@ def header(): Debian NEW and BYHAND Packages
@@ -222,6 +225,7 @@ def header(): def footer(): print "Timestamp: %s (UTC)
" % (time.strftime("%d.%m.%Y / %H:%M:%S", time.gmtime())) + print "There are graphs about the queues available.
" print """Click to toggle all/binary-NEW packages
%s | " % (source) + if "sourceNEW" in trclass: + print "%s | " % (source) + else: + print "%(source)s | " % {'source': source} print ""
for vers in version.split():
print "%s " % (source, utils.html_escape(vers), utils.html_escape(vers)) @@ -309,8 +320,8 @@ def table_row(source, version, arch, last_mod, maint, distribution, closes, fing if sponsor: try: (login, domain) = sponsor.split("@", 1) - print "Sponsor: %s@debian.org " % (utils.html_escape(login), utils.html_escape(login)) - except Exception, e: + print "Sponsor: %s@debian.org " % (utils.html_escape(login), utils.html_escape(login)) + except Exception as e: pass print "Fingerprint: %s" % (fingerprint) @@ -360,42 +371,38 @@ RRA:MAX:0.5:288:795 ############################################################ -def process_changes_files(changes_files, type, log, rrd_dir): +def process_queue(queue, log, rrd_dir): msg = "" - cache = {} - # Read in all the .changes files - for filename in changes_files: - try: - u = Upload() - u.load_changes(filename) - cache[filename] = copy(u.pkg.changes) - cache[filename]["filename"] = filename - except Exception, e: - print "WARNING: Exception %s" % e - continue + type = queue.queue_name + session = DBConn().session() + # Divide the .changes into per-source groups per_source = {} - for filename in cache.keys(): - source = cache[filename]["source"] - if not per_source.has_key(source): + for upload in queue.uploads: + source = upload.changes.source + if source not in per_source: per_source[source] = {} per_source[source]["list"] = [] - per_source[source]["list"].append(cache[filename]) + per_source[source]["processed"] = "" + handler = PolicyQueueUploadHandler(upload, session) + if handler.get_action(): + per_source[source]["processed"] = " | PENDING %s" % handler.get_action() + per_source[source]["list"].append(upload) # Determine oldest time and have note status for each source group for source in per_source.keys(): source_list = per_source[source]["list"] first = source_list[0] - oldest = os.stat(first["filename"])[stat.ST_MTIME] + oldest = time.mktime(first.changes.created.timetuple()) have_note = 0 for d in per_source[source]["list"]: - mtime = os.stat(d["filename"])[stat.ST_MTIME] + mtime = time.mktime(d.changes.created.timetuple()) if Cnf.has_key("Queue-Report::Options::New"): if mtime > oldest: oldest = mtime else: if mtime < oldest: oldest = mtime - have_note += has_new_comment(d["source"], d["version"]) + have_note += has_new_comment(d.changes.source, d.changes.version) per_source[source]["oldest"] = oldest if not have_note: per_source[source]["note_state"] = 0; # none @@ -406,7 +413,7 @@ def process_changes_files(changes_files, type, log, rrd_dir): per_source_items = per_source.items() per_source_items.sort(sg_compare) - update_graph_database(rrd_dir, type, len(per_source_items), len(changes_files)) + update_graph_database(rrd_dir, type, len(per_source_items), len(queue.uploads)) entries = [] max_source_len = 0 @@ -421,31 +428,25 @@ def process_changes_files(changes_files, type, log, rrd_dir): changeby = {} changedby="" sponsor="" - filename=i[1]["list"][0]["filename"] + filename=i[1]["list"][0].changes.changesname last_modified = time.time()-i[1]["oldest"] - source = i[1]["list"][0]["source"] + source = i[1]["list"][0].changes.source if len(source) > max_source_len: max_source_len = len(source) - binary_list = i[1]["list"][0]["binary"].keys() - binary = ', '.join(binary_list) - arches = {} - versions = {} + binary_list = i[1]["list"][0].binaries + binary = ', '.join([ b.package for b in binary_list ]) + arches = set() + versions = set() for j in i[1]["list"]: - changesbase = os.path.basename(j["filename"]) - try: - session = DBConn().session() - dbc = session.query(DBChange).filter_by(changesname=changesbase).one() - session.close() - except Exception, e: - print "Can't find changes file in NEW for %s (%s)" % (changesbase, e) - dbc = None + dbc = j.changes + changesbase = dbc.changesname if Cnf.has_key("Queue-Report::Options::New") or Cnf.has_key("Queue-Report::Options::822"): try: (maintainer["maintainer822"], maintainer["maintainer2047"], maintainer["maintainername"], maintainer["maintaineremail"]) = \ - fix_maintainer (j["maintainer"]) - except ParseMaintError, msg: + fix_maintainer (dbc.maintainer) + except ParseMaintError as msg: print "Problems while parsing maintainer address\n" maintainer["maintainername"] = "Unknown" maintainer["maintaineremail"] = "Unknown" @@ -454,31 +455,30 @@ def process_changes_files(changes_files, type, log, rrd_dir): try: (changeby["changedby822"], changeby["changedby2047"], changeby["changedbyname"], changeby["changedbyemail"]) = \ - fix_maintainer (j["changed-by"]) - except ParseMaintError, msg: + fix_maintainer (dbc.changedby) + except ParseMaintError as msg: (changeby["changedby822"], changeby["changedby2047"], changeby["changedbyname"], changeby["changedbyemail"]) = \ ("", "", "", "") changedby="%s:%s" % (changeby["changedbyname"], changeby["changedbyemail"]) - distribution=j["distribution"].keys() - closes=j["closes"].keys() - if dbc: - fingerprint = dbc.fingerprint - sponsor_name = get_uid_from_fingerprint(fingerprint).name - sponsor_email = get_uid_from_fingerprint(fingerprint).uid + "@debian.org" - if sponsor_name != maintainer["maintainername"] and sponsor_name != changeby["changedbyname"] and \ - sponsor_email != maintainer["maintaineremail"] and sponsor_name != changeby["changedbyemail"]: - sponsor = sponsor_email - - for arch in j["architecture"].keys(): - arches[arch] = "" - version = j["version"] - versions[version] = "" - arches_list = arches.keys() + distribution=dbc.distribution.split() + closes=dbc.closes + + fingerprint = dbc.fingerprint + sponsor_name = get_uid_from_fingerprint(fingerprint).name + sponsor_email = get_uid_from_fingerprint(fingerprint).uid + "@debian.org" + if sponsor_name != maintainer["maintainername"] and sponsor_name != changeby["changedbyname"] and \ + sponsor_email != maintainer["maintaineremail"] and sponsor_name != changeby["changedbyemail"]: + sponsor = sponsor_email + + for arch in dbc.architecture.split(): + arches.add(arch) + versions.add(dbc.version) + arches_list = list(arches) arches_list.sort(utils.arch_compare_sw) arch_list = " ".join(arches_list) - version_list = " ".join(versions.keys()) + version_list = " ".join(versions) if len(version_list) > max_version_len: max_version_len = len(version_list) if len(arch_list) > max_arch_len: @@ -487,7 +487,7 @@ def process_changes_files(changes_files, type, log, rrd_dir): note = " | [N]" else: note = "" - entries.append([source, binary, version_list, arch_list, note, last_modified, maint, distribution, closes, fingerprint, sponsor, changedby, filename]) + entries.append([source, binary, version_list, arch_list, per_source[source]["processed"], note, last_modified, maint, distribution, closes, fingerprint, sponsor, changedby, filename]) # direction entry consists of "Which field, which direction, time-consider" where # time-consider says how we should treat last_modified. Thats all. @@ -498,16 +498,16 @@ def process_changes_files(changes_files, type, log, rrd_dir): age = Cnf["Queue-Report::Options::Age"] if Cnf.has_key("Queue-Report::Options::New"): # If we produce html we always have oldest first. - direction.append([5,-1,"ao"]) + direction.append([6,-1,"ao"]) else: if Cnf.has_key("Queue-Report::Options::Sort"): for i in Cnf["Queue-Report::Options::Sort"].split(","): if i == "ao": # Age, oldest first. - direction.append([5,-1,age]) + direction.append([6,-1,age]) elif i == "an": # Age, newest first. - direction.append([5,1,age]) + direction.append([6,1,age]) elif i == "na": # Name, Ascending. direction.append([0,1,0]) @@ -516,10 +516,10 @@ def process_changes_files(changes_files, type, log, rrd_dir): direction.append([0,-1,0]) elif i == "nl": # Notes last. - direction.append([4,1,0]) + direction.append([5,1,0]) elif i == "nf": # Notes first. - direction.append([4,-1,0]) + direction.append([5,-1,0]) entries.sort(lambda x, y: sortfunc(x, y)) # Yes, in theory you can add several sort options at the commandline with. But my mind is to small # at the moment to come up with a real good sorting function that considers all the sidesteps you @@ -529,7 +529,7 @@ def process_changes_files(changes_files, type, log, rrd_dir): if Cnf.has_key("Queue-Report::Options::822"): # print stuff out in 822 format for entry in entries: - (source, binary, version_list, arch_list, note, last_modified, maint, distribution, closes, fingerprint, sponsor, changedby, changes_file) = entry + (source, binary, version_list, arch_list, processed, note, last_modified, maint, distribution, closes, fingerprint, sponsor, changedby, changes_file) = entry # We'll always have Source, Version, Arch, Mantainer, and Dist # For the rest, check to see if we have them, then print them out @@ -548,7 +548,7 @@ def process_changes_files(changes_files, type, log, rrd_dir): (name, mail) = changedby.split(":", 1) log.write("Changed-By: " + name + " <"+mail+">" + "\n") if sponsor: - log.write("Sponsored-By: " + sponsor + "\n") + log.write("Sponsored-By: " + "@".join(sponsor.split("@")[:2]) + "\n") log.write("Distribution:") for dist in distribution: log.write(" " + dist) @@ -562,31 +562,30 @@ def process_changes_files(changes_files, type, log, rrd_dir): log.write("Changes-File: " + os.path.basename(changes_file) + "\n") log.write("\n") + total_count = len(queue.uploads) + source_count = len(per_source_items) + if Cnf.has_key("Queue-Report::Options::New"): - direction.append([5,1,"ao"]) + direction.append([6,1,"ao"]) entries.sort(lambda x, y: sortfunc(x, y)) # Output for a html file. First table header. then table_footer. # Any line between them is then a |