X-Git-Url: https://git.decadent.org.uk/gitweb/?a=blobdiff_plain;f=dak%2Fqueue_report.py;h=1bef3e55915bd9de7b1b43e04120a69ffe377190;hb=362ea497d3c9ee5891fca772132269d809a0547a;hp=98c80f0af85dd0b50bcbed1e6e8b856b80a2d11e;hpb=50db22ea5f288daa39f81138a41a509d9a41cc3e;p=dak.git diff --git a/dak/queue_report.py b/dak/queue_report.py index 98c80f0a..1bef3e55 100755 --- a/dak/queue_report.py +++ b/dak/queue_report.py @@ -43,9 +43,11 @@ except ImportError: pass from daklib import utils -from daklib.queue import Upload -from daklib.dbconn import DBConn, has_new_comment, DBChange, DBSource, get_uid_from_fingerprint +from daklib.dbconn import DBConn, DBSource, has_new_comment, PolicyQueue, \ + get_uid_from_fingerprint +from daklib.policy import PolicyQueueUploadHandler from daklib.textutils import fix_maintainer +from daklib.utils import get_logins_from_ldap from daklib.dak_exceptions import * Cnf = None @@ -56,7 +58,7 @@ row_number = 0 def usage(exit_code=0): print """Usage: dak queue-report -Prints a report of packages in queue directories (usually new and byhand). +Prints a report of packages in queues (usually new and byhand). -h, --help show this help and exit. -8, --822 writes 822 formated output to the location set in dak.conf @@ -178,6 +180,7 @@ def header(): Debian NEW and BYHAND Packages
@@ -242,12 +246,12 @@ def footer(): """ def table_header(type, source_count, total_count): - print "Click to toggle all/binary-NEW packages
%s | " % (source) + if "sourceNEW" in trclass: + print "%s | " % (source) + else: + print "%(source)s | " % {'source': source} print ""
for vers in version.split():
print "%s " % (source, utils.html_escape(vers), utils.html_escape(vers)) @@ -308,11 +319,7 @@ def table_row(source, version, arch, last_mod, maint, distribution, closes, fing print "Changed-By: %s " % (utils.html_escape(mail), utils.html_escape(name)) if sponsor: - try: - (login, domain) = sponsor.split("@", 1) - print "Sponsor: %s@debian.org " % (utils.html_escape(login), utils.html_escape(login)) - except Exception, e: - pass + print "Sponsor: %s@debian.org " % (utils.html_escape(sponsor), utils.html_escape(sponsor)) print "Fingerprint: %s" % (fingerprint) print " | "
@@ -354,49 +361,45 @@ RRA:MAX:0.5:288:795
try:
rc = rrdtool.create(*create)
ru = rrdtool.update(*update)
- except rrdtool.error, e:
+ except rrdtool.error as e:
print('warning: queue_report: rrdtool error, skipping %s.rrd: %s' % (type, e))
except NameError:
pass
############################################################
-def process_changes_files(changes_files, type, log, rrd_dir):
+def process_queue(queue, log, rrd_dir):
msg = ""
- cache = {}
- # Read in all the .changes files
- for filename in changes_files:
- try:
- u = Upload()
- u.load_changes(filename)
- cache[filename] = copy(u.pkg.changes)
- cache[filename]["filename"] = filename
- except Exception, e:
- print "WARNING: Exception %s" % e
- continue
+ type = queue.queue_name
+ session = DBConn().session()
+
# Divide the .changes into per-source groups
per_source = {}
- for filename in cache.keys():
- source = cache[filename]["source"]
- if not per_source.has_key(source):
+ for upload in queue.uploads:
+ source = upload.changes.source
+ if source not in per_source:
per_source[source] = {}
per_source[source]["list"] = []
- per_source[source]["list"].append(cache[filename])
+ per_source[source]["processed"] = ""
+ handler = PolicyQueueUploadHandler(upload, session)
+ if handler.get_action():
+ per_source[source]["processed"] = " | PENDING %s" % handler.get_action()
+ per_source[source]["list"].append(upload)
# Determine oldest time and have note status for each source group
for source in per_source.keys():
source_list = per_source[source]["list"]
first = source_list[0]
- oldest = os.stat(first["filename"])[stat.ST_MTIME]
+ oldest = time.mktime(first.changes.created.timetuple())
have_note = 0
for d in per_source[source]["list"]:
- mtime = os.stat(d["filename"])[stat.ST_MTIME]
+ mtime = time.mktime(d.changes.created.timetuple())
if Cnf.has_key("Queue-Report::Options::New"):
if mtime > oldest:
oldest = mtime
else:
if mtime < oldest:
oldest = mtime
- have_note += has_new_comment(d["source"], d["version"])
+ have_note += has_new_comment(d.policy_queue, d.changes.source, d.changes.version)
per_source[source]["oldest"] = oldest
if not have_note:
per_source[source]["note_state"] = 0; # none
@@ -407,12 +410,16 @@ def process_changes_files(changes_files, type, log, rrd_dir):
per_source_items = per_source.items()
per_source_items.sort(sg_compare)
- update_graph_database(rrd_dir, type, len(per_source_items), len(changes_files))
+ update_graph_database(rrd_dir, type, len(per_source_items), len(queue.uploads))
entries = []
max_source_len = 0
max_version_len = 0
max_arch_len = 0
+ try:
+ logins = get_logins_from_ldap()
+ except:
+ logins = dict()
for i in per_source_items:
maintainer = {}
maint=""
@@ -422,31 +429,25 @@ def process_changes_files(changes_files, type, log, rrd_dir):
changeby = {}
changedby=""
sponsor=""
- filename=i[1]["list"][0]["filename"]
+ filename=i[1]["list"][0].changes.changesname
last_modified = time.time()-i[1]["oldest"]
- source = i[1]["list"][0]["source"]
+ source = i[1]["list"][0].changes.source
if len(source) > max_source_len:
max_source_len = len(source)
- binary_list = i[1]["list"][0]["binary"].keys()
- binary = ', '.join(binary_list)
- arches = {}
- versions = {}
+ binary_list = i[1]["list"][0].binaries
+ binary = ', '.join([ b.package for b in binary_list ])
+ arches = set()
+ versions = set()
for j in i[1]["list"]:
- changesbase = os.path.basename(j["filename"])
- try:
- session = DBConn().session()
- dbc = session.query(DBChange).filter_by(changesname=changesbase).one()
- session.close()
- except Exception, e:
- print "Can't find changes file in NEW for %s (%s)" % (changesbase, e)
- dbc = None
+ dbc = j.changes
+ changesbase = dbc.changesname
if Cnf.has_key("Queue-Report::Options::New") or Cnf.has_key("Queue-Report::Options::822"):
try:
(maintainer["maintainer822"], maintainer["maintainer2047"],
maintainer["maintainername"], maintainer["maintaineremail"]) = \
- fix_maintainer (j["maintainer"])
- except ParseMaintError, msg:
+ fix_maintainer (dbc.maintainer)
+ except ParseMaintError as msg:
print "Problems while parsing maintainer address\n"
maintainer["maintainername"] = "Unknown"
maintainer["maintaineremail"] = "Unknown"
@@ -455,31 +456,35 @@ def process_changes_files(changes_files, type, log, rrd_dir):
try:
(changeby["changedby822"], changeby["changedby2047"],
changeby["changedbyname"], changeby["changedbyemail"]) = \
- fix_maintainer (j["changed-by"])
- except ParseMaintError, msg:
+ fix_maintainer (dbc.changedby)
+ except ParseMaintError as msg:
(changeby["changedby822"], changeby["changedby2047"],
changeby["changedbyname"], changeby["changedbyemail"]) = \
("", "", "", "")
changedby="%s:%s" % (changeby["changedbyname"], changeby["changedbyemail"])
- distribution=j["distribution"].keys()
- closes=j["closes"].keys()
- if dbc:
- fingerprint = dbc.fingerprint
- sponsor_name = get_uid_from_fingerprint(fingerprint).name
- sponsor_email = get_uid_from_fingerprint(fingerprint).uid + "@debian.org"
- if sponsor_name != maintainer["maintainername"] and sponsor_name != changeby["changedbyname"] and \
- sponsor_email != maintainer["maintaineremail"] and sponsor_name != changeby["changedbyemail"]:
- sponsor = sponsor_email
-
- for arch in j["architecture"].keys():
- arches[arch] = ""
- version = j["version"]
- versions[version] = ""
- arches_list = arches.keys()
+ distribution=dbc.distribution.split()
+ closes=dbc.closes
+
+ fingerprint = dbc.fingerprint
+ sponsor_name = get_uid_from_fingerprint(fingerprint).name
+ sponsor_login = get_uid_from_fingerprint(fingerprint).uid
+ if '@' in sponsor_login:
+ if fingerprint in logins:
+ sponsor_login = logins[fingerprint]
+ if (sponsor_name != maintainer["maintainername"] and
+ sponsor_name != changeby["changedbyname"] and
+ sponsor_login + '@debian.org' != maintainer["maintaineremail"] and
+ sponsor_name != changeby["changedbyemail"]):
+ sponsor = sponsor_login
+
+ for arch in dbc.architecture.split():
+ arches.add(arch)
+ versions.add(dbc.version)
+ arches_list = list(arches)
arches_list.sort(utils.arch_compare_sw)
arch_list = " ".join(arches_list)
- version_list = " ".join(versions.keys())
+ version_list = " ".join(versions)
if len(version_list) > max_version_len:
max_version_len = len(version_list)
if len(arch_list) > max_arch_len:
@@ -488,7 +493,7 @@ def process_changes_files(changes_files, type, log, rrd_dir):
note = " | [N]"
else:
note = ""
- entries.append([source, binary, version_list, arch_list, note, last_modified, maint, distribution, closes, fingerprint, sponsor, changedby, filename])
+ entries.append([source, binary, version_list, arch_list, per_source[source]["processed"], note, last_modified, maint, distribution, closes, fingerprint, sponsor, changedby, filename])
# direction entry consists of "Which field, which direction, time-consider" where
# time-consider says how we should treat last_modified. Thats all.
@@ -499,16 +504,16 @@ def process_changes_files(changes_files, type, log, rrd_dir):
age = Cnf["Queue-Report::Options::Age"]
if Cnf.has_key("Queue-Report::Options::New"):
# If we produce html we always have oldest first.
- direction.append([5,-1,"ao"])
+ direction.append([6,-1,"ao"])
else:
if Cnf.has_key("Queue-Report::Options::Sort"):
for i in Cnf["Queue-Report::Options::Sort"].split(","):
if i == "ao":
# Age, oldest first.
- direction.append([5,-1,age])
+ direction.append([6,-1,age])
elif i == "an":
# Age, newest first.
- direction.append([5,1,age])
+ direction.append([6,1,age])
elif i == "na":
# Name, Ascending.
direction.append([0,1,0])
@@ -517,10 +522,10 @@ def process_changes_files(changes_files, type, log, rrd_dir):
direction.append([0,-1,0])
elif i == "nl":
# Notes last.
- direction.append([4,1,0])
+ direction.append([5,1,0])
elif i == "nf":
# Notes first.
- direction.append([4,-1,0])
+ direction.append([5,-1,0])
entries.sort(lambda x, y: sortfunc(x, y))
# Yes, in theory you can add several sort options at the commandline with. But my mind is to small
# at the moment to come up with a real good sorting function that considers all the sidesteps you
@@ -530,7 +535,7 @@ def process_changes_files(changes_files, type, log, rrd_dir):
if Cnf.has_key("Queue-Report::Options::822"):
# print stuff out in 822 format
for entry in entries:
- (source, binary, version_list, arch_list, note, last_modified, maint, distribution, closes, fingerprint, sponsor, changedby, changes_file) = entry
+ (source, binary, version_list, arch_list, processed, note, last_modified, maint, distribution, closes, fingerprint, sponsor, changedby, changes_file) = entry
# We'll always have Source, Version, Arch, Mantainer, and Dist
# For the rest, check to see if we have them, then print them out
@@ -549,7 +554,7 @@ def process_changes_files(changes_files, type, log, rrd_dir):
(name, mail) = changedby.split(":", 1)
log.write("Changed-By: " + name + " <"+mail+">" + "\n")
if sponsor:
- log.write("Sponsored-By: " + sponsor + "\n")
+ log.write("Sponsored-By: %s@debian.org\n" % sponsor)
log.write("Distribution:")
for dist in distribution:
log.write(" " + dist)
@@ -563,31 +568,30 @@ def process_changes_files(changes_files, type, log, rrd_dir):
log.write("Changes-File: " + os.path.basename(changes_file) + "\n")
log.write("\n")
+ total_count = len(queue.uploads)
+ source_count = len(per_source_items)
+
if Cnf.has_key("Queue-Report::Options::New"):
- direction.append([5,1,"ao"])
+ direction.append([6,1,"ao"])
entries.sort(lambda x, y: sortfunc(x, y))
# Output for a html file. First table header. then table_footer.
# Any line between them is then a