X-Git-Url: https://git.decadent.org.uk/gitweb/?a=blobdiff_plain;f=dak%2Fqueue_report.py;h=1bef3e55915bd9de7b1b43e04120a69ffe377190;hb=362ea497d3c9ee5891fca772132269d809a0547a;hp=8e338e526181c9d5d16dd4737ae7b36e6dce725e;hpb=19e99ee6a161707062fd7ac5ea5ac4781a28d0bc;p=dak.git diff --git a/dak/queue_report.py b/dak/queue_report.py index 8e338e52..1bef3e55 100755 --- a/dak/queue_report.py +++ b/dak/queue_report.py @@ -37,11 +37,17 @@ from copy import copy import glob, os, stat, sys, time import apt_pkg +try: + import rrdtool +except ImportError: + pass from daklib import utils -from daklib.changes import Changes -from daklib.dbconn import DBConn, has_new_comment +from daklib.dbconn import DBConn, DBSource, has_new_comment, PolicyQueue, \ + get_uid_from_fingerprint +from daklib.policy import PolicyQueueUploadHandler from daklib.textutils import fix_maintainer +from daklib.utils import get_logins_from_ldap from daklib.dak_exceptions import * Cnf = None @@ -52,7 +58,7 @@ row_number = 0 def usage(exit_code=0): print """Usage: dak queue-report -Prints a report of packages in queue directories (usually new and byhand). +Prints a report of packages in queues (usually new and byhand). -h, --help show this help and exit. -8, --822 writes 822 formated output to the location set in dak.conf @@ -60,6 +66,7 @@ Prints a report of packages in queue directories (usually new and byhand). -s, --sort=key sort output according to key, see below. -a, --age=key if using sort by age, how should time be treated? If not given a default of hours will be used. + -r, --rrd=key Directory where rrd files to be updated are stored -d, --directories=key A comma seperated list of queues to be scanned Sorting Keys: ao=age, oldest first. an=age, newest first. @@ -172,6 +179,25 @@ def header():
Timestamp: %s (UTC)
" % (time.strftime("%d.%m.%Y / %H:%M:%S", time.gmtime())) + print "There are graphs about the queues available.
" print """Click to toggle all/binary-NEW packages
%s | " % (source) + if "sourceNEW" in trclass: + print "%s | " % (source) + else: + print "%(source)s | " % {'source': source} print ""
for vers in version.split():
- print "%s " % (source, utils.html_escape(vers), utils.html_escape(vers)) + print "%s " % (source, utils.html_escape(vers), utils.html_escape(vers)) print " | "
print "%s | " % (arch) print ""
@@ -277,12 +318,8 @@ def table_row(source, version, arch, last_mod, maint, distribution, closes, fing
(name, mail) = changedby.split(":", 1)
print "Changed-By: %s " % (utils.html_escape(mail), utils.html_escape(name)) - try: - (login, domain) = sponsor.split("@", 1) - print "Sponsor: %s@debian.org " % (utils.html_escape(login), utils.html_escape(login)) - except Exception, e: - print "WARNING: Exception %s" % e - pass + if sponsor: + print "Sponsor: %s@debian.org " % (utils.html_escape(sponsor), utils.html_escape(sponsor)) print "Fingerprint: %s" % (fingerprint) print " | "
@@ -295,42 +332,74 @@ def table_row(source, version, arch, last_mod, maint, distribution, closes, fing
############################################################
-def process_changes_files(changes_files, type, log):
- msg = ""
- cache = {}
- # Read in all the .changes files
- for filename in changes_files:
+def update_graph_database(rrd_dir, type, n_source, n_binary):
+ if not rrd_dir:
+ return
+
+ rrd_file = os.path.join(rrd_dir, type.lower()+'.rrd')
+ update = [rrd_file, "N:%s:%s" % (n_source, n_binary)]
+
+ try:
+ rrdtool.update(*update)
+ except rrdtool.error:
+ create = [rrd_file]+"""
+--step
+300
+--start
+0
+DS:ds0:GAUGE:7200:0:1000
+DS:ds1:GAUGE:7200:0:1000
+RRA:AVERAGE:0.5:1:599
+RRA:AVERAGE:0.5:6:700
+RRA:AVERAGE:0.5:24:775
+RRA:AVERAGE:0.5:288:795
+RRA:MAX:0.5:1:600
+RRA:MAX:0.5:6:700
+RRA:MAX:0.5:24:775
+RRA:MAX:0.5:288:795
+""".strip().split("\n")
try:
- c = Changes()
- c.load_dot_dak(filename)
- cache[filename] = copy(c.changes)
- cache[filename]["filename"] = filename
- except Exception, e:
- print "WARNING: Exception %s" % e
- continue
+ rc = rrdtool.create(*create)
+ ru = rrdtool.update(*update)
+ except rrdtool.error as e:
+ print('warning: queue_report: rrdtool error, skipping %s.rrd: %s' % (type, e))
+ except NameError:
+ pass
+
+############################################################
+
+def process_queue(queue, log, rrd_dir):
+ msg = ""
+ type = queue.queue_name
+ session = DBConn().session()
+
# Divide the .changes into per-source groups
per_source = {}
- for filename in cache.keys():
- source = cache[filename]["source"]
- if not per_source.has_key(source):
+ for upload in queue.uploads:
+ source = upload.changes.source
+ if source not in per_source:
per_source[source] = {}
per_source[source]["list"] = []
- per_source[source]["list"].append(cache[filename])
+ per_source[source]["processed"] = ""
+ handler = PolicyQueueUploadHandler(upload, session)
+ if handler.get_action():
+ per_source[source]["processed"] = " | PENDING %s" % handler.get_action()
+ per_source[source]["list"].append(upload)
# Determine oldest time and have note status for each source group
for source in per_source.keys():
source_list = per_source[source]["list"]
first = source_list[0]
- oldest = os.stat(first["filename"])[stat.ST_MTIME]
+ oldest = time.mktime(first.changes.created.timetuple())
have_note = 0
for d in per_source[source]["list"]:
- mtime = os.stat(d["filename"])[stat.ST_MTIME]
+ mtime = time.mktime(d.changes.created.timetuple())
if Cnf.has_key("Queue-Report::Options::New"):
if mtime > oldest:
oldest = mtime
else:
if mtime < oldest:
oldest = mtime
- have_note += has_new_comment(d["source"], d["version"])
+ have_note += has_new_comment(d.policy_queue, d.changes.source, d.changes.version)
per_source[source]["oldest"] = oldest
if not have_note:
per_source[source]["note_state"] = 0; # none
@@ -341,10 +410,16 @@ def process_changes_files(changes_files, type, log):
per_source_items = per_source.items()
per_source_items.sort(sg_compare)
+ update_graph_database(rrd_dir, type, len(per_source_items), len(queue.uploads))
+
entries = []
max_source_len = 0
max_version_len = 0
max_arch_len = 0
+ try:
+ logins = get_logins_from_ldap()
+ except:
+ logins = dict()
for i in per_source_items:
maintainer = {}
maint=""
@@ -354,20 +429,25 @@ def process_changes_files(changes_files, type, log):
changeby = {}
changedby=""
sponsor=""
- filename=i[1]["list"][0]["filename"]
+ filename=i[1]["list"][0].changes.changesname
last_modified = time.time()-i[1]["oldest"]
- source = i[1]["list"][0]["source"]
+ source = i[1]["list"][0].changes.source
if len(source) > max_source_len:
max_source_len = len(source)
- arches = {}
- versions = {}
+ binary_list = i[1]["list"][0].binaries
+ binary = ', '.join([ b.package for b in binary_list ])
+ arches = set()
+ versions = set()
for j in i[1]["list"]:
+ dbc = j.changes
+ changesbase = dbc.changesname
+
if Cnf.has_key("Queue-Report::Options::New") or Cnf.has_key("Queue-Report::Options::822"):
try:
(maintainer["maintainer822"], maintainer["maintainer2047"],
maintainer["maintainername"], maintainer["maintaineremail"]) = \
- fix_maintainer (j["maintainer"])
- except ParseMaintError, msg:
+ fix_maintainer (dbc.maintainer)
+ except ParseMaintError as msg:
print "Problems while parsing maintainer address\n"
maintainer["maintainername"] = "Unknown"
maintainer["maintaineremail"] = "Unknown"
@@ -376,26 +456,35 @@ def process_changes_files(changes_files, type, log):
try:
(changeby["changedby822"], changeby["changedby2047"],
changeby["changedbyname"], changeby["changedbyemail"]) = \
- fix_maintainer (j["changed-by"])
- except ParseMaintError, msg:
+ fix_maintainer (dbc.changedby)
+ except ParseMaintError as msg:
(changeby["changedby822"], changeby["changedby2047"],
changeby["changedbyname"], changeby["changedbyemail"]) = \
("", "", "", "")
changedby="%s:%s" % (changeby["changedbyname"], changeby["changedbyemail"])
- distribution=j["distribution"].keys()
- closes=j["closes"].keys()
- fingerprint=j["fingerprint"]
- if j.has_key("sponsoremail"):
- sponsor=j["sponsoremail"]
- for arch in j["architecture"].keys():
- arches[arch] = ""
- version = j["version"]
- versions[version] = ""
- arches_list = arches.keys()
+ distribution=dbc.distribution.split()
+ closes=dbc.closes
+
+ fingerprint = dbc.fingerprint
+ sponsor_name = get_uid_from_fingerprint(fingerprint).name
+ sponsor_login = get_uid_from_fingerprint(fingerprint).uid
+ if '@' in sponsor_login:
+ if fingerprint in logins:
+ sponsor_login = logins[fingerprint]
+ if (sponsor_name != maintainer["maintainername"] and
+ sponsor_name != changeby["changedbyname"] and
+ sponsor_login + '@debian.org' != maintainer["maintaineremail"] and
+ sponsor_name != changeby["changedbyemail"]):
+ sponsor = sponsor_login
+
+ for arch in dbc.architecture.split():
+ arches.add(arch)
+ versions.add(dbc.version)
+ arches_list = list(arches)
arches_list.sort(utils.arch_compare_sw)
arch_list = " ".join(arches_list)
- version_list = " ".join(versions.keys())
+ version_list = " ".join(versions)
if len(version_list) > max_version_len:
max_version_len = len(version_list)
if len(arch_list) > max_arch_len:
@@ -404,7 +493,7 @@ def process_changes_files(changes_files, type, log):
note = " | [N]"
else:
note = ""
- entries.append([source, version_list, arch_list, note, last_modified, maint, distribution, closes, fingerprint, sponsor, changedby, filename])
+ entries.append([source, binary, version_list, arch_list, per_source[source]["processed"], note, last_modified, maint, distribution, closes, fingerprint, sponsor, changedby, filename])
# direction entry consists of "Which field, which direction, time-consider" where
# time-consider says how we should treat last_modified. Thats all.
@@ -415,16 +504,16 @@ def process_changes_files(changes_files, type, log):
age = Cnf["Queue-Report::Options::Age"]
if Cnf.has_key("Queue-Report::Options::New"):
# If we produce html we always have oldest first.
- direction.append([4,-1,"ao"])
+ direction.append([6,-1,"ao"])
else:
if Cnf.has_key("Queue-Report::Options::Sort"):
for i in Cnf["Queue-Report::Options::Sort"].split(","):
if i == "ao":
# Age, oldest first.
- direction.append([4,-1,age])
+ direction.append([6,-1,age])
elif i == "an":
# Age, newest first.
- direction.append([4,1,age])
+ direction.append([6,1,age])
elif i == "na":
# Name, Ascending.
direction.append([0,1,0])
@@ -433,10 +522,10 @@ def process_changes_files(changes_files, type, log):
direction.append([0,-1,0])
elif i == "nl":
# Notes last.
- direction.append([3,1,0])
+ direction.append([5,1,0])
elif i == "nf":
# Notes first.
- direction.append([3,-1,0])
+ direction.append([5,-1,0])
entries.sort(lambda x, y: sortfunc(x, y))
# Yes, in theory you can add several sort options at the commandline with. But my mind is to small
# at the moment to come up with a real good sorting function that considers all the sidesteps you
@@ -446,11 +535,12 @@ def process_changes_files(changes_files, type, log):
if Cnf.has_key("Queue-Report::Options::822"):
# print stuff out in 822 format
for entry in entries:
- (source, version_list, arch_list, note, last_modified, maint, distribution, closes, fingerprint, sponsor, changedby, changes_file) = entry
+ (source, binary, version_list, arch_list, processed, note, last_modified, maint, distribution, closes, fingerprint, sponsor, changedby, changes_file) = entry
# We'll always have Source, Version, Arch, Mantainer, and Dist
# For the rest, check to see if we have them, then print them out
log.write("Source: " + source + "\n")
+ log.write("Binary: " + binary + "\n")
log.write("Version: " + version_list + "\n")
log.write("Architectures: ")
log.write( (", ".join(arch_list.split(" "))) + "\n")
@@ -464,7 +554,7 @@ def process_changes_files(changes_files, type, log):
(name, mail) = changedby.split(":", 1)
log.write("Changed-By: " + name + " <"+mail+">" + "\n")
if sponsor:
- log.write("Sponsored-By: " + sponsor + "\n")
+ log.write("Sponsored-By: %s@debian.org\n" % sponsor)
log.write("Distribution:")
for dist in distribution:
log.write(" " + dist)
@@ -478,31 +568,30 @@ def process_changes_files(changes_files, type, log):
log.write("Changes-File: " + os.path.basename(changes_file) + "\n")
log.write("\n")
+ total_count = len(queue.uploads)
+ source_count = len(per_source_items)
+
if Cnf.has_key("Queue-Report::Options::New"):
- direction.append([4,1,"ao"])
+ direction.append([6,1,"ao"])
entries.sort(lambda x, y: sortfunc(x, y))
# Output for a html file. First table header. then table_footer.
# Any line between them is then a