X-Git-Url: https://git.decadent.org.uk/gitweb/?a=blobdiff_plain;f=dak%2Fqueue_report.py;h=8152bfa63e7e9c4cabe85754dc8a7aa1059bc334;hb=b79fdbd11acc8e8f4bd1d7a1a001524c645f0714;hp=be4787219e9fde8a5963fa3fb665637c90a0fd86;hpb=54a1e63df2a81a8f5e6f3148d59afcfeaa7b3495;p=dak.git diff --git a/dak/queue_report.py b/dak/queue_report.py index be478721..8152bfa6 100755 --- a/dak/queue_report.py +++ b/dak/queue_report.py @@ -37,10 +37,14 @@ from copy import copy import glob, os, stat, sys, time import apt_pkg +try: + import rrdtool +except ImportError: + pass from daklib import utils -from daklib.queue import Upload -from daklib.dbconn import DBConn, has_new_comment, DBChange, get_uid_from_fingerprint +from daklib.dbconn import DBConn, DBSource, has_new_comment, PolicyQueue, \ + get_uid_from_fingerprint from daklib.textutils import fix_maintainer from daklib.dak_exceptions import * @@ -52,7 +56,7 @@ row_number = 0 def usage(exit_code=0): print """Usage: dak queue-report -Prints a report of packages in queue directories (usually new and byhand). +Prints a report of packages in queues (usually new and byhand). -h, --help show this help and exit. -8, --822 writes 822 formated output to the location set in dak.conf @@ -60,6 +64,7 @@ Prints a report of packages in queue directories (usually new and byhand). -s, --sort=key sort output according to key, see below. -a, --age=key if using sort by age, how should time be treated? If not given a default of hours will be used. + -r, --rrd=key Directory where rrd files to be updated are stored -d, --directories=key A comma seperated list of queues to be scanned Sorting Keys: ao=age, oldest first. an=age, newest first. @@ -172,6 +177,25 @@ def header():
Timestamp: %s (UTC)
" % (time.strftime("%d.%m.%Y / %H:%M:%S", time.gmtime())) + print "There are graphs about the queues available.
" print """Click to toggle all/binary-NEW packages
%s | " % (source) + if "binNEW" in trclass: + print "%s | " % (source) + else: + print "%(source)s | " % {'source': source} print ""
for vers in version.split():
- print "%s " % (source, utils.html_escape(vers), utils.html_escape(vers)) + print "%s " % (source, utils.html_escape(vers), utils.html_escape(vers)) print " | "
print "%s | " % (arch) print ""
@@ -280,8 +315,8 @@ def table_row(source, version, arch, last_mod, maint, distribution, closes, fing
if sponsor:
try:
(login, domain) = sponsor.split("@", 1)
- print "Sponsor: %s@debian.org " % (utils.html_escape(login), utils.html_escape(login)) - except Exception, e: + print "Sponsor: %s@debian.org " % (utils.html_escape(login), utils.html_escape(login)) + except Exception as e: pass print "Fingerprint: %s" % (fingerprint) @@ -295,43 +330,69 @@ def table_row(source, version, arch, last_mod, maint, distribution, closes, fing ############################################################ -def process_changes_files(changes_files, type, log): - session = DBConn().session() - msg = "" - cache = {} - # Read in all the .changes files - for filename in changes_files: +def update_graph_database(rrd_dir, type, n_source, n_binary): + if not rrd_dir: + return + + rrd_file = os.path.join(rrd_dir, type.lower()+'.rrd') + update = [rrd_file, "N:%s:%s" % (n_source, n_binary)] + + try: + rrdtool.update(*update) + except rrdtool.error: + create = [rrd_file]+""" +--step +300 +--start +0 +DS:ds0:GAUGE:7200:0:1000 +DS:ds1:GAUGE:7200:0:1000 +RRA:AVERAGE:0.5:1:599 +RRA:AVERAGE:0.5:6:700 +RRA:AVERAGE:0.5:24:775 +RRA:AVERAGE:0.5:288:795 +RRA:MAX:0.5:1:600 +RRA:MAX:0.5:6:700 +RRA:MAX:0.5:24:775 +RRA:MAX:0.5:288:795 +""".strip().split("\n") try: - u = Upload() - u.load_changes(filename) - cache[filename] = copy(u.pkg.changes) - cache[filename]["filename"] = filename - except Exception, e: - print "WARNING: Exception %s" % e - continue + rc = rrdtool.create(*create) + ru = rrdtool.update(*update) + except rrdtool.error as e: + print('warning: queue_report: rrdtool error, skipping %s.rrd: %s' % (type, e)) + except NameError: + pass + +############################################################ + +def process_queue(queue, log, rrd_dir): + msg = "" + type = queue.queue_name + # Divide the .changes into per-source groups per_source = {} - for filename in cache.keys(): - source = cache[filename]["source"] - if not per_source.has_key(source): + for upload in queue.uploads: + source = upload.changes.source + if source not in per_source: per_source[source] = {} per_source[source]["list"] = [] - per_source[source]["list"].append(cache[filename]) + per_source[source]["list"].append(upload) # Determine oldest time and have note status for each source group for source in per_source.keys(): source_list = per_source[source]["list"] first = source_list[0] - oldest = os.stat(first["filename"])[stat.ST_MTIME] + oldest = time.mktime(first.changes.created.timetuple()) have_note = 0 for d in per_source[source]["list"]: - mtime = os.stat(d["filename"])[stat.ST_MTIME] + mtime = time.mktime(d.changes.created.timetuple()) if Cnf.has_key("Queue-Report::Options::New"): if mtime > oldest: oldest = mtime else: if mtime < oldest: oldest = mtime - have_note += has_new_comment(d["source"], d["version"]) + have_note += has_new_comment(d.changes.source, d.changes.version) per_source[source]["oldest"] = oldest if not have_note: per_source[source]["note_state"] = 0; # none @@ -342,6 +403,8 @@ def process_changes_files(changes_files, type, log): per_source_items = per_source.items() per_source_items.sort(sg_compare) + update_graph_database(rrd_dir, type, len(per_source_items), len(queue.uploads)) + entries = [] max_source_len = 0 max_version_len = 0 @@ -355,29 +418,25 @@ def process_changes_files(changes_files, type, log): changeby = {} changedby="" sponsor="" - filename=i[1]["list"][0]["filename"] + filename=i[1]["list"][0].changes.changesname last_modified = time.time()-i[1]["oldest"] - source = i[1]["list"][0]["source"] + source = i[1]["list"][0].changes.source if len(source) > max_source_len: max_source_len = len(source) - binary_list = i[1]["list"][0]["binary"].keys() - binary = ', '.join(binary_list) - arches = {} - versions = {} + binary_list = i[1]["list"][0].binaries + binary = ', '.join([ b.package for b in binary_list ]) + arches = set() + versions = set() for j in i[1]["list"]: - changesbase = os.path.basename(j["filename"]) - try: - dbc = session.query(DBChange).filter_by(changesname=changesbase).one() - except Exception, e: - print "Can't find changes file in NEW for %s (%s)" % (changesbase, e) - dbc = None + dbc = j.changes + changesbase = dbc.changesname if Cnf.has_key("Queue-Report::Options::New") or Cnf.has_key("Queue-Report::Options::822"): try: (maintainer["maintainer822"], maintainer["maintainer2047"], maintainer["maintainername"], maintainer["maintaineremail"]) = \ - fix_maintainer (j["maintainer"]) - except ParseMaintError, msg: + fix_maintainer (dbc.maintainer) + except ParseMaintError as msg: print "Problems while parsing maintainer address\n" maintainer["maintainername"] = "Unknown" maintainer["maintaineremail"] = "Unknown" @@ -386,31 +445,30 @@ def process_changes_files(changes_files, type, log): try: (changeby["changedby822"], changeby["changedby2047"], changeby["changedbyname"], changeby["changedbyemail"]) = \ - fix_maintainer (j["changed-by"]) - except ParseMaintError, msg: + fix_maintainer (dbc.changedby) + except ParseMaintError as msg: (changeby["changedby822"], changeby["changedby2047"], changeby["changedbyname"], changeby["changedbyemail"]) = \ ("", "", "", "") changedby="%s:%s" % (changeby["changedbyname"], changeby["changedbyemail"]) - distribution=j["distribution"].keys() - closes=j["closes"].keys() - if dbc: - fingerprint = dbc.fingerprint - sponsor_name = get_uid_from_fingerprint(fingerprint).name - sponsor_email = get_uid_from_fingerprint(fingerprint).uid + "@debian.org" - if sponsor_name != maintainer["maintainername"] and sponsor_name != changeby["changedbyname"] and \ - sponsor_email != maintainer["maintaineremail"] and sponsor_name != changeby["changedbyemail"]: - sponsor = sponsor_email - - for arch in j["architecture"].keys(): - arches[arch] = "" - version = j["version"] - versions[version] = "" - arches_list = arches.keys() + distribution=dbc.distribution.split() + closes=dbc.closes + + fingerprint = dbc.fingerprint + sponsor_name = get_uid_from_fingerprint(fingerprint).name + sponsor_email = get_uid_from_fingerprint(fingerprint).uid + "@debian.org" + if sponsor_name != maintainer["maintainername"] and sponsor_name != changeby["changedbyname"] and \ + sponsor_email != maintainer["maintaineremail"] and sponsor_name != changeby["changedbyemail"]: + sponsor = sponsor_email + + for arch in dbc.architecture.split(): + arches.add(arch) + versions.add(dbc.version) + arches_list = list(arches) arches_list.sort(utils.arch_compare_sw) arch_list = " ".join(arches_list) - version_list = " ".join(versions.keys()) + version_list = " ".join(versions) if len(version_list) > max_version_len: max_version_len = len(version_list) if len(arch_list) > max_arch_len: @@ -480,7 +538,7 @@ def process_changes_files(changes_files, type, log): (name, mail) = changedby.split(":", 1) log.write("Changed-By: " + name + " <"+mail+">" + "\n") if sponsor: - log.write("Sponsored-By: " + sponsor + "\n") + log.write("Sponsored-By: " + "@".join(sponsor.split("@")[:2]) + "\n") log.write("Distribution:") for dist in distribution: log.write(" " + dist) @@ -495,13 +553,13 @@ def process_changes_files(changes_files, type, log): log.write("\n") if Cnf.has_key("Queue-Report::Options::New"): + total_count = len(queue.uploads) + source_count = len(per_source_items) direction.append([5,1,"ao"]) entries.sort(lambda x, y: sortfunc(x, y)) # Output for a html file. First table header. then table_footer. # Any line between them is then a |