X-Git-Url: https://git.decadent.org.uk/gitweb/?p=dak.git;a=blobdiff_plain;f=dak%2Fqueue_report.py;h=a4b997af90e17e26cf242057c3a658d33437e281;hp=84f6a22f23946139349199b27afa46f3597a7f52;hb=519c1dbf89c13557afc15a429164616ac563d379;hpb=e15117cd260692b56d3b44f1415fe13f8f1482c1 diff --git a/dak/queue_report.py b/dak/queue_report.py index 84f6a22f..a4b997af 100755 --- a/dak/queue_report.py +++ b/dak/queue_report.py @@ -37,11 +37,17 @@ from copy import copy import glob, os, stat, sys, time import apt_pkg +try: + import rrdtool +except ImportError: + pass from daklib import utils -from daklib.queue import Upload -from daklib.dbconn import DBConn, has_new_comment, DBChange, DBSource, get_uid_from_fingerprint +from daklib.dbconn import DBConn, DBSource, has_new_comment, PolicyQueue, \ + get_uid_from_fingerprint +from daklib.policy import PolicyQueueUploadHandler from daklib.textutils import fix_maintainer +from daklib.utils import get_logins_from_ldap from daklib.dak_exceptions import * Cnf = None @@ -52,7 +58,7 @@ row_number = 0 def usage(exit_code=0): print """Usage: dak queue-report -Prints a report of packages in queue directories (usually new and byhand). +Prints a report of packages in queues (usually new and byhand). -h, --help show this help and exit. -8, --822 writes 822 formated output to the location set in dak.conf @@ -60,6 +66,7 @@ Prints a report of packages in queue directories (usually new and byhand). -s, --sort=key sort output according to key, see below. -a, --age=key if using sort by age, how should time be treated? If not given a default of hours will be used. + -r, --rrd=key Directory where rrd files to be updated are stored -d, --directories=key A comma seperated list of queues to be scanned Sorting Keys: ao=age, oldest first. an=age, newest first. @@ -110,7 +117,15 @@ def time_pp(x): def sg_compare (a, b): a = a[1] b = b[1] - """Sort by have note, time of oldest upload.""" + """Sort by have pending action, have note, time of oldest upload.""" + # Sort by have pending action + a_note_state = a["processed"] + b_note_state = b["processed"] + if a_note_state < b_note_state: + return -1 + elif a_note_state > b_note_state: + return 1 + # Sort by have note a_note_state = a["note_state"] b_note_state = b["note_state"] @@ -168,11 +183,12 @@ def header(): - + Debian NEW and BYHAND Packages
- corner image - corner image - corner image - corner image Debian NEW and BYHAND Packages @@ -217,31 +234,24 @@ def header(): def footer(): print "

Timestamp: %s (UTC)

" % (time.strftime("%d.%m.%Y / %H:%M:%S", time.gmtime())) + print "

There are graphs about the queues available.

" print """ """ def table_header(type, source_count, total_count): - print "

Summary for: %s

" % (type) - print "

Summary for: binary-%s only

" % (type) + print "

Summary for: %s

" % (type) + print "

Summary for: binary-%s only

" % (type) print """ +

Click to toggle all/binary-NEW packages

-

Click to toggle all/binary-NEW packages

- " % (trclass) - print "" % (source) + if "sourceNEW" in trclass: + print "" % (source) + else: + print "" % {'source': source} print "" print "" % (arch) print "" print "" row_number+=1 ############################################################ -def process_changes_files(changes_files, type, log): - session = DBConn().session() - msg = "" - cache = {} - # Read in all the .changes files - for filename in changes_files: +def update_graph_database(rrd_dir, type, n_source, n_binary): + if not rrd_dir: + return + + rrd_file = os.path.join(rrd_dir, type.lower()+'.rrd') + update = [rrd_file, "N:%s:%s" % (n_source, n_binary)] + + try: + rrdtool.update(*update) + except rrdtool.error: + create = [rrd_file]+""" +--step +300 +--start +0 +DS:ds0:GAUGE:7200:0:1000 +DS:ds1:GAUGE:7200:0:1000 +RRA:AVERAGE:0.5:1:599 +RRA:AVERAGE:0.5:6:700 +RRA:AVERAGE:0.5:24:775 +RRA:AVERAGE:0.5:288:795 +RRA:MAX:0.5:1:600 +RRA:MAX:0.5:6:700 +RRA:MAX:0.5:24:775 +RRA:MAX:0.5:288:795 +""".strip().split("\n") try: - u = Upload() - u.load_changes(filename) - cache[filename] = copy(u.pkg.changes) - cache[filename]["filename"] = filename - except Exception, e: - print "WARNING: Exception %s" % e - continue + rc = rrdtool.create(*create) + ru = rrdtool.update(*update) + except rrdtool.error as e: + print('warning: queue_report: rrdtool error, skipping %s.rrd: %s' % (type, e)) + except NameError: + pass + +############################################################ + +def process_queue(queue, log, rrd_dir): + msg = "" + type = queue.queue_name + session = DBConn().session() + # Divide the .changes into per-source groups per_source = {} - for filename in cache.keys(): - source = cache[filename]["source"] - if not per_source.has_key(source): + total_pending = 0 + for upload in queue.uploads: + source = upload.changes.source + if source not in per_source: per_source[source] = {} per_source[source]["list"] = [] - per_source[source]["list"].append(cache[filename]) + per_source[source]["processed"] = "" + handler = PolicyQueueUploadHandler(upload, session) + if handler.get_action(): + per_source[source]["processed"] = "PENDING %s" % handler.get_action() + total_pending += 1 + per_source[source]["list"].append(upload) + per_source[source]["list"].sort(lambda x, y: cmp(x.changes.created, y.changes.created), reverse=True) # Determine oldest time and have note status for each source group for source in per_source.keys(): source_list = per_source[source]["list"] first = source_list[0] - oldest = os.stat(first["filename"])[stat.ST_MTIME] + oldest = time.mktime(first.changes.created.timetuple()) have_note = 0 for d in per_source[source]["list"]: - mtime = os.stat(d["filename"])[stat.ST_MTIME] + mtime = time.mktime(d.changes.created.timetuple()) if Cnf.has_key("Queue-Report::Options::New"): if mtime > oldest: oldest = mtime else: if mtime < oldest: oldest = mtime - have_note += has_new_comment(d["source"], d["version"]) + have_note += has_new_comment(d.policy_queue, d.changes.source, d.changes.version) per_source[source]["oldest"] = oldest if not have_note: per_source[source]["note_state"] = 0; # none @@ -366,10 +413,16 @@ def process_changes_files(changes_files, type, log): per_source_items = per_source.items() per_source_items.sort(sg_compare) + update_graph_database(rrd_dir, type, len(per_source_items), len(queue.uploads)) + entries = [] max_source_len = 0 max_version_len = 0 max_arch_len = 0 + try: + logins = get_logins_from_ldap() + except: + logins = dict() for i in per_source_items: maintainer = {} maint="" @@ -379,29 +432,25 @@ def process_changes_files(changes_files, type, log): changeby = {} changedby="" sponsor="" - filename=i[1]["list"][0]["filename"] + filename=i[1]["list"][0].changes.changesname last_modified = time.time()-i[1]["oldest"] - source = i[1]["list"][0]["source"] + source = i[1]["list"][0].changes.source if len(source) > max_source_len: max_source_len = len(source) - binary_list = i[1]["list"][0]["binary"].keys() - binary = ', '.join(binary_list) - arches = {} - versions = {} + binary_list = i[1]["list"][0].binaries + binary = ', '.join([ b.package for b in binary_list ]) + arches = set() + versions = set() for j in i[1]["list"]: - changesbase = os.path.basename(j["filename"]) - try: - dbc = session.query(DBChange).filter_by(changesname=changesbase).one() - except Exception, e: - print "Can't find changes file in NEW for %s (%s)" % (changesbase, e) - dbc = None + dbc = j.changes + changesbase = dbc.changesname if Cnf.has_key("Queue-Report::Options::New") or Cnf.has_key("Queue-Report::Options::822"): try: (maintainer["maintainer822"], maintainer["maintainer2047"], maintainer["maintainername"], maintainer["maintaineremail"]) = \ - fix_maintainer (j["maintainer"]) - except ParseMaintError, msg: + fix_maintainer (dbc.maintainer) + except ParseMaintError as msg: print "Problems while parsing maintainer address\n" maintainer["maintainername"] = "Unknown" maintainer["maintaineremail"] = "Unknown" @@ -410,31 +459,35 @@ def process_changes_files(changes_files, type, log): try: (changeby["changedby822"], changeby["changedby2047"], changeby["changedbyname"], changeby["changedbyemail"]) = \ - fix_maintainer (j["changed-by"]) - except ParseMaintError, msg: + fix_maintainer (dbc.changedby) + except ParseMaintError as msg: (changeby["changedby822"], changeby["changedby2047"], changeby["changedbyname"], changeby["changedbyemail"]) = \ ("", "", "", "") changedby="%s:%s" % (changeby["changedbyname"], changeby["changedbyemail"]) - distribution=j["distribution"].keys() - closes=j["closes"].keys() - if dbc: - fingerprint = dbc.fingerprint - sponsor_name = get_uid_from_fingerprint(fingerprint).name - sponsor_email = get_uid_from_fingerprint(fingerprint).uid + "@debian.org" - if sponsor_name != maintainer["maintainername"] and sponsor_name != changeby["changedbyname"] and \ - sponsor_email != maintainer["maintaineremail"] and sponsor_name != changeby["changedbyemail"]: - sponsor = sponsor_email - - for arch in j["architecture"].keys(): - arches[arch] = "" - version = j["version"] - versions[version] = "" - arches_list = arches.keys() + distribution=dbc.distribution.split() + closes=dbc.closes + + fingerprint = dbc.fingerprint + sponsor_name = get_uid_from_fingerprint(fingerprint).name + sponsor_login = get_uid_from_fingerprint(fingerprint).uid + if '@' in sponsor_login: + if fingerprint in logins: + sponsor_login = logins[fingerprint] + if (sponsor_name != maintainer["maintainername"] and + sponsor_name != changeby["changedbyname"] and + sponsor_login + '@debian.org' != maintainer["maintaineremail"] and + sponsor_name != changeby["changedbyemail"]): + sponsor = sponsor_login + + for arch in dbc.architecture.split(): + arches.add(arch) + versions.add(dbc.version) + arches_list = list(arches) arches_list.sort(utils.arch_compare_sw) arch_list = " ".join(arches_list) - version_list = " ".join(versions.keys()) + version_list = " ".join(sorted(versions, reverse=True)) if len(version_list) > max_version_len: max_version_len = len(version_list) if len(arch_list) > max_arch_len: @@ -443,7 +496,7 @@ def process_changes_files(changes_files, type, log): note = " | [N]" else: note = "" - entries.append([source, binary, version_list, arch_list, note, last_modified, maint, distribution, closes, fingerprint, sponsor, changedby, filename]) + entries.append([source, binary, version_list, arch_list, per_source[source]["processed"], note, last_modified, maint, distribution, closes, fingerprint, sponsor, changedby, filename]) # direction entry consists of "Which field, which direction, time-consider" where # time-consider says how we should treat last_modified. Thats all. @@ -454,16 +507,16 @@ def process_changes_files(changes_files, type, log): age = Cnf["Queue-Report::Options::Age"] if Cnf.has_key("Queue-Report::Options::New"): # If we produce html we always have oldest first. - direction.append([5,-1,"ao"]) + direction.append([6,-1,"ao"]) else: if Cnf.has_key("Queue-Report::Options::Sort"): for i in Cnf["Queue-Report::Options::Sort"].split(","): if i == "ao": # Age, oldest first. - direction.append([5,-1,age]) + direction.append([6,-1,age]) elif i == "an": # Age, newest first. - direction.append([5,1,age]) + direction.append([6,1,age]) elif i == "na": # Name, Ascending. direction.append([0,1,0]) @@ -472,10 +525,10 @@ def process_changes_files(changes_files, type, log): direction.append([0,-1,0]) elif i == "nl": # Notes last. - direction.append([4,1,0]) + direction.append([5,1,0]) elif i == "nf": # Notes first. - direction.append([4,-1,0]) + direction.append([5,-1,0]) entries.sort(lambda x, y: sortfunc(x, y)) # Yes, in theory you can add several sort options at the commandline with. But my mind is to small # at the moment to come up with a real good sorting function that considers all the sidesteps you @@ -485,7 +538,7 @@ def process_changes_files(changes_files, type, log): if Cnf.has_key("Queue-Report::Options::822"): # print stuff out in 822 format for entry in entries: - (source, binary, version_list, arch_list, note, last_modified, maint, distribution, closes, fingerprint, sponsor, changedby, changes_file) = entry + (source, binary, version_list, arch_list, processed, note, last_modified, maint, distribution, closes, fingerprint, sponsor, changedby, changes_file) = entry # We'll always have Source, Version, Arch, Mantainer, and Dist # For the rest, check to see if we have them, then print them out @@ -504,7 +557,7 @@ def process_changes_files(changes_files, type, log): (name, mail) = changedby.split(":", 1) log.write("Changed-By: " + name + " <"+mail+">" + "\n") if sponsor: - log.write("Sponsored-By: " + sponsor + "\n") + log.write("Sponsored-By: %s@debian.org\n" % sponsor) log.write("Distribution:") for dist in distribution: log.write(" " + dist) @@ -518,39 +571,43 @@ def process_changes_files(changes_files, type, log): log.write("Changes-File: " + os.path.basename(changes_file) + "\n") log.write("\n") + total_count = len(queue.uploads) + source_count = len(per_source_items) + if Cnf.has_key("Queue-Report::Options::New"): - direction.append([5,1,"ao"]) + direction.append([6,1,"ao"]) entries.sort(lambda x, y: sortfunc(x, y)) # Output for a html file. First table header. then table_footer. # Any line between them is then a printed from subroutine table_row. if len(entries) > 0: - total_count = len(changes_files) - source_count = len(per_source_items) table_header(type.upper(), source_count, total_count) for entry in entries: - (source, binary, version_list, arch_list, note, last_modified, maint, distribution, closes, fingerprint, sponsor, changedby, undef) = entry + (source, binary, version_list, arch_list, processed, note, last_modified, maint, distribution, closes, fingerprint, sponsor, changedby, undef) = entry table_row(source, version_list, arch_list, time_pp(last_modified), maint, distribution, closes, fingerprint, sponsor, changedby) table_footer(type.upper()) elif not Cnf.has_key("Queue-Report::Options::822"): # The "normal" output without any formatting. - format="%%-%ds | %%-%ds | %%-%ds%%s | %%s old\n" % (max_source_len, max_version_len, max_arch_len) - msg = "" for entry in entries: - (source, binary, version_list, arch_list, note, last_modified, undef, undef, undef, undef, undef, undef, undef) = entry - msg += format % (source, version_list, arch_list, note, time_pp(last_modified)) + (source, binary, version_list, arch_list, processed, note, last_modified, undef, undef, undef, undef, undef, undef, undef) = entry + if processed: + format="%%-%ds | %%-%ds | %%-%ds | %%s\n" % (max_source_len, max_version_len, max_arch_len) + msg += format % (source, version_list, arch_list, processed) + else: + format="%%-%ds | %%-%ds | %%-%ds%%s | %%s old\n" % (max_source_len, max_version_len, max_arch_len) + msg += format % (source, version_list, arch_list, note, time_pp(last_modified)) if msg: - total_count = len(changes_files) - source_count = len(per_source_items) print type.upper() print "-"*len(type) print print msg - print "%s %s source package%s / %s %s package%s in total." % (source_count, type, plural(source_count), total_count, type, plural(total_count)) + print ("%s %s source package%s / %s %s package%s in total / %s %s package%s to be processed." % + (source_count, type, plural(source_count), + total_count, type, plural(total_count), + total_pending, type, plural(total_pending))) print - ################################################################################ def main(): @@ -562,41 +619,54 @@ def main(): ('8','822',"Queue-Report::Options::822"), ('s',"sort","Queue-Report::Options::Sort", "HasArg"), ('a',"age","Queue-Report::Options::Age", "HasArg"), + ('r',"rrd","Queue-Report::Options::Rrd", "HasArg"), ('d',"directories","Queue-Report::Options::Directories", "HasArg")] for i in [ "help" ]: if not Cnf.has_key("Queue-Report::Options::%s" % (i)): Cnf["Queue-Report::Options::%s" % (i)] = "" - apt_pkg.ParseCommandLine(Cnf, Arguments, sys.argv) + apt_pkg.parse_commandline(Cnf, Arguments, sys.argv) - Options = Cnf.SubTree("Queue-Report::Options") + Options = Cnf.subtree("Queue-Report::Options") if Options["Help"]: usage() if Cnf.has_key("Queue-Report::Options::New"): header() - # Initialize db so we can get the NEW comments - dbconn = DBConn() - - directories = [ ] + queue_names = [] if Cnf.has_key("Queue-Report::Options::Directories"): for i in Cnf["Queue-Report::Options::Directories"].split(","): - directories.append(i) + queue_names.append(i) elif Cnf.has_key("Queue-Report::Directories"): - directories = Cnf.ValueList("Queue-Report::Directories") + queue_names = Cnf.value_list("Queue-Report::Directories") + else: + queue_names = [ "byhand", "new" ] + + if Cnf.has_key("Queue-Report::Options::Rrd"): + rrd_dir = Cnf["Queue-Report::Options::Rrd"] + elif Cnf.has_key("Dir::Rrd"): + rrd_dir = Cnf["Dir::Rrd"] else: - directories = [ "byhand", "new" ] + rrd_dir = None f = None if Cnf.has_key("Queue-Report::Options::822"): # Open the report file - f = open(Cnf["Queue-Report::ReportLocations::822Location"], "w") + f = sys.stdout + filename822 = Cnf.get("Queue-Report::ReportLocations::822Location") + if filename822: + f = open(filename822, "w") - for directory in directories: - changes_files = glob.glob("%s/*.changes" % (Cnf["Dir::Queue::%s" % (directory)])) - process_changes_files(changes_files, directory, f) + session = DBConn().session() + + for queue_name in queue_names: + queue = session.query(PolicyQueue).filter_by(queue_name=queue_name).first() + if queue is not None: + process_queue(queue, f, rrd_dir) + else: + utils.warn("Cannot find queue %s" % queue_name) if Cnf.has_key("Queue-Report::Options::822"): f.close()
+ """ print "Package count in %s: %s |  Total Package count: %s" % (type, source_count, total_count) print """ @@ -274,8 +284,12 @@ def table_row(source, version, arch, last_mod, maint, distribution, closes, fing if dist == "experimental": trclass = "exp" - if not len(session.query(DBSource).filter_by(source = source).all()): - trclass += " binNEW" + query = '''SELECT source + FROM source_suite + WHERE source = :source + AND suite_name IN ('unstable', 'experimental')''' + if not session.execute(query, {'source': source}).rowcount: + trclass += " sourceNEW" session.commit() if row_number % 2 != 0: @@ -283,10 +297,13 @@ def table_row(source, version, arch, last_mod, maint, distribution, closes, fing else: print "
%s%s%(source)s" for vers in version.split(): - print "%s
" % (source, utils.html_escape(vers), utils.html_escape(vers)) + print "%s
" % (source, utils.html_escape(vers), utils.html_escape(vers)) print "
%s" @@ -297,65 +314,95 @@ def table_row(source, version, arch, last_mod, maint, distribution, closes, fing (name, mail) = maint.split(":", 1) print "" - print "Maintainer: %s
" % (utils.html_escape(mail), utils.html_escape(name)) + print "Maintainer: %s
" % (utils.html_escape(mail), utils.html_escape(name)) (name, mail) = changedby.split(":", 1) - print "Changed-By: %s
" % (utils.html_escape(mail), utils.html_escape(name)) + print "Changed-By: %s
" % (utils.html_escape(mail), utils.html_escape(name)) if sponsor: - try: - (login, domain) = sponsor.split("@", 1) - print "Sponsor: %s@debian.org
" % (utils.html_escape(login), utils.html_escape(login)) - except Exception, e: - pass + print "Sponsor: %s@debian.org
" % (utils.html_escape(sponsor), utils.html_escape(sponsor)) print "Fingerprint: %s" % (fingerprint) print "
" for close in closes: - print "#%s
" % (utils.html_escape(close), utils.html_escape(close)) + print "#%s
" % (utils.html_escape(close), utils.html_escape(close)) print "