from copy import copy
import glob, os, stat, sys, time
import apt_pkg
+try:
+ import rrdtool
+except ImportError:
+ pass
from daklib import utils
from daklib.queue import Upload
-from daklib.dbconn import DBConn, has_new_comment, DBChange, DBSource, get_uid_from_fingerprint
+from daklib.dbconn import DBConn, has_new_comment, DBChange, DBSource, \
+ get_uid_from_fingerprint, get_policy_queue
from daklib.textutils import fix_maintainer
from daklib.dak_exceptions import *
-import debianbts
Cnf = None
direction = []
def usage(exit_code=0):
print """Usage: dak queue-report
-Prints a report of packages in queue directories (usually new and byhand).
+Prints a report of packages in queues (usually new and byhand).
-h, --help show this help and exit.
-8, --822 writes 822 formated output to the location set in dak.conf
-s, --sort=key sort output according to key, see below.
-a, --age=key if using sort by age, how should time be treated?
If not given a default of hours will be used.
+ -r, --rrd=key Directory where rrd files to be updated are stored
-d, --directories=key A comma seperated list of queues to be scanned
Sorting Keys: ao=age, oldest first. an=age, newest first.
def footer():
print "<p class=\"timestamp\">Timestamp: %s (UTC)</p>" % (time.strftime("%d.%m.%Y / %H:%M:%S", time.gmtime()))
+ print "<p class=\"timestamp\">There are <a href=\"/stat.html\">graphs about the queues</a> available.</p>"
print """
<div class="footer">
print "</tbody></table>"
-def find_bug_severities(bugnrs):
- """Take a list of bug numbers and
- return a dictionary with bug number - severity pairs."""
- statuses = None
- try: statuses = debianbts.get_status(bugnrs)
- except Exception:
- return None
- severities = {}
- for bug_status in statuses:
- severities[bug_status.bug_num] = bug_status.severity
- return severities
-
-def rcbug_close_check(closes):
- """Take a list of bug numbers and determine if at least one of them
- is release critical."""
- severity_list = find_bug_severities(closes)
- if severity_list:
- severities = severity_list.values()
- return ('critical' in severities) or \
- ('serious' in severities) or \
- ('grave' in severities)
- return False
-
-def table_row(source, version, arch, last_mod, maint, distribution, closes, fingerprint, sponsor, changedby, fixes_rcbug):
+def table_row(source, version, arch, last_mod, maint, distribution, closes, fingerprint, sponsor, changedby):
global row_number
if dist == "experimental":
trclass = "exp"
- if fixes_rcbug:
- trclass = "rcfix"
-
if not len(session.query(DBSource).filter_by(source = source).all()):
trclass += " binNEW"
session.commit()
try:
(login, domain) = sponsor.split("@", 1)
print "<span class=\"sponsor\">Sponsor: <a href=\"http://qa.debian.org/developer.php?login=%s\">%s</a></span>@debian.org<br/>" % (utils.html_escape(login), utils.html_escape(login))
- except Exception, e:
+ except Exception as e:
pass
print "<span class=\"signature\">Fingerprint: %s</span>" % (fingerprint)
############################################################
-def process_changes_files(changes_files, type, log):
- session = DBConn().session()
+def update_graph_database(rrd_dir, type, n_source, n_binary):
+ if not rrd_dir:
+ return
+
+ rrd_file = os.path.join(rrd_dir, type.lower()+'.rrd')
+ update = [rrd_file, "N:%s:%s" % (n_source, n_binary)]
+
+ try:
+ rrdtool.update(*update)
+ except rrdtool.error:
+ create = [rrd_file]+"""
+--step
+300
+--start
+0
+DS:ds0:GAUGE:7200:0:1000
+DS:ds1:GAUGE:7200:0:1000
+RRA:AVERAGE:0.5:1:599
+RRA:AVERAGE:0.5:6:700
+RRA:AVERAGE:0.5:24:775
+RRA:AVERAGE:0.5:288:795
+RRA:MAX:0.5:1:600
+RRA:MAX:0.5:6:700
+RRA:MAX:0.5:24:775
+RRA:MAX:0.5:288:795
+""".strip().split("\n")
+ try:
+ rc = rrdtool.create(*create)
+ ru = rrdtool.update(*update)
+ except rrdtool.error as e:
+ print('warning: queue_report: rrdtool error, skipping %s.rrd: %s' % (type, e))
+ except NameError:
+ pass
+
+############################################################
+
+def process_changes_files(changes_files, type, log, rrd_dir):
msg = ""
cache = {}
# Read in all the .changes files
u.load_changes(filename)
cache[filename] = copy(u.pkg.changes)
cache[filename]["filename"] = filename
- except Exception, e:
+ except Exception as e:
print "WARNING: Exception %s" % e
continue
# Divide the .changes into per-source groups
per_source_items = per_source.items()
per_source_items.sort(sg_compare)
+ update_graph_database(rrd_dir, type, len(per_source_items), len(changes_files))
+
entries = []
max_source_len = 0
max_version_len = 0
for j in i[1]["list"]:
changesbase = os.path.basename(j["filename"])
try:
+ session = DBConn().session()
dbc = session.query(DBChange).filter_by(changesname=changesbase).one()
- except Exception, e:
+ session.close()
+ except Exception as e:
print "Can't find changes file in NEW for %s (%s)" % (changesbase, e)
dbc = None
(maintainer["maintainer822"], maintainer["maintainer2047"],
maintainer["maintainername"], maintainer["maintaineremail"]) = \
fix_maintainer (j["maintainer"])
- except ParseMaintError, msg:
+ except ParseMaintError as msg:
print "Problems while parsing maintainer address\n"
maintainer["maintainername"] = "Unknown"
maintainer["maintaineremail"] = "Unknown"
(changeby["changedby822"], changeby["changedby2047"],
changeby["changedbyname"], changeby["changedbyemail"]) = \
fix_maintainer (j["changed-by"])
- except ParseMaintError, msg:
+ except ParseMaintError as msg:
(changeby["changedby822"], changeby["changedby2047"],
changeby["changedbyname"], changeby["changedbyemail"]) = \
("", "", "", "")
note = " | [N]"
else:
note = ""
- fixes_rcbug = rcbug_close_check(closes)
- entries.append([source, binary, version_list, arch_list, note, last_modified, maint, distribution, closes, fingerprint, sponsor, changedby, filename, fixes_rcbug])
+ entries.append([source, binary, version_list, arch_list, note, last_modified, maint, distribution, closes, fingerprint, sponsor, changedby, filename])
# direction entry consists of "Which field, which direction, time-consider" where
# time-consider says how we should treat last_modified. Thats all.
if Cnf.has_key("Queue-Report::Options::822"):
# print stuff out in 822 format
for entry in entries:
- (source, binary, version_list, arch_list, note, last_modified, maint, distribution, closes, fingerprint, sponsor, changedby, changes_file, fixes_rcbug) = entry
+ (source, binary, version_list, arch_list, note, last_modified, maint, distribution, closes, fingerprint, sponsor, changedby, changes_file) = entry
# We'll always have Source, Version, Arch, Mantainer, and Dist
# For the rest, check to see if we have them, then print them out
source_count = len(per_source_items)
table_header(type.upper(), source_count, total_count)
for entry in entries:
- (source, binary, version_list, arch_list, note, last_modified, maint, distribution, closes, fingerprint, sponsor, changedby, undef, fixes_rcbug) = entry
- table_row(source, version_list, arch_list, time_pp(last_modified), maint, distribution, closes, fingerprint, sponsor, changedby, entries, fixes_rcbug)
+ (source, binary, version_list, arch_list, note, last_modified, maint, distribution, closes, fingerprint, sponsor, changedby, undef) = entry
+ table_row(source, version_list, arch_list, time_pp(last_modified), maint, distribution, closes, fingerprint, sponsor, changedby)
table_footer(type.upper())
elif not Cnf.has_key("Queue-Report::Options::822"):
# The "normal" output without any formatting.
msg = ""
for entry in entries:
- (source, binary, version_list, arch_list, note, last_modified, undef, undef, undef, undef, undef, undef, undef, fixes_rcbug) = entry
+ (source, binary, version_list, arch_list, note, last_modified, undef, undef, undef, undef, undef, undef, undef) = entry
msg += format % (source, version_list, arch_list, note, time_pp(last_modified))
if msg:
('8','822',"Queue-Report::Options::822"),
('s',"sort","Queue-Report::Options::Sort", "HasArg"),
('a',"age","Queue-Report::Options::Age", "HasArg"),
+ ('r',"rrd","Queue-Report::Options::Rrd", "HasArg"),
('d',"directories","Queue-Report::Options::Directories", "HasArg")]
for i in [ "help" ]:
if not Cnf.has_key("Queue-Report::Options::%s" % (i)):
# Initialize db so we can get the NEW comments
dbconn = DBConn()
- directories = [ ]
+ queue_names = [ ]
if Cnf.has_key("Queue-Report::Options::Directories"):
for i in Cnf["Queue-Report::Options::Directories"].split(","):
- directories.append(i)
+ queue_names.append(i)
elif Cnf.has_key("Queue-Report::Directories"):
- directories = Cnf.ValueList("Queue-Report::Directories")
+ queue_names = Cnf.ValueList("Queue-Report::Directories")
+ else:
+ queue_names = [ "byhand", "new" ]
+
+ if Cnf.has_key("Queue-Report::Options::Rrd"):
+ rrd_dir = Cnf["Queue-Report::Options::Rrd"]
+ elif Cnf.has_key("Dir::Rrd"):
+ rrd_dir = Cnf["Dir::Rrd"]
else:
- directories = [ "byhand", "new" ]
+ rrd_dir = None
f = None
if Cnf.has_key("Queue-Report::Options::822"):
# Open the report file
f = open(Cnf["Queue-Report::ReportLocations::822Location"], "w")
- for directory in directories:
- changes_files = glob.glob("%s/*.changes" % (Cnf["Dir::Queue::%s" % (directory)]))
- process_changes_files(changes_files, directory, f)
+ session = dbconn.session()
+
+ for queue_name in queue_names:
+ queue = get_policy_queue(queue_name, session)
+ if queue:
+ directory = os.path.abspath(queue.path)
+ changes_files = glob.glob("%s/*.changes" % (directory))
+ process_changes_files(changes_files, os.path.basename(directory), f, rrd_dir)
+ else:
+ utils.warn("Cannot find queue %s" % queue_name)
if Cnf.has_key("Queue-Report::Options::822"):
f.close()