X-Git-Url: https://git.decadent.org.uk/gitweb/?a=blobdiff_plain;f=jennifer;h=2106c34f11e6894fcb21036d4d3c7ef8d23be4e8;hb=cb2f76d9a2039de8dcc73c7c302043dff5132eee;hp=1f1016bcc285eb8f60d100e070d57adda097ecf1;hpb=cd4e8043f530e38f975d877438bdd2707477487e;p=dak.git diff --git a/jennifer b/jennifer index 1f1016bc..2106c34f 100755 --- a/jennifer +++ b/jennifer @@ -2,7 +2,7 @@ # Checks Debian packages from Incoming # Copyright (C) 2000, 2001, 2002 James Troup -# $Id: jennifer,v 1.16 2002-05-08 11:52:31 troup Exp $ +# $Id: jennifer,v 1.29 2002-12-10 21:48:30 troup Exp $ # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by @@ -29,7 +29,7 @@ ################################################################################ -import FCNTL, errno, fcntl, gzip, os, re, select, shutil, stat, string, sys, time, traceback; +import FCNTL, errno, fcntl, gzip, os, re, select, shutil, stat, sys, time, traceback; import apt_inst, apt_pkg; import db_access, katie, logging, utils; @@ -38,12 +38,14 @@ from types import *; ################################################################################ re_bad_diff = re.compile("^[\-\+][\-\+][\-\+] /dev/null"); -re_is_changes = re.compile (r"(.+?)_(.+?)_(.+?)\.changes$"); +re_is_changes = re.compile(r"(.+?)_(.+?)_(.+?)\.changes$"); +re_valid_version = re.compile(r"^([0-9]+:)?[0-9A-Za-z\.\-\+:]+$"); +re_valid_pkg_name = re.compile(r"^[\dA-Za-z][\dA-Za-z\+\-\.]+$"); ################################################################################ # Globals -jennifer_version = "$Revision: 1.16 $"; +jennifer_version = "$Revision: 1.29 $"; Cnf = None; Options = None; @@ -150,9 +152,9 @@ def get_status_output(cmd, status_read, status_write): if len(r) > 0: more_data.append(fd); if fd == c2pwrite or fd == errin: - output = output + r; + output += r; elif fd == status_read: - status = status + r; + status += r; else: utils.fubar("Unexpected file descriptor [%s] returned from select\n" % (fd)); if not more_data: @@ -160,9 +162,11 @@ def get_status_output(cmd, status_read, status_write): try: os.close(status_write); os.close(status_read); + os.close(c2pread); os.close(c2pwrite); os.close(p2cwrite); os.close(errin); + os.close(errout); except: pass; break; @@ -176,7 +180,7 @@ def Dict(**dict): return dict def reject (str, prefix="Rejected: "): global reject_message; if str: - reject_message = reject_message + prefix + str + "\n"; + reject_message += prefix + str + "\n"; ######################################################################################### @@ -193,21 +197,21 @@ def check_signature (filename): # Process the status-fd output keywords = {}; bad = internal_error = ""; - for line in string.split(status, '\n'): - line = string.strip(line); + for line in status.split('\n'): + line = line.strip(); if line == "": continue; - split = string.split(line); + split = line.split(); if len(split) < 2: - internal_error = internal_error + "gpgv status line is malformed (< 2 atoms) ['%s'].\n" % (line); + internal_error += "gpgv status line is malformed (< 2 atoms) ['%s'].\n" % (line); continue; (gnupg, keyword) = split[:2]; if gnupg != "[GNUPG:]": - internal_error = internal_error + "gpgv status line is malformed (incorrect prefix '%s').\n" % (gnupg); + internal_error += "gpgv status line is malformed (incorrect prefix '%s').\n" % (gnupg); continue; args = split[2:]; if keywords.has_key(keyword) and keyword != "NODATA": - internal_error = internal_error + "found duplicate status token ('%s')." % (keyword); + internal_error += "found duplicate status token ('%s').\n" % (keyword); continue; else: keywords[keyword] = args; @@ -248,7 +252,7 @@ def check_signature (filename): # Next check gpgv exited with a zero return code if exit_status: reject("gpgv failed while checking %s." % (filename)); - if string.strip(status): + if status.strip(): reject(utils.prefix_multi_line_string(status, " [GPG status-fd output:] "), ""); else: reject(utils.prefix_multi_line_string(output, " [GPG output:] "), ""); @@ -279,7 +283,7 @@ def check_signature (filename): for keyword in keywords.keys(): if not known_keywords.has_key(keyword): - reject("found unknown status token '%s' from gpgv with args '%s' in %s." % (keyword, repr(keywords[keyword]), filename)); + reject("found unknown status token '%s' from gpgv with args '%r' in %s." % (keyword, keywords[keyword], filename)); bad = 1; if bad: @@ -309,7 +313,7 @@ def copy_to_holding(filename): try: shutil.copy(filename, dest); except IOError, e: - # In either case (ENOENT or EPERM) we want to remove the + # In either case (ENOENT or EACCES) we want to remove the # O_CREAT | O_EXCLed ghost file, so add the file to the list # of 'in holding' even if it's not the real file. if errno.errorcode[e.errno] == 'ENOENT': @@ -323,7 +327,6 @@ def copy_to_holding(filename): raise; in_holding[base_filename] = ""; - return dest; ################################################################################ @@ -334,7 +337,7 @@ def clean_holding(): os.chdir(Cnf["Dir::Queue::Holding"]); for file in in_holding.keys(): if os.path.exists(file): - if string.find(file, '/') != -1: + if file.find('/') != -1: utils.fubar("WTF? clean_holding() got a file ('%s') with / in it!" % (file)); else: os.unlink(file); @@ -353,7 +356,7 @@ def check_changes(): # Parse the .changes field into a dictionary try: - changes.update(utils.parse_changes(filename, 0)); + changes.update(utils.parse_changes(filename)); except utils.cant_open_exc: reject("can't read changes file '%s'." % (filename)); return 0; @@ -363,7 +366,7 @@ def check_changes(): # Parse the Files field from the .changes into another dictionary try: - files.update(utils.build_file_list(changes, "")); + files.update(utils.build_file_list(changes)); except utils.changes_parse_error_exc, line: reject("error parsing changes file '%s', can't grok: %s." % (filename, line)); except utils.nk_format_exc, format: @@ -382,7 +385,7 @@ def check_changes(): if o != "": del changes[i] changes[i] = {} - for j in string.split(o): + for j in o.split(): changes[i][j] = 1 # Fix the Maintainer: field to be RFC822 compatible @@ -417,30 +420,30 @@ def check_distributions(): "Check and map the Distribution field of a .changes file." # Handle suite mappings - if Cnf.has_key("SuiteMappings"): - for map in Cnf.ValueList("SuiteMappings"): - args = string.split(map); - type = args[0]; - if type == "map": - (source, dest) = args[1:3]; - if changes["distribution"].has_key(source): - del changes["distribution"][source] - changes["distribution"][dest] = 1; + for map in Cnf.ValueList("SuiteMappings"): + args = map.split(); + type = args[0]; + if type == "map" or type == "silent-map": + (source, dest) = args[1:3]; + if changes["distribution"].has_key(source): + del changes["distribution"][source] + changes["distribution"][dest] = 1; + if type != "silent-map": reject("Mapping %s to %s." % (source, dest),""); - elif type == "map-unreleased": - (source, dest) = args[1:3]; - if changes["distribution"].has_key(source): - for arch in changes["architecture"].keys(): - if not Cnf.has_key("Suite::%s::Architectures::%s" % (source, arch)): - reject("Mapping %s to %s for unreleased architecture %s." % (source, dest, arch),""); - del changes["distribution"][source]; - changes["distribution"][dest] = 1; - break; - elif type == "ignore": - suite = args[1]; - if changes["distribution"].has_key(suite): - del changes["distribution"][suite]; - reject("Ignoring %s as a target suite." % (suite), "Warning: "); + elif type == "map-unreleased": + (source, dest) = args[1:3]; + if changes["distribution"].has_key(source): + for arch in changes["architecture"].keys(): + if arch not in Cnf.ValueList("Suite::%s::Architectures" % (source)): + reject("Mapping %s to %s for unreleased architecture %s." % (source, dest, arch),""); + del changes["distribution"][source]; + changes["distribution"][dest] = 1; + break; + elif type == "ignore": + suite = args[1]; + if changes["distribution"].has_key(suite): + del changes["distribution"][suite]; + reject("Ignoring %s as a target suite." % (suite), "Warning: "); # Ensure there is (still) a target distribution if changes["distribution"].keys() == []: @@ -512,20 +515,38 @@ def check_files(): for field in [ "Package", "Architecture", "Version" ]: if control.Find(field) == None: reject("%s: No %s field in control." % (file, field)); + # Can't continue + continue; # Ensure the package name matches the one give in the .changes if not changes["binary"].has_key(control.Find("Package", "")): reject("%s: control file lists name as `%s', which isn't in changes file." % (file, control.Find("Package", ""))); + # Validate the package field + package = control.Find("Package"); + if not re_valid_pkg_name.match(package): + reject("%s: invalid package name '%s'." % (file, package)); + + # Validate the version field + version = control.Find("Version"); + if not re_valid_version.match(version): + reject("%s: invalid version number '%s'." % (file, version)); + # Ensure the architecture of the .deb is one we know about. default_suite = Cnf.get("Dinstall::DefaultSuite", "Unstable") - if not Cnf.has_key("Suite::%s::Architectures::%s" % (default_suite, control.Find("Architecture", ""))): - reject("Unknown architecture '%s'." % (control.Find("Architecture", ""))); + architecture = control.Find("Architecture"); + if architecture not in Cnf.ValueList("Suite::%s::Architectures" % (default_suite)): + reject("Unknown architecture '%s'." % (architecture)); # Ensure the architecture of the .deb is one of the ones # listed in the .changes. - if not changes["architecture"].has_key(control.Find("Architecture", "")): - reject("%s: control file lists arch as `%s', which isn't in changes file." % (file, control.Find("Architecture", ""))); + if not changes["architecture"].has_key(architecture): + reject("%s: control file lists arch as `%s', which isn't in changes file." % (file, architecture)); + + # Sanity-check the Depends field + depends = control.Find("Depends"); + if depends == '': + reject("%s: Depends field is empty." % (file)); # Check the section & priority match those given in the .changes (non-fatal) if control.Find("Section") != None and files[file]["section"] != "" and files[file]["section"] != control.Find("Section"): @@ -533,23 +554,21 @@ def check_files(): if control.Find("Priority") != None and files[file]["priority"] != "" and files[file]["priority"] != control.Find("Priority"): reject("%s control file lists priority as `%s', but changes file has `%s'." % (file, control.Find("Priority", ""), files[file]["priority"]),"Warning: "); - files[file]["package"] = control.Find("Package"); - files[file]["architecture"] = control.Find("Architecture"); - files[file]["version"] = control.Find("Version"); + files[file]["package"] = package; + files[file]["architecture"] = architecture; + files[file]["version"] = version; files[file]["maintainer"] = control.Find("Maintainer", ""); - if file[-5:] == ".udeb": + if file.endswith(".udeb"): files[file]["dbtype"] = "udeb"; - elif file[-4:] == ".deb": + elif file.endswith(".deb"): files[file]["dbtype"] = "deb"; else: reject("%s is neither a .deb or a .udeb." % (file)); - files[file]["source"] = control.Find("Source", ""); - if files[file]["source"] == "": - files[file]["source"] = files[file]["package"]; + files[file]["source"] = control.Find("Source", files[file]["package"]); # Get the source version source = files[file]["source"]; source_version = "" - if string.find(source, "(") != -1: + if source.find("(") != -1: m = utils.re_extract_src_version.match(source) source = m.group(1) source_version = m.group(2) @@ -564,7 +583,7 @@ def check_files(): file_package = m.group(1); if files[file]["package"] != file_package: reject("%s: package part of filename (%s) does not match package name in the %s (%s)." % (file, file_package, files[file]["dbtype"], files[file]["package"])); - epochless_version = utils.re_no_epoch.sub('', control.Find("Version", "")) + epochless_version = utils.re_no_epoch.sub('', control.Find("Version")); # version file_version = m.group(2); if epochless_version != file_version: @@ -592,6 +611,8 @@ def check_files(): files[file]["new"] = 1; elif not os.path.exists(Cnf["Dir::Queue::Accepted"] + '/' + dsc_filename): reject("no source found for %s %s (%s)." % (source_package, source_version, file)); + # Check the version and for file overwrites + reject(Katie.check_binary_against_db(file),""); # Checks for a source package... else: @@ -633,10 +654,17 @@ def check_files(): for suite in changes["distribution"].keys(): # Skip byhand if files[file].has_key("byhand"): - continue + continue; + # Handle component mappings + for map in Cnf.ValueList("ComponentMappings"): + (source, dest) = map.split(); + if files[file]["component"] == source: + files[file]["original component"] = source; + files[file]["component"] = dest; # Ensure the component is valid for the target suite - if Cnf.has_key("Suite:%s::Components" % (suite)) and not Cnf.has_key("Suite::%s::Components::%s" % (suite, files[file]["component"])): + if Cnf.has_key("Suite:%s::Components" % (suite)) and \ + files[file]["component"] not in Cnf.ValueList("Suite::%s::Components" % (suite)): reject("unknown component `%s' for suite `%s'." % (files[file]["component"], suite)); continue @@ -644,9 +672,6 @@ def check_files(): if not Katie.in_override_p(files[file]["package"], files[file]["component"], suite, files[file].get("dbtype",""), file): files[file]["new"] = 1; - if files[file]["type"] == "deb": - reject(Katie.check_binaries_against_db(file, suite),""); - # Validate the component component = files[file]["component"]; component_id = db_access.get_component_id(component); @@ -655,7 +680,7 @@ def check_files(): continue; # Validate the priority - if string.find(files[file]["priority"],'/') != -1: + if files[file]["priority"].find('/') != -1: reject("file '%s' has invalid priority '%s' [contains '/']." % (file, files[file]["priority"])); # Determine the location @@ -675,8 +700,18 @@ def check_files(): files[file]["files id"] = files_id # Check for packages that have moved from one component to another - if files[file]["oldfiles"].has_key(suite) and files[file]["oldfiles"][suite]["name"] != files[file]["component"]: - files[file]["othercomponents"] = files[file]["oldfiles"][suite]["name"]; + q = Katie.projectB.query(""" +SELECT c.name FROM binaries b, bin_associations ba, suite s, location l, + component c, architecture a, files f + WHERE b.package = '%s' AND s.suite_name = '%s' + AND (a.arch_string = '%s' OR a.arch_string = 'all') + AND ba.bin = b.id AND ba.suite = s.id AND b.architecture = a.id + AND f.location = l.id AND l.component = c.id AND b.file = f.id""" + % (files[file]["package"], suite, + files[file]["architecture"])); + ql = q.getresult(); + if ql: + files[file]["othercomponents"] = ql[0][0]; # If the .changes file says it has source, it must have source. if changes["architecture"].has_key("source"): @@ -693,10 +728,14 @@ def check_dsc (): global reprocess; for file in files.keys(): + # The .orig.tar.gz can disappear out from under us is it's a + # duplicate of one in the archive. + if not files.has_key(file): + continue; if files[file]["type"] == "dsc": # Parse the .dsc file try: - dsc.update(utils.parse_changes(file, 1)); + dsc.update(utils.parse_changes(file, dsc_whitespace_rules=1)); except utils.cant_open_exc: # if not -n copy_to_holding() will have done this for us... if Options["No-Action"]: @@ -707,7 +746,7 @@ def check_dsc (): reject("syntax error in .dsc file '%s', line %s." % (file, line)); # Build up the file list of files mentioned by the .dsc try: - dsc_files.update(utils.build_file_list(dsc, 1)); + dsc_files.update(utils.build_file_list(dsc, is_a_dsc=1)); except utils.no_files_exc: reject("no Files: field in .dsc file."); continue; @@ -720,6 +759,12 @@ def check_dsc (): if not dsc.has_key(i): reject("Missing field `%s' in dsc file." % (i)); + # Validate the source and version fields + if dsc.has_key("source") and not re_valid_pkg_name.match(dsc["source"]): + reject("%s: invalid source name '%s'." % (file, dsc["source"])); + if dsc.has_key("version") and not re_valid_version.match(dsc["version"]): + reject("%s: invalid version number '%s'." % (file, dsc["version"])); + # The dpkg maintainer from hell strikes again! Bumping the # version number of the .dsc breaks extraction by stable's # dpkg-source. @@ -789,10 +834,10 @@ def check_urgency (): if changes["architecture"].has_key("source"): if not changes.has_key("urgency"): changes["urgency"] = Cnf["Urgency::Default"]; - if not Cnf.has_key("Urgency::Valid::%s" % changes["urgency"]): + if changes["urgency"] not in Cnf.ValueList("Urgency::Valid"): reject("%s is not a valid urgency; it will be treated as %s by testing." % (changes["urgency"], Cnf["Urgency::Default"]), "Warning: "); changes["urgency"] = Cnf["Urgency::Default"]; - changes["urgency"] = string.lower(changes["urgency"]); + changes["urgency"] = changes["urgency"].lower(); ################################################################################ @@ -801,17 +846,41 @@ def check_md5sums (): try: file_handle = utils.open_file(file); except utils.cant_open_exc: - pass; - else: - if apt_pkg.md5sum(file_handle) != files[file]["md5sum"]: - reject("md5sum check failed for %s." % (file)); - file_handle.close(); + continue; + + # Check md5sum + if apt_pkg.md5sum(file_handle) != files[file]["md5sum"]: + reject("%s: md5sum check failed." % (file)); + file_handle.close(); + # Check size + actual_size = os.stat(file)[stat.ST_SIZE]; + size = int(files[file]["size"]); + if size != actual_size: + reject("%s: actual file size (%s) does not match size (%s) in .changes" + % (file, actual_size, size)); + + for file in dsc_files.keys(): + try: + file_handle = utils.open_file(file); + except utils.cant_open_exc: + continue; + + # Check md5sum + if apt_pkg.md5sum(file_handle) != dsc_files[file]["md5sum"]: + reject("%s: md5sum check failed." % (file)); + file_handle.close(); + # Check size + actual_size = os.stat(file)[stat.ST_SIZE]; + size = int(dsc_files[file]["size"]); + if size != actual_size: + reject("%s: actual file size (%s) does not match size (%s) in .dsc" + % (file, actual_size, size)); ################################################################################ # Sanity check the time stamps of files inside debs. # [Files in the near future cause ugly warnings and extreme time -# travel can causes errors on extraction] +# travel can cause errors on extraction] def check_timestamps(): class Tar: @@ -910,7 +979,7 @@ def action (): if Options["No-Action"] or Options["Automatic"]: answer = 'S' - if string.find(reject_message, "Rejected") != -1: + if reject_message.find("Rejected") != -1: if upload_too_new(): print "SKIP (too new)\n" + reject_message,; prompt = "[S]kip, Quit ?"; @@ -920,7 +989,7 @@ def action (): if Options["Automatic"]: answer = 'R'; elif new: - print "NEW to %s\n%s%s" % (string.join(changes["distribution"].keys(), ", "), reject_message, summary),; + print "NEW to %s\n%s%s" % (", ".join(changes["distribution"].keys()), reject_message, summary),; prompt = "[N]ew, Skip, Quit ?"; if Options["Automatic"]: answer = 'N'; @@ -935,12 +1004,12 @@ def action (): if Options["Automatic"]: answer = 'A'; - while string.find(prompt, answer) == -1: + while prompt.find(answer) == -1: answer = utils.our_raw_input(prompt); m = katie.re_default_answer.match(prompt); if answer == "": answer = m.group(1); - answer = string.upper(answer[:1]); + answer = answer[:1].upper(); if answer == 'R': os.chdir (pkg.directory); @@ -1062,9 +1131,9 @@ def process_it (changes_file): while reprocess: check_distributions(); check_files(); - check_md5sums(); check_dsc(); check_diff(); + check_md5sums(); check_urgency(); check_timestamps(); Katie.update_subst(reject_message); @@ -1099,7 +1168,7 @@ def main(): # Ensure all the arguments we were given are .changes files for file in changes_files: - if file[-8:] != ".changes": + if not file.endswith(".changes"): utils.warn("Ignoring '%s' because it's not a .changes file." % (file)); changes_files.remove(file);