X-Git-Url: https://git.decadent.org.uk/gitweb/?a=blobdiff_plain;f=dak%2Fimport_archive.py;h=028239b0d4b3b374107ab52550e40e3ce1a02237;hb=b43ed3ff3738940ce46caa836d88b6937a76582c;hp=13343c2f5aa93954c94d091dad425e5afca881ef;hpb=7aaaad3135c9164390af5897925660842368660b;p=dak.git diff --git a/dak/import_archive.py b/dak/import_archive.py index 13343c2f..028239b0 100755 --- a/dak/import_archive.py +++ b/dak/import_archive.py @@ -1,6 +1,6 @@ #!/usr/bin/env python -# Populate the DB +""" Populate the DB """ # Copyright (C) 2000, 2001, 2002, 2003, 2004, 2006 James Troup # This program is free software; you can redistribute it and/or modify @@ -36,14 +36,13 @@ ############################################################################### -import commands, os, pg, re, sys, time +import commands, os, pg, sys, time import apt_pkg -import dak.lib.database as database -import dak.lib.utils as utils - -############################################################################### - -re_arch_from_filename = re.compile(r"binary-[^/]+") +from daklib import database +from daklib import utils +from daklib.dak_exceptions import * +from daklib.regexes import re_arch_from_filename, re_taint_free, re_no_epoch, \ + re_extract_src_version ############################################################################### @@ -93,13 +92,13 @@ def reject (str, prefix="Rejected: "): ############################################################################### def check_signature (filename): - if not utils.re_taint_free.match(os.path.basename(filename)): + if not re_taint_free.match(os.path.basename(filename)): reject("!!WARNING!! tainted filename: '%s'." % (filename)) return None status_read, status_write = os.pipe() - cmd = "gpgv --status-fd %s --keyring %s --keyring %s %s" \ - % (status_write, Cnf["Dinstall::PGPKeyring"], Cnf["Dinstall::GPGKeyring"], filename) + cmd = "gpgv --status-fd %s %s %s" \ + % (status_write, utils.gpg_keyring_args(), filename) (output, status, exit_status) = utils.gpgv_get_status_output(cmd, status_read, status_write) # Process the status-fd output @@ -234,13 +233,10 @@ def update_locations (): SubSec = Cnf.SubTree("Location::%s" % (location)) archive_id = database.get_archive_id(SubSec["archive"]) type = SubSec.Find("type") - if type == "legacy-mixed": - projectB.query("INSERT INTO location (path, archive, type) VALUES ('%s', %d, '%s')" % (location, archive_id, SubSec["type"])) - else: - for component in Cnf.SubTree("Component").List(): - component_id = database.get_component_id(component) - projectB.query("INSERT INTO location (path, component, archive, type) VALUES ('%s', %d, %d, '%s')" % - (location, component_id, archive_id, SubSec["type"])) + for component in Cnf.SubTree("Component").List(): + component_id = database.get_component_id(component) + projectB.query("INSERT INTO location (path, component, archive, type) VALUES ('%s', %d, %d, '%s')" % + (location, component_id, archive_id, SubSec["type"])) def update_architectures (): projectB.query("DELETE FROM architecture") @@ -255,7 +251,7 @@ def update_suites (): for i in ("Version", "Origin", "Description"): if SubSec.has_key(i): projectB.query("UPDATE suite SET %s = '%s' WHERE suite_name = '%s'" % (i.lower(), SubSec[i], suite.lower())) - for architecture in Cnf.ValueList("Suite::%s::Architectures" % (suite)): + for architecture in database.get_suite_architectures(suite): architecture_id = database.get_architecture_id (architecture) projectB.query("INSERT INTO suite_architectures (suite, architecture) VALUES (currval('suite_id_seq'), %d)" % (architecture_id)) @@ -280,7 +276,6 @@ def update_section(): prefix = "" else: prefix = "" - component = component.replace("non-US/", "") if component != 'main': suffix = '/' + component else: @@ -307,7 +302,7 @@ def get_location_path(directory): def get_or_set_files_id (filename, size, md5sum, location_id): global files_id_cache, files_id_serial, files_query_cache - cache_key = "~".join((filename, size, md5sum, repr(location_id))) + cache_key = "_".join((filename, size, md5sum, repr(location_id))) if not files_id_cache.has_key(cache_key): files_id_serial += 1 files_query_cache.write("%d\t%s\t%s\t%s\t%d\t\\N\n" % (files_id_serial, filename, size, md5sum, location_id)) @@ -324,7 +319,7 @@ def process_sources (filename, suite, component, archive): suite_id = database.get_suite_id(suite) try: file = utils.open_file (filename) - except utils.cant_open_exc: + except CantOpenError: utils.warn("can't open '%s'" % (filename)) return Scanner = apt_pkg.ParseTagFile(file) @@ -332,12 +327,12 @@ def process_sources (filename, suite, component, archive): package = Scanner.Section["package"] version = Scanner.Section["version"] directory = Scanner.Section["directory"] - dsc_file = os.path.join(Cnf["Dir::Root"], directory, "%s_%s.dsc" % (package, utils.re_no_epoch.sub('', version))) + dsc_file = os.path.join(Cnf["Dir::Root"], directory, "%s_%s.dsc" % (package, re_no_epoch.sub('', version))) # Sometimes the Directory path is a lie; check in the pool if not os.path.exists(dsc_file): if directory.split('/')[0] == "dists": directory = Cnf["Dir::PoolRoot"] + utils.poolify(package, component) - dsc_file = os.path.join(Cnf["Dir::Root"], directory, "%s_%s.dsc" % (package, utils.re_no_epoch.sub('', version))) + dsc_file = os.path.join(Cnf["Dir::Root"], directory, "%s_%s.dsc" % (package, re_no_epoch.sub('', version))) if not os.path.exists(dsc_file): utils.fubar("%s not found." % (dsc_file)) install_date = time.strftime("%Y-%m-%d", time.localtime(os.path.getmtime(dsc_file))) @@ -355,7 +350,7 @@ def process_sources (filename, suite, component, archive): directory = poolify (directory, location) if directory != "" and not directory.endswith("/"): directory += '/' - no_epoch_version = utils.re_no_epoch.sub('', version) + no_epoch_version = re_no_epoch.sub('', version) # Add all files referenced by the .dsc to the files table ids = [] for line in Scanner.Section["files"].split('\n'): @@ -363,7 +358,7 @@ def process_sources (filename, suite, component, archive): (md5sum, size, filename) = line.strip().split() # Don't duplicate .orig.tar.gz's if filename.endswith(".orig.tar.gz"): - cache_key = "%s~%s~%s" % (filename, size, md5sum) + cache_key = "%s_%s_%s" % (filename, size, md5sum) if orig_tar_gz_cache.has_key(cache_key): id = orig_tar_gz_cache[cache_key] else: @@ -376,9 +371,9 @@ def process_sources (filename, suite, component, archive): if filename.endswith(".dsc"): files_id = id filename = directory + package + '_' + no_epoch_version + '.dsc' - cache_key = "%s~%s" % (package, version) + cache_key = "%s_%s" % (package, version) if not source_cache.has_key(cache_key): - nasty_key = "%s~%s" % (package, version) + nasty_key = "%s_%s" % (package, version) source_id_serial += 1 if not source_cache_for_binaries.has_key(nasty_key): source_cache_for_binaries[nasty_key] = source_id_serial @@ -407,7 +402,7 @@ def process_packages (filename, suite, component, archive): suite_id = database.get_suite_id(suite) try: file = utils.open_file (filename) - except utils.cant_open_exc: + except CantOpenError: utils.warn("can't open '%s'" % (filename)) return Scanner = apt_pkg.ParseTagFile(file) @@ -427,27 +422,30 @@ def process_packages (filename, suite, component, archive): source = Scanner.Section["source"] source_version = "" if source.find("(") != -1: - m = utils.re_extract_src_version.match(source) + m = re_extract_src_version.match(source) source = m.group(1) source_version = m.group(2) if not source_version: source_version = version filename = Scanner.Section["filename"] + if filename.endswith(".deb"): + type = "deb" + else: + type = "udeb" location = get_location_path(filename.split('/')[0]) - location_id = database.get_location_id (location, component, archive) + location_id = database.get_location_id (location, component.replace("/debian-installer", ""), archive) filename = poolify (filename, location) if architecture == "all": filename = re_arch_from_filename.sub("binary-all", filename) - cache_key = "%s~%s" % (source, source_version) + cache_key = "%s_%s" % (source, source_version) source_id = source_cache_for_binaries.get(cache_key, None) size = Scanner.Section["size"] md5sum = Scanner.Section["md5sum"] files_id = get_or_set_files_id (filename, size, md5sum, location_id) - type = "deb"; # FIXME - cache_key = "%s~%s~%s~%d~%d~%d~%d" % (package, version, repr(source_id), architecture_id, location_id, files_id, suite_id) + cache_key = "%s_%s_%s_%d_%d_%d_%d" % (package, version, repr(source_id), architecture_id, location_id, files_id, suite_id) if not arch_all_cache.has_key(cache_key): arch_all_cache[cache_key] = 1 - cache_key = "%s~%s~%s~%d" % (package, version, repr(source_id), architecture_id) + cache_key = "%s_%s_%s_%d" % (package, version, repr(source_id), architecture_id) if not binary_cache.has_key(cache_key): if not source_id: source_id = "\N" @@ -474,7 +472,7 @@ def process_packages (filename, suite, component, archive): ############################################################################### def do_sources(sources, suite, component, server): - temp_filename = utils.temp_filename() + (fd, temp_filename) = utils.temp_filename() (result, output) = commands.getstatusoutput("gunzip -c %s > %s" % (sources, temp_filename)) if (result != 0): utils.fubar("Gunzip invocation failed!\n%s" % (output), result) @@ -491,14 +489,14 @@ def do_da_do_da (): Arguments = [('a', "action", "Import-Archive::Options::Action"), ('h', "help", "Import-Archive::Options::Help")] for i in [ "action", "help" ]: - if not Cnf.has_key("Import-Archive::Options::%s" % (i)): - Cnf["Import-Archive::Options::%s" % (i)] = "" + if not Cnf.has_key("Import-Archive::Options::%s" % (i)): + Cnf["Import-Archive::Options::%s" % (i)] = "" apt_pkg.ParseCommandLine(Cnf, Arguments, sys.argv) Options = Cnf.SubTree("Import-Archive::Options") if Options["Help"]: - usage() + usage() if not Options["Action"]: utils.warn("""no -a/--action given; not doing anything. @@ -541,11 +539,7 @@ Please read the documentation before running this script. SubSec = Cnf.SubTree("Location::%s" % (location)) server = SubSec["Archive"] type = Cnf.Find("Location::%s::Type" % (location)) - if type == "legacy-mixed": - sources = location + 'Sources.gz' - suite = Cnf.Find("Location::%s::Suite" % (location)) - do_sources(sources, suite, "", server) - elif type == "legacy" or type == "pool": + if type == "pool": for suite in Cnf.ValueList("Location::%s::Suites" % (location)): for component in Cnf.SubTree("Component").List(): sources = Cnf["Dir::Root"] + "dists/" + Cnf["Suite::%s::CodeName" % (suite)] + '/' + component + '/source/' + 'Sources.gz' @@ -559,16 +553,12 @@ Please read the documentation before running this script. SubSec = Cnf.SubTree("Location::%s" % (location)) server = SubSec["Archive"] type = Cnf.Find("Location::%s::Type" % (location)) - if type == "legacy-mixed": - packages = location + 'Packages' - suite = Cnf.Find("Location::%s::Suite" % (location)) - print 'Processing '+location+'...' - process_packages (packages, suite, "", server) - elif type == "legacy" or type == "pool": + if type == "pool": for suite in Cnf.ValueList("Location::%s::Suites" % (location)): - for component in Cnf.SubTree("Component").List(): - architectures = filter(utils.real_arch, - Cnf.ValueList("Suite::%s::Architectures" % (suite))) + udeb_components = map(lambda x: x+"/debian-installer", + Cnf.ValueList("Suite::%s::UdebComponents" % suite)) + for component in Cnf.SubTree("Component").List() + udeb_components: + architectures = filter(utils.real_arch, database.get_suite_architectures(suite)) for architecture in architectures: packages = Cnf["Dir::Root"] + "dists/" + Cnf["Suite::%s::CodeName" % (suite)] + '/' + component + '/binary-' + architecture + '/Packages' print 'Processing '+packages+'...'