]> git.decadent.org.uk Git - dak.git/blobdiff - daklib/dbconn.py
Merge branch 'master' into deb-src-3.0-sqla
[dak.git] / daklib / dbconn.py
index f64a1fec144ee914a6a2a7de697d333505156736..98ebcaa1ca43aaaa065cf8f5ce350776caf8a1eb 100755 (executable)
@@ -37,610 +37,2371 @@ import os
 import psycopg2
 import traceback
 
+from inspect import getargspec
+
 from sqlalchemy import create_engine, Table, MetaData, select
-from sqlalchemy.orm import sessionmaker
+from sqlalchemy.orm import sessionmaker, mapper, relation
 
-from singleton import Singleton
+# Don't remove this, we re-export the exceptions to scripts which import us
+from sqlalchemy.exc import *
+from sqlalchemy.orm.exc import NoResultFound
+
+# Only import Config until Queue stuff is changed to store its config
+# in the database
 from config import Config
+from singleton import Singleton
+from textutils import fix_maintainer
 
 ################################################################################
 
-class Cache(object):
-    def __init__(self, hashfunc=None):
-        if hashfunc:
-            self.hashfunc = hashfunc
-        else:
-            self.hashfunc = lambda x: str(x)
+__all__ = ['IntegrityError', 'SQLAlchemyError']
+
+################################################################################
+
+def session_wrapper(fn):
+    def wrapped(*args, **kwargs):
+        private_transaction = False
+        session = kwargs.get('session')
 
-        self.data = {}
+        # No session specified as last argument or in kwargs, create one.
+        if session is None and len(args) <= len(getargspec(fn)[0]) - 1:
+            private_transaction = True
+            kwargs['session'] = DBConn().session()
 
-    def SetValue(self, keys, value):
-        self.data[self.hashfunc(keys)] = value
+        try:
+            return fn(*args, **kwargs)
+        finally:
+            if private_transaction:
+                # We created a session; close it.
+                kwargs['session'].close()
+
+    wrapped.__doc__ = fn.__doc__
+    wrapped.func_name = fn.func_name
 
-    def GetValue(self, keys):
-        return self.data.get(self.hashfunc(keys))
+    return wrapped
 
 ################################################################################
 
-class DBConn(Singleton):
+class Architecture(object):
+    def __init__(self, *args, **kwargs):
+        pass
+
+    def __eq__(self, val):
+        if isinstance(val, str):
+            return (self.arch_string== val)
+        # This signals to use the normal comparison operator
+        return NotImplemented
+
+    def __ne__(self, val):
+        if isinstance(val, str):
+            return (self.arch_string != val)
+        # This signals to use the normal comparison operator
+        return NotImplemented
+
+    def __repr__(self):
+        return '<Architecture %s>' % self.arch_string
+
+__all__.append('Architecture')
+
+@session_wrapper
+def get_architecture(architecture, session=None):
     """
-    database module init.
+    Returns database id for given C{architecture}.
+
+    @type architecture: string
+    @param architecture: The name of the architecture
+
+    @type session: Session
+    @param session: Optional SQLA session object (a temporary one will be
+    generated if not supplied)
+
+    @rtype: Architecture
+    @return: Architecture object for the given arch (None if not present)
+    """
+
+    q = session.query(Architecture).filter_by(arch_string=architecture)
+
+    try:
+        return q.one()
+    except NoResultFound:
+        return None
+
+__all__.append('get_architecture')
+
+@session_wrapper
+def get_architecture_suites(architecture, session=None):
+    """
+    Returns list of Suite objects for given C{architecture} name
+
+    @type source: str
+    @param source: Architecture name to search for
+
+    @type session: Session
+    @param session: Optional SQL session object (a temporary one will be
+    generated if not supplied)
+
+    @rtype: list
+    @return: list of Suite objects for the given name (may be empty)
     """
+
+    q = session.query(Suite)
+    q = q.join(SuiteArchitecture)
+    q = q.join(Architecture).filter_by(arch_string=architecture).order_by('suite_name')
+
+    ret = q.all()
+
+    return ret
+
+__all__.append('get_architecture_suites')
+
+################################################################################
+
+class Archive(object):
     def __init__(self, *args, **kwargs):
-        super(DBConn, self).__init__(*args, **kwargs)
+        pass
 
-    def _startup(self, *args, **kwargs):
-        self.__createconn()
-        self.__init_caches()
+    def __repr__(self):
+        return '<Archive %s>' % self.archive_name
 
-    def __setuptables(self):
-        self.tbl_architecture = Table('architecture', self.db_meta, autoload=True)
-        self.tbl_archive = Table('archive', self.db_meta, autoload=True)
-        self.tbl_bin_associations = Table('bin_associations', self.db_meta, autoload=True)
-        self.tbl_binaries = Table('binaries', self.db_meta, autoload=True)
-        self.tbl_component = Table('component', self.db_meta, autoload=True)
-        self.tbl_config = Table('config', self.db_meta, autoload=True)
-        self.tbl_content_associations = Table('content_associations', self.db_meta, autoload=True)
-        self.tbl_content_file_names = Table('content_file_names', self.db_meta, autoload=True)
-        self.tbl_content_file_paths = Table('content_file_paths', self.db_meta, autoload=True)
-        self.tbl_dsc_files = Table('dsc_files', self.db_meta, autoload=True)
-        self.tbl_files = Table('files', self.db_meta, autoload=True)
-        self.tbl_fingerprint = Table('fingerprint', self.db_meta, autoload=True)
-        self.tbl_keyrings = Table('keyrings', self.db_meta, autoload=True)
-        self.tbl_location = Table('location', self.db_meta, autoload=True)
-        self.tbl_maintainer = Table('maintainer', self.db_meta, autoload=True)
-        self.tbl_override = Table('override', self.db_meta, autoload=True)
-        self.tbl_override_type = Table('override_type', self.db_meta, autoload=True)
-        self.tbl_pending_content_associations = Table('pending_content_associations', self.db_meta, autoload=True)
-        self.tbl_priority = Table('priority', self.db_meta, autoload=True)
-        self.tbl_queue = Table('queue', self.db_meta, autoload=True)
-        self.tbl_queue_build = Table('queue_build', self.db_meta, autoload=True)
-        self.tbl_section = Table('section', self.db_meta, autoload=True)
-        self.tbl_source = Table('source', self.db_meta, autoload=True)
-        self.tbl_src_associations = Table('src_associations', self.db_meta, autoload=True)
-        self.tbl_src_uploaders = Table('src_uploaders', self.db_meta, autoload=True)
-        self.tbl_suite = Table('suite', self.db_meta, autoload=True)
-        self.tbl_suite_architectures = Table('suite_architectures', self.db_meta, autoload=True)
-        self.tbl_uid = Table('uid', self.db_meta, autoload=True)
+__all__.append('Archive')
 
-    ## Connection functions
-    def __createconn(self):
-        cnf = Config()
-        if cnf["DB::Host"]:
-            # TCP/IP
-            connstr = "postgres://%s" % cnf["DB::Host"]
-            if cnf["DB::Port"] and cnf["DB::Port"] != "-1":
-                connstr += ":%s" % cnf["DB::Port"]
-            connstr += "/%s" % cnf["DB::Name"]
-        else:
-            # Unix Socket
-            connstr = "postgres:///%s" % cnf["DB::Name"]
-            if cnf["DB::Port"] and cnf["DB::Port"] != "-1":
-                connstr += "?port=%s" % cnf["DB::Port"]
+@session_wrapper
+def get_archive(archive, session=None):
+    """
+    returns database id for given c{archive}.
 
-        self.db_pg   = create_engine(connstr)
-        self.db_meta = MetaData()
-        self.db_meta.bind = self.db_pg
-        self.db_smaker = sessionmaker(bind=self.db_pg,
-                                      autoflush=True,
-                                      transactional=True)
+    @type archive: string
+    @param archive: the name of the arhive
 
-        self.__setuptables()
+    @type session: Session
+    @param session: Optional SQLA session object (a temporary one will be
+    generated if not supplied)
 
-    def session(self):
-        return self.db_smaker()
+    @rtype: Archive
+    @return: Archive object for the given name (None if not present)
 
-    ## Cache functions
-    def __init_caches(self):
-        self.caches = {'suite':         Cache(),
-                       'section':       Cache(),
-                       'priority':      Cache(),
-                       'override_type': Cache(),
-                       'architecture':  Cache(),
-                       'archive':       Cache(),
-                       'component':     Cache(),
-                       'content_path_names':     Cache(),
-                       'content_file_names':     Cache(),
-                       'location':      Cache(lambda x: '%s_%s_%s' % (x['location'], x['component'], x['location'])),
-                       'maintainer':    {}, # TODO
-                       'keyring':       {}, # TODO
-                       'source':        Cache(lambda x: '%s_%s_' % (x['source'], x['version'])),
-                       'files':         Cache(lambda x: '%s_%s_' % (x['filename'], x['location'])),
-                       'maintainer':    {}, # TODO
-                       'fingerprint':   {}, # TODO
-                       'queue':         {}, # TODO
-                       'uid':           {}, # TODO
-                       'suite_version': Cache(lambda x: '%s_%s' % (x['source'], x['suite'])),
-                      }
-
-        self.prepared_statements = {}
-
-    def prepare(self,name,statement):
-        if not self.prepared_statements.has_key(name):
-            pgc.execute(statement)
-            self.prepared_statements[name] = statement
-
-    def clear_caches(self):
-        self.__init_caches()
-
-    ## Get functions
-    def __get_id(self, retfield, selectobj, cachekey, cachename=None):
-        # This is a bit of a hack but it's an internal function only
-        if cachename is not None:
-            res = self.caches[cachename].GetValue(cachekey)
-            if res:
-                return res
-
-        c = selectobj.execute()
-
-        if c.rowcount != 1:
-            return None
-
-        res = c.fetchone()
-
-        if retfield not in res.keys():
-            return None
-
-        res = res[retfield]
-
-        if cachename is not None:
-            self.caches[cachename].SetValue(cachekey, res)
-
-        return res
-
-    def get_suite_id(self, suite):
-        """
-        Returns database id for given C{suite}.
-        Results are kept in a cache during runtime to minimize database queries.
+    """
+    archive = archive.lower()
 
-        @type suite: string
-        @param suite: The name of the suite
+    q = session.query(Archive).filter_by(archive_name=archive)
 
-        @rtype: int
-        @return: the database id for the given suite
+    try:
+        return q.one()
+    except NoResultFound:
+        return None
 
-        """
-        return int(self.__get_id('id',
-                                 self.tbl_suite.select(self.tbl_suite.columns.suite_name == suite),
-                                 suite,
-                                 'suite'))
+__all__.append('get_archive')
 
-    def get_section_id(self, section):
-        """
-        Returns database id for given C{section}.
-        Results are kept in a cache during runtime to minimize database queries.
+################################################################################
 
-        @type section: string
-        @param section: The name of the section
+class BinAssociation(object):
+    def __init__(self, *args, **kwargs):
+        pass
 
-        @rtype: int
-        @return: the database id for the given section
+    def __repr__(self):
+        return '<BinAssociation %s (%s, %s)>' % (self.ba_id, self.binary, self.suite)
 
-        """
-        return self.__get_id('id',
-                             self.tbl_section.select(self.tbl_section.columns.section == section),
-                             section,
-                             'section')
+__all__.append('BinAssociation')
 
-    def get_priority_id(self, priority):
-        """
-        Returns database id for given C{priority}.
-        Results are kept in a cache during runtime to minimize database queries.
+################################################################################
 
-        @type priority: string
-        @param priority: The name of the priority
+class DBBinary(object):
+    def __init__(self, *args, **kwargs):
+        pass
 
-        @rtype: int
-        @return: the database id for the given priority
+    def __repr__(self):
+        return '<DBBinary %s (%s, %s)>' % (self.package, self.version, self.architecture)
 
-        """
-        return self.__get_id('id',
-                             self.tbl_priority.select(self.tbl_priority.columns.priority == priority),
-                             priority,
-                             'priority')
+__all__.append('DBBinary')
 
-    def get_override_type_id(self, override_type):
-        """
-        Returns database id for given override C{type}.
-        Results are kept in a cache during runtime to minimize database queries.
+@session_wrapper
+def get_suites_binary_in(package, session=None):
+    """
+    Returns list of Suite objects which given C{package} name is in
 
-        @type override_type: string
-        @param override_type: The name of the override type
+    @type source: str
+    @param source: DBBinary package name to search for
 
-        @rtype: int
-        @return: the database id for the given override type
+    @rtype: list
+    @return: list of Suite objects for the given package
+    """
 
-        """
-        return self.__get_id('id',
-                             self.tbl_override_type.select(self.tbl_override_type.columns.type == override_type),
-                             override_type,
-                             'override_type')
+    return session.query(Suite).join(BinAssociation).join(DBBinary).filter_by(package=package).all()
 
-    def get_architecture_id(self, architecture):
-        """
-        Returns database id for given C{architecture}.
-        Results are kept in a cache during runtime to minimize database queries.
+__all__.append('get_suites_binary_in')
 
-        @type architecture: string
-        @param architecture: The name of the override type
+@session_wrapper
+def get_binary_from_id(id, session=None):
+    """
+    Returns DBBinary object for given C{id}
 
-        @rtype: int
-        @return: the database id for the given architecture
+    @type id: int
+    @param id: Id of the required binary
 
-        """
-        return self.__get_id('id',
-                             self.tbl_architecture.select(self.tbl_architecture.columns.arch_string == architecture),
-                             architecture,
-                             'architecture')
+    @type session: Session
+    @param session: Optional SQLA session object (a temporary one will be
+    generated if not supplied)
 
-    def get_archive_id(self, archive):
-        """
-        returns database id for given c{archive}.
-        results are kept in a cache during runtime to minimize database queries.
+    @rtype: DBBinary
+    @return: DBBinary object for the given binary (None if not present)
+    """
 
-        @type archive: string
-        @param archive: the name of the override type
+    q = session.query(DBBinary).filter_by(binary_id=id)
 
-        @rtype: int
-        @return: the database id for the given archive
+    try:
+        return q.one()
+    except NoResultFound:
+        return None
 
-        """
-        archive = archive.lower()
-        return self.__get_id('id',
-                             self.tbl_archive.select(self.tbl_archive.columns.name == archive),
-                             archive,
-                             'archive')
+__all__.append('get_binary_from_id')
 
-    def get_component_id(self, component):
-        """
-        Returns database id for given C{component}.
-        Results are kept in a cache during runtime to minimize database queries.
+@session_wrapper
+def get_binaries_from_name(package, version=None, architecture=None, session=None):
+    """
+    Returns list of DBBinary objects for given C{package} name
 
-        @type component: string
-        @param component: The name of the override type
+    @type package: str
+    @param package: DBBinary package name to search for
 
-        @rtype: int
-        @return: the database id for the given component
+    @type version: str or None
+    @param version: Version to search for (or None)
 
-        """
-        component = component.lower()
-        return self.__get_id('id',
-                             self.tbl_component.select(self.tbl_component.columns.name == component),
-                             component.lower(),
-                             'component')
+    @type package: str, list or None
+    @param package: Architectures to limit to (or None if no limit)
 
-    def get_location_id(self, location, component, archive):
-        """
-        Returns database id for the location behind the given combination of
-          - B{location} - the path of the location, eg. I{/srv/ftp.debian.org/ftp/pool/}
-          - B{component} - the id of the component as returned by L{get_component_id}
-          - B{archive} - the id of the archive as returned by L{get_archive_id}
-        Results are kept in a cache during runtime to minimize database queries.
+    @type session: Session
+    @param session: Optional SQL session object (a temporary one will be
+    generated if not supplied)
 
-        @type location: string
-        @param location: the path of the location
+    @rtype: list
+    @return: list of DBBinary objects for the given name (may be empty)
+    """
 
-        @type component: string
-        @param component: the name of the component
+    q = session.query(DBBinary).filter_by(package=package)
 
-        @type archive: string
-        @param archive: the name of the archive
+    if version is not None:
+        q = q.filter_by(version=version)
 
-        @rtype: int
-        @return: the database id for the location
+    if architecture is not None:
+        if not isinstance(architecture, list):
+            architecture = [architecture]
+        q = q.join(Architecture).filter(Architecture.arch_string.in_(architecture))
 
-        """
+    ret = q.all()
 
-        archive = archive.lower()
-        component = component.lower()
+    return ret
 
-        values = {'archive': archive, 'location': location, 'component': component}
+__all__.append('get_binaries_from_name')
 
-        s = self.tbl_location.join(self.tbl_archive).join(self.tbl_component)
+@session_wrapper
+def get_binaries_from_source_id(source_id, session=None):
+    """
+    Returns list of DBBinary objects for given C{source_id}
 
-        s = s.select(self.tbl_location.columns.path == location)
-        s = s.where(self.tbl_archive.columns.name == archive)
-        s = s.where(self.tbl_component.columns.name == component)
+    @type source_id: int
+    @param source_id: source_id to search for
 
-        return self.__get_id('location.id', s, values, 'location')
+    @type session: Session
+    @param session: Optional SQL session object (a temporary one will be
+    generated if not supplied)
 
-    def get_source_id(self, source, version):
-        """
-        Returns database id for the combination of C{source} and C{version}
-          - B{source} - source package name, eg. I{mailfilter}, I{bbdb}, I{glibc}
-          - B{version}
-        Results are kept in a cache during runtime to minimize database queries.
+    @rtype: list
+    @return: list of DBBinary objects for the given name (may be empty)
+    """
 
-        @type source: string
-        @param source: source package name
+    return session.query(DBBinary).filter_by(source_id=source_id).all()
 
-        @type version: string
-        @param version: the source version
+__all__.append('get_binaries_from_source_id')
 
-        @rtype: int
-        @return: the database id for the source
+@session_wrapper
+def get_binary_from_name_suite(package, suitename, session=None):
+    ### For dak examine-package
+    ### XXX: Doesn't use object API yet
 
-        """
-        s = self.tbl_source.select()
-        s = s.where(self.tbl_source.columns.source  == source)
-        s = s.where(self.tbl_source.columns.version == version)
-
-        return self.__get_id('id', s, {'source': source, 'version': version}, 'source')
-
-    def get_suite(self, suite):
-        if isinstance(suite, str):
-            suite_id = self.get_suite_id(suite.lower())
-        elif type(suite) == int:
-            suite_id = suite
-
-        s = self.tbl_suite.select(self.tbl_suite.columns.id == suite_id)
-        c = s.execute()
-        if c.rowcount < 1:
-            return None
-        else:
-            return c.fetchone()
+    sql = """SELECT DISTINCT(b.package), b.version, c.name, su.suite_name
+             FROM binaries b, files fi, location l, component c, bin_associations ba, suite su
+             WHERE b.package=:package
+               AND b.file = fi.id
+               AND fi.location = l.id
+               AND l.component = c.id
+               AND ba.bin=b.id
+               AND ba.suite = su.id
+               AND su.suite_name=:suitename
+          ORDER BY b.version DESC"""
 
-    def get_suite_version(self, source, suite):
-        """
-        Returns database id for a combination of C{source} and C{suite}.
+    return session.execute(sql, {'package': package, 'suitename': suitename})
 
-          - B{source} - source package name, eg. I{mailfilter}, I{bbdb}, I{glibc}
-          - B{suite} - a suite name, eg. I{unstable}
+__all__.append('get_binary_from_name_suite')
 
-        Results are kept in a cache during runtime to minimize database queries.
+@session_wrapper
+def get_binary_components(package, suitename, arch, session=None):
+    # Check for packages that have moved from one component to another
+    query = """SELECT c.name FROM binaries b, bin_associations ba, suite s, location l, component c, architecture a, files f
+    WHERE b.package=:package AND s.suite_name=:suitename
+      AND (a.arch_string = :arch OR a.arch_string = 'all')
+      AND ba.bin = b.id AND ba.suite = s.id AND b.architecture = a.id
+      AND f.location = l.id
+      AND l.component = c.id
+      AND b.file = f.id"""
 
-        @type source: string
-        @param source: source package name
+    vals = {'package': package, 'suitename': suitename, 'arch': arch}
 
-        @type suite: string
-        @param suite: the suite name
+    return session.execute(query, vals)
 
-        @rtype: string
-        @return: the version for I{source} in I{suite}
+__all__.append('get_binary_components')
 
-        """
-        s = select([self.tbl_source.columns.source, self.tbl_source.columns.version])
-#        s = self.tbl_source.join(self.tbl_src_associations).join(self.tbl_suite)
+################################################################################
 
-        s = s.select(self.tbl_suite.columns.suite_name == suite, use_labels=True)
-        s = s.select(self.tbl_source.columns.source == source)
+class Component(object):
+    def __init__(self, *args, **kwargs):
+        pass
 
-        return self.__get_id('source.version', s, {'suite': suite, 'source': source}, 'suite_version')
+    def __eq__(self, val):
+        if isinstance(val, str):
+            return (self.component_name == val)
+        # This signals to use the normal comparison operator
+        return NotImplemented
 
+    def __ne__(self, val):
+        if isinstance(val, str):
+            return (self.component_name != val)
+        # This signals to use the normal comparison operator
+        return NotImplemented
 
-    def get_files_id (self, filename, size, md5sum, location_id):
-        """
-        Returns -1, -2 or the file_id for filename, if its C{size} and C{md5sum} match an
-        existing copy.
+    def __repr__(self):
+        return '<Component %s>' % self.component_name
 
-        The database is queried using the C{filename} and C{location_id}. If a file does exist
-        at that location, the existing size and md5sum are checked against the provided
-        parameters. A size or checksum mismatch returns -2. If more than one entry is
-        found within the database, a -1 is returned, no result returns None, otherwise
-        the file id.
 
-        Results are kept in a cache during runtime to minimize database queries.
+__all__.append('Component')
 
-        @type filename: string
-        @param filename: the filename of the file to check against the DB
+@session_wrapper
+def get_component(component, session=None):
+    """
+    Returns database id for given C{component}.
 
-        @type size: int
-        @param size: the size of the file to check against the DB
+    @type component: string
+    @param component: The name of the override type
 
-        @type md5sum: string
-        @param md5sum: the md5sum of the file to check against the DB
+    @rtype: int
+    @return: the database id for the given component
 
-        @type location_id: int
-        @param location_id: the id of the location as returned by L{get_location_id}
+    """
+    component = component.lower()
 
-        @rtype: int / None
-        @return: Various return values are possible:
-                   - -2: size/checksum error
-                   - -1: more than one file found in database
-                   - None: no file found in database
-                   - int: file id
+    q = session.query(Component).filter_by(component_name=component)
 
-        """
-        values = {'filename' : filename,
-                  'location' : location_id}
+    try:
+        return q.one()
+    except NoResultFound:
+        return None
 
-        res = self.caches['files'].GetValue( values )
+__all__.append('get_component')
 
-        if not res:
-            query = """SELECT id, size, md5sum
-                       FROM files
-                       WHERE filename = %(filename)s AND location = %(location)s"""
+################################################################################
 
-            cursor = self.db_con.cursor()
-            cursor.execute( query, values )
+class DBConfig(object):
+    def __init__(self, *args, **kwargs):
+        pass
 
-            if cursor.rowcount == 0:
-                res = None
+    def __repr__(self):
+        return '<DBConfig %s>' % self.name
 
-            elif cursor.rowcount != 1:
-                res = -1
+__all__.append('DBConfig')
 
-            else:
-                row = cursor.fetchone()
+################################################################################
 
-                if row[1] != int(size) or row[2] != md5sum:
-                    res =  -2
+class ContentFilename(object):
+    def __init__(self, *args, **kwargs):
+        pass
 
-                else:
-                    self.caches['files'].SetValue(values, row[0])
-                    res = row[0]
+    def __repr__(self):
+        return '<ContentFilename %s>' % self.filename
 
-        return res
+__all__.append('ContentFilename')
 
+def get_or_set_contents_file_id(filename, session=None):
+    """
+    Returns database id for given filename.
 
-    def get_or_set_contents_file_id(self, filename):
-        """
-        Returns database id for given filename.
+    If no matching file is found, a row is inserted.
 
-        Results are kept in a cache during runtime to minimize database queries.
-        If no matching file is found, a row is inserted.
+    @type filename: string
+    @param filename: The filename
+    @type session: SQLAlchemy
+    @param session: Optional SQL session object (a temporary one will be
+    generated if not supplied).  If not passed, a commit will be performed at
+    the end of the function, otherwise the caller is responsible for commiting.
 
-        @type filename: string
-        @param filename: The filename
+    @rtype: int
+    @return: the database id for the given component
+    """
+    privatetrans = False
+    if session is None:
+        session = DBConn().session()
+        privatetrans = True
+
+    q = session.query(ContentFilename).filter_by(filename=filename)
+
+    try:
+        ret = q.one().cafilename_id
+    except NoResultFound:
+        cf = ContentFilename()
+        cf.filename = filename
+        session.add(cf)
+        if privatetrans:
+            session.commit()
+        else:
+            session.flush()
+        ret = cf.cafilename_id
 
-        @rtype: int
-        @return: the database id for the given component
-        """
-        try:
-            values={'value': filename}
-            query = "SELECT id FROM content_file_names WHERE file = %(value)s"
-            id = self.__get_single_id(query, values, cachename='content_file_names')
-            if not id:
-                c = self.db_con.cursor()
-                c.execute( "INSERT INTO content_file_names VALUES (DEFAULT, %(value)s) RETURNING id",
-                           values )
-
-                id = c.fetchone()[0]
-                self.caches['content_file_names'].SetValue(values, id)
-
-            return id
-        except:
-            traceback.print_exc()
-            raise
-
-    def get_or_set_contents_path_id(self, path):
-        """
-        Returns database id for given path.
+    if privatetrans:
+        session.close()
 
-        Results are kept in a cache during runtime to minimize database queries.
-        If no matching file is found, a row is inserted.
+    return ret
 
-        @type path: string
-        @param path: The filename
+__all__.append('get_or_set_contents_file_id')
 
-        @rtype: int
-        @return: the database id for the given component
-        """
-        try:
-            values={'value': path}
-            query = "SELECT id FROM content_file_paths WHERE path = %(value)s"
-            id = self.__get_single_id(query, values, cachename='content_path_names')
-            if not id:
-                c = self.db_con.cursor()
-                c.execute( "INSERT INTO content_file_paths VALUES (DEFAULT, %(value)s) RETURNING id",
-                           values )
-
-                id = c.fetchone()[0]
-                self.caches['content_path_names'].SetValue(values, id)
-
-            return id
-        except:
-            traceback.print_exc()
-            raise
-
-    def get_suite_architectures(self, suite):
-        """
-        Returns list of architectures for C{suite}.
+@session_wrapper
+def get_contents(suite, overridetype, section=None, session=None):
+    """
+    Returns contents for a suite / overridetype combination, limiting
+    to a section if not None.
 
-        @type suite: string, int
-        @param suite: the suite name or the suite_id
+    @type suite: Suite
+    @param suite: Suite object
 
-        @rtype: list
-        @return: the list of architectures for I{suite}
-        """
+    @type overridetype: OverrideType
+    @param overridetype: OverrideType object
+
+    @type section: Section
+    @param section: Optional section object to limit results to
+
+    @type session: SQLAlchemy
+    @param session: Optional SQL session object (a temporary one will be
+    generated if not supplied)
+
+    @rtype: ResultsProxy
+    @return: ResultsProxy object set up to return tuples of (filename, section,
+    package, arch_id)
+    """
+
+    # find me all of the contents for a given suite
+    contents_q = """SELECT (p.path||'/'||n.file) AS fn,
+                            s.section,
+                            b.package,
+                            b.architecture
+                   FROM content_associations c join content_file_paths p ON (c.filepath=p.id)
+                   JOIN content_file_names n ON (c.filename=n.id)
+                   JOIN binaries b ON (b.id=c.binary_pkg)
+                   JOIN override o ON (o.package=b.package)
+                   JOIN section s ON (s.id=o.section)
+                   WHERE o.suite = :suiteid AND o.type = :overridetypeid
+                   AND b.type=:overridetypename"""
+
+    vals = {'suiteid': suite.suite_id,
+            'overridetypeid': overridetype.overridetype_id,
+            'overridetypename': overridetype.overridetype}
+
+    if section is not None:
+        contents_q += " AND s.id = :sectionid"
+        vals['sectionid'] = section.section_id
+
+    contents_q += " ORDER BY fn"
+
+    return session.execute(contents_q, vals)
+
+__all__.append('get_contents')
+
+################################################################################
+
+class ContentFilepath(object):
+    def __init__(self, *args, **kwargs):
+        pass
+
+    def __repr__(self):
+        return '<ContentFilepath %s>' % self.filepath
+
+__all__.append('ContentFilepath')
+
+def get_or_set_contents_path_id(filepath, session=None):
+    """
+    Returns database id for given path.
+
+    If no matching file is found, a row is inserted.
 
-        suite_id = None
-        if type(suite) == str:
-            suite_id = self.get_suite_id(suite)
-        elif type(suite) == int:
-            suite_id = suite
+    @type filename: string
+    @param filename: The filepath
+    @type session: SQLAlchemy
+    @param session: Optional SQL session object (a temporary one will be
+    generated if not supplied).  If not passed, a commit will be performed at
+    the end of the function, otherwise the caller is responsible for commiting.
+
+    @rtype: int
+    @return: the database id for the given path
+    """
+    privatetrans = False
+    if session is None:
+        session = DBConn().session()
+        privatetrans = True
+
+    q = session.query(ContentFilepath).filter_by(filepath=filepath)
+
+    try:
+        ret = q.one().cafilepath_id
+    except NoResultFound:
+        cf = ContentFilepath()
+        cf.filepath = filepath
+        session.add(cf)
+        if privatetrans:
+            session.commit()
         else:
-            return None
+            session.flush()
+        ret = cf.cafilepath_id
 
-        c = self.db_con.cursor()
-        c.execute( """SELECT a.arch_string FROM suite_architectures sa
-                      JOIN architecture a ON (a.id = sa.architecture)
-                      WHERE suite='%s'""" % suite_id )
+    if privatetrans:
+        session.close()
 
-        return map(lambda x: x[0], c.fetchall())
+    return ret
 
-    def insert_content_paths(self, bin_id, fullpaths):
-        """
-        Make sure given path is associated with given binary id
+__all__.append('get_or_set_contents_path_id')
+
+################################################################################
 
-        @type bin_id: int
-        @param bin_id: the id of the binary
-        @type fullpaths: list
-        @param fullpaths: the list of paths of the file being associated with the binary
+class ContentAssociation(object):
+    def __init__(self, *args, **kwargs):
+        pass
 
-        @return: True upon success
-        """
+    def __repr__(self):
+        return '<ContentAssociation %s>' % self.ca_id
 
-        c = self.db_con.cursor()
+__all__.append('ContentAssociation')
 
-        c.execute("BEGIN WORK")
-        try:
+def insert_content_paths(binary_id, fullpaths, session=None):
+    """
+    Make sure given path is associated with given binary id
+
+    @type binary_id: int
+    @param binary_id: the id of the binary
+    @type fullpaths: list
+    @param fullpaths: the list of paths of the file being associated with the binary
+    @type session: SQLAlchemy session
+    @param session: Optional SQLAlchemy session.  If this is passed, the caller
+    is responsible for ensuring a transaction has begun and committing the
+    results or rolling back based on the result code.  If not passed, a commit
+    will be performed at the end of the function, otherwise the caller is
+    responsible for commiting.
+
+    @return: True upon success
+    """
 
-            for fullpath in fullpaths:
-                (path, file) = os.path.split(fullpath)
+    privatetrans = False
+    if session is None:
+        session = DBConn().session()
+        privatetrans = True
+
+    try:
+        # Insert paths
+        pathcache = {}
+        for fullpath in fullpaths:
+            # Get the necessary IDs ...
+            (path, file) = os.path.split(fullpath)
+
+            filepath_id = get_or_set_contents_path_id(path, session)
+            filename_id = get_or_set_contents_file_id(file, session)
+
+            pathcache[fullpath] = (filepath_id, filename_id)
+
+        for fullpath, dat in pathcache.items():
+            ca = ContentAssociation()
+            ca.binary_id = binary_id
+            ca.filepath_id = dat[0]
+            ca.filename_id = dat[1]
+            session.add(ca)
+
+        # Only commit if we set up the session ourself
+        if privatetrans:
+            session.commit()
+            session.close()
+        else:
+            session.flush()
 
-                if path.startswith( "./" ):
-                    path = path[2:]
-                # Get the necessary IDs ...
-                file_id = self.get_or_set_contents_file_id(file)
-                path_id = self.get_or_set_contents_path_id(path)
+        return True
 
-                c.execute("""INSERT INTO content_associations
-                               (binary_pkg, filepath, filename)
-                           VALUES ( '%d', '%d', '%d')""" % (bin_id, path_id, file_id) )
+    except:
+        traceback.print_exc()
 
-            c.execute("COMMIT")
-            return True
-        except:
-            traceback.print_exc()
-            c.execute("ROLLBACK")
-            return False
+        # Only rollback if we set up the session ourself
+        if privatetrans:
+            session.rollback()
+            session.close()
 
-    def insert_pending_content_paths(self, package, fullpaths):
-        """
-        Make sure given paths are temporarily associated with given
-        package
+        return False
 
-        @type package: dict
-        @param package: the package to associate with should have been read in from the binary control file
-        @type fullpaths: list
-        @param fullpaths: the list of paths of the file being associated with the binary
+__all__.append('insert_content_paths')
 
-        @return: True upon success
-        """
+################################################################################
 
-        c = self.db_con.cursor()
+class DSCFile(object):
+    def __init__(self, *args, **kwargs):
+        pass
+
+    def __repr__(self):
+        return '<DSCFile %s>' % self.dscfile_id
+
+__all__.append('DSCFile')
+
+@session_wrapper
+def get_dscfiles(dscfile_id=None, source_id=None, poolfile_id=None, session=None):
+    """
+    Returns a list of DSCFiles which may be empty
+
+    @type dscfile_id: int (optional)
+    @param dscfile_id: the dscfile_id of the DSCFiles to find
+
+    @type source_id: int (optional)
+    @param source_id: the source id related to the DSCFiles to find
+
+    @type poolfile_id: int (optional)
+    @param poolfile_id: the poolfile id related to the DSCFiles to find
+
+    @rtype: list
+    @return: Possibly empty list of DSCFiles
+    """
+
+    q = session.query(DSCFile)
+
+    if dscfile_id is not None:
+        q = q.filter_by(dscfile_id=dscfile_id)
+
+    if source_id is not None:
+        q = q.filter_by(source_id=source_id)
+
+    if poolfile_id is not None:
+        q = q.filter_by(poolfile_id=poolfile_id)
+
+    return q.all()
+
+__all__.append('get_dscfiles')
+
+################################################################################
+
+class PoolFile(object):
+    def __init__(self, *args, **kwargs):
+        pass
+
+    def __repr__(self):
+        return '<PoolFile %s>' % self.filename
+
+__all__.append('PoolFile')
+
+@session_wrapper
+def check_poolfile(filename, filesize, md5sum, location_id, session=None):
+    """
+    Returns a tuple:
+     (ValidFileFound [boolean or None], PoolFile object or None)
+
+    @type filename: string
+    @param filename: the filename of the file to check against the DB
+
+    @type filesize: int
+    @param filesize: the size of the file to check against the DB
+
+    @type md5sum: string
+    @param md5sum: the md5sum of the file to check against the DB
+
+    @type location_id: int
+    @param location_id: the id of the location to look in
+
+    @rtype: tuple
+    @return: Tuple of length 2.
+             If more than one file found with that name:
+                    (None,  None)
+             If valid pool file found: (True, PoolFile object)
+             If valid pool file not found:
+                    (False, None) if no file found
+                    (False, PoolFile object) if file found with size/md5sum mismatch
+    """
+
+    q = session.query(PoolFile).filter_by(filename=filename)
+    q = q.join(Location).filter_by(location_id=location_id)
+
+    ret = None
+
+    if q.count() > 1:
+        ret = (None, None)
+    elif q.count() < 1:
+        ret = (False, None)
+    else:
+        obj = q.one()
+        if obj.md5sum != md5sum or obj.filesize != filesize:
+            ret = (False, obj)
+
+    if ret is None:
+        ret = (True, obj)
+
+    return ret
+
+__all__.append('check_poolfile')
+
+@session_wrapper
+def get_poolfile_by_id(file_id, session=None):
+    """
+    Returns a PoolFile objects or None for the given id
+
+    @type file_id: int
+    @param file_id: the id of the file to look for
+
+    @rtype: PoolFile or None
+    @return: either the PoolFile object or None
+    """
+
+    q = session.query(PoolFile).filter_by(file_id=file_id)
+
+    try:
+        return q.one()
+    except NoResultFound:
+        return None
+
+__all__.append('get_poolfile_by_id')
+
+
+@session_wrapper
+def get_poolfile_by_name(filename, location_id=None, session=None):
+    """
+    Returns an array of PoolFile objects for the given filename and
+    (optionally) location_id
+
+    @type filename: string
+    @param filename: the filename of the file to check against the DB
+
+    @type location_id: int
+    @param location_id: the id of the location to look in (optional)
+
+    @rtype: array
+    @return: array of PoolFile objects
+    """
+
+    q = session.query(PoolFile).filter_by(filename=filename)
+
+    if location_id is not None:
+        q = q.join(Location).filter_by(location_id=location_id)
+
+    return q.all()
+
+__all__.append('get_poolfile_by_name')
+
+@session_wrapper
+def get_poolfile_like_name(filename, session=None):
+    """
+    Returns an array of PoolFile objects which are like the given name
+
+    @type filename: string
+    @param filename: the filename of the file to check against the DB
+
+    @rtype: array
+    @return: array of PoolFile objects
+    """
+
+    # TODO: There must be a way of properly using bind parameters with %FOO%
+    q = session.query(PoolFile).filter(PoolFile.filename.like('%%%s%%' % filename))
+
+    return q.all()
+
+__all__.append('get_poolfile_like_name')
+
+################################################################################
+
+class Fingerprint(object):
+    def __init__(self, *args, **kwargs):
+        pass
+
+    def __repr__(self):
+        return '<Fingerprint %s>' % self.fingerprint
+
+__all__.append('Fingerprint')
+
+def get_or_set_fingerprint(fpr, session=None):
+    """
+    Returns Fingerprint object for given fpr.
+
+    If no matching fpr is found, a row is inserted.
+
+    @type fpr: string
+    @param fpr: The fpr to find / add
+
+    @type session: SQLAlchemy
+    @param session: Optional SQL session object (a temporary one will be
+    generated if not supplied).  If not passed, a commit will be performed at
+    the end of the function, otherwise the caller is responsible for commiting.
+    A flush will be performed either way.
+
+    @rtype: Fingerprint
+    @return: the Fingerprint object for the given fpr
+    """
+    privatetrans = False
+    if session is None:
+        session = DBConn().session()
+        privatetrans = True
+
+    q = session.query(Fingerprint).filter_by(fingerprint=fpr)
+
+    try:
+        ret = q.one()
+    except NoResultFound:
+        fingerprint = Fingerprint()
+        fingerprint.fingerprint = fpr
+        session.add(fingerprint)
+        if privatetrans:
+            session.commit()
+        else:
+            session.flush()
+        ret = fingerprint
+
+    if privatetrans:
+        session.close()
+
+    return ret
+
+__all__.append('get_or_set_fingerprint')
+
+################################################################################
+
+class Keyring(object):
+    def __init__(self, *args, **kwargs):
+        pass
+
+    def __repr__(self):
+        return '<Keyring %s>' % self.keyring_name
+
+__all__.append('Keyring')
+
+def get_or_set_keyring(keyring, session=None):
+    """
+    If C{keyring} does not have an entry in the C{keyrings} table yet, create one
+    and return the new Keyring
+    If C{keyring} already has an entry, simply return the existing Keyring
+
+    @type keyring: string
+    @param keyring: the keyring name
+
+    @rtype: Keyring
+    @return: the Keyring object for this keyring
+
+    """
+    privatetrans = False
+    if session is None:
+        session = DBConn().session()
+        privatetrans = True
+
+    try:
+        obj = session.query(Keyring).filter_by(keyring_name=keyring).first()
+
+        if obj is None:
+            obj = Keyring(keyring_name=keyring)
+            session.add(obj)
+            if privatetrans:
+                session.commit()
+            else:
+                session.flush()
+
+        return obj
+    finally:
+        if privatetrans:
+            session.close()
+
+__all__.append('get_or_set_keyring')
+
+################################################################################
+
+class Location(object):
+    def __init__(self, *args, **kwargs):
+        pass
+
+    def __repr__(self):
+        return '<Location %s (%s)>' % (self.path, self.location_id)
+
+__all__.append('Location')
+
+@session_wrapper
+def get_location(location, component=None, archive=None, session=None):
+    """
+    Returns Location object for the given combination of location, component
+    and archive
+
+    @type location: string
+    @param location: the path of the location, e.g. I{/srv/ftp.debian.org/ftp/pool/}
+
+    @type component: string
+    @param component: the component name (if None, no restriction applied)
+
+    @type archive: string
+    @param archive_id: the archive name (if None, no restriction applied)
+
+    @rtype: Location / None
+    @return: Either a Location object or None if one can't be found
+    """
+
+    q = session.query(Location).filter_by(path=location)
+
+    if archive is not None:
+        q = q.join(Archive).filter_by(archive_name=archive)
+
+    if component is not None:
+        q = q.join(Component).filter_by(component_name=component)
+
+    try:
+        return q.one()
+    except NoResultFound:
+        return None
+
+__all__.append('get_location')
+
+################################################################################
+
+class Maintainer(object):
+    def __init__(self, *args, **kwargs):
+        pass
+
+    def __repr__(self):
+        return '''<Maintainer '%s' (%s)>''' % (self.name, self.maintainer_id)
+
+    def get_split_maintainer(self):
+        if not hasattr(self, 'name') or self.name is None:
+            return ('', '', '', '')
+
+        return fix_maintainer(self.name.strip())
+
+__all__.append('Maintainer')
+
+def get_or_set_maintainer(name, session=None):
+    """
+    Returns Maintainer object for given maintainer name.
+
+    If no matching maintainer name is found, a row is inserted.
+
+    @type name: string
+    @param name: The maintainer name to add
+
+    @type session: SQLAlchemy
+    @param session: Optional SQL session object (a temporary one will be
+    generated if not supplied).  If not passed, a commit will be performed at
+    the end of the function, otherwise the caller is responsible for commiting.
+    A flush will be performed either way.
+
+    @rtype: Maintainer
+    @return: the Maintainer object for the given maintainer
+    """
+    privatetrans = False
+    if session is None:
+        session = DBConn().session()
+        privatetrans = True
+
+    q = session.query(Maintainer).filter_by(name=name)
+    try:
+        ret = q.one()
+    except NoResultFound:
+        maintainer = Maintainer()
+        maintainer.name = name
+        session.add(maintainer)
+        if privatetrans:
+            session.commit()
+        else:
+            session.flush()
+        ret = maintainer
+
+    if privatetrans:
+        session.close()
+
+    return ret
+
+__all__.append('get_or_set_maintainer')
+
+def get_maintainer(maintainer_id, session=None):
+    """
+    Return the name of the maintainer behind C{maintainer_id} or None if that
+    maintainer_id is invalid.
+
+    @type maintainer_id: int
+    @param maintainer_id: the id of the maintainer
+
+    @rtype: Maintainer
+    @return: the Maintainer with this C{maintainer_id}
+    """
+
+    privatetrans = False
+    if session is None:
+        session = DBConn().session()
+        privatetrans = True
+
+    try:
+        return session.query(Maintainer).get(maintainer_id)
+    finally:
+        if privatetrans:
+            session.close()
+
+__all__.append('get_maintainer')
+
+################################################################################
+
+class NewComment(object):
+    def __init__(self, *args, **kwargs):
+        pass
+
+    def __repr__(self):
+        return '''<NewComment for '%s %s' (%s)>''' % (self.package, self.version, self.comment_id)
+
+__all__.append('NewComment')
+
+@session_wrapper
+def has_new_comment(package, version, session=None):
+    """
+    Returns true if the given combination of C{package}, C{version} has a comment.
+
+    @type package: string
+    @param package: name of the package
+
+    @type version: string
+    @param version: package version
+
+    @type session: Session
+    @param session: Optional SQLA session object (a temporary one will be
+    generated if not supplied)
+
+    @rtype: boolean
+    @return: true/false
+    """
+
+    q = session.query(NewComment)
+    q = q.filter_by(package=package)
+    q = q.filter_by(version=version)
+
+    return bool(q.count() > 0)
+
+__all__.append('has_new_comment')
+
+@session_wrapper
+def get_new_comments(package=None, version=None, comment_id=None, session=None):
+    """
+    Returns (possibly empty) list of NewComment objects for the given
+    parameters
+
+    @type package: string (optional)
+    @param package: name of the package
+
+    @type version: string (optional)
+    @param version: package version
+
+    @type comment_id: int (optional)
+    @param comment_id: An id of a comment
+
+    @type session: Session
+    @param session: Optional SQLA session object (a temporary one will be
+    generated if not supplied)
+
+    @rtype: list
+    @return: A (possibly empty) list of NewComment objects will be returned
+    """
+
+    q = session.query(NewComment)
+    if package is not None: q = q.filter_by(package=package)
+    if version is not None: q = q.filter_by(version=version)
+    if comment_id is not None: q = q.filter_by(comment_id=comment_id)
+
+    return q.all()
+
+__all__.append('get_new_comments')
+
+################################################################################
+
+class Override(object):
+    def __init__(self, *args, **kwargs):
+        pass
+
+    def __repr__(self):
+        return '<Override %s (%s)>' % (self.package, self.suite_id)
+
+__all__.append('Override')
+
+@session_wrapper
+def get_override(package, suite=None, component=None, overridetype=None, session=None):
+    """
+    Returns Override object for the given parameters
+
+    @type package: string
+    @param package: The name of the package
+
+    @type suite: string, list or None
+    @param suite: The name of the suite (or suites if a list) to limit to.  If
+                  None, don't limit.  Defaults to None.
+
+    @type component: string, list or None
+    @param component: The name of the component (or components if a list) to
+                      limit to.  If None, don't limit.  Defaults to None.
+
+    @type overridetype: string, list or None
+    @param overridetype: The name of the overridetype (or overridetypes if a list) to
+                         limit to.  If None, don't limit.  Defaults to None.
+
+    @type session: Session
+    @param session: Optional SQLA session object (a temporary one will be
+    generated if not supplied)
+
+    @rtype: list
+    @return: A (possibly empty) list of Override objects will be returned
+    """
+
+    q = session.query(Override)
+    q = q.filter_by(package=package)
+
+    if suite is not None:
+        if not isinstance(suite, list): suite = [suite]
+        q = q.join(Suite).filter(Suite.suite_name.in_(suite))
+
+    if component is not None:
+        if not isinstance(component, list): component = [component]
+        q = q.join(Component).filter(Component.component_name.in_(component))
+
+    if overridetype is not None:
+        if not isinstance(overridetype, list): overridetype = [overridetype]
+        q = q.join(OverrideType).filter(OverrideType.overridetype.in_(overridetype))
+
+    return q.all()
+
+__all__.append('get_override')
+
+
+################################################################################
+
+class OverrideType(object):
+    def __init__(self, *args, **kwargs):
+        pass
+
+    def __repr__(self):
+        return '<OverrideType %s>' % self.overridetype
+
+__all__.append('OverrideType')
+
+@session_wrapper
+def get_override_type(override_type, session=None):
+    """
+    Returns OverrideType object for given C{override type}.
+
+    @type override_type: string
+    @param override_type: The name of the override type
+
+    @type session: Session
+    @param session: Optional SQLA session object (a temporary one will be
+    generated if not supplied)
+
+    @rtype: int
+    @return: the database id for the given override type
+    """
+
+    q = session.query(OverrideType).filter_by(overridetype=override_type)
+
+    try:
+        return q.one()
+    except NoResultFound:
+        return None
+
+__all__.append('get_override_type')
+
+################################################################################
+
+class PendingContentAssociation(object):
+    def __init__(self, *args, **kwargs):
+        pass
+
+    def __repr__(self):
+        return '<PendingContentAssociation %s>' % self.pca_id
+
+__all__.append('PendingContentAssociation')
+
+def insert_pending_content_paths(package, fullpaths, session=None):
+    """
+    Make sure given paths are temporarily associated with given
+    package
+
+    @type package: dict
+    @param package: the package to associate with should have been read in from the binary control file
+    @type fullpaths: list
+    @param fullpaths: the list of paths of the file being associated with the binary
+    @type session: SQLAlchemy session
+    @param session: Optional SQLAlchemy session.  If this is passed, the caller
+    is responsible for ensuring a transaction has begun and committing the
+    results or rolling back based on the result code.  If not passed, a commit
+    will be performed at the end of the function
+
+    @return: True upon success, False if there is a problem
+    """
+
+    privatetrans = False
+
+    if session is None:
+        session = DBConn().session()
+        privatetrans = True
+
+    try:
+        arch = get_architecture(package['Architecture'], session)
+        arch_id = arch.arch_id
+
+        # Remove any already existing recorded files for this package
+        q = session.query(PendingContentAssociation)
+        q = q.filter_by(package=package['Package'])
+        q = q.filter_by(version=package['Version'])
+        q = q.filter_by(architecture=arch_id)
+        q.delete()
+
+        # Insert paths
+        pathcache = {}
+        for fullpath in fullpaths:
+            (path, file) = os.path.split(fullpath)
+
+            if path.startswith( "./" ):
+                path = path[2:]
+
+            filepath_id = get_or_set_contents_path_id(path, session)
+            filename_id = get_or_set_contents_file_id(file, session)
+
+            pathcache[fullpath] = (filepath_id, filename_id)
+
+        for fullpath, dat in pathcache.items():
+            pca = PendingContentAssociation()
+            pca.package = package['Package']
+            pca.version = package['Version']
+            pca.filepath_id = dat[0]
+            pca.filename_id = dat[1]
+            pca.architecture = arch_id
+            session.add(pca)
+
+        # Only commit if we set up the session ourself
+        if privatetrans:
+            session.commit()
+            session.close()
+        else:
+            session.flush()
+
+        return True
+    except Exception, e:
+        traceback.print_exc()
+
+        # Only rollback if we set up the session ourself
+        if privatetrans:
+            session.rollback()
+            session.close()
+
+        return False
+
+__all__.append('insert_pending_content_paths')
+
+################################################################################
+
+class Priority(object):
+    def __init__(self, *args, **kwargs):
+        pass
+
+    def __eq__(self, val):
+        if isinstance(val, str):
+            return (self.priority == val)
+        # This signals to use the normal comparison operator
+        return NotImplemented
+
+    def __ne__(self, val):
+        if isinstance(val, str):
+            return (self.priority != val)
+        # This signals to use the normal comparison operator
+        return NotImplemented
+
+    def __repr__(self):
+        return '<Priority %s (%s)>' % (self.priority, self.priority_id)
+
+__all__.append('Priority')
+
+@session_wrapper
+def get_priority(priority, session=None):
+    """
+    Returns Priority object for given C{priority name}.
+
+    @type priority: string
+    @param priority: The name of the priority
+
+    @type session: Session
+    @param session: Optional SQLA session object (a temporary one will be
+    generated if not supplied)
+
+    @rtype: Priority
+    @return: Priority object for the given priority
+    """
+
+    q = session.query(Priority).filter_by(priority=priority)
+
+    try:
+        return q.one()
+    except NoResultFound:
+        return None
+
+__all__.append('get_priority')
+
+@session_wrapper
+def get_priorities(session=None):
+    """
+    Returns dictionary of priority names -> id mappings
+
+    @type session: Session
+    @param session: Optional SQL session object (a temporary one will be
+    generated if not supplied)
+
+    @rtype: dictionary
+    @return: dictionary of priority names -> id mappings
+    """
+
+    ret = {}
+    q = session.query(Priority)
+    for x in q.all():
+        ret[x.priority] = x.priority_id
+
+    return ret
+
+__all__.append('get_priorities')
+
+################################################################################
+
+class Queue(object):
+    def __init__(self, *args, **kwargs):
+        pass
+
+    def __repr__(self):
+        return '<Queue %s>' % self.queue_name
+
+    def autobuild_upload(self, changes, srcpath, session=None):
+        """
+        Update queue_build database table used for incoming autobuild support.
+
+        @type changes: Changes
+        @param changes: changes object for the upload to process
+
+        @type srcpath: string
+        @param srcpath: path for the queue file entries/link destinations
+
+        @type session: SQLAlchemy session
+        @param session: Optional SQLAlchemy session.  If this is passed, the
+        caller is responsible for ensuring a transaction has begun and
+        committing the results or rolling back based on the result code.  If
+        not passed, a commit will be performed at the end of the function,
+        otherwise the caller is responsible for commiting.
+
+        @rtype: NoneType or string
+        @return: None if the operation failed, a string describing the error if not
+        """
+
+        privatetrans = False
+        if session is None:
+            session = DBConn().session()
+            privatetrans = True
+
+        # TODO: Remove by moving queue config into the database
+        conf = Config()
+
+        for suitename in changes.changes["distribution"].keys():
+            # TODO: Move into database as:
+            #       buildqueuedir TEXT DEFAULT NULL (i.e. NULL is no build)
+            #       buildqueuecopy BOOLEAN NOT NULL DEFAULT FALSE (i.e. default is symlink)
+            #       This also gets rid of the SecurityQueueBuild hack below
+            if suitename not in conf.ValueList("Dinstall::QueueBuildSuites"):
+                continue
+
+            # Find suite object
+            s = get_suite(suitename, session)
+            if s is None:
+                return "INTERNAL ERROR: Could not find suite %s" % suitename
+
+            # TODO: Get from database as above
+            dest_dir = conf["Dir::QueueBuild"]
+
+            # TODO: Move into database as above
+            if conf.FindB("Dinstall::SecurityQueueBuild"):
+                dest_dir = os.path.join(dest_dir, suitename)
+
+            for file_entry in changes.files.keys():
+                src = os.path.join(srcpath, file_entry)
+                dest = os.path.join(dest_dir, file_entry)
+
+                # TODO: Move into database as above
+                if conf.FindB("Dinstall::SecurityQueueBuild"):
+                    # Copy it since the original won't be readable by www-data
+                    import utils
+                    utils.copy(src, dest)
+                else:
+                    # Create a symlink to it
+                    os.symlink(src, dest)
+
+                qb = QueueBuild()
+                qb.suite_id = s.suite_id
+                qb.queue_id = self.queue_id
+                qb.filename = dest
+                qb.in_queue = True
+
+                session.add(qb)
+
+            # If the .orig tarballs are in the pool, create a symlink to
+            # them (if one doesn't already exist)
+            for dsc_file in changes.dsc_files.keys():
+                # Skip all files except orig tarballs
+                if not re_is_orig_source.match(dsc_file):
+                    continue
+                # Skip orig files not identified in the pool
+                if not (changes.orig_files.has_key(dsc_file) and
+                        changes.orig_files[dsc_file].has_key("id")):
+                    continue
+                orig_file_id = changes.orig_files[dsc_file]["id"]
+                dest = os.path.join(dest_dir, dsc_file)
+
+                # If it doesn't exist, create a symlink
+                if not os.path.exists(dest):
+                    q = session.execute("SELECT l.path, f.filename FROM location l, files f WHERE f.id = :id and f.location = l.id",
+                                        {'id': orig_file_id})
+                    res = q.fetchone()
+                    if not res:
+                        return "[INTERNAL ERROR] Couldn't find id %s in files table." % (orig_file_id)
+
+                    src = os.path.join(res[0], res[1])
+                    os.symlink(src, dest)
+
+                    # Add it to the list of packages for later processing by apt-ftparchive
+                    qb = QueueBuild()
+                    qb.suite_id = s.suite_id
+                    qb.queue_id = self.queue_id
+                    qb.filename = dest
+                    qb.in_queue = True
+                    session.add(qb)
+
+                # If it does, update things to ensure it's not removed prematurely
+                else:
+                    qb = get_queue_build(dest, s.suite_id, session)
+                    if qb is None:
+                        qb.in_queue = True
+                        qb.last_used = None
+                        session.add(qb)
+
+        if privatetrans:
+            session.commit()
+            session.close()
+
+        return None
+
+__all__.append('Queue')
+
+@session_wrapper
+def get_queue(queuename, session=None):
+    """
+    Returns Queue object for given C{queue name}.
+
+    @type queuename: string
+    @param queuename: The name of the queue
+
+    @type session: Session
+    @param session: Optional SQLA session object (a temporary one will be
+    generated if not supplied)
+
+    @rtype: Queue
+    @return: Queue object for the given queue
+    """
+
+    q = session.query(Queue).filter_by(queue_name=queuename)
+
+    try:
+        return q.one()
+    except NoResultFound:
+        return None
+
+__all__.append('get_queue')
+
+################################################################################
+
+class QueueBuild(object):
+    def __init__(self, *args, **kwargs):
+        pass
+
+    def __repr__(self):
+        return '<QueueBuild %s (%s)>' % (self.filename, self.queue_id)
+
+__all__.append('QueueBuild')
+
+@session_wrapper
+def get_queue_build(filename, suite, session=None):
+    """
+    Returns QueueBuild object for given C{filename} and C{suite}.
+
+    @type filename: string
+    @param filename: The name of the file
+
+    @type suiteid: int or str
+    @param suiteid: Suite name or ID
+
+    @type session: Session
+    @param session: Optional SQLA session object (a temporary one will be
+    generated if not supplied)
+
+    @rtype: Queue
+    @return: Queue object for the given queue
+    """
+
+    if isinstance(suite, int):
+        q = session.query(QueueBuild).filter_by(filename=filename).filter_by(suite_id=suite)
+    else:
+        q = session.query(QueueBuild).filter_by(filename=filename)
+        q = q.join(Suite).filter_by(suite_name=suite)
+
+    try:
+        return q.one()
+    except NoResultFound:
+        return None
+
+__all__.append('get_queue_build')
+
+################################################################################
+
+class Section(object):
+    def __init__(self, *args, **kwargs):
+        pass
+
+    def __eq__(self, val):
+        if isinstance(val, str):
+            return (self.section == val)
+        # This signals to use the normal comparison operator
+        return NotImplemented
+
+    def __ne__(self, val):
+        if isinstance(val, str):
+            return (self.section != val)
+        # This signals to use the normal comparison operator
+        return NotImplemented
+
+    def __repr__(self):
+        return '<Section %s>' % self.section
+
+__all__.append('Section')
+
+@session_wrapper
+def get_section(section, session=None):
+    """
+    Returns Section object for given C{section name}.
+
+    @type section: string
+    @param section: The name of the section
+
+    @type session: Session
+    @param session: Optional SQLA session object (a temporary one will be
+    generated if not supplied)
+
+    @rtype: Section
+    @return: Section object for the given section name
+    """
+
+    q = session.query(Section).filter_by(section=section)
+
+    try:
+        return q.one()
+    except NoResultFound:
+        return None
+
+__all__.append('get_section')
+
+@session_wrapper
+def get_sections(session=None):
+    """
+    Returns dictionary of section names -> id mappings
+
+    @type session: Session
+    @param session: Optional SQL session object (a temporary one will be
+    generated if not supplied)
+
+    @rtype: dictionary
+    @return: dictionary of section names -> id mappings
+    """
+
+    ret = {}
+    q = session.query(Section)
+    for x in q.all():
+        ret[x.section] = x.section_id
+
+    return ret
+
+__all__.append('get_sections')
+
+################################################################################
+
+class DBSource(object):
+    def __init__(self, *args, **kwargs):
+        pass
+
+    def __repr__(self):
+        return '<DBSource %s (%s)>' % (self.source, self.version)
+
+__all__.append('DBSource')
+
+@session_wrapper
+def source_exists(source, source_version, suites = ["any"], session=None):
+    """
+    Ensure that source exists somewhere in the archive for the binary
+    upload being processed.
+      1. exact match     => 1.0-3
+      2. bin-only NMU    => 1.0-3+b1 , 1.0-3.1+b1
+
+    @type package: string
+    @param package: package source name
+
+    @type source_version: string
+    @param source_version: expected source version
+
+    @type suites: list
+    @param suites: list of suites to check in, default I{any}
+
+    @type session: Session
+    @param session: Optional SQLA session object (a temporary one will be
+    generated if not supplied)
+
+    @rtype: int
+    @return: returns 1 if a source with expected version is found, otherwise 0
+
+    """
+
+    cnf = Config()
+    ret = 1
+
+    for suite in suites:
+        q = session.query(DBSource).filter_by(source=source)
+        if suite != "any":
+            # source must exist in suite X, or in some other suite that's
+            # mapped to X, recursively... silent-maps are counted too,
+            # unreleased-maps aren't.
+            maps = cnf.ValueList("SuiteMappings")[:]
+            maps.reverse()
+            maps = [ m.split() for m in maps ]
+            maps = [ (x[1], x[2]) for x in maps
+                            if x[0] == "map" or x[0] == "silent-map" ]
+            s = [suite]
+            for x in maps:
+                if x[1] in s and x[0] not in s:
+                    s.append(x[0])
+
+            q = q.join(SrcAssociation).join(Suite)
+            q = q.filter(Suite.suite_name.in_(s))
+
+        # Reduce the query results to a list of version numbers
+        ql = [ j.version for j in q.all() ]
+
+        # Try (1)
+        if source_version in ql:
+            continue
+
+        # Try (2)
+        from daklib.regexes import re_bin_only_nmu
+        orig_source_version = re_bin_only_nmu.sub('', source_version)
+        if orig_source_version in ql:
+            continue
+
+        # No source found so return not ok
+        ret = 0
+
+    return ret
+
+__all__.append('source_exists')
+
+@session_wrapper
+def get_suites_source_in(source, session=None):
+    """
+    Returns list of Suite objects which given C{source} name is in
+
+    @type source: str
+    @param source: DBSource package name to search for
+
+    @rtype: list
+    @return: list of Suite objects for the given source
+    """
+
+    return session.query(Suite).join(SrcAssociation).join(DBSource).filter_by(source=source).all()
+
+__all__.append('get_suites_source_in')
+
+@session_wrapper
+def get_sources_from_name(source, version=None, dm_upload_allowed=None, session=None):
+    """
+    Returns list of DBSource objects for given C{source} name and other parameters
+
+    @type source: str
+    @param source: DBSource package name to search for
+
+    @type source: str or None
+    @param source: DBSource version name to search for or None if not applicable
+
+    @type dm_upload_allowed: bool
+    @param dm_upload_allowed: If None, no effect.  If True or False, only
+    return packages with that dm_upload_allowed setting
+
+    @type session: Session
+    @param session: Optional SQL session object (a temporary one will be
+    generated if not supplied)
+
+    @rtype: list
+    @return: list of DBSource objects for the given name (may be empty)
+    """
+
+    q = session.query(DBSource).filter_by(source=source)
+
+    if version is not None:
+        q = q.filter_by(version=version)
+
+    if dm_upload_allowed is not None:
+        q = q.filter_by(dm_upload_allowed=dm_upload_allowed)
+
+    return q.all()
+
+__all__.append('get_sources_from_name')
+
+@session_wrapper
+def get_source_in_suite(source, suite, session=None):
+    """
+    Returns list of DBSource objects for a combination of C{source} and C{suite}.
+
+      - B{source} - source package name, eg. I{mailfilter}, I{bbdb}, I{glibc}
+      - B{suite} - a suite name, eg. I{unstable}
+
+    @type source: string
+    @param source: source package name
+
+    @type suite: string
+    @param suite: the suite name
+
+    @rtype: string
+    @return: the version for I{source} in I{suite}
+
+    """
+
+    q = session.query(SrcAssociation)
+    q = q.join('source').filter_by(source=source)
+    q = q.join('suite').filter_by(suite_name=suite)
+
+    try:
+        return q.one().source
+    except NoResultFound:
+        return None
+
+__all__.append('get_source_in_suite')
+
+################################################################################
+
+class SrcAssociation(object):
+    def __init__(self, *args, **kwargs):
+        pass
+
+    def __repr__(self):
+        return '<SrcAssociation %s (%s, %s)>' % (self.sa_id, self.source, self.suite)
+
+__all__.append('SrcAssociation')
+
+################################################################################
+
+class SrcFormat(object):
+    def __init__(self, *args, **kwargs):
+        pass
+
+    def __repr__(self):
+        return '<SrcFormat %s>' % (self.format_name)
+
+__all__.append('SrcFormat')
+
+################################################################################
+
+class SrcUploader(object):
+    def __init__(self, *args, **kwargs):
+        pass
+
+    def __repr__(self):
+        return '<SrcUploader %s>' % self.uploader_id
+
+__all__.append('SrcUploader')
+
+################################################################################
+
+SUITE_FIELDS = [ ('SuiteName', 'suite_name'),
+                 ('SuiteID', 'suite_id'),
+                 ('Version', 'version'),
+                 ('Origin', 'origin'),
+                 ('Label', 'label'),
+                 ('Description', 'description'),
+                 ('Untouchable', 'untouchable'),
+                 ('Announce', 'announce'),
+                 ('Codename', 'codename'),
+                 ('OverrideCodename', 'overridecodename'),
+                 ('ValidTime', 'validtime'),
+                 ('Priority', 'priority'),
+                 ('NotAutomatic', 'notautomatic'),
+                 ('CopyChanges', 'copychanges'),
+                 ('CopyDotDak', 'copydotdak'),
+                 ('CommentsDir', 'commentsdir'),
+                 ('OverrideSuite', 'overridesuite'),
+                 ('ChangelogBase', 'changelogbase')]
+
+
+class Suite(object):
+    def __init__(self, *args, **kwargs):
+        pass
+
+    def __repr__(self):
+        return '<Suite %s>' % self.suite_name
+
+    def __eq__(self, val):
+        if isinstance(val, str):
+            return (self.suite_name == val)
+        # This signals to use the normal comparison operator
+        return NotImplemented
+
+    def __ne__(self, val):
+        if isinstance(val, str):
+            return (self.suite_name != val)
+        # This signals to use the normal comparison operator
+        return NotImplemented
+
+    def details(self):
+        ret = []
+        for disp, field in SUITE_FIELDS:
+            val = getattr(self, field, None)
+            if val is not None:
+                ret.append("%s: %s" % (disp, val))
+
+        return "\n".join(ret)
+
+__all__.append('Suite')
+
+@session_wrapper
+def get_suite_architecture(suite, architecture, session=None):
+    """
+    Returns a SuiteArchitecture object given C{suite} and ${arch} or None if it
+    doesn't exist
+
+    @type suite: str
+    @param suite: Suite name to search for
+
+    @type architecture: str
+    @param architecture: Architecture name to search for
+
+    @type session: Session
+    @param session: Optional SQL session object (a temporary one will be
+    generated if not supplied)
+
+    @rtype: SuiteArchitecture
+    @return: the SuiteArchitecture object or None
+    """
+
+    q = session.query(SuiteArchitecture)
+    q = q.join(Architecture).filter_by(arch_string=architecture)
+    q = q.join(Suite).filter_by(suite_name=suite)
+
+    try:
+        return q.one()
+    except NoResultFound:
+        return None
+
+__all__.append('get_suite_architecture')
+
+@session_wrapper
+def get_suite(suite, session=None):
+    """
+    Returns Suite object for given C{suite name}.
+
+    @type suite: string
+    @param suite: The name of the suite
+
+    @type session: Session
+    @param session: Optional SQLA session object (a temporary one will be
+    generated if not supplied)
+
+    @rtype: Suite
+    @return: Suite object for the requested suite name (None if not presenT)
+    """
+
+    q = session.query(Suite).filter_by(suite_name=suite)
+
+    try:
+        return q.one()
+    except NoResultFound:
+        return None
+
+__all__.append('get_suite')
+
+################################################################################
+
+class SuiteArchitecture(object):
+    def __init__(self, *args, **kwargs):
+        pass
+
+    def __repr__(self):
+        return '<SuiteArchitecture (%s, %s)>' % (self.suite_id, self.arch_id)
+
+__all__.append('SuiteArchitecture')
+
+@session_wrapper
+def get_suite_architectures(suite, skipsrc=False, skipall=False, session=None):
+    """
+    Returns list of Architecture objects for given C{suite} name
+
+    @type source: str
+    @param source: Suite name to search for
+
+    @type skipsrc: boolean
+    @param skipsrc: Whether to skip returning the 'source' architecture entry
+    (Default False)
+
+    @type skipall: boolean
+    @param skipall: Whether to skip returning the 'all' architecture entry
+    (Default False)
+
+    @type session: Session
+    @param session: Optional SQL session object (a temporary one will be
+    generated if not supplied)
+
+    @rtype: list
+    @return: list of Architecture objects for the given name (may be empty)
+    """
+
+    q = session.query(Architecture)
+    q = q.join(SuiteArchitecture)
+    q = q.join(Suite).filter_by(suite_name=suite)
+
+    if skipsrc:
+        q = q.filter(Architecture.arch_string != 'source')
+
+    if skipall:
+        q = q.filter(Architecture.arch_string != 'all')
+
+    q = q.order_by('arch_string')
+
+    return q.all()
+
+__all__.append('get_suite_architectures')
+
+################################################################################
+
+class SuiteSrcFormat(object):
+    def __init__(self, *args, **kwargs):
+        pass
+
+    def __repr__(self):
+        return '<SuiteSrcFormat (%s, %s)>' % (self.suite_id, self.src_format_id)
+
+__all__.append('SuiteSrcFormat')
+
+def get_suite_src_formats(suite, session=None):
+    """
+    Returns list of allowed SrcFormat for C{suite}.
+
+    @type suite: str
+    @param suite: Suite name to search for
+
+    @type session: Session
+    @param session: Optional SQL session object (a temporary one will be
+    generated if not supplied)
+
+    @rtype: list
+    @return: the list of allowed source formats for I{suite}
+    """
+
+    privatetrans = False
+    if session is None:
+        session = DBConn().session()
+        privatetrans = True
+
+    q = session.query(SrcFormat)
+    q = q.join(SuiteSrcFormat)
+    q = q.join(Suite).filter_by(suite_name=suite)
+    q = q.order_by('format_name')
+
+    ret = q.all()
+
+    if privatetrans:
+        session.close()
+
+    return ret
+
+__all__.append('get_suite_src_formats')
+
+################################################################################
+
+class Uid(object):
+    def __init__(self, *args, **kwargs):
+        pass
+
+    def __eq__(self, val):
+        if isinstance(val, str):
+            return (self.uid == val)
+        # This signals to use the normal comparison operator
+        return NotImplemented
+
+    def __ne__(self, val):
+        if isinstance(val, str):
+            return (self.uid != val)
+        # This signals to use the normal comparison operator
+        return NotImplemented
+
+    def __repr__(self):
+        return '<Uid %s (%s)>' % (self.uid, self.name)
+
+__all__.append('Uid')
+
+def add_database_user(uidname, session=None):
+    """
+    Adds a database user
+
+    @type uidname: string
+    @param uidname: The uid of the user to add
+
+    @type session: SQLAlchemy
+    @param session: Optional SQL session object (a temporary one will be
+    generated if not supplied).  If not passed, a commit will be performed at
+    the end of the function, otherwise the caller is responsible for commiting.
+
+    @rtype: Uid
+    @return: the uid object for the given uidname
+    """
+
+    privatetrans = False
+    if session is None:
+        session = DBConn().session()
+        privatetrans = True
+
+    session.execute("CREATE USER :uid", {'uid': uidname})
+
+    if privatetrans:
+        session.commit()
+        session.close()
+
+__all__.append('add_database_user')
+
+def get_or_set_uid(uidname, session=None):
+    """
+    Returns uid object for given uidname.
+
+    If no matching uidname is found, a row is inserted.
+
+    @type uidname: string
+    @param uidname: The uid to add
+
+    @type session: SQLAlchemy
+    @param session: Optional SQL session object (a temporary one will be
+    generated if not supplied).  If not passed, a commit will be performed at
+    the end of the function, otherwise the caller is responsible for commiting.
+
+    @rtype: Uid
+    @return: the uid object for the given uidname
+    """
+
+    privatetrans = False
+    if session is None:
+        session = DBConn().session()
+        privatetrans = True
+
+    q = session.query(Uid).filter_by(uid=uidname)
+
+    try:
+        ret = q.one()
+    except NoResultFound:
+        uid = Uid()
+        uid.uid = uidname
+        session.add(uid)
+        if privatetrans:
+            session.commit()
+        else:
+            session.flush()
+        ret = uid
+
+    if privatetrans:
+        session.close()
+
+    return ret
+
+__all__.append('get_or_set_uid')
+
+@session_wrapper
+def get_uid_from_fingerprint(fpr, session=None):
+    q = session.query(Uid)
+    q = q.join(Fingerprint).filter_by(fingerprint=fpr)
+
+    try:
+        return q.one()
+    except NoResultFound:
+        return None
+
+__all__.append('get_uid_from_fingerprint')
+
+################################################################################
+
+class DBConn(Singleton):
+    """
+    database module init.
+    """
+    def __init__(self, *args, **kwargs):
+        super(DBConn, self).__init__(*args, **kwargs)
+
+    def _startup(self, *args, **kwargs):
+        self.debug = False
+        if kwargs.has_key('debug'):
+            self.debug = True
+        self.__createconn()
+
+    def __setuptables(self):
+        self.tbl_architecture = Table('architecture', self.db_meta, autoload=True)
+        self.tbl_archive = Table('archive', self.db_meta, autoload=True)
+        self.tbl_bin_associations = Table('bin_associations', self.db_meta, autoload=True)
+        self.tbl_binaries = Table('binaries', self.db_meta, autoload=True)
+        self.tbl_component = Table('component', self.db_meta, autoload=True)
+        self.tbl_config = Table('config', self.db_meta, autoload=True)
+        self.tbl_content_associations = Table('content_associations', self.db_meta, autoload=True)
+        self.tbl_content_file_names = Table('content_file_names', self.db_meta, autoload=True)
+        self.tbl_content_file_paths = Table('content_file_paths', self.db_meta, autoload=True)
+        self.tbl_dsc_files = Table('dsc_files', self.db_meta, autoload=True)
+        self.tbl_files = Table('files', self.db_meta, autoload=True)
+        self.tbl_fingerprint = Table('fingerprint', self.db_meta, autoload=True)
+        self.tbl_keyrings = Table('keyrings', self.db_meta, autoload=True)
+        self.tbl_location = Table('location', self.db_meta, autoload=True)
+        self.tbl_maintainer = Table('maintainer', self.db_meta, autoload=True)
+        self.tbl_new_comments = Table('new_comments', self.db_meta, autoload=True)
+        self.tbl_override = Table('override', self.db_meta, autoload=True)
+        self.tbl_override_type = Table('override_type', self.db_meta, autoload=True)
+        self.tbl_pending_content_associations = Table('pending_content_associations', self.db_meta, autoload=True)
+        self.tbl_priority = Table('priority', self.db_meta, autoload=True)
+        self.tbl_queue = Table('queue', self.db_meta, autoload=True)
+        self.tbl_queue_build = Table('queue_build', self.db_meta, autoload=True)
+        self.tbl_section = Table('section', self.db_meta, autoload=True)
+        self.tbl_source = Table('source', self.db_meta, autoload=True)
+        self.tbl_src_associations = Table('src_associations', self.db_meta, autoload=True)
+        self.tbl_src_format = Table('src_format', self.db_meta, autoload=True)
+        self.tbl_src_uploaders = Table('src_uploaders', self.db_meta, autoload=True)
+        self.tbl_suite = Table('suite', self.db_meta, autoload=True)
+        self.tbl_suite_architectures = Table('suite_architectures', self.db_meta, autoload=True)
+        self.tbl_suite_src_formats = Table('suite_src_formats', self.db_meta, autoload=True)
+        self.tbl_uid = Table('uid', self.db_meta, autoload=True)
+
+    def __setupmappers(self):
+        mapper(Architecture, self.tbl_architecture,
+               properties = dict(arch_id = self.tbl_architecture.c.id))
+
+        mapper(Archive, self.tbl_archive,
+               properties = dict(archive_id = self.tbl_archive.c.id,
+                                 archive_name = self.tbl_archive.c.name))
+
+        mapper(BinAssociation, self.tbl_bin_associations,
+               properties = dict(ba_id = self.tbl_bin_associations.c.id,
+                                 suite_id = self.tbl_bin_associations.c.suite,
+                                 suite = relation(Suite),
+                                 binary_id = self.tbl_bin_associations.c.bin,
+                                 binary = relation(DBBinary)))
+
+        mapper(DBBinary, self.tbl_binaries,
+               properties = dict(binary_id = self.tbl_binaries.c.id,
+                                 package = self.tbl_binaries.c.package,
+                                 version = self.tbl_binaries.c.version,
+                                 maintainer_id = self.tbl_binaries.c.maintainer,
+                                 maintainer = relation(Maintainer),
+                                 source_id = self.tbl_binaries.c.source,
+                                 source = relation(DBSource),
+                                 arch_id = self.tbl_binaries.c.architecture,
+                                 architecture = relation(Architecture),
+                                 poolfile_id = self.tbl_binaries.c.file,
+                                 poolfile = relation(PoolFile),
+                                 binarytype = self.tbl_binaries.c.type,
+                                 fingerprint_id = self.tbl_binaries.c.sig_fpr,
+                                 fingerprint = relation(Fingerprint),
+                                 install_date = self.tbl_binaries.c.install_date,
+                                 binassociations = relation(BinAssociation,
+                                                            primaryjoin=(self.tbl_binaries.c.id==self.tbl_bin_associations.c.bin))))
+
+        mapper(Component, self.tbl_component,
+               properties = dict(component_id = self.tbl_component.c.id,
+                                 component_name = self.tbl_component.c.name))
+
+        mapper(DBConfig, self.tbl_config,
+               properties = dict(config_id = self.tbl_config.c.id))
+
+        mapper(ContentAssociation, self.tbl_content_associations,
+               properties = dict(ca_id = self.tbl_content_associations.c.id,
+                                 filename_id = self.tbl_content_associations.c.filename,
+                                 filename    = relation(ContentFilename),
+                                 filepath_id = self.tbl_content_associations.c.filepath,
+                                 filepath    = relation(ContentFilepath),
+                                 binary_id   = self.tbl_content_associations.c.binary_pkg,
+                                 binary      = relation(DBBinary)))
+
+
+        mapper(ContentFilename, self.tbl_content_file_names,
+               properties = dict(cafilename_id = self.tbl_content_file_names.c.id,
+                                 filename = self.tbl_content_file_names.c.file))
+
+        mapper(ContentFilepath, self.tbl_content_file_paths,
+               properties = dict(cafilepath_id = self.tbl_content_file_paths.c.id,
+                                 filepath = self.tbl_content_file_paths.c.path))
+
+        mapper(DSCFile, self.tbl_dsc_files,
+               properties = dict(dscfile_id = self.tbl_dsc_files.c.id,
+                                 source_id = self.tbl_dsc_files.c.source,
+                                 source = relation(DBSource),
+                                 poolfile_id = self.tbl_dsc_files.c.file,
+                                 poolfile = relation(PoolFile)))
+
+        mapper(PoolFile, self.tbl_files,
+               properties = dict(file_id = self.tbl_files.c.id,
+                                 filesize = self.tbl_files.c.size,
+                                 location_id = self.tbl_files.c.location,
+                                 location = relation(Location)))
+
+        mapper(Fingerprint, self.tbl_fingerprint,
+               properties = dict(fingerprint_id = self.tbl_fingerprint.c.id,
+                                 uid_id = self.tbl_fingerprint.c.uid,
+                                 uid = relation(Uid),
+                                 keyring_id = self.tbl_fingerprint.c.keyring,
+                                 keyring = relation(Keyring)))
+
+        mapper(Keyring, self.tbl_keyrings,
+               properties = dict(keyring_name = self.tbl_keyrings.c.name,
+                                 keyring_id = self.tbl_keyrings.c.id))
+
+        mapper(Location, self.tbl_location,
+               properties = dict(location_id = self.tbl_location.c.id,
+                                 component_id = self.tbl_location.c.component,
+                                 component = relation(Component),
+                                 archive_id = self.tbl_location.c.archive,
+                                 archive = relation(Archive),
+                                 archive_type = self.tbl_location.c.type))
+
+        mapper(Maintainer, self.tbl_maintainer,
+               properties = dict(maintainer_id = self.tbl_maintainer.c.id))
+
+        mapper(NewComment, self.tbl_new_comments,
+               properties = dict(comment_id = self.tbl_new_comments.c.id))
+
+        mapper(Override, self.tbl_override,
+               properties = dict(suite_id = self.tbl_override.c.suite,
+                                 suite = relation(Suite),
+                                 component_id = self.tbl_override.c.component,
+                                 component = relation(Component),
+                                 priority_id = self.tbl_override.c.priority,
+                                 priority = relation(Priority),
+                                 section_id = self.tbl_override.c.section,
+                                 section = relation(Section),
+                                 overridetype_id = self.tbl_override.c.type,
+                                 overridetype = relation(OverrideType)))
+
+        mapper(OverrideType, self.tbl_override_type,
+               properties = dict(overridetype = self.tbl_override_type.c.type,
+                                 overridetype_id = self.tbl_override_type.c.id))
+
+        mapper(PendingContentAssociation, self.tbl_pending_content_associations,
+               properties = dict(pca_id = self.tbl_pending_content_associations.c.id,
+                                 filepath_id = self.tbl_pending_content_associations.c.filepath,
+                                 filepath = relation(ContentFilepath),
+                                 filename_id = self.tbl_pending_content_associations.c.filename,
+                                 filename = relation(ContentFilename)))
+
+        mapper(Priority, self.tbl_priority,
+               properties = dict(priority_id = self.tbl_priority.c.id))
+
+        mapper(Queue, self.tbl_queue,
+               properties = dict(queue_id = self.tbl_queue.c.id))
+
+        mapper(QueueBuild, self.tbl_queue_build,
+               properties = dict(suite_id = self.tbl_queue_build.c.suite,
+                                 queue_id = self.tbl_queue_build.c.queue,
+                                 queue = relation(Queue, backref='queuebuild')))
+
+        mapper(Section, self.tbl_section,
+               properties = dict(section_id = self.tbl_section.c.id))
+
+        mapper(DBSource, self.tbl_source,
+               properties = dict(source_id = self.tbl_source.c.id,
+                                 version = self.tbl_source.c.version,
+                                 maintainer_id = self.tbl_source.c.maintainer,
+                                 maintainer = relation(Maintainer,
+                                                       primaryjoin=(self.tbl_source.c.maintainer==self.tbl_maintainer.c.id)),
+                                 poolfile_id = self.tbl_source.c.file,
+                                 poolfile = relation(PoolFile),
+                                 fingerprint_id = self.tbl_source.c.sig_fpr,
+                                 fingerprint = relation(Fingerprint),
+                                 changedby_id = self.tbl_source.c.changedby,
+                                 changedby = relation(Maintainer,
+                                                      primaryjoin=(self.tbl_source.c.changedby==self.tbl_maintainer.c.id)),
+                                 srcfiles = relation(DSCFile,
+                                                     primaryjoin=(self.tbl_source.c.id==self.tbl_dsc_files.c.source)),
+                                 srcassociations = relation(SrcAssociation,
+                                                            primaryjoin=(self.tbl_source.c.id==self.tbl_src_associations.c.source))))
+
+        mapper(SrcAssociation, self.tbl_src_associations,
+               properties = dict(sa_id = self.tbl_src_associations.c.id,
+                                 suite_id = self.tbl_src_associations.c.suite,
+                                 suite = relation(Suite),
+                                 source_id = self.tbl_src_associations.c.source,
+                                 source = relation(DBSource)))
+
+        mapper(SrcFormat, self.tbl_src_format,
+               properties = dict(src_format_id = self.tbl_src_format.c.id,
+                                 format_name = self.tbl_src_format.c.format_name))
+
+        mapper(SrcUploader, self.tbl_src_uploaders,
+               properties = dict(uploader_id = self.tbl_src_uploaders.c.id,
+                                 source_id = self.tbl_src_uploaders.c.source,
+                                 source = relation(DBSource,
+                                                   primaryjoin=(self.tbl_src_uploaders.c.source==self.tbl_source.c.id)),
+                                 maintainer_id = self.tbl_src_uploaders.c.maintainer,
+                                 maintainer = relation(Maintainer,
+                                                       primaryjoin=(self.tbl_src_uploaders.c.maintainer==self.tbl_maintainer.c.id))))
+
+        mapper(Suite, self.tbl_suite,
+               properties = dict(suite_id = self.tbl_suite.c.id))
+
+        mapper(SuiteArchitecture, self.tbl_suite_architectures,
+               properties = dict(suite_id = self.tbl_suite_architectures.c.suite,
+                                 suite = relation(Suite, backref='suitearchitectures'),
+                                 arch_id = self.tbl_suite_architectures.c.architecture,
+                                 architecture = relation(Architecture)))
+
+        mapper(SuiteSrcFormat, self.tbl_suite_src_formats,
+               properties = dict(suite_id = self.tbl_suite_src_formats.c.suite,
+                                 suite = relation(Suite, backref='suitesrcformats'),
+                                 src_format_id = self.tbl_suite_src_formats.c.src_format,
+                                 src_format = relation(SrcFormat)))
+
+        mapper(Uid, self.tbl_uid,
+               properties = dict(uid_id = self.tbl_uid.c.id,
+                                 fingerprint = relation(Fingerprint)))
+
+    ## Connection functions
+    def __createconn(self):
+        from config import Config
+        cnf = Config()
+        if cnf["DB::Host"]:
+            # TCP/IP
+            connstr = "postgres://%s" % cnf["DB::Host"]
+            if cnf["DB::Port"] and cnf["DB::Port"] != "-1":
+                connstr += ":%s" % cnf["DB::Port"]
+            connstr += "/%s" % cnf["DB::Name"]
+        else:
+            # Unix Socket
+            connstr = "postgres:///%s" % cnf["DB::Name"]
+            if cnf["DB::Port"] and cnf["DB::Port"] != "-1":
+                connstr += "?port=%s" % cnf["DB::Port"]
+
+        self.db_pg   = create_engine(connstr, echo=self.debug)
+        self.db_meta = MetaData()
+        self.db_meta.bind = self.db_pg
+        self.db_smaker = sessionmaker(bind=self.db_pg,
+                                      autoflush=True,
+                                      autocommit=False)
+
+        self.__setuptables()
+        self.__setupmappers()
+
+    def session(self):
+        return self.db_smaker()
+
+__all__.append('DBConn')
 
-        c.execute("BEGIN WORK")
-        try:
-            arch_id = self.get_architecture_id(package['Architecture'])
-
-            # Remove any already existing recorded files for this package
-            c.execute("""DELETE FROM pending_content_associations
-                         WHERE package=%(Package)s
-                         AND version=%(Version)s
-                         AND architecture=%(ArchID)s""", {'Package': package['Package'],
-                                                          'Version': package['Version'],
-                                                          'ArchID':  arch_id})
-
-            for fullpath in fullpaths:
-                (path, file) = os.path.split(fullpath)
-
-                if path.startswith( "./" ):
-                    path = path[2:]
-                # Get the necessary IDs ...
-                file_id = self.get_or_set_contents_file_id(file)
-                path_id = self.get_or_set_contents_path_id(path)
-
-                c.execute("""INSERT INTO pending_content_associations
-                               (package, version, architecture, filepath, filename)
-                            VALUES (%%(Package)s, %%(Version)s, '%d', '%d', '%d')"""
-                    % (arch_id, path_id, file_id), package )
-
-            c.execute("COMMIT")
-            return True
-        except:
-            traceback.print_exc()
-            c.execute("ROLLBACK")
-            return False