X-Git-Url: https://git.decadent.org.uk/gitweb/?a=blobdiff_plain;f=config%2Fdebian%2Fdinstall.functions;h=6e80479cdf5fe5821738b5102dbf362cba71585c;hb=610b08292f816e81e7dea5446bbdd22effb434d7;hp=a29ae26eb0faaf44219e119f79c2bc272c49be4b;hpb=32f7684f5d703c6de86e1f1601fdb7fd14193a55;p=dak.git diff --git a/config/debian/dinstall.functions b/config/debian/dinstall.functions index a29ae26e..6e80479c 100644 --- a/config/debian/dinstall.functions +++ b/config/debian/dinstall.functions @@ -32,20 +32,20 @@ function onerror() { # pushing merkels QA user, part one function merkel1() { log "Telling merkels QA user that we start dinstall" - ssh -2 -i ~dak/.ssh/push_merkel_qa -o BatchMode=yes -o SetupTimeOut=90 -o ConnectTimeout=90 qa@merkel.debian.org sleep 1 + ssh -2 -i ~dak/.ssh/push_merkel_qa -o BatchMode=yes -o SetupTimeOut=90 -o ConnectTimeout=90 qa@qa.debian.org sleep 1 } # Create the postgres dump files function pgdump_pre() { - log "Creating pre-daily-cron-job backup of projectb database..." - pg_dump projectb > $base/backup/dump_pre_$(date +%Y.%m.%d-%H:%M:%S) + log "Creating pre-daily-cron-job backup of $PGDATABASE database..." + pg_dump > $base/backup/dump_pre_$(date +%Y.%m.%d-%H:%M:%S) } function pgdump_post() { - log "Creating post-daily-cron-job backup of projectb database..." + log "Creating post-daily-cron-job backup of $PGDATABASE database..." cd $base/backup POSTDUMP=$(date +%Y.%m.%d-%H:%M:%S) - pg_dump projectb > $base/backup/dump_$POSTDUMP + pg_dump > $base/backup/dump_$POSTDUMP #pg_dumpall --globals-only > $base/backup/dumpall_$POSTDUMP ln -sf $base/backup/dump_$POSTDUMP current #ln -sf $base/backup/dumpall_$POSTDUMP currentall @@ -53,11 +53,21 @@ function pgdump_post() { # Load the dak-dev projectb function pgdakdev() { + # Make sure to unset any possible psql variables so we don't drop the wrong + # f****** database by accident + local PGDATABASE + unset PGDATABASE + local PGHOST + unset PGHOST + local PGPORT + unset PGPORT + local PGUSER + unset PGUSER cd $base/backup - echo "drop database projectb" | psql -p 5433 template1 + echo "drop database projectb" | psql -p 5434 template1 #cat currentall | psql -p 5433 template1 - createdb -p 5433 -T template1 projectb - fgrep -v '\connect' current | psql -p 5433 projectb + createdb -p 5434 -T template1 projectb + fgrep -v '\connect' current | psql -p 5434 projectb } # Updating various files @@ -189,7 +199,8 @@ function packages() { log "Generating Packages and Sources files" cd $configdir GZIP='--rsyncable' ; export GZIP - apt-ftparchive generate apt.conf + #apt-ftparchive generate apt.conf + dak generate-packages-sources } function pdiff() { @@ -312,8 +323,8 @@ function mkfilesindices() { ARCHLIST=$(tempfile) - log "Querying projectb..." - echo 'SELECT l.path, f.filename, a.arch_string FROM location l JOIN files f ON (f.location = l.id) LEFT OUTER JOIN (binaries b JOIN architecture a ON (b.architecture = a.id)) ON (f.id = b.file)' | psql projectb -At | sed 's/|//;s,^/srv/ftp-master.debian.org/ftp,.,' | sort >$ARCHLIST + log "Querying $PGDATABASE..." + echo 'SELECT l.path, f.filename, a.arch_string FROM location l JOIN files f ON (f.location = l.id) LEFT OUTER JOIN (binaries b JOIN architecture a ON (b.architecture = a.id)) ON (f.id = b.file)' | psql -At | sed 's/|//;s,^/srv/ftp-master.debian.org/ftp,.,' | sort >$ARCHLIST includedirs () { perl -ne 'print; while (m,/[^/]+$,) { $_=$`; print $_ . "\n" unless $d{$_}++; }' @@ -346,12 +357,12 @@ function mkfilesindices() { log "Generating suite lists" suite_list () { - printf 'SELECT DISTINCT l.path, f.filename FROM (SELECT sa.source AS source FROM src_associations sa WHERE sa.suite = %d UNION SELECT b.source AS source FROM bin_associations ba JOIN binaries b ON (ba.bin = b.id) WHERE ba.suite = %d) s JOIN dsc_files df ON (s.source = df.source) JOIN files f ON (df.file = f.id) JOIN location l ON (f.location = l.id)\n' $1 $1 | psql -F' ' -A -t projectb + printf 'SELECT DISTINCT l.path, f.filename FROM (SELECT sa.source AS source FROM src_associations sa WHERE sa.suite = %d UNION SELECT b.source AS source FROM bin_associations ba JOIN binaries b ON (ba.bin = b.id) WHERE ba.suite = %d) s JOIN dsc_files df ON (s.source = df.source) JOIN files f ON (df.file = f.id) JOIN location l ON (f.location = l.id)\n' $1 $1 | psql -F' ' -A -t - printf 'SELECT l.path, f.filename FROM bin_associations ba JOIN binaries b ON (ba.bin = b.id) JOIN files f ON (b.file = f.id) JOIN location l ON (f.location = l.id) WHERE ba.suite = %d\n' $1 | psql -F' ' -A -t projectb + printf 'SELECT l.path, f.filename FROM bin_associations ba JOIN binaries b ON (ba.bin = b.id) JOIN files f ON (b.file = f.id) JOIN location l ON (f.location = l.id) WHERE ba.suite = %d\n' $1 | psql -F' ' -A -t } - printf 'SELECT id, suite_name FROM suite\n' | psql -F' ' -At projectb | + printf 'SELECT id, suite_name FROM suite\n' | psql -F' ' -At | while read id suite; do [ -e $base/ftp/dists/$suite ] || continue ( @@ -458,7 +469,7 @@ function bts() { function merkel2() { # Push dak@merkel so it syncs the projectb there. Returns immediately, the sync runs detached - log "Trigger merkel/flotows projectb sync" + log "Trigger merkel/flotows $PGDATABASE sync" ssh -2 -o BatchMode=yes -o SetupTimeOut=30 -o ConnectTimeout=30 -i ~/.ssh/push_merkel_projectb dak@merkel.debian.org sleep 1 # Also trigger flotow, the ftpmaster test box ssh -2 -o BatchMode=yes -o SetupTimeOut=30 -o ConnectTimeout=30 -i ~/.ssh/push_flotow_projectb dak@flotow.debconf.org sleep 1 @@ -472,9 +483,9 @@ function merkel3() { function mirrorpush() { log "Starting the mirrorpush" - date -u > /srv/ftp-master.debian.org/web/mirrorstart - echo "Using dak v1" >> /srv/ftp-master.debian.org/web/mirrorstart - echo "Running on host $(hostname -f)" >> /srv/ftp-master.debian.org/web/mirrorstart + date -u > /srv/ftp.debian.org/web/mirrorstart + echo "Using dak v1" >> /srv/ftp.debian.org/web/mirrorstart + echo "Running on host $(hostname -f)" >> /srv/ftp.debian.org/web/mirrorstart sudo -H -u archvsync /home/archvsync/runmirrors > ~dak/runmirrors.log 2>&1 & } @@ -593,3 +604,11 @@ Current action: ${1} Action start: ${RIGHTNOW} EOF } + +# extract changelogs and stuff +function changelogs() { + log "Extracting changelogs" + dak make-changelog -e + cd ${exportpublic}/changelogs + rsync -aHW --delete --delete-after --ignore-errors ${exportdir}/changelogs/. . +}