X-Git-Url: http://git.hcoop.net/clinton/scripts.git/blobdiff_plain/a5a60f0c2c38227fdd12a90c2cd81871baf8d41f..a2983bf8f9241b733bb827cf58ca54768026e986:/s3.common diff --git a/s3.common b/s3.common index 09d1e77..0365beb 100644 --- a/s3.common +++ b/s3.common @@ -1,3 +1,4 @@ +# -*- Shell-Script -*- # Common functions for dealing with Amazon S3. # units for BWLIMIT are KB/s @@ -7,7 +8,6 @@ CHUNKSIZE=5000 BUCKET=hcoop.net-backups BACKUPDIR=full -SUBDIR=$(date +%Y.%m.%d) S3CMD=$(dirname $0)/s3 IFS=$'\n' @@ -33,19 +33,21 @@ function s3_cmd () { function move_over () { # Move file to its offsite destination. # Expects the file to come from STDIN. - # $1: filename, $2: relative directory (optional) - if test -z "$1" || test -n "$3"; then + # $1: date subdirectory + # $2: filename + if test -z "$2" || test -n "$3"; then echo "Bad programming" exit 1 fi - local FILE=$1 - local DEST=$BACKUPDIR/$SUBDIR - if test -n "$2"; then - DEST=$DEST/$2 - fi - split -d -b ${CHUNKSIZE}m - ${FILE}. - for i in ${FILE}.*; do - s3_cmd put $DEST/$i $i + local subdir=$1 + local file=$2 + local dest=$BACKUPDIR/$subdir + local ret + split -d -b ${CHUNKSIZE}m - ${file}. +# for i in ${file}.*; do + for i in *.[0-9][0-9]; do + echo "Transferring $i to S3 ..." + s3_cmd put $dest/$i $i rm -f $i done } @@ -53,6 +55,7 @@ function move_over () { function prune_old_backups () { # Implement me + local subdir=$1 local oldpwd=$PWD cd $BACKUPDIR find . -mindepth 1 -maxdepth 1 -type d -ctime +7 \