#!/bin/sh

CLIENT=client1
OBJECT=domain.nl
BASEDIR=/var/lib/psa/backupscript
BASEDIROBJECT=/var/lib/psa/dumps/.discovered/*_
DATADIR=/var/lib/psa/dumps/resellers/beheer/clients

cd /$BASEDIR
file=”$BASEDIROBJECT$OBJECT”

### bash check if file exists

if [ -f $file ];

### first backup the new stuff

then python s3cmd put -r $DATADIR/$CLIENT/domains/ s3://bucketname/”$CLIENT”_`date +%d%m%Y`/
rm -rf $BASEDIROBJECT$OBJECT
touch “$BASEDIR”/logs/”$CLIENT”_backup-transferred`date +_%d%m%Y`.txt

else echo “nothing to backup”

fi
echo “ready”

So, this is the modified backup script, where I have removed the delete part of the domain, I have created a logfile after ‘successful’ uploading the content, so that with this file you can delete finally the content on the server in another script.
script has not run in crontab yet and I have only adapted it for 1 domain yet. Comments are welcome.

First thanks to Uli for the first step towards some idea’s and Armijn giving good guidelines how to use some variables together with commands.
cause I am a newbie on bin/bash scripting.. but with lots of testing tonight I was able to backup several ‘clients’ content towards buckets and objects on Amazon S3. Probably this is a first good step towards the usage of a simple script for the backup of your server’s content to Amazon S3

Needed: Python, S3cmd, data and the correct settings

note: this script is not finished yet as I am manually executing it and the removal of data from Amazon S3 is not well tested (it works, but I need to check it during the time the stuff is running.

#!/bin/sh

###################################
### ###
### BACKUP FOR