# Backup for a personal website: Hello BackupMonitor !

In the previous post I explained why BackupMonitor has become obsolete on my Ubuntu desktop where DejaDup solves the backup problem.

However there’s still one backup problem that I hadn’t automatized yet: backing up my website !

You’ll tell me that this kind of backup can certainly be handled (usually with an additional fee) by my hosting provider, and that is actually quite right ! However, I still fancy the idea to have my data at home – physically – and doing that doesn’t exclude other backup solutions anyway (the more, the better).

So I’m going to explain how I set up this kind of backup.

So the main idea here is to throw away almost everything that was at the heart of BackupMonitor and just keep the following trivial features:

• timely notifications of a need for backup
• easy registration of custom scripts
• a simplistic GUI with a big button to start launching all those scripts in a row, and reporting success or errors for each script

The custom scripts to be registered as backup “items” can actually be very simple such as the one provided below, that makes it possible to copy files from a distant server to the convention “backup” directory as defined by BackupMonitor.

#!/bin/bash   # Backup data from a remote server # where backups are stored # in a directory indicated # by $HOME/backups/Latest # # WARNING: requires that you set a # ssh key allowing to access your remote # server without a password ! # # Author: Thibauld Nion # License: Public Domain # A name that will identify the data once backed up BKP_NAME=example.com # Server access configuration USERNAME=myusername SERVER_NETLOC=server.example.com # Direcotry where stuff will be downloaded BACKUP_DOWNLOAD_DIR=$BACKUP_MONITOR_DESTINATION_FOLDER/$BKP_NAME if [ -d$BACKUP_DOWNLOAD_DIR ] then echo "Removing previously created download directory $BACKUP_DOWNLOAD_DIR" rm -r$BACKUP_DOWNLOAD_DIR fi   echo "Preparing to download remote backup from $SERVER_NETLOC to$BACKUP_DOWNLOAD_DIR" scp -rq $USERNAME@$SERVER_NETLOC:/home/$USERNAME/backups/Latest$BACKUP_DOWNLOAD_DIR echo "Download finished"

This script is simple but has two requirements:

1. that a ssh-key (and also potentially a ssh-agent) is used to have access to the website’s hosting server without having to type the password
2. that, on the server side, a script regularly makes a backup of the website’s files and databases and stores all the backed up data in a folder named $HOME/backups/Latest In my case, to fulfill the second requirement I adapted a script fromDockStreet media that did a good job in backing up a mysql db and a website’s folder as they are set-up on Dreamhost. A “template” of my adapted script is show below and my main modifications are as follow: • handle several folders (not just a single one) • handle several databases • create and update a symbolic link “$HOME/backups/Latest” pointing to the latest folder where the latest backups actually are
#!/bin/bash   # Site backup # Adapted from http://www.dockstreetmedia.com/2011/03/automatic-site-backups-using-cron-jobs-in-dreamhost/   # specific config variables (EDIT THESE) # # Space separated list of the names of folders # below your home directory, that you want to backup DIRNAMES="topFolderA topFolderB topFolderC" # host for your site's database(s) DBHOST=mysql.example.com # user name for this mysql server DBUSER=myusername # corresponding password (the same you gave # to the applications powering your site (like wordpress etc) DBPASS=xxx # Space separated list of the names of dbs # within DBHOST that you want to backup DBNAMES="sitedb_one sitedb_two sitedb_three"   # other config variables(DO NOT EDIT THESE) NOWDATE=$(date +"%y%m%d") NOWDAY=$(date +"%d") BACKUPDIR=backups MYSQLDUMP="$(which mysqldump)" LOGFILE=$HOME/$BACKUPDIR/log.txt TARGETPATH=$HOME/$BACKUPDIR/$NOWDAY LINK_TO_LATEST=$HOME/$BACKUPDIR/Latest   # check to see if target path exists – if so, delete the old one and create a new one, otherwise just create it if [ -d $TARGETPATH ] then # print a message for the logfile / output email echo "$NOWDATE - Removing previous backup $TARGETPATH" | tee -a$LOGFILE rm -r $TARGETPATH mkdir -p$TARGETPATH else mkdir -p $TARGETPATH fi for SITEDIR in$DIRNAMES do   # create a GZIP of the directory inside the target path tar -zcf $TARGETPATH/${SITEDIR}_$NOWDATE.tar.gz$HOME/$SITEDIR # print a message for the logfile / output email echo "$NOWDATE - $SITEDIR has been backed up" | tee -a$LOGFILE   done   for DBNAME in $DBNAMES do # dump the data into a SQL file inside the target path$MYSQLDUMP -u $DBUSER -h$DBHOST -p$DBPASS$DBNAME | gzip &gt; $TARGETPATH/${DBNAME}_$NOWDATE.sql.gz # print a message for the logfile / output email echo "$NOWDATE - $DBNAME has been backed up" | tee -a$LOGFILE done   # Put a link to the latest backup if [ -e $LINK_TO_LATEST ] then rm$LINK_TO_LATEST fi   ln -s $TARGETPATH$LINK_TO_LATEST

The following package backup_monitor_servers.tar.gz gather both scripts as well as the custom configuration file of BackupMonitor that must be put in ~/.config/backup-monitor/.

Of course, I could still use BackupMonitor for both kinds of backups (fetching remote backups and backing up local folders). That would be a little more elegant and I’m not pretending that it’s a lot of fun to have two tools doing basically the same stuff.

However I’m definitely hoping that a polished tool will surface (or maybe just new functionalities to DejaDup) to help doing all those backups and to let me stop maintaining the couple of scripts that make BackupMonitor.

Until then, the same trick (but with different scripts) should make it possible to backup personal data from other sources like social networks (Delicious, Google, Facebook, LastFM etc).