For a while now, I’ve been backing up the few WordPress blogs that I run for various people with a very simple script that followed this algorithm:

  1. Copy files to a temporary directory.
  2. Dump the MySQL data into a file in that directory.
  3. Tarball it up.
  4. Scp that file to another server that I run.
At the time, I did this because it was the simplest thing that could possibly work.  It didn't depend on any external facility other than mysqldump, tar, and scp.

Well, running that script on a nightly cron filled up my disk allocation on that remote server a couple times, so I got clever with the backup organization so I could quickly remove old backups while keeping sparser (monthly) backups for longer.  This only helped a little, because I was still nervous about deleting backups because I didn’t know what they contained.

I also have been using git more and more recently and I liked the idea of version control that can go in any direction.  So, in the spare bits of time I’ve had in the past few weeks, I wrote  It takes a git repository and does the following:

  • git add <any new or modified files>
  • git rm <any deleted files>
  • git commit
  • git push backup
Now, when the backup is run, only the small changes are sent to the remote server and I can look at the differences by examining the git log.

There are options for dumping database tables, changing the commit message and the remote that gets the push.  Running “ –man” will show all the options.

The source is (of course) in a git repo:

Update 2012-07-31: Project repo moved to github.