Hey guys,
i tried to create a simple rsync script that should create daily backups
from a ZFS storage and put them into a timestamp folder.
After creating the initial full backup, the following backups should
only contain "new data" and the rest will be referenced via hardlinks
(-link-dest)
This was at least a simple enough scenario to achieve it with my
pathetic scripting skills. This is what i came up with:
#!/bin/sh
# rsync copy script for rsync pull from FreeNAS to BackupNAS for Buero
dataset
# Set variables
EXPIRED=`date +"%d-%m-%Y" -d "14 days ago"`
# Copy previous timefile to timeold.txt if it exists
if [ -f "/volume1/rsync/Buero/timenow.txt" ]
then
yes | cp /volume1/rsync/Buero/timenow.txt
/volume1/rsync/Buero/timeold.txt
fi
# Create current timefile
echo `date +"%d-%m-%Y-%H%M"` > /volume1/rsync/Buero/timenow.txt
# rsync command
if [ -f "/volume1/rsync/Buero/timeold.txt" ]
then
rsync -aqzh \
--delete --stats --exclude-from=/volume1/rsync/Buero/exclude.txt \
--log-file=/volume1/Backup_Test/logs/rsync-`date
+"%d-%m-%Y-%H%M"`.log \
--link-dest=/volume1/Backup_Test/`cat
/volume1/rsync/Buero/timeold.txt` \
Test@192.168.2.2::Test/volume1/Backup_Test/`date +"%d-%m-%Y-%H%M"`
else
rsync -aqzh \
--delete --stats --exclude-from=/volume1/rsync/Buero/exclude.txt \
--log-file=/volume1/Backup_Buero/logs/rsync-`date
+"%d-%m-%Y-%H%M"`.log \
Test@192.168.2.2::Test/volume1/Backup_Test/`date +"%d-%m-%Y-%H%M"`
fi
# Delete expired snapshots (2 weeks old)
if [ -d /volume1/Backup_Buero/$EXPIRED-* ]
then
rm -Rf /volume1/Backup_Buero/$EXPIRED-*
fi
Well, it works but there is a huge flaw with his approach and i am not
able to solve it on my own unfortunately.
As long as the backups are finishing properly, everything is fine but as
soon as one backup job couldn`t be finished for some reason, (like it
will be aborted accidently or a power cut occurs)
the whole backup chain is messed up and usually the script creates a new
full backup which fills up my backup storage.
What i would like to achieve is, to improve the script so that a backup
run that wasn`t finished properly will be resumed, next time the script
triggers.
Only if that was successful should the next incremental backup be
created so that the files that didn`t changed from the previous backup
can be hardlinked properly.
I did a little bit of research and i am not sure if i am on the right
track here but apparently this can be done with return codes, but i
honestly don`t know how to do this.
Thank you in advance for your help and sorry if this question may seem
foolish to most of you people.
Regards
Dennis
--
Please use reply-all for most replies to avoid omitting the mailing list.
To unsubscribe or change options: https://lists.samba.org/mailman/listinfo/rsync
Before posting, read: http://www.catb.org/~esr/faqs/smart-questions.html