X-Git-Url: http://git.pippins.net/embedvideo/.git/?a=blobdiff_plain;f=zfs-replicate;h=b3f7cffd16d31dc3bbfa44df3ce5b7cdd2840b5f;hb=f1a0eaf6cb15faf514577513f17fa9a06c34d129;hp=09cb3babb7600f0526a14b4b2c2f945f41660ea4;hpb=984d48e66a916fd9ba533ae19b43aafd449dad77;p=zfs-ubuntu%2F.git diff --git a/zfs-replicate b/zfs-replicate index 09cb3ba..b3f7cff 100755 --- a/zfs-replicate +++ b/zfs-replicate @@ -1,61 +1,113 @@ -#/bin/bash +#!/bin/bash +# Author: Carl Baldwin & Alan Pippin +# Description: This script replicates a remote zfs filesystem to a local zfs pool. +# This script will keep all snapshots in sync, removing the ones +# that have been deleted since the last replicate was performed. +# This script will only send the new, or missing, snapshots since +# the last replicate was performed. # Usage: replicate + +# source our configuration +config="${0%/*}/zfs-scripts.conf" +[ -e "${config}.dist" ] && . ${config}.dist +[ -e "${config}" ] && . ${config} + +# command line arg parsing remote=$1 remote_fs=$2 remote_pool=${2%%/*} -remote_lockdir="/tmp/zfs-admin-lock" - -# Set the name of the local pool used to store the backup of the remote -local_pool=backups - -# Set the email address to send notification to -mailto=root@pippins.net - -# When this variable is set, local filesystems will be destroyed -# before receiving a full streams into them from the remote source. -destroy_local_filesystem_on_full_replicate=0 - -# The ssh connection doesn't find zfs without this. -zfs=/usr/sbin/zfs # Setup our cleanup and exit trap cleanup() { - ssh $remote rm -rf "$remote_lockdir" - exit + if [[ -e "$local_list" ]]; then + rm -f $local_list + fi + if [[ -e "$remote_list" ]]; then + rm -f $remote_list + fi + $ssh $remote ls -d "$lockdir" > /dev/null 2>&1 + if [[ $? == 0 ]]; then + $ssh $remote rm -rf "$lockdir" + fi +} +fatal_and_exit() { + echo -e 2>&1 "$1" + # Destroy the current backup markers from the local backup_pool and remote_pool if they exist + if [[ -n "$current_backup_marker" ]]; then + # Local backup pool current backup marker + $zfs list -t snapshot ${backup_pool}/${current_backup_marker} > /dev/null 2>&1 + if [ $? == 0 ]; then + $zfs destroy ${backup_pool}/${current_backup_marker} + fi + # Remote pool current backup marker + $ssh $remote zfs list -t snapshot ${current_backup_marker} > /dev/null 2>&1 + if [ $? == 0 ]; then + $ssh $remote $zfs destroy ${current_backup_marker} + fi + fi + # send email notification + if [[ -n "$2" ]]; then + echo -e "$1" | $mailx -s "zfs replicate on $hostname failed" "$2" + fi + # exit + exit 1 } -trap cleanup INT +trap fatal_and_exit INT trap cleanup EXIT +# Declare a function to handle the replicate operation +replicate() { + zfs_send="$1" + zfs_recv="$zfs receive -vF -d ${backup_pool}/${remote_pool}" + glue="$throttle $throttle_opt" + if [[ $throttle_enable == 1 && -e $throttle ]]; then + # handle using the glue in the local and remote host case properly + if [[ -z "$ssh" ]]; then + # local host glue case + $zfs_send | $glue | $zfs_recv + else + # remote host glue case + $ssh $remote "$zfs_send | $glue" | $zfs_recv + fi + else + # no glue case - works for both the local and remote host cases + $ssh $remote $zfs_send | $zfs_recv + fi + # The return code of the zfs_send | zfs_recv operation will be returned to the caller +} + # Make sure we have valid arguments if [[ -z "$remote" ]] || [[ -z "$remote_fs" ]]; then - echo "Usage: $0 " - exit 1 + fatal_and_exit "Usage: $0 " +fi + +# check for localhost +if [[ $remote = "localhost" ]]; then + remote="" + ssh="" fi -# Make sure the local pool and local receiving filesystem exist, or print some errors -zpool list -H "$local_pool" >/dev/null 2>&1 +# Make sure the local backup pool and local receiving filesystem exist, or print some errors +zpool list -H "$backup_pool" >/dev/null 2>&1 if [ $? != 0 ]; then - echo >&2 "-E- The local pool, '$local_pool' doesn't seem to exist." - exit 1 + fatal_and_exit "-E- The local backup pool on $hostname, '$backup_pool' doesn't seem to exist." $mailto fi -zfs list "$local_pool/$remote_pool" >/dev/null 2>&1 +zfs list "$backup_pool/$remote_pool" >/dev/null 2>&1 if [ $? != 0 ]; then - echo >&2 "-I- The local filesystem for the remote pool, '$local_pool/$remote_pool' doesn't seem to exist." - echo >&2 " Creating the local filesystem to receive the remote pool into: $local_pool/$remote_pool" - $zfs create $local_pool/$remote_pool + echo >&2 "-I- The local filesystem for the remote pool, '$backup_pool/$remote_pool' doesn't seem to exist." + echo >&2 " Creating the local filesystem to receive the remote pool into: $backup_pool/$remote_pool" + $zfs create $backup_pool/$remote_pool if [ $? != 0 ]; then - echo "-E- remote $zfs create command failed" - exit 1 + fatal_and_exit "-E- remote $zfs on $hostname create command failed" $mailto fi fi -# Obtain the zpool guid for the local pool -local_pool_guid=`zpool get guid $local_pool 2>&1 | grep $local_pool | awk '{ print $3 }'` -zpool get guid $local_pool > /dev/null 2>&1 +# Obtain the zpool guid for the local backup pool +backup_pool_guid=`zpool get guid $backup_pool 2>&1 | grep $backup_pool | awk '{ print $3 }'` +zpool get guid $backup_pool > /dev/null 2>&1 if [ $? != 0 ]; then - echo >&2 "-E- Unable to extract the guid for the local pool: $local_pool" - exit 1 + fatal_and_exit "-E- Unable to extract the guid for the local backup pool on $hostname: $backup_pool" $mailto fi # Turn on shell verbosity @@ -64,10 +116,10 @@ set -x # Create the remote lockdir before continuing with the replicate # Spinlock on creating the lock maxsleeptime=60 -maxattempts=100 +maxattempts=500 attempts=0 while true; do - ssh $remote mkdir "$remote_lockdir" >/dev/null 2>&1 + $ssh $remote mkdir "$lockdir" >/dev/null 2>&1 if [ $? != 0 ]; then # Another zfs admin tool is running. # Wait a random amount of time and try again @@ -80,25 +132,24 @@ while true; do fi if [[ $attempts -gt $maxattempts ]]; then # We've exceeded our maximum while loop count - echo "-W- The zfs filesystem has been locked down. Skipping replicate operation." - ssh $remote ls -ld $remote_lockdir | /usr/bin/mailx -s "zfs-replicate-all unable to obtain zfs admin lock" $mailto - exit 1 + echo "-E- The zfs filesystem has been locked down. Skipping replicate operation." + fail_msg=`$ssh $remote ls -ld $lockdir 2>&1` + fatal_and_exit "zfs-replicate-all on $hostname unable to obtain zfs admin lock:\n$fail_msg" $mailto fi done # Setup our backup marker names -current_backup_marker=${remote_fs}@current-backup-${local_pool_guid} -previous_backup_marker=${remote_fs}@previous-backup-${local_pool_guid} +current_backup_marker=${remote_fs}@current-backup-${backup_pool_guid} +previous_backup_marker=${remote_fs}@previous-backup-${backup_pool_guid} # List the snapshots on the remote machine. remote_list=$(mktemp /tmp/replicate.XXXXXX) -ssh $remote \ +$ssh $remote \ $zfs list -H -t snapshot | grep ^${remote_fs}@ | awk '{print$1}' > $remote_list if [ $? != 0 ]; then - echo "-E- remote $zfs list command failed" - exit 1 + fatal_and_exit "-E- remote $zfs list on $hostname command failed" $mailto fi # List the snapshots on the local machine. @@ -106,29 +157,26 @@ fi # If you do, it can mess up the common finding algorithm below. local_list=$(mktemp /tmp/replicate.XXXXXX) $zfs list -H -t snapshot | - grep ^${local_pool}/${remote_fs}@ | - grep -v ^${local_pool}/${current_backup_marker} | - awk "{gsub(/^$local_pool./,\"\",\$1); print\$1}" > $local_list + grep ^${backup_pool}/${remote_fs}@ | + grep -v ^${backup_pool}/${current_backup_marker} | + awk "{gsub(/^$backup_pool./,\"\",\$1); print\$1}" > $local_list if [ $? != 0 ]; then - echo "-E- local $zfs list command failed" - exit 1 + fatal_and_exit "-E- local $zfs list on $hostname command failed" $mailto fi # Destroy the current backup marker snapshot on the remote system if it exists grep -q ${current_backup_marker} $remote_list if [ $? == 0 ]; then - ssh $remote $zfs destroy ${current_backup_marker} + $ssh $remote $zfs destroy ${current_backup_marker} if [ $? != 0 ]; then - echo "-E- remote $zfs destroy command failed" - exit 1 + fatal_and_exit "-E- remote $zfs destroy on $hostname command failed" $mailto fi fi # Create the current backup marker snapshot on the remote system -ssh $remote $zfs snapshot ${current_backup_marker} +$ssh $remote $zfs snapshot ${current_backup_marker} if [ $? != 0 ]; then - echo "-E- remote $zfs snapshot command failed" - exit 1 + fatal_and_exit "-E- remote $zfs snapshot on $hostname command failed" $mailto fi # Check to see if the previous backup marker exists in the remote snapshot list. @@ -144,21 +192,18 @@ no_markers=$(($no_markers || $?)) if [ $no_markers == 0 ]; then # We found backup markers, incrementally send the new snaps - # First, rollback the local pool to the previous backup marker in case the previous + # First, rollback the local backup pool to the previous backup marker in case the previous # backup was interrupted for some reason. If we don't do this, the zfs send -R command # below may complain about snaps already existing as it tries to resend from the # previous backup marker again from a previously interrupted replicate. - $zfs rollback -r ${local_pool}/${previous_backup_marker} + $zfs rollback -r ${backup_pool}/${previous_backup_marker} if [ $? != 0 ]; then - echo "-E- remote incremental $zfs rollback command failed" - exit 1 + fatal_and_exit "-E- remote incremental $zfs rollback command failed on $hostname" $mailto fi # Now it should be safe to send the snaps - ssh $remote $zfs send -R -I${previous_backup_marker} ${current_backup_marker} | - $zfs receive -vF -d ${local_pool}/${remote_pool} + replicate "$zfs send -Rc -I${previous_backup_marker} ${current_backup_marker}" if [ $? != 0 ]; then - echo "-E- remote incremental $zfs send command failed" - exit 1 + fatal_and_exit "-E- remote incremental $zfs send command failed on $hostname" $mailto fi else # We didn't find any backup markers, next check to see if we have a common snapshot. @@ -175,58 +220,46 @@ else if [[ -n "$common" ]]; then # We found a common snapshot, incrementally send the new snaps - ssh $remote $zfs send -R -I${common/*@/@} ${current_backup_marker} | - $zfs receive -vF -d ${local_pool}/${remote_pool} + replicate "$zfs send -Rc -I${common/*@/@} ${current_backup_marker}" if [ $? != 0 ]; then - echo "-E- remote incremental $zfs send command failed" - exit 1 + fatal_and_exit "-E- remote incremental $zfs send command failed on $hostname" $mailto fi else # We did not find any markers or a common snapshot # At this point, we'll have to send the entire filesystem # Destroy the local filesystem if it exists before receving the full replicate - zfs list ${local_pool}/${remote_fs} > /dev/null 2>&1 + zfs list ${backup_pool}/${remote_fs} > /dev/null 2>&1 if [ $? == 0 ]; then if [[ $destroy_local_filesystem_on_full_replicate == 1 ]]; then - $zfs destroy -r ${local_pool}/${remote_fs} + $zfs destroy -r ${backup_pool}/${remote_fs} if [ $? != 0 ]; then - echo "-E- remote full $zfs destroy command failed" - exit 1 + fatal_and_exit "-E- remote full $zfs destroy command failed on $hostname" $mailto fi else echo "-W- We need to destroy a local filesystem before receiving a full stream." echo " However, since the option is set to prevent this, skipping replicate operation." - echo "$zfs destroy -r ${local_pool}/${remote_fs} not able to run" | /usr/bin/mailx -s "zfs-replicate-all unable to destroy local filesystem" $mailto - exit 1 + fatal_and_exit "unable to destroy local filesystem:\n$zfs destroy -r ${backup_pool}/${remote_fs} not able to run on $hostname" $mailto fi fi # Send the full filesystem - ssh $remote $zfs send -R ${current_backup_marker} | - $zfs receive -vF -d ${local_pool}/${remote_pool} + replicate "$zfs send -Rc ${current_backup_marker}" if [ $? != 0 ]; then - echo "-E- remote full $zfs send command failed" - exit 1 + fatal_and_exit "-E- remote full $zfs send command failed on $hostname" $mailto fi fi fi # destroy the previous backup markers now that we've replicated past them # don't check the return codes here because these may not exist, and that is ok -$zfs destroy ${local_pool}/${previous_backup_marker} > /dev/null 2>&1 -ssh $remote $zfs destroy ${previous_backup_marker} > /dev/null 2>&1 +$zfs destroy ${backup_pool}/${previous_backup_marker} > /dev/null 2>&1 +$ssh $remote $zfs destroy ${previous_backup_marker} > /dev/null 2>&1 # Rename the current backup marker to be the previous backup marker -$zfs rename ${local_pool}/${current_backup_marker} ${local_pool}/${previous_backup_marker} +$zfs rename ${backup_pool}/${current_backup_marker} ${backup_pool}/${previous_backup_marker} if [ $? != 0 ]; then - echo "-E- local $zfs rename command failed" - exit 1 + fatal_and_exit "-E- local $zfs rename command failed on $hostname" $mailto fi -ssh $remote $zfs rename ${current_backup_marker} ${previous_backup_marker} +$ssh $remote $zfs rename ${current_backup_marker} ${previous_backup_marker} if [ $? != 0 ]; then - echo "-E- remote $zfs rename command failed" - exit 1 + fatal_and_exit "-E- remote $zfs rename command failed on $hostname" $mailto fi - -# Remove tmp files -rm -f $local_list $remote_list -