mirror of
https://github.com/offen/docker-volume-backup.git
synced 2025-12-06 09:38:01 +01:00
Compare commits
9 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
4c84674650 | ||
|
|
6fe81cdf2d | ||
|
|
b7ba0e08df | ||
|
|
b558a57de9 | ||
|
|
278df9b2f7 | ||
|
|
0782af88f4 | ||
|
|
f82577fcb5 | ||
|
|
7f261a54b6 | ||
|
|
0069faa7fd |
@@ -11,6 +11,10 @@ jobs:
|
|||||||
name: Build
|
name: Build
|
||||||
command: |
|
command: |
|
||||||
docker build . -t offen/docker-volume-backup:canary
|
docker build . -t offen/docker-volume-backup:canary
|
||||||
|
- run:
|
||||||
|
name: Install gnupg
|
||||||
|
command: |
|
||||||
|
sudo apt-get install -y gnupg
|
||||||
- run:
|
- run:
|
||||||
name: Run tests
|
name: Run tests
|
||||||
working_directory: ~/docker-volume-backup/test
|
working_directory: ~/docker-volume-backup/test
|
||||||
|
|||||||
14
README.md
14
README.md
@@ -51,11 +51,12 @@ AWS_S3_BUCKET_NAME="<xxx>"
|
|||||||
|
|
||||||
# **IMPORTANT, PLEASE READ THIS BEFORE USING THIS FEATURE**:
|
# **IMPORTANT, PLEASE READ THIS BEFORE USING THIS FEATURE**:
|
||||||
# The mechanism used for pruning backups is not very sophisticated
|
# The mechanism used for pruning backups is not very sophisticated
|
||||||
# and applies its rules to **all files in the target directory**,
|
# and applies its rules to **all files in the target directory** by default,
|
||||||
# which means that if you are storing your backups next to other files,
|
# which means that if you are storing your backups next to other files,
|
||||||
# these might become subject to deletion too. When using this option
|
# these might become subject to deletion too. When using this option
|
||||||
# make sure the backup files are stored in a directory used exclusively
|
# make sure the backup files are stored in a directory used exclusively
|
||||||
# for storing them or you might lose data.
|
# for storing them or to configure BACKUP_PRUNING_PREFIX to limit
|
||||||
|
# removal to certain files.
|
||||||
|
|
||||||
# Define this value to enable automatic pruning of old backups. The value
|
# Define this value to enable automatic pruning of old backups. The value
|
||||||
# declares the number of days for which a backup is kept.
|
# declares the number of days for which a backup is kept.
|
||||||
@@ -71,6 +72,15 @@ AWS_S3_BUCKET_NAME="<xxx>"
|
|||||||
|
|
||||||
# BACKUP_PRUNING_LEEWAY="10m"
|
# BACKUP_PRUNING_LEEWAY="10m"
|
||||||
|
|
||||||
|
# In case your target bucket or directory contains other files than the ones
|
||||||
|
# managed by this container, you can limit the scope of rotation by setting
|
||||||
|
# a prefix value. This would usually be the non-parametrized part of your
|
||||||
|
# BACKUP_FILENAME. E.g. if BACKUP_FILENAME is `db-backup-%Y-%m-%dT%H-%M-%S.tar.gz`,
|
||||||
|
# you can set BACKUP_PRUNING_PREFIX to `db-backup-` and make sure
|
||||||
|
# unrelated files are not affected.
|
||||||
|
|
||||||
|
# BACKUP_PRUNING_PREFIX="backup-"
|
||||||
|
|
||||||
########### BACKUP ENCRYPTION
|
########### BACKUP ENCRYPTION
|
||||||
|
|
||||||
# Backups can be encrypted using gpg in case a passphrase is given
|
# Backups can be encrypted using gpg in case a passphrase is given
|
||||||
|
|||||||
@@ -103,32 +103,67 @@ fi
|
|||||||
info "Backup finished"
|
info "Backup finished"
|
||||||
echo "Will wait for next scheduled backup."
|
echo "Will wait for next scheduled backup."
|
||||||
|
|
||||||
prune () {
|
probe_expired () {
|
||||||
target=$1
|
local target=$1
|
||||||
rule_applies_to=$(
|
local is_local=$2
|
||||||
|
if [ -z "$is_local" ]; then
|
||||||
|
if [ ! -z "$BACKUP_PRUNING_PREFIX" ]; then
|
||||||
|
target="${target}/${BACKUP_PRUNING_PREFIX}"
|
||||||
|
fi
|
||||||
mc rm $MC_GLOBAL_OPTIONS --fake --recursive --force \
|
mc rm $MC_GLOBAL_OPTIONS --fake --recursive --force \
|
||||||
--older-than "${BACKUP_RETENTION_DAYS}d" \
|
--older-than "${BACKUP_RETENTION_DAYS}d" \
|
||||||
"$target" \
|
"$target"
|
||||||
| wc -l
|
else
|
||||||
)
|
find $target -name "${BACKUP_PRUNING_PREFIX:-*}" -type f -mtime "+${BACKUP_RETENTION_DAYS}"
|
||||||
|
fi
|
||||||
|
}
|
||||||
|
|
||||||
|
probe_all () {
|
||||||
|
local target=$1
|
||||||
|
local is_local=$2
|
||||||
|
if [ -z "$is_local" ]; then
|
||||||
|
if [ ! -z "$BACKUP_PRUNING_PREFIX" ]; then
|
||||||
|
target="${target}/${BACKUP_PRUNING_PREFIX}"
|
||||||
|
fi
|
||||||
|
mc ls $MC_GLOBAL_OPTIONS "$target"
|
||||||
|
else
|
||||||
|
find $target -name "${BACKUP_PRUNING_PREFIX:-*}" -type f
|
||||||
|
fi
|
||||||
|
}
|
||||||
|
|
||||||
|
delete () {
|
||||||
|
local target=$1
|
||||||
|
local is_local=$2
|
||||||
|
if [ -z "$is_local" ]; then
|
||||||
|
if [ ! -z "$BACKUP_PRUNING_PREFIX" ]; then
|
||||||
|
target="${target}/${BACKUP_PRUNING_PREFIX}"
|
||||||
|
fi
|
||||||
|
mc rm $MC_GLOBAL_OPTIONS --recursive --force \
|
||||||
|
--older-than "${BACKUP_RETENTION_DAYS}d" \
|
||||||
|
"$target"
|
||||||
|
else
|
||||||
|
find $target -name "${BACKUP_PRUNING_PREFIX:-*}" -type f -mtime "+${BACKUP_RETENTION_DAYS}" -delete
|
||||||
|
fi
|
||||||
|
}
|
||||||
|
|
||||||
|
prune () {
|
||||||
|
local target=$1
|
||||||
|
local is_local=$2
|
||||||
|
rule_applies_to=$(probe_expired "$target" "$is_local" | wc -l)
|
||||||
if [ "$rule_applies_to" == "0" ]; then
|
if [ "$rule_applies_to" == "0" ]; then
|
||||||
echo "No backups found older than the configured retention period of $BACKUP_RETENTION_DAYS days."
|
echo "No backups found older than the configured retention period of ${BACKUP_RETENTION_DAYS} days."
|
||||||
echo "Doing nothing."
|
echo "Doing nothing."
|
||||||
exit 0
|
else
|
||||||
|
total=$(probe_all "$target" "$is_local" | wc -l)
|
||||||
|
|
||||||
|
if [ "$rule_applies_to" == "$total" ]; then
|
||||||
|
echo "Using a retention of ${BACKUP_RETENTION_DAYS} days would prune all currently existing backups, will not continue."
|
||||||
|
echo "If this is what you want, please remove files manually instead of using this script."
|
||||||
|
else
|
||||||
|
delete "$target" "$is_local"
|
||||||
|
echo "Successfully pruned ${rule_applies_to} backups older than ${BACKUP_RETENTION_DAYS} days."
|
||||||
|
fi
|
||||||
fi
|
fi
|
||||||
|
|
||||||
total=$(mc ls $MC_GLOBAL_OPTIONS "$target" | wc -l)
|
|
||||||
|
|
||||||
if [ "$rule_applies_to" == "$total" ]; then
|
|
||||||
echo "Using a retention of ${BACKUP_RETENTION_DAYS} days would prune all currently existing backups, will not continue."
|
|
||||||
echo "If this is what you want, please remove files manually instead of using this script."
|
|
||||||
exit 1
|
|
||||||
fi
|
|
||||||
|
|
||||||
mc rm $MC_GLOBAL_OPTIONS \
|
|
||||||
--recursive --force \
|
|
||||||
--older-than "${BACKUP_RETENTION_DAYS}d" "$target"
|
|
||||||
echo "Successfully pruned ${rule_applies_to} backups older than ${BACKUP_RETENTION_DAYS} days."
|
|
||||||
}
|
}
|
||||||
|
|
||||||
if [ ! -z "$BACKUP_RETENTION_DAYS" ]; then
|
if [ ! -z "$BACKUP_RETENTION_DAYS" ]; then
|
||||||
@@ -137,10 +172,10 @@ if [ ! -z "$BACKUP_RETENTION_DAYS" ]; then
|
|||||||
sleep "$BACKUP_PRUNING_LEEWAY"
|
sleep "$BACKUP_PRUNING_LEEWAY"
|
||||||
if [ ! -z "$AWS_S3_BUCKET_NAME" ]; then
|
if [ ! -z "$AWS_S3_BUCKET_NAME" ]; then
|
||||||
info "Pruning old backups from remote storage"
|
info "Pruning old backups from remote storage"
|
||||||
prune "backup-target/$bucket"
|
prune "backup-target/$AWS_S3_BUCKET_NAME"
|
||||||
fi
|
fi
|
||||||
if [ -d "$BACKUP_ARCHIVE" ]; then
|
if [ -d "$BACKUP_ARCHIVE" ]; then
|
||||||
info "Pruning old backups from local archive"
|
info "Pruning old backups from local archive"
|
||||||
prune "$BACKUP_ARCHIVE"
|
prune "$BACKUP_ARCHIVE" "local"
|
||||||
fi
|
fi
|
||||||
fi
|
fi
|
||||||
|
|||||||
@@ -17,6 +17,7 @@ BACKUP_ARCHIVE="${BACKUP_ARCHIVE:-/archive}"
|
|||||||
|
|
||||||
BACKUP_RETENTION_DAYS="${BACKUP_RETENTION_DAYS:-}"
|
BACKUP_RETENTION_DAYS="${BACKUP_RETENTION_DAYS:-}"
|
||||||
BACKUP_PRUNING_LEEWAY="${BACKUP_PRUNING_LEEWAY:-10m}"
|
BACKUP_PRUNING_LEEWAY="${BACKUP_PRUNING_LEEWAY:-10m}"
|
||||||
|
BACKUP_PRUNING_PREFIX="${BACKUP_PRUNING_PREFIX:-}"
|
||||||
BACKUP_STOP_CONTAINER_LABEL="${BACKUP_STOP_CONTAINER_LABEL:-true}"
|
BACKUP_STOP_CONTAINER_LABEL="${BACKUP_STOP_CONTAINER_LABEL:-true}"
|
||||||
|
|
||||||
AWS_S3_BUCKET_NAME="${AWS_S3_BUCKET_NAME:-}"
|
AWS_S3_BUCKET_NAME="${AWS_S3_BUCKET_NAME:-}"
|
||||||
|
|||||||
@@ -25,6 +25,10 @@ services:
|
|||||||
AWS_S3_BUCKET_NAME: backup
|
AWS_S3_BUCKET_NAME: backup
|
||||||
BACKUP_FILENAME: test.tar.gz
|
BACKUP_FILENAME: test.tar.gz
|
||||||
BACKUP_CRON_EXPRESSION: 0 0 5 31 2 ?
|
BACKUP_CRON_EXPRESSION: 0 0 5 31 2 ?
|
||||||
|
BACKUP_RETENTION_DAYS: ${BACKUP_RETENTION_DAYS:-7}
|
||||||
|
BACKUP_PRUNING_LEEWAY: 5s
|
||||||
|
BACKUP_PRUNING_PREFIX: test
|
||||||
|
GPG_PASSPHRASE: 1234secret
|
||||||
volumes:
|
volumes:
|
||||||
- ./local:/archive
|
- ./local:/archive
|
||||||
- app_data:/backup/app_data:ro
|
- app_data:/backup/app_data:ro
|
||||||
|
|||||||
@@ -13,11 +13,13 @@ docker-compose exec backup backup
|
|||||||
|
|
||||||
docker run --rm -it \
|
docker run --rm -it \
|
||||||
-v compose_backup_data:/data alpine \
|
-v compose_backup_data:/data alpine \
|
||||||
ash -c 'tar -xf /data/backup/test.tar.gz && test -f /backup/app_data/offen.db'
|
ash -c 'apk add gnupg && echo 1234secret | gpg -d --pinentry-mode loopback --passphrase-fd 0 --yes /data/backup/test.tar.gz.gpg > /tmp/test.tar.gz && tar -xf /tmp/test.tar.gz -C /tmp && test -f /tmp/backup/app_data/offen.db'
|
||||||
|
|
||||||
echo "[TEST:PASS] Found relevant files in untared remote backup."
|
echo "[TEST:PASS] Found relevant files in untared remote backup."
|
||||||
|
|
||||||
tar -xf ./local/test.tar.gz -C /tmp && test -f /tmp/backup/app_data/offen.db
|
echo 1234secret | gpg -d --yes --passphrase-fd 0 ./local/test.tar.gz.gpg > ./local/decrypted.tar.gz
|
||||||
|
tar -xf ./local/decrypted.tar.gz -C /tmp && test -f /tmp/backup/app_data/offen.db
|
||||||
|
rm ./local/decrypted.tar.gz
|
||||||
|
|
||||||
echo "[TEST:PASS] Found relevant files in untared local backup."
|
echo "[TEST:PASS] Found relevant files in untared local backup."
|
||||||
|
|
||||||
@@ -29,4 +31,25 @@ fi
|
|||||||
|
|
||||||
echo "[TEST:PASS] All containers running post backup."
|
echo "[TEST:PASS] All containers running post backup."
|
||||||
|
|
||||||
|
# The second part of this test checks if backups get deleted when the retention
|
||||||
|
# is set to 0 days (which it should not as it would mean all backups get deleted)
|
||||||
|
# TODO: find out if we can test actual deletion without having to wait for a day
|
||||||
|
BACKUP_RETENTION_DAYS="0" docker-compose up -d
|
||||||
|
sleep 5
|
||||||
|
|
||||||
|
docker-compose exec backup backup
|
||||||
|
|
||||||
|
docker run --rm -it \
|
||||||
|
-v compose_backup_data:/data alpine \
|
||||||
|
ash -c '[ $(find /data/backup/ -type f | wc -l) = "1" ]'
|
||||||
|
|
||||||
|
echo "[TEST:PASS] Remote backups have not been deleted."
|
||||||
|
|
||||||
|
if [ "$(find ./local -type f | wc -l)" != "1" ]; then
|
||||||
|
echo "[TEST:FAIL] Backups should not have been deleted, instead seen:"
|
||||||
|
find ./local -type f
|
||||||
|
fi
|
||||||
|
|
||||||
|
echo "[TEST:PASS] Local backups have not been deleted."
|
||||||
|
|
||||||
docker-compose down --volumes
|
docker-compose down --volumes
|
||||||
|
|||||||
@@ -33,6 +33,8 @@ services:
|
|||||||
AWS_S3_BUCKET_NAME: backup
|
AWS_S3_BUCKET_NAME: backup
|
||||||
BACKUP_FILENAME: test.tar.gz
|
BACKUP_FILENAME: test.tar.gz
|
||||||
BACKUP_CRON_EXPRESSION: 0 0 5 31 2 ?
|
BACKUP_CRON_EXPRESSION: 0 0 5 31 2 ?
|
||||||
|
BACKUP_RETENTION_DAYS: 7
|
||||||
|
BACKUP_PRUNING_LEEWAY: 5s
|
||||||
volumes:
|
volumes:
|
||||||
- pg_data:/backup/pg_data:ro
|
- pg_data:/backup/pg_data:ro
|
||||||
- /var/run/docker.sock:/var/run/docker.sock
|
- /var/run/docker.sock:/var/run/docker.sock
|
||||||
|
|||||||
Reference in New Issue
Block a user