Backups on Exoscale
To configure the backups, you first need to create an Exoscale bucket.
For our backups, we have one named ‘swiss-backups’.
In this instance, I chose the CH-GVA-2
zone so that the data remains in Geneva.
Exoscale has great instructions on how to connect to a bucket.
First, you must install s3cmd
to connect to the Exoscale Bucket (as root
):
apt-get update && apt-get install s3cmd
s3cmd
enables us to create and use Exoscale buckets. Our configuration file (located at /home/ubuntu/.s3cfg
) looks like this:
[default]
host_base = sos-ch-gva-2.exo.io
host_bucket = %(bucket)s.sos-ch-gva-2.exo.io
access_key = $EXO_SOS_KEY
secret_key = $EXO_SOS_SECRET
use_https = True
(Note: $EXO_SOS_KEY and $EXO_SOS_SECRET are hidden for this guide)
I created an IAM Role named ‘Create Backups’ that will be used to connect the compute instances to the bucket. Once this was done, I generated an API key pair for that role.
Now that this is configured, you can place files into the bucket by simply running s3cmd put ${file_name} s3://swiss-backups/${file_name}
:
ubuntu@sp7cloud-swiss-1:~$ touch hello-world.txt
ubuntu@sp7cloud-swiss-1:~$ s3cmd put hello-world.txt s3://swiss-backups/hello-world.txt
upload: 'hello-world.txt' -> 's3://swiss-backups/hello-world.txt' [1 of 1]
12 of 12 100% in 0s 110.91 B/s done
Now that this is configured, we can use a script to backup the databases nightly:
#!/bin/bash
# This script is used to clean up old backups in an S3 bucket based on a retention policy.
# The retention policy is as follows:
#
# - Backups less than a week old are kept.
# - Friday backups within the last month are kept.
# - The first Friday backup of each month is kept forever.
# - All other backups are deleted.
# Set the S3 bucket directory
bucket_dir="s3://swiss-backups/"
# Get the current date in YYYY-MM-DD format
current_date=$(date +"%Y-%m-%d")
# Convert the current date to a timestamp
current_timestamp=$(date -d "$current_date" +%s)
# Array to keep track of the first Friday of each month
declare -A first_friday_backups
# Loop through each backup directory in the S3 bucket
for backup_dir in $(s3cmd ls "$bucket_dir" | awk '{print $2}' | sed 's/\/$//'); do
# Extract the date from the directory name
backup_date=$(basename "$backup_dir")
# Convert the backup date to a timestamp
# Assuming the backup date format is YYYY-MM-DD or YYYY_MM_DD
backup_date_formatted=${backup_date//_/ } # Replace underscores with spaces
backup_date_formatted=${backup_date_formatted// /-} # Replace spaces with hyphens
backup_timestamp=$(date -d "$backup_date_formatted" +%s 2>/dev/null)
# Check if the timestamp was successfully created
if [ -z "$backup_timestamp" ]; then
echo "Skipping invalid date format for backup directory: $backup_dir"
continue
fi
# Calculate the difference in days
diff_days=$(( (current_timestamp - backup_timestamp) / 86400 ))
# Determine the day of the week (1=Monday, ..., 7=Sunday)
day_of_week=$(date -d "$backup_date_formatted" +%u)
# Determine the month and year for the backup date
backup_month=$(date -d "$backup_date_formatted" +%Y-%m)
# Check if the backup is less than a week old
if [ "$diff_days" -lt 7 ]; then
echo "Keeping backup: $backup_dir (less than a week old)"
continue
fi
# Check if it's a Friday backup (day_of_week == 5)
if [ "$day_of_week" -eq 5 ]; then
# If it's the first Friday of the month, keep it
if [ -z "${first_friday_backups[$backup_month]}" ]; then
first_friday_backups[$backup_month]="$backup_dir"
echo "Keeping first Friday backup: $backup_dir"
continue # Skip deletion since it's kept indefinitely
else
# If it's a Friday backup within the last month, keep it
if [ "$diff_days" -lt 30 ]; then
echo "Keeping Friday backup: $backup_dir (within the last month)"
continue # Skip deletion since it's kept
fi
fi
fi
# If none of the conditions are met, delete the backup
echo "Deleting backup: $backup_dir (does not meet retention criteria)"
s3cmd del "$backup_dir" --recursive
done
# Final output of the first Friday backups kept indefinitely
echo "First Friday backups kept indefinitely:"
for backup in "${first_friday_backups[@]}"; do
echo "$backup"
done
See the process after running the script at /home/ubuntu/.backup/backup_script.sh
:
ubuntu@sp7cloud-swiss-1:~/.backup$ sh backup_script.sh
Backing up database: geo_swiss
upload: '<stdin>' -> 's3://swiss-backups/2024_08_14/geo_swiss_2024_08_14.sql.gz' [part 1 of -, 6MB] [1 of 1]
6365386 of 6365386 100% in 0s 26.16 MB/s done
Backup of geo_swiss completed and uploaded successfully to s3://swiss-backups/2024_08_14.
Backing up database: mcsn
upload: '<stdin>' -> 's3://swiss-backups/2024_08_14/mcsn_2024_08_14.sql.gz' [part 1 of -, 4MB] [1 of 1]
4602206 of 4602206 100% in 0s 20.33 MB/s done
Backup of mcsn completed and uploaded successfully to s3://swiss-backups/2024_08_14.
Backing up database: mhnc
upload: '<stdin>' -> 's3://swiss-backups/2024_08_14/mhnc_2024_08_14.sql.gz' [part 1 of -, 4MB] [1 of 1]
4588230 of 4588230 100% in 0s 16.26 MB/s done
Backup of mhnc completed and uploaded successfully to s3://swiss-backups/2024_08_14.
Backing up database: mhnf
upload: '<stdin>' -> 's3://swiss-backups/2024_08_14/mhnf_2024_08_14.sql.gz' [part 1 of -, 5MB] [1 of 1]
5861026 of 5861026 100% in 0s 25.46 MB/s done
Backup of mhnf completed and uploaded successfully to s3://swiss-backups/2024_08_14.
Backing up database: naag
upload: '<stdin>' -> 's3://swiss-backups/2024_08_14/naag_2024_08_14.sql.gz' [part 1 of -, 10MB] [1 of 1]
11213948 of 11213948 100% in 0s 24.78 MB/s done
Backup of naag completed and uploaded successfully to s3://swiss-backups/2024_08_14.
Backing up database: nmb_rinvert
upload: '<stdin>' -> 's3://swiss-backups/2024_08_14/nmb_rinvert_2024_08_14.sql.gz' [part 1 of -, 6MB] [1 of 1]
6450282 of 6450282 100% in 0s 30.53 MB/s done
Backup of nmb_rinvert completed and uploaded successfully to s3://swiss-backups/2024_08_14.
Backing up database: sp7demofish_swiss
upload: '<stdin>' -> 's3://swiss-backups/2024_08_14/sp7demofish_swiss_2024_08_14.sql.gz' [part 1 of -, 15MB] [1 of 1]
15728640 of 15728640 100% in 0s 30.45 MB/s done
upload: '<stdin>' -> 's3://swiss-backups/2024_08_14/sp7demofish_swiss_2024_08_14.sql.gz' [part 2 of -, 6MB] [1 of 1]
6776807 of 6776807 100% in 0s 23.11 MB/s done
Backup of sp7demofish_swiss completed and uploaded successfully to s3://swiss-backups/2024_08_14.
ubuntu@sp7cloud-swiss-1:~/.backup$
I configured a cron job to run this at 2 AM CEST (12 AM UTC) every day:
# m h dom mon dow command
0 0 * * * /home/ubuntu/.backup/backup_script.sh
Now this backup will run and be available within the bucket for future retrieval as needed!
ubuntu@sp7cloud-swiss-1:~/.backup$ s3cmd ls s3://swiss-backups/
DIR s3://swiss-backups/2024_08_14/
ubuntu@sp7cloud-swiss-1:~/.backup$ s3cmd ls s3://swiss-backups/2024_08_14/
2024-08-14 00:44 6365385 s3://swiss-backups/2024_08_14/geo_swiss_2024_08_14.sql.gz
2024-08-14 00:44 4602206 s3://swiss-backups/2024_08_14/mcsn_2024_08_14.sql.gz
2024-08-14 00:44 4588230 s3://swiss-backups/2024_08_14/mhnc_2024_08_14.sql.gz
2024-08-14 00:45 5861026 s3://swiss-backups/2024_08_14/mhnf_2024_08_14.sql.gz
2024-08-14 00:45 11213948 s3://swiss-backups/2024_08_14/naag_2024_08_14.sql.gz
2024-08-14 00:45 6450282 s3://swiss-backups/2024_08_14/nmb_rinvert_2024_08_14.sql.gz
2024-08-14 00:45 22505447 s3://swiss-backups/2024_08_14/sp7demofish_swiss_2024_08_14.sql.gz
ubuntu@sp7cloud-swiss-1:~/.backup$
Then the backups are cleaned with another cron job:
# m h dom mon dow command
0 1 * * * /home/ubuntu/.backup/cleanup_script.sh
#!/bin/bash
# This script is used to clean up old backups in an S3 bucket based on a retention policy.
# The retention policy is as follows:
#
# - Backups less than a week old are kept.
# - Friday backups within the last month are kept.
# - The first Friday backup of each month is kept forever.
# - All other backups are deleted.
# Set the S3 bucket directory
bucket_dir="s3://swiss-backups/"
# Get the current date in YYYY-MM-DD format
current_date=$(date +"%Y-%m-%d")
# Convert the current date to a timestamp
current_timestamp=$(date -d "$current_date" +%s)
# Array to keep track of the first Friday of each month
declare -A first_friday_backups
# Loop through each backup directory in the S3 bucket
for backup_dir in $(s3cmd ls "$bucket_dir" | awk '{print $2}' | sed 's/\/$//'); do
# Extract the date from the directory name
backup_date=$(basename "$backup_dir")
# Convert the backup date to a timestamp
# Assuming the backup date format is YYYY-MM-DD or YYYY_MM_DD
backup_date_formatted=${backup_date//_/ } # Replace underscores with spaces
backup_date_formatted=${backup_date_formatted// /-} # Replace spaces with hyphens
backup_timestamp=$(date -d "$backup_date_formatted" +%s 2>/dev/null)
# Check if the timestamp was successfully created
if [ -z "$backup_timestamp" ]; then
echo "Skipping invalid date format for backup directory: $backup_dir"
continue
fi
# Calculate the difference in days
diff_days=$(( (current_timestamp - backup_timestamp) / 86400 ))
# Determine the day of the week (1=Monday, ..., 7=Sunday)
day_of_week=$(date -d "$backup_date_formatted" +%u)
# Determine the month and year for the backup date
backup_month=$(date -d "$backup_date_formatted" +%Y-%m)
# Check if the backup is less than a week old
if [ "$diff_days" -lt 7 ]; then
echo "Keeping backup: $backup_dir (less than a week old)"
continue
fi
# Check if it's a Friday backup (day_of_week == 5)
if [ "$day_of_week" -eq 5 ]; then
# If it's the first Friday of the month, keep it
if [ -z "${first_friday_backups[$backup_month]}" ]; then
first_friday_backups[$backup_month]="$backup_dir"
echo "Keeping first Friday backup: $backup_dir"
continue # Skip deletion since it's kept indefinitely
else
# If it's a Friday backup within the last month, keep it
if [ "$diff_days" -lt 30 ]; then
echo "Keeping Friday backup: $backup_dir (within the last month)"
continue # Skip deletion since it's kept
fi
fi
fi
# If none of the conditions are met, delete the backup
echo "Deleting backup: $backup_dir (does not meet retention criteria)"
s3cmd del "$backup_dir" --recursive
done
# Final output of the first Friday backups kept indefinitely
echo "First Friday backups kept indefinitely:"
for backup in "${first_friday_backups[@]}"; do
echo "$backup"
done