mirror of
https://github.com/tiredofit/docker-db-backup.git
synced 2025-12-21 13:23:12 +01:00
Release 3.8.0 - See CHANGELOG.md
This commit is contained in:
@@ -1,3 +1,10 @@
|
||||
## 3.8.0 2023-03-27 <dave at tiredofit dot ca>
|
||||
|
||||
### Added
|
||||
- Introduce DB_DUMP_TARGET_ARCHIVE which works with DB_ARCHIVE_TIME to move backups older than (x) minutes from DB_DUMP_TARGET to DB_DUMP_TARGET_ARCHIVE for use with external backup systems and custom exclude rules
|
||||
- Introduce CREATE_LATEST_SYMLINK which creates a symbolic link in DB_DUMP_TARGET of `latest-(DB_TYPE)-(DB_NAME)-(DB_HOST)`
|
||||
|
||||
|
||||
## 3.7.7 2023-03-20 <codemonium@github>
|
||||
|
||||
### Changed
|
||||
|
||||
14
README.md
14
README.md
@@ -123,7 +123,7 @@ The following directories are used for configuration and can be mapped for persi
|
||||
|
||||
#### Base Images used
|
||||
|
||||
This image relies on an [Alpine Linux](https://hub.docker.com/r/tiredofit/alpine) or [Debian Linux](https://hub.docker.com/r/tiredofit/debian) base image that relies on an [init system](https://github.com/just-containers/s6-overlay) for added capabilities. Outgoing SMTP capabilities are handlded via `msmtp`. Individual container performance monitoring is performed by [zabbix-agent](https://zabbix.org). Additional tools include: `bash`,`curl`,`less`,`logrotate`, `nano`.
|
||||
This image relies on an [Alpine Linux](https://hub.docker.com/r/tiredofit/alpine) base image that relies on an [init system](https://github.com/just-containers/s6-overlay) for added capabilities. Outgoing SMTP capabilities are handlded via `msmtp`. Individual container performance monitoring is performed by [zabbix-agent](https://zabbix.org). Additional tools include: `bash`,`curl`,`less`,`logrotate`, `nano`.
|
||||
|
||||
Be sure to view the following repositories to understand all the customizable options:
|
||||
|
||||
@@ -134,12 +134,13 @@ Be sure to view the following repositories to understand all the customizable op
|
||||
#### Container Options
|
||||
|
||||
| Parameter | Description | Default |
|
||||
| -------------------- | -------------------------------------------------------------------------------------------------------------------------------- | --------------- |
|
||||
| ----------------------- | -------------------------------------------------------------------------------------------------------------------------------- | --------------- |
|
||||
| `BACKUP_LOCATION` | Backup to `FILESYSTEM` or `S3` compatible services like S3, Minio, Wasabi | `FILESYSTEM` |
|
||||
| `MODE` | `AUTO` mode to use internal scheduling routines or `MANUAL` to simply use this as manual backups only executed by your own means | `AUTO` |
|
||||
| `MANUAL_RUN_FOREVER` | `TRUE` or `FALSE` if you wish to try to make the container exit after the backup | `TRUE` |
|
||||
| `TEMP_LOCATION` | Perform Backups and Compression in this temporary directory | `/tmp/backups/` |
|
||||
| `DEBUG_MODE` | If set to `true`, print copious shell script messages to the container log. Otherwise only basic messages are printed. | `FALSE` |
|
||||
| `CREATE_LATEST_SYMLINK` | Create a symbolic link pointing to last backup in this format: `latest-(DB_TYPE)-(DB_NAME)-(DB_HOST) | `TRUE` |
|
||||
| `PRE_SCRIPT` | Fill this variable in with a command to execute pre backing up | |
|
||||
| `POST_SCRIPT` | Fill this variable in with a command to execute post backing up | |
|
||||
| `SPLIT_DB` | For each backup, create a new archive. `TRUE` or `FALSE` (MySQL and Postgresql Only) | `TRUE` |
|
||||
@@ -164,20 +165,21 @@ Your Organization will be mapped to `DB_USER` and your root token will need to b
|
||||
|
||||
### Scheduling Options
|
||||
| Parameter | Description | Default |
|
||||
| ----------------- | -------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- | --------- |
|
||||
| ------------------------ | -------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- | ---------------------------- |
|
||||
| `DB_DUMP_FREQ` | How often to do a dump, in minutes after the first backup. Defaults to 1440 minutes, or once per day. | `1440` |
|
||||
| `DB_DUMP_BEGIN` | What time to do the first dump. Defaults to immediate. Must be in one of two formats | |
|
||||
| | Absolute HHMM, e.g. `2330` or `0415` | |
|
||||
| | Relative +MM, i.e. how many minutes after starting the container, e.g. `+0` (immediate), `+10` (in 10 minutes), or `+90` in an hour and a half | |
|
||||
| `DB_DUMP_TARGET` | Directory where the database dumps are kept. | `/backup` |
|
||||
| `DB_DUMP_TARGET` | Directory where the database dumps are kept. | `${DB_DUMP_TARGET}/archive/` |
|
||||
| `DB_DUMP_TARGET_ARCHIVE` | Optional Directory where the database dumps archivs are kept. |
|
||||
| `DB_CLEANUP_TIME` | Value in minutes to delete old backups (only fired when dump freqency fires). 1440 would delete anything above 1 day old. You don't need to set this variable if you want to hold onto everything. | `FALSE` |
|
||||
|
||||
| `DB_ARCHIVE_TIME` | Value in minutes to move all files from `DB_DUMP_TARGET` to `DB_DUMP_TARGET_ARCHIVE` - which is useful when pairing against an external backup system. |
|
||||
|
||||
- You may need to wrap your `DB_DUMP_BEGIN` value in quotes for it to properly parse. There have been reports of backups that start with a `0` get converted into a different format which will not allow the timer to start at the correct time.
|
||||
|
||||
### Backup Options
|
||||
| Parameter | Description | Default |
|
||||
| ------------------------------ | ------------------------------------------------------------------------------------------------------------------------------------- | ------------------------- |
|
||||
| ------------------------------ | ---------------------------------------------------------------------------------------------------------------------------- | ------------------------- |
|
||||
| `COMPRESSION` | Use either Gzip `GZ`, Bzip2 `BZ`, XZip `XZ`, ZSTD `ZSTD` or none `NONE` | `ZSTD` |
|
||||
| `COMPRESSION_LEVEL` | Numberical value of what level of compression to use, most allow `1` to `9` except for `ZSTD` which allows for `1` to `19` - | `3` |
|
||||
| `ENABLE_PARALLEL_COMPRESSION` | Use multiple cores when compressing backups `TRUE` or `FALSE` | `TRUE` |
|
||||
|
||||
@@ -8,6 +8,7 @@ COMPRESSION_LEVEL=${COMPRESSION_LEVEL:-"3"}
|
||||
DB_DUMP_BEGIN=${DB_DUMP_BEGIN:-+0}
|
||||
DB_DUMP_FREQ=${DB_DUMP_FREQ:-1440}
|
||||
DB_DUMP_TARGET=${DB_DUMP_TARGET:-"/backup"}
|
||||
DB_DUMP_TARGET_ARCHIVE=${DB_DUMP_TARGET_ARCHIVE:-"${DB_DUMP_TARGET}/archive/"}
|
||||
ENABLE_CHECKSUM=${ENABLE_CHECKSUM:-"TRUE"}
|
||||
ENABLE_PARALLEL_COMPRESSION=${ENABLE_PARALLEL_COMPRESSION:-"TRUE"}
|
||||
MANUAL_RUN_FOREVER=${MANUAL_RUN_FOREVER:-"TRUE"}
|
||||
@@ -28,3 +29,4 @@ SIZE_VALUE=${SIZE_VALUE:-"bytes"}
|
||||
SKIP_AVAILABILITY_CHECK=${SKIP_AVAILABILITY_CHECK:-"FALSE"}
|
||||
SPLIT_DB=${SPLIT_DB:-"TRUE"}
|
||||
TEMP_LOCATION=${TEMP_LOCATION:-"/tmp/backups"}
|
||||
CREATE_LATEST_SYMLINK=${CREATE_LATEST_SYMLINK:-"TRUE"}
|
||||
@@ -114,6 +114,7 @@ bootstrap_variables() {
|
||||
backup_couch() {
|
||||
prepare_dbbackup
|
||||
target=couch_${DB_NAME}_${DB_HOST#*//}_${now}.txt
|
||||
ltarget=couch_${DB_NAME}_${DB_HOST#*//}
|
||||
compression
|
||||
pre_dbbackup ${DB_NAME}
|
||||
print_notice "Dumping CouchDB database: '${DB_NAME}' ${compression_string}"
|
||||
@@ -139,6 +140,7 @@ backup_influx() {
|
||||
prepare_dbbackup
|
||||
if [ "${db}" != "justbackupeverything" ] ; then bucket="-db ${db}" ; else db=all ; fi
|
||||
target=influx_${db}_${DB_HOST#*//}_${now}
|
||||
ltarget=influx_${db}_${DB_HOST#*//}
|
||||
compression
|
||||
pre_dbbackup $db
|
||||
print_notice "Dumping Influx database: '${db}'"
|
||||
@@ -148,6 +150,7 @@ backup_influx() {
|
||||
print_notice "Creating archive file of '${target_dir}' with tar ${compression_string}"
|
||||
tar cf - "${TEMP_LOCATION}"/"${target_dir}" | $dir_compress_cmd > "${TEMP_LOCATION}"/"${target_dir}".tar"${extension}"
|
||||
target=influx_${db}_${DB_HOST#*//}_${now}.tar${extension}
|
||||
target=influx_${db}_${DB_HOST#*//}
|
||||
generate_checksum
|
||||
move_dbbackup
|
||||
post_dbbackup $db
|
||||
@@ -158,6 +161,7 @@ backup_influx() {
|
||||
prepare_dbbackup
|
||||
if [ "${db}" != "justbackupeverything" ] ; then bucket="--bucket $db" ; else db=all ; fi
|
||||
target=influx2_${db}_${DB_HOST#*//}_${now}
|
||||
ltarget=influx2_${db}_${DB_HOST#*//}
|
||||
compression
|
||||
pre_dbbackup $db
|
||||
print_notice "Dumping Influx2 database: '${db}'"
|
||||
@@ -178,8 +182,10 @@ backup_mongo() {
|
||||
prepare_dbbackup
|
||||
if [ "${ENABLE_COMPRESSION,,}" = "none" ] || [ "${ENABLE_COMPRESSION,,}" = "false" ] ; then
|
||||
target=${dbtype}_${DB_NAME,,}_${DB_HOST,,}_${now}.archive
|
||||
ltarget=${dbtype}_${DB_NAME,,}_${DB_HOST,,}
|
||||
else
|
||||
target=${dbtype}_${DB_NAME,,}_${DB_HOST,,}_${now}.archive.gz
|
||||
ltarget=${dbtype}_${DB_NAME,,}_${DB_HOST,,}
|
||||
mongo_compression="--gzip"
|
||||
compression_string="and compressing with gzip"
|
||||
fi
|
||||
@@ -201,6 +207,7 @@ backup_mongo() {
|
||||
backup_mssql() {
|
||||
prepare_dbbackup
|
||||
target=mssql_${DB_NAME,,}_${DB_HOST,,}_${now}.bak
|
||||
target=mssql_${DB_NAME,,}_${DB_HOST,,}
|
||||
compression
|
||||
pre_dbbackup "${DB_NAME}"
|
||||
print_notice "Dumping MSSQL database: '${DB_NAME}'"
|
||||
@@ -240,6 +247,7 @@ backup_mysql() {
|
||||
for db in ${db_names} ; do
|
||||
prepare_dbbackup
|
||||
target=mysql_${db}_${DB_HOST,,}_${now}.sql
|
||||
ltarget=mysql_${db}_${DB_HOST,,}
|
||||
compression
|
||||
pre_dbbackup $db
|
||||
print_notice "Dumping MySQL/MariaDB database: '${db}' ${compression_string}"
|
||||
@@ -254,6 +262,7 @@ backup_mysql() {
|
||||
print_debug "Not splitting database dumps into their own files"
|
||||
prepare_dbbackup
|
||||
target=mysql_all_${DB_HOST,,}_${now}.sql
|
||||
ltarget=mysql_all_${DB_HOST,,}
|
||||
compression
|
||||
pre_dbbackup all
|
||||
print_notice "Dumping all MySQL / MariaDB databases: '$(echo ${db_names} | xargs | tr ' ' ',')' ${compression_string}"
|
||||
@@ -289,6 +298,7 @@ backup_pgsql() {
|
||||
for db in ${db_names} ; do
|
||||
prepare_dbbackup
|
||||
target=pgsql_${db}_${DB_HOST,,}_${now}.sql
|
||||
ltarget=pgsql_${db}_${DB_HOST,,}
|
||||
compression
|
||||
pre_dbbackup $db
|
||||
print_notice "Dumping PostgresSQL database: '${db}' ${compression_string}"
|
||||
@@ -303,6 +313,7 @@ backup_pgsql() {
|
||||
print_debug "Not splitting database dumps into their own files"
|
||||
prepare_dbbackup
|
||||
target=pgsql_all_${DB_HOST,,}_${now}.sql
|
||||
ltarget=pgsql_${db}_${DB_HOST,,}
|
||||
compression
|
||||
pre_dbbackup all
|
||||
print_notice "Dumping all PostgreSQL databases: '$(echo ${db_names} | xargs | tr ' ' ',')' ${compression_string}"
|
||||
@@ -327,6 +338,7 @@ backup_redis() {
|
||||
prepare_dbbackup
|
||||
print_notice "Dumping Redis - Flushing Redis Cache First"
|
||||
target=redis_all_${DB_HOST,,}_${now}.rdb
|
||||
ltarget=redis_${DB_HOST,,}
|
||||
echo bgsave | silent redis-cli -h ${DB_HOST} -p ${DB_PORT} ${REDIS_PASS_STR} --rdb ${TEMP_LOCATION}/${target} ${EXTRA_OPTS}
|
||||
sleep 10
|
||||
try=5
|
||||
@@ -357,6 +369,7 @@ backup_sqlite3() {
|
||||
db=$(basename "${DB_HOST}")
|
||||
db="${db%.*}"
|
||||
target=sqlite3_${db}_${now}.sqlite3
|
||||
target=sqlite3_${db}.sqlite3
|
||||
compression
|
||||
pre_dbbackup $db
|
||||
print_notice "Dumping sqlite3 database: '${DB_HOST}' ${compression_string}"
|
||||
@@ -653,6 +666,14 @@ move_dbbackup() {
|
||||
mkdir -p "${DB_DUMP_TARGET}"
|
||||
mv "${TEMP_LOCATION}"/*."${checksum_extension}" "${DB_DUMP_TARGET}"/
|
||||
mv "${TEMP_LOCATION}"/"${target}" "${DB_DUMP_TARGET}"/"${target}"
|
||||
if var_true "${CREATE_LATEST_SYMLINK}" ; then
|
||||
ln -sf "${DB_DUMP_TARGET}"/"${target}" "${DB_DUMP_TARGET}"/latest-"${ltarget}"
|
||||
fi
|
||||
if [ -n "${DB_ARCHIVE_TIME}" ] ; then
|
||||
mkdir -p "${DB_DUMP_TARGET_ARCHIVE}"
|
||||
find "${DB_DUMP_TARGET}"/ -maxdepth 1 -mmin +"${DB_ARCHIVE_TIME}" -iname "*" -exec mv {} foo \;
|
||||
find "${DB_DUMP_TARGET}"/ -maxdepth 1 -mmin +"${DB_CLEANUP_TIME}" -iname "*" -exec rm {} \;
|
||||
fi
|
||||
;;
|
||||
"s3" | "minio" )
|
||||
print_debug "Moving backup to S3 Bucket"
|
||||
@@ -707,6 +728,7 @@ prepare_dbbackup() {
|
||||
now=$(date +"%Y%m%d-%H%M%S")
|
||||
now_time=$(date +"%H:%M:%S")
|
||||
now_date=$(date +"%Y-%m-%d")
|
||||
ltarget=${dbtype}_${DB_NAME,,}_${DB_HOST,,}
|
||||
target=${dbtype}_${DB_NAME,,}_${DB_HOST,,}_${now}.sql
|
||||
}
|
||||
|
||||
@@ -836,8 +858,7 @@ setup_mode() {
|
||||
mkdir -p /etc/services.d/99-run_forever
|
||||
cat <<EOF > /etc/services.d/99-run_forever/run
|
||||
#!/bin/bash
|
||||
while true
|
||||
do
|
||||
while true; do
|
||||
sleep 86400
|
||||
done
|
||||
EOF
|
||||
|
||||
Reference in New Issue
Block a user