mirror of
https://github.com/tiredofit/docker-db-backup.git
synced 2025-12-22 05:33:53 +01:00
Compare commits
14 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
68174c061f | ||
|
|
fd1d95090e | ||
|
|
7befba0d96 | ||
|
|
583253fce7 | ||
|
|
068577001e | ||
|
|
7781542816 | ||
|
|
9283b5440e | ||
|
|
5e62485e7f | ||
|
|
f224571448 | ||
|
|
01620fec00 | ||
|
|
18a38b4f1d | ||
|
|
150f356275 | ||
|
|
e838ed0027 | ||
|
|
8329b4c065 |
50
CHANGELOG.md
50
CHANGELOG.md
@@ -1,3 +1,53 @@
|
|||||||
|
## 3.9.2 2023-05-10 <dave at tiredofit dot ca>
|
||||||
|
|
||||||
|
### Changed
|
||||||
|
- Alpine 3.18 base
|
||||||
|
|
||||||
|
|
||||||
|
## 3.9.1 2023-05-03 <dave at tiredofit dot ca>
|
||||||
|
|
||||||
|
### Changed
|
||||||
|
- Properly allow multiple _FILE environment variables to execute solving an issue with MySQL backups
|
||||||
|
- Fix _FILE functionality for DB_NAME variable
|
||||||
|
|
||||||
|
|
||||||
|
## 3.9.0 2023-04-26 <dave at tiredofit dot ca>
|
||||||
|
|
||||||
|
### Added
|
||||||
|
- Add support for _FILE environment variables
|
||||||
|
|
||||||
|
|
||||||
|
## 3.8.5 2023-04-11 <tpansino@github>
|
||||||
|
|
||||||
|
### Changed
|
||||||
|
- Fix SQLite3, Influx, and MSSQL backups failing due to malformed/non existent ltarget
|
||||||
|
|
||||||
|
|
||||||
|
## 3.8.4 2023-04-06 <dave at tiredofit dot ca>
|
||||||
|
|
||||||
|
### Changed
|
||||||
|
- Fix issue with Influx2 and MSSQL clients not installing properly
|
||||||
|
|
||||||
|
|
||||||
|
## 3.8.3 2023-03-30 <dave at tiredofit dot ca>
|
||||||
|
|
||||||
|
### Changed
|
||||||
|
- Patchup for 3.8.2
|
||||||
|
|
||||||
|
|
||||||
|
## 3.8.2 2023-03-30 <dave at tiredofit dot ca>
|
||||||
|
|
||||||
|
### Changed
|
||||||
|
- Patchup for 3.8.1
|
||||||
|
|
||||||
|
|
||||||
|
## 3.8.1 2023-03-30 <dave at tiredofit dot ca>
|
||||||
|
|
||||||
|
### Changed
|
||||||
|
- Cleanup Dockerfile
|
||||||
|
- Fix issue with DB_ARCHIVE_TIME not firing correctly
|
||||||
|
|
||||||
|
|
||||||
## 3.8.0 2023-03-27 <dave at tiredofit dot ca>
|
## 3.8.0 2023-03-27 <dave at tiredofit dot ca>
|
||||||
|
|
||||||
### Added
|
### Added
|
||||||
|
|||||||
29
Dockerfile
29
Dockerfile
@@ -1,4 +1,7 @@
|
|||||||
FROM docker.io/tiredofit/alpine:3.17
|
ARG DISTRO=alpine
|
||||||
|
ARG DISTRO_VARIANT=3.18
|
||||||
|
|
||||||
|
FROM docker.io/tiredofit/${DISTRO}:${DISTRO_VARIANT}
|
||||||
LABEL maintainer="Dave Conroy (github.com/tiredofit)"
|
LABEL maintainer="Dave Conroy (github.com/tiredofit)"
|
||||||
|
|
||||||
### Set Environment Variables
|
### Set Environment Variables
|
||||||
@@ -13,9 +16,9 @@ ENV INFLUX2_VERSION=2.4.0 \
|
|||||||
### Dependencies
|
### Dependencies
|
||||||
RUN source /assets/functions/00-container && \
|
RUN source /assets/functions/00-container && \
|
||||||
set -ex && \
|
set -ex && \
|
||||||
apk update && \
|
package update && \
|
||||||
apk upgrade && \
|
package upgrade && \
|
||||||
apk add -t .db-backup-build-deps \
|
package install .db-backup-build-deps \
|
||||||
build-base \
|
build-base \
|
||||||
bzip2-dev \
|
bzip2-dev \
|
||||||
git \
|
git \
|
||||||
@@ -27,7 +30,7 @@ RUN source /assets/functions/00-container && \
|
|||||||
xz-dev \
|
xz-dev \
|
||||||
&& \
|
&& \
|
||||||
\
|
\
|
||||||
apk add --no-cache -t .db-backup-run-deps \
|
package install .db-backup-run-deps \
|
||||||
aws-cli \
|
aws-cli \
|
||||||
bzip2 \
|
bzip2 \
|
||||||
influxdb \
|
influxdb \
|
||||||
@@ -75,12 +78,14 @@ RUN source /assets/functions/00-container && \
|
|||||||
\
|
\
|
||||||
pip3 install blobxfer && \
|
pip3 install blobxfer && \
|
||||||
\
|
\
|
||||||
### Cleanup
|
package remove .db-backup-build-deps && \
|
||||||
apk del .db-backup-build-deps && \
|
package cleanup && \
|
||||||
rm -rf /usr/src/* && \
|
rm -rf \
|
||||||
rm -rf /*.apk && \
|
/*.apk \
|
||||||
rm -rf /etc/logrotate.d/redis && \
|
/etc/logrotate.d/* \
|
||||||
rm -rf /root/.cache /tmp/* /var/cache/apk/*
|
/root/.cache \
|
||||||
|
/tmp/* \
|
||||||
|
/usr/src/*
|
||||||
|
|
||||||
|
|
||||||
### S6 Setup
|
|
||||||
COPY install /
|
COPY install /
|
||||||
|
|||||||
107
README.md
107
README.md
@@ -140,25 +140,25 @@ Be sure to view the following repositories to understand all the customizable op
|
|||||||
| `MANUAL_RUN_FOREVER` | `TRUE` or `FALSE` if you wish to try to make the container exit after the backup | `TRUE` |
|
| `MANUAL_RUN_FOREVER` | `TRUE` or `FALSE` if you wish to try to make the container exit after the backup | `TRUE` |
|
||||||
| `TEMP_LOCATION` | Perform Backups and Compression in this temporary directory | `/tmp/backups/` |
|
| `TEMP_LOCATION` | Perform Backups and Compression in this temporary directory | `/tmp/backups/` |
|
||||||
| `DEBUG_MODE` | If set to `true`, print copious shell script messages to the container log. Otherwise only basic messages are printed. | `FALSE` |
|
| `DEBUG_MODE` | If set to `true`, print copious shell script messages to the container log. Otherwise only basic messages are printed. | `FALSE` |
|
||||||
| `CREATE_LATEST_SYMLINK` | Create a symbolic link pointing to last backup in this format: `latest-(DB_TYPE)-(DB_NAME)-(DB_HOST) | `TRUE` |
|
| `CREATE_LATEST_SYMLINK` | Create a symbolic link pointing to last backup in this format: `latest-(DB_TYPE)-(DB_NAME)-(DB_HOST)` | `TRUE` |
|
||||||
| `PRE_SCRIPT` | Fill this variable in with a command to execute pre backing up | |
|
| `PRE_SCRIPT` | Fill this variable in with a command to execute pre backing up | |
|
||||||
| `POST_SCRIPT` | Fill this variable in with a command to execute post backing up | |
|
| `POST_SCRIPT` | Fill this variable in with a command to execute post backing up | |
|
||||||
| `SPLIT_DB` | For each backup, create a new archive. `TRUE` or `FALSE` (MySQL and Postgresql Only) | `TRUE` |
|
| `SPLIT_DB` | For each backup, create a new archive. `TRUE` or `FALSE` (MySQL and Postgresql Only) | `TRUE` |
|
||||||
|
|
||||||
### Database Specific Options
|
### Database Specific Options
|
||||||
| Parameter | Description | Default |
|
| Parameter | Description | Default | `_FILE` |
|
||||||
| ------------------ | ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ | ------- |
|
| ------------------ | ------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------ | ------- | ------- |
|
||||||
| `DB_AUTH` | (Mongo Only - Optional) Authentication Database | |
|
| `DB_AUTH` | (Mongo Only - Optional) Authentication Database | | |
|
||||||
| `DB_TYPE` | Type of DB Server to backup `couch` `influx` `mysql` `pgsql` `mongo` `redis` `sqlite3` | |
|
| `DB_TYPE` | Type of DB Server to backup `couch` `influx` `mysql` `pgsql` `mongo` `redis` `sqlite3` | | |
|
||||||
| `DB_HOST` | Server Hostname e.g. `mariadb`. For `sqlite3`, full path to DB file e.g. `/backup/db.sqlite3` | |
|
| `DB_HOST` | Server Hostname e.g. `mariadb`. For `sqlite3`, full path to DB file e.g. `/backup/db.sqlite3` | | x |
|
||||||
| `DB_NAME` | Schema Name e.g. `database` or `ALL` to backup all databases the user has access to. Backup multiple by seperating with commas eg `db1,db2` | |
|
| `DB_NAME` | Schema Name e.g. `database` or `ALL` to backup all databases the user has access to. Backup multiple by seperating with commas eg `db1,db2` | | x |
|
||||||
| `DB_NAME_EXCLUDE` | If using `ALL` - use this as to exclude databases seperated via commas from being backed up | |
|
| `DB_NAME_EXCLUDE` | If using `ALL` - use this as to exclude databases seperated via commas from being backed up | | x |
|
||||||
| `DB_USER` | username for the database(s) - Can use `root` for MySQL | |
|
| `DB_USER` | username for the database(s) - Can use `root` for MySQL | | x |
|
||||||
| `DB_PASS` | (optional if DB doesn't require it) password for the database | |
|
| `DB_PASS` | (optional if DB doesn't require it) password for the database | | x |
|
||||||
| `DB_PORT` | (optional) Set port to connect to DB_HOST. Defaults are provided | varies |
|
| `DB_PORT` | (optional) Set port to connect to DB_HOST. Defaults are provided | varies | x |
|
||||||
| `INFLUX_VERSION` | What Version of Influx are you backing up from `1`.x or `2` series - AMD64 and ARM64 only for `2` | |
|
| `INFLUX_VERSION` | What Version of Influx are you backing up from `1`.x or `2` series - AMD64 and ARM64 only for `2` | | |
|
||||||
| `MONGO_CUSTOM_URI` | If you wish to override the MongoDB Connection string enter it here e.g. `mongodb+srv://username:password@cluster.id.mongodb.net` | |
|
| `MONGO_CUSTOM_URI` | If you wish to override the MongoDB Connection string enter it here e.g. `mongodb+srv://username:password@cluster.id.mongodb.net` | | x |
|
||||||
| | This environment variable will be parsed and populate the `DB_NAME` and `DB_HOST` variables to properly build your backup filenames. You can overrde them by making your own entries |
|
| | This environment variable will be parsed and populate the `DB_NAME` and `DB_HOST` variables to properly build your backup filenames. You can overrde them by making your own entries | | |
|
||||||
|
|
||||||
#### For Influx DB2:
|
#### For Influx DB2:
|
||||||
Your Organization will be mapped to `DB_USER` and your root token will need to be mapped to `DB_PASS`. You may use `DB_NAME=ALL` to backup the entire set of databases. For `DB_HOST` use syntax of `http(s)://db-name`
|
Your Organization will be mapped to `DB_USER` and your root token will need to be mapped to `DB_PASS`. You may use `DB_NAME=ALL` to backup the entire set of databases. For `DB_HOST` use syntax of `http(s)://db-name`
|
||||||
@@ -171,32 +171,32 @@ Your Organization will be mapped to `DB_USER` and your root token will need to b
|
|||||||
| | Absolute HHMM, e.g. `2330` or `0415` | |
|
| | Absolute HHMM, e.g. `2330` or `0415` | |
|
||||||
| | Relative +MM, i.e. how many minutes after starting the container, e.g. `+0` (immediate), `+10` (in 10 minutes), or `+90` in an hour and a half | |
|
| | Relative +MM, i.e. how many minutes after starting the container, e.g. `+0` (immediate), `+10` (in 10 minutes), or `+90` in an hour and a half | |
|
||||||
| `DB_DUMP_TARGET` | Directory where the database dumps are kept. | `${DB_DUMP_TARGET}/archive/` |
|
| `DB_DUMP_TARGET` | Directory where the database dumps are kept. | `${DB_DUMP_TARGET}/archive/` |
|
||||||
| `DB_DUMP_TARGET_ARCHIVE` | Optional Directory where the database dumps archivs are kept. |
|
| `DB_DUMP_TARGET_ARCHIVE` | Optional Directory where the database dumps archives are kept. | |
|
||||||
| `DB_CLEANUP_TIME` | Value in minutes to delete old backups (only fired when dump freqency fires). 1440 would delete anything above 1 day old. You don't need to set this variable if you want to hold onto everything. | `FALSE` |
|
| `DB_CLEANUP_TIME` | Value in minutes to delete old backups (only fired when dump freqency fires). 1440 would delete anything above 1 day old. You don't need to set this variable if you want to hold onto everything. | `FALSE` |
|
||||||
| `DB_ARCHIVE_TIME` | Value in minutes to move all files from `DB_DUMP_TARGET` to `DB_DUMP_TARGET_ARCHIVE` - which is useful when pairing against an external backup system. |
|
| `DB_ARCHIVE_TIME` | Value in minutes to move all files files older than (x) from `DB_DUMP_TARGET` to `DB_DUMP_TARGET_ARCHIVE` - which is useful when pairing against an external backup system. | |
|
||||||
|
|
||||||
- You may need to wrap your `DB_DUMP_BEGIN` value in quotes for it to properly parse. There have been reports of backups that start with a `0` get converted into a different format which will not allow the timer to start at the correct time.
|
- You may need to wrap your `DB_DUMP_BEGIN` value in quotes for it to properly parse. There have been reports of backups that start with a `0` get converted into a different format which will not allow the timer to start at the correct time.
|
||||||
|
|
||||||
### Backup Options
|
### Backup Options
|
||||||
| Parameter | Description | Default |
|
| Parameter | Description | Default | `_FILE` |
|
||||||
| ------------------------------ | ---------------------------------------------------------------------------------------------------------------------------- | ------------------------- |
|
| ------------------------------ | ---------------------------------------------------------------------------------------------------------------------------- | ------------------------- | ------- |
|
||||||
| `COMPRESSION` | Use either Gzip `GZ`, Bzip2 `BZ`, XZip `XZ`, ZSTD `ZSTD` or none `NONE` | `ZSTD` |
|
| `COMPRESSION` | Use either Gzip `GZ`, Bzip2 `BZ`, XZip `XZ`, ZSTD `ZSTD` or none `NONE` | `ZSTD` | |
|
||||||
| `COMPRESSION_LEVEL` | Numberical value of what level of compression to use, most allow `1` to `9` except for `ZSTD` which allows for `1` to `19` - | `3` |
|
| `COMPRESSION_LEVEL` | Numberical value of what level of compression to use, most allow `1` to `9` except for `ZSTD` which allows for `1` to `19` - | `3` | |
|
||||||
| `ENABLE_PARALLEL_COMPRESSION` | Use multiple cores when compressing backups `TRUE` or `FALSE` | `TRUE` |
|
| `ENABLE_PARALLEL_COMPRESSION` | Use multiple cores when compressing backups `TRUE` or `FALSE` | `TRUE` | |
|
||||||
| `PARALLEL_COMPRESSION_THREADS` | Maximum amount of threads to use when compressing - Integer value e.g. `8` | `autodetected` |
|
| `PARALLEL_COMPRESSION_THREADS` | Maximum amount of threads to use when compressing - Integer value e.g. `8` | `autodetected` | |
|
||||||
| `GZ_RSYNCABLE` | Use `--rsyncable` (gzip only) for faster rsync transfers and incremental backup deduplication. e.g. `TRUE` | `FALSE` |
|
| `GZ_RSYNCABLE` | Use `--rsyncable` (gzip only) for faster rsync transfers and incremental backup deduplication. e.g. `TRUE` | `FALSE` | |
|
||||||
| `ENABLE_CHECKSUM` | Generate either a MD5 or SHA1 in Directory, `TRUE` or `FALSE` | `TRUE` |
|
| `ENABLE_CHECKSUM` | Generate either a MD5 or SHA1 in Directory, `TRUE` or `FALSE` | `TRUE` | |
|
||||||
| `CHECKSUM` | Either `MD5` or `SHA1` | `MD5` |
|
| `CHECKSUM` | Either `MD5` or `SHA1` | `MD5` | |
|
||||||
| `EXTRA_OPTS` | If you need to pass extra arguments to the backup command, add them here e.g. `--extra-command` | |
|
| `EXTRA_OPTS` | If you need to pass extra arguments to the backup command, add them here e.g. `--extra-command` | | |
|
||||||
| `MYSQL_MAX_ALLOWED_PACKET` | Max allowed packet if backing up MySQL / MariaDB | `512M` |
|
| `MYSQL_MAX_ALLOWED_PACKET` | Max allowed packet if backing up MySQL / MariaDB | `512M` | |
|
||||||
| `MYSQL_SINGLE_TRANSACTION` | Backup in a single transaction with MySQL / MariaDB | `TRUE` |
|
| `MYSQL_SINGLE_TRANSACTION` | Backup in a single transaction with MySQL / MariaDB | `TRUE` | |
|
||||||
| `MYSQL_STORED_PROCEDURES` | Backup stored procedures with MySQL / MariaDB | `TRUE` |
|
| `MYSQL_STORED_PROCEDURES` | Backup stored procedures with MySQL / MariaDB | `TRUE` | |
|
||||||
| `MYSQL_ENABLE_TLS` | Enable TLS functionality for MySQL client | `FALSE` |
|
| `MYSQL_ENABLE_TLS` | Enable TLS functionality for MySQL client | `FALSE` | |
|
||||||
| `MYSQL_TLS_VERIFY` | (optional) If using TLS (by means of MYSQL_TLS_* variables) verify remote host | `FALSE` |
|
| `MYSQL_TLS_VERIFY` | (optional) If using TLS (by means of MYSQL_TLS_* variables) verify remote host | `FALSE` | |
|
||||||
| `MYSQL_TLS_VERSION` | What TLS `v1.1` `v1.2` `v1.3` version to utilize | `TLSv1.1,TLSv1.2,TLSv1.3` |
|
| `MYSQL_TLS_VERSION` | What TLS `v1.1` `v1.2` `v1.3` version to utilize | `TLSv1.1,TLSv1.2,TLSv1.3` | |
|
||||||
| `MYSQL_TLS_CA_FILE` | Filename to load custom CA certificate for connecting via TLS | `/etc/ssl/cert.pem` |
|
| `MYSQL_TLS_CA_FILE` | Filename to load custom CA certificate for connecting via TLS | `/etc/ssl/cert.pem` | x |
|
||||||
| `MYSQL_TLS_CERT_FILE` | Filename to load client certificate for connecting via TLS | |
|
| `MYSQL_TLS_CERT_FILE` | Filename to load client certificate for connecting via TLS | | x |
|
||||||
| `MYSQL_TLS_KEY_FILE` | Filename to load client key for connecting via TLS | |
|
| `MYSQL_TLS_KEY_FILE` | Filename to load client key for connecting via TLS | | x |
|
||||||
|
|
||||||
- When using compression with MongoDB, only `GZ` compression is possible.
|
- When using compression with MongoDB, only `GZ` compression is possible.
|
||||||
|
|
||||||
@@ -204,19 +204,19 @@ Your Organization will be mapped to `DB_USER` and your root token will need to b
|
|||||||
|
|
||||||
If `BACKUP_LOCATION` = `S3` then the following options are used.
|
If `BACKUP_LOCATION` = `S3` then the following options are used.
|
||||||
|
|
||||||
| Parameter | Description | Default |
|
| Parameter | Description | Default | `_FILE` |
|
||||||
| --------------------- | ----------------------------------------------------------------------------------------- | ------- |
|
| --------------------- | ----------------------------------------------------------------------------------------- | ------- | ------- |
|
||||||
| `S3_BUCKET` | S3 Bucket name e.g. `mybucket` | |
|
| `S3_BUCKET` | S3 Bucket name e.g. `mybucket` | | x |
|
||||||
| `S3_KEY_ID` | S3 Key ID (Optional) | |
|
| `S3_KEY_ID` | S3 Key ID (Optional) | | x |
|
||||||
| `S3_KEY_SECRET` | S3 Key Secret (Optional) | |
|
| `S3_KEY_SECRET` | S3 Key Secret (Optional) | | x |
|
||||||
| `S3_PATH` | S3 Pathname to save to (must NOT end in a trailing slash e.g. '`backup`') | |
|
| `S3_PATH` | S3 Pathname to save to (must NOT end in a trailing slash e.g. '`backup`') | | x |
|
||||||
| `S3_REGION` | Define region in which bucket is defined. Example: `ap-northeast-2` | |
|
| `S3_REGION` | Define region in which bucket is defined. Example: `ap-northeast-2` | | x |
|
||||||
| `S3_HOST` | Hostname (and port) of S3-compatible service, e.g. `minio:8080`. Defaults to AWS. | |
|
| `S3_HOST` | Hostname (and port) of S3-compatible service, e.g. `minio:8080`. Defaults to AWS. | | x |
|
||||||
| `S3_PROTOCOL` | Protocol to connect to `S3_HOST`. Either `http` or `https`. Defaults to `https`. | `https` |
|
| `S3_PROTOCOL` | Protocol to connect to `S3_HOST`. Either `http` or `https`. Defaults to `https`. | `https` | x |
|
||||||
| `S3_EXTRA_OPTS` | Add any extra options to the end of the `aws-cli` process execution | |
|
| `S3_EXTRA_OPTS` | Add any extra options to the end of the `aws-cli` process execution | | x |
|
||||||
| `S3_CERT_CA_FILE` | Map a volume and point to your custom CA Bundle for verification e.g. `/certs/bundle.pem` | |
|
| `S3_CERT_CA_FILE` | Map a volume and point to your custom CA Bundle for verification e.g. `/certs/bundle.pem` | | x |
|
||||||
| _*OR*_ | | |
|
| _*OR*_ | | | |
|
||||||
| `S3_CERT_SKIP_VERIFY` | Skip verifying self signed certificates when connecting | `TRUE` |
|
| `S3_CERT_SKIP_VERIFY` | Skip verifying self signed certificates when connecting | `TRUE` | |
|
||||||
|
|
||||||
- When `S3_KEY_ID` and/or `S3_KEY_SECRET` is not set, will try to use IAM role assigned (if any) for uploading the backup files to S3 bucket.
|
- When `S3_KEY_ID` and/or `S3_KEY_SECRET` is not set, will try to use IAM role assigned (if any) for uploading the backup files to S3 bucket.
|
||||||
|
|
||||||
@@ -224,14 +224,13 @@ If `BACKUP_LOCATION` = `S3` then the following options are used.
|
|||||||
|
|
||||||
Support to upload backup files with [blobxfer](https://github.com/Azure/blobxfer) to the Azure fileshare storage.
|
Support to upload backup files with [blobxfer](https://github.com/Azure/blobxfer) to the Azure fileshare storage.
|
||||||
|
|
||||||
|
|
||||||
If `BACKUP_LOCATION` = `blobxfer` then the following options are used.
|
If `BACKUP_LOCATION` = `blobxfer` then the following options are used.
|
||||||
|
|
||||||
| Parameter | Description | Default |
|
| Parameter | Description | Default | `_FILE` |
|
||||||
| ------------------------------ | ------------------------------------------- | ------------------- |
|
| ------------------------------ | ------------------------------------------- | ------------------- | ------- |
|
||||||
| `BLOBXFER_STORAGE_ACCOUNT` | Microsoft Azure Cloud storage account name. | |
|
| `BLOBXFER_STORAGE_ACCOUNT` | Microsoft Azure Cloud storage account name. | | x |
|
||||||
| `BLOBXFER_STORAGE_ACCOUNT_KEY` | Microsoft Azure Cloud storage account key. | |
|
| `BLOBXFER_STORAGE_ACCOUNT_KEY` | Microsoft Azure Cloud storage account key. | | x |
|
||||||
| `BLOBXFER_REMOTE_PATH` | Remote Azure path | `/docker-db-backup` |
|
| `BLOBXFER_REMOTE_PATH` | Remote Azure path | `/docker-db-backup` | x |
|
||||||
|
|
||||||
> This service uploads files from backup targed directory `DB_DUMP_TARGET`.
|
> This service uploads files from backup targed directory `DB_DUMP_TARGET`.
|
||||||
> If the a cleanup configuration in `DB_CLEANUP_TIME` is defined, the remote directory on Azure storage will also be cleaned automatically.
|
> If the a cleanup configuration in `DB_CLEANUP_TIME` is defined, the remote directory on Azure storage will also be cleaned automatically.
|
||||||
|
|||||||
@@ -5,6 +5,7 @@ BLOBXFER_REMOTE_PATH=${BLOBXFER_REMOTE_PATH:-"/docker-db-backup"}
|
|||||||
CHECKSUM=${CHECKSUM:-"MD5"}
|
CHECKSUM=${CHECKSUM:-"MD5"}
|
||||||
COMPRESSION=${COMPRESSION:-"ZSTD"}
|
COMPRESSION=${COMPRESSION:-"ZSTD"}
|
||||||
COMPRESSION_LEVEL=${COMPRESSION_LEVEL:-"3"}
|
COMPRESSION_LEVEL=${COMPRESSION_LEVEL:-"3"}
|
||||||
|
CREATE_LATEST_SYMLINK=${CREATE_LATEST_SYMLINK:-"TRUE"}
|
||||||
DB_DUMP_BEGIN=${DB_DUMP_BEGIN:-+0}
|
DB_DUMP_BEGIN=${DB_DUMP_BEGIN:-+0}
|
||||||
DB_DUMP_FREQ=${DB_DUMP_FREQ:-1440}
|
DB_DUMP_FREQ=${DB_DUMP_FREQ:-1440}
|
||||||
DB_DUMP_TARGET=${DB_DUMP_TARGET:-"/backup"}
|
DB_DUMP_TARGET=${DB_DUMP_TARGET:-"/backup"}
|
||||||
@@ -29,4 +30,3 @@ SIZE_VALUE=${SIZE_VALUE:-"bytes"}
|
|||||||
SKIP_AVAILABILITY_CHECK=${SKIP_AVAILABILITY_CHECK:-"FALSE"}
|
SKIP_AVAILABILITY_CHECK=${SKIP_AVAILABILITY_CHECK:-"FALSE"}
|
||||||
SPLIT_DB=${SPLIT_DB:-"TRUE"}
|
SPLIT_DB=${SPLIT_DB:-"TRUE"}
|
||||||
TEMP_LOCATION=${TEMP_LOCATION:-"/tmp/backups"}
|
TEMP_LOCATION=${TEMP_LOCATION:-"/tmp/backups"}
|
||||||
CREATE_LATEST_SYMLINK=${CREATE_LATEST_SYMLINK:-"TRUE"}
|
|
||||||
@@ -2,12 +2,19 @@
|
|||||||
|
|
||||||
bootstrap_variables() {
|
bootstrap_variables() {
|
||||||
sanity_var DB_TYPE "Set appropriate DB_TYPE"
|
sanity_var DB_TYPE "Set appropriate DB_TYPE"
|
||||||
|
transform_file_var \
|
||||||
|
DB_HOST \
|
||||||
|
DB_NAME \
|
||||||
|
DB_PORT \
|
||||||
|
DB_USER \
|
||||||
|
DB_PASS
|
||||||
|
|
||||||
case "${DB_TYPE,,}" in
|
case "${DB_TYPE,,}" in
|
||||||
couch* )
|
couch* )
|
||||||
dbtype=couch
|
dbtype=couch
|
||||||
DB_PORT=${DB_PORT:-5984}
|
DB_PORT=${DB_PORT:-5984}
|
||||||
file_env 'DB_USER'
|
sanity_var DB_USER
|
||||||
file_env 'DB_PASS'
|
sanity_var DB_PASS
|
||||||
;;
|
;;
|
||||||
influx* )
|
influx* )
|
||||||
dbtype=influx
|
dbtype=influx
|
||||||
@@ -15,31 +22,31 @@ bootstrap_variables() {
|
|||||||
1) DB_PORT=${DB_PORT:-8088} ;;
|
1) DB_PORT=${DB_PORT:-8088} ;;
|
||||||
2) DB_PORT=${DB_PORT:-8086} ;;
|
2) DB_PORT=${DB_PORT:-8086} ;;
|
||||||
esac
|
esac
|
||||||
file_env 'DB_USER'
|
sanity_var DB_USER
|
||||||
file_env 'DB_PASS'
|
sanity_var DB_PASS
|
||||||
sanity_var INFLUX_VERSION "What InfluxDB version you are backing up from '1' or '2'"
|
sanity_var INFLUX_VERSION "What InfluxDB version you are backing up from '1' or '2'"
|
||||||
;;
|
;;
|
||||||
mongo* )
|
mongo* )
|
||||||
dbtype=mongo
|
dbtype=mongo
|
||||||
|
transform_file_var MONGO_CUSTOM_URI
|
||||||
if [ -n "${MONGO_CUSTOM_URI}" ] ; then
|
if [ -n "${MONGO_CUSTOM_URI}" ] ; then
|
||||||
mongo_uri_proto=$(echo ${MONGO_CUSTOM_URI} | grep :// | sed -e's,^\(.*://\).*,\1,g')
|
mongo_uri_proto=$(echo "${MONGO_CUSTOM_URI}" | grep :// | sed -e's,^\(.*://\).*,\1,g')
|
||||||
mongo_uri_scratch="${MONGO_CUSTOM_URI/${mongo_uri_proto}/}"
|
mongo_uri_scratch="${MONGO_CUSTOM_URI/${mongo_uri_proto}/}"
|
||||||
mongo_uri_username_password=$(echo ${mongo_uri_scratch} | grep @ | rev | cut -d@ -f2- | rev)
|
mongo_uri_username_password="$(echo "${mongo_uri_scratch}" | grep @ | rev | cut -d@ -f2- | rev)"
|
||||||
if [ -n "${mongo_uri_username_password}" ]; then mongo_uri_scratch=$(echo ${mongo_uri_scratch} | rev | cut -d@ -f1 | rev) ; fi
|
if [ -n "${mongo_uri_username_password}" ]; then mongo_uri_scratch="$(echo "${mongo_uri_scratch}" | rev | cut -d@ -f1 | rev)" ; fi
|
||||||
mongo_uri_port=$(echo ${mongo_uri_scratch} | grep : | rev | cut -d: -f2- | rev)
|
mongo_uri_port="$(echo "${mongo_uri_scratch}" | grep : | rev | cut -d: -f2- | rev)"
|
||||||
if [ -n "${mongo_uri_port}" ]; then mongo_uri_port=$(echo ${mongo_uri_scratch} | rev | cut -d: -f1 | cut -d/ -f2 | rev) ; fi
|
if [ -n "${mongo_uri_port}" ]; then mongo_uri_port="$(echo "${mongo_uri_scratch}" | rev | cut -d: -f1 | cut -d/ -f2 | rev)" ; fi
|
||||||
mongo_uri_hostname=$(echo ${mongo_uri_scratch} | cut -d/ -f1 | cut -d: -f1 )
|
mongo_uri_hostname="$(echo "${mongo_uri_scratch}" | cut -d/ -f1 | cut -d: -f1 )"
|
||||||
mongo_uri_database=$(echo ${mongo_uri_scratch} | cut -d/ -f2 | cut -d? -f1 )
|
mongo_uri_database="$(echo "${mongo_uri_scratch}" | cut -d/ -f2 | cut -d? -f1 )"
|
||||||
mongo_uri_options=$(echo ${mongo_uri_scratch} | cut -d/ -f2 | cut -d? -f2 )
|
mongo_uri_options="$(echo "${mongo_uri_scratch}" | cut -d/ -f2 | cut -d? -f2 )"
|
||||||
DB_NAME=${DB_NAME:-"${mongo_uri_database,,}"}
|
DB_NAME=${DB_NAME:-"${mongo_uri_database,,}"}
|
||||||
DB_HOST=${DB_HOST:-"${mongo_uri_hostname,,}"}
|
DB_HOST=${DB_HOST:-"${mongo_uri_hostname,,}"}
|
||||||
else
|
else
|
||||||
DB_PORT=${DB_PORT:-27017}
|
DB_PORT=${DB_PORT:-27017}
|
||||||
[[ ( -n "${DB_USER}" ) || ( -n "${DB_USER_FILE}" ) ]] && file_env 'DB_USER'
|
|
||||||
[[ ( -n "${DB_PASS}" ) || ( -n "${DB_PASS_FILE}" ) ]] && file_env 'DB_PASS'
|
|
||||||
[[ ( -n "${DB_USER}" ) ]] && MONGO_USER_STR=" --username ${DB_USER}"
|
[[ ( -n "${DB_USER}" ) ]] && MONGO_USER_STR=" --username ${DB_USER}"
|
||||||
[[ ( -n "${DB_PASS}" ) ]] && MONGO_PASS_STR=" --password ${DB_PASS}"
|
[[ ( -n "${DB_PASS}" ) ]] && MONGO_PASS_STR=" --password ${DB_PASS}"
|
||||||
[[ ( -n "${DB_NAME}" ) ]] && MONGO_DB_STR=" --db ${DB_NAME}"
|
[[ ( -n "${DB_NAME}" ) ]] && MONGO_DB_STR=" --db ${DB_NAME}"
|
||||||
|
transform_file_var DB_AUTH
|
||||||
[[ ( -n "${DB_AUTH}" ) ]] && MONGO_AUTH_STR=" --authenticationDatabase ${DB_AUTH}"
|
[[ ( -n "${DB_AUTH}" ) ]] && MONGO_AUTH_STR=" --authenticationDatabase ${DB_AUTH}"
|
||||||
fi
|
fi
|
||||||
;;
|
;;
|
||||||
@@ -47,8 +54,8 @@ bootstrap_variables() {
|
|||||||
dbtype=mysql
|
dbtype=mysql
|
||||||
DB_PORT=${DB_PORT:-3306}
|
DB_PORT=${DB_PORT:-3306}
|
||||||
sanity_var DB_NAME "Database Name to backup. Multiple seperated by commas"
|
sanity_var DB_NAME "Database Name to backup. Multiple seperated by commas"
|
||||||
[[ ( -n "${DB_PASS}" ) || ( -n "${DB_PASS_FILE}" ) ]] && file_env 'DB_PASS'
|
transform_file_var DB_PASS
|
||||||
[[ ( -n "${DB_PASS}" ) ]] && export MYSQL_PWD=${DB_PASS}
|
if [ -n "${DB_PASS}" ] ; then export MYSQL_PWD=${DB_PASS} ; fi
|
||||||
if var_true "${MYSQL_ENABLE_TLS}" ; then
|
if var_true "${MYSQL_ENABLE_TLS}" ; then
|
||||||
if [ -n "${MYSQL_TLS_CA_FILE}" ] ; then
|
if [ -n "${MYSQL_TLS_CA_FILE}" ] ; then
|
||||||
mysql_tls_args="--ssl_ca=${MYSQL_TLS_CA_FILE}"
|
mysql_tls_args="--ssl_ca=${MYSQL_TLS_CA_FILE}"
|
||||||
@@ -81,14 +88,12 @@ bootstrap_variables() {
|
|||||||
postgres* | "pgsql" )
|
postgres* | "pgsql" )
|
||||||
dbtype=pgsql
|
dbtype=pgsql
|
||||||
DB_PORT=${DB_PORT:-5432}
|
DB_PORT=${DB_PORT:-5432}
|
||||||
[[ ( -n "${DB_PASS}" ) || ( -n "${DB_PASS_FILE}" ) ]] && file_env 'DB_PASS'
|
|
||||||
[[ ( -n "${DB_PASS}" ) ]] && POSTGRES_PASS_STR="PGPASSWORD=${DB_PASS}"
|
[[ ( -n "${DB_PASS}" ) ]] && POSTGRES_PASS_STR="PGPASSWORD=${DB_PASS}"
|
||||||
sanity_var DB_NAME "Database Name to backup. Multiple seperated by commas"
|
sanity_var DB_NAME "Database Name to backup. Multiple seperated by commas"
|
||||||
;;
|
;;
|
||||||
"redis" )
|
"redis" )
|
||||||
dbtype=redis
|
dbtype=redis
|
||||||
DB_PORT=${DB_PORT:-6379}
|
DB_PORT=${DB_PORT:-6379}
|
||||||
[[ ( -n "${DB_PASS}" || ( -n "${DB_PASS_FILE}" ) ) ]] && file_env 'DB_PASS'
|
|
||||||
[[ ( -n "${DB_PASS}" ) ]] && REDIS_PASS_STR=" -a ${DB_PASS}"
|
[[ ( -n "${DB_PASS}" ) ]] && REDIS_PASS_STR=" -a ${DB_PASS}"
|
||||||
;;
|
;;
|
||||||
sqlite* )
|
sqlite* )
|
||||||
@@ -100,14 +105,23 @@ bootstrap_variables() {
|
|||||||
;;
|
;;
|
||||||
esac
|
esac
|
||||||
|
|
||||||
if [ "${BACKUP_LOCATION,,}" = "s3" ] || [ "${BACKUP_LOCATION,,}" = "minio" ] && [ -n "${S3_KEY_ID}" ] && [ -n "${S3_KEY_SECRET}" ]; then
|
if [ "${BACKUP_LOCATION,,}" = "s3" ] || [ "${BACKUP_LOCATION,,}" = "minio" ] ; then
|
||||||
file_env 'S3_KEY_ID'
|
transform_file_var \
|
||||||
file_env 'S3_KEY_SECRET'
|
S3_BUCKET \
|
||||||
|
S3_KEY_ID \
|
||||||
|
S3_KEY_SECRET \
|
||||||
|
S3_PATH \
|
||||||
|
S3_REGION \
|
||||||
|
S3_HOST \
|
||||||
|
S3_PROTOCOL \
|
||||||
|
S3_EXTRA_OPTS \
|
||||||
|
S3_CERT_CA_FILE
|
||||||
fi
|
fi
|
||||||
|
|
||||||
if [ "${BACKUP_LOCATION,,}" = "blobxfer" ] && [ -n "${BLOBXFER_STORAGE_ACCOUNT_FILE}" ] && [ -n "${BLOBXFER_STORAGE_ACCOUNT_KEY_FILE}" ]; then
|
if [ "${BACKUP_LOCATION,,}" = "blobxfer" ] ; then
|
||||||
file_env 'BLOBXFER_STORAGE_ACCOUNT_FILE'
|
transform_file_var \
|
||||||
file_env 'BLOBXFER_STORAGE_ACCOUNT_KEY_FILE'
|
BLOBXFER_STORAGE_ACCOUNT \
|
||||||
|
BLOBXFER_STORAGE_KEY
|
||||||
fi
|
fi
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -150,7 +164,7 @@ backup_influx() {
|
|||||||
print_notice "Creating archive file of '${target_dir}' with tar ${compression_string}"
|
print_notice "Creating archive file of '${target_dir}' with tar ${compression_string}"
|
||||||
tar cf - "${TEMP_LOCATION}"/"${target_dir}" | $dir_compress_cmd > "${TEMP_LOCATION}"/"${target_dir}".tar"${extension}"
|
tar cf - "${TEMP_LOCATION}"/"${target_dir}" | $dir_compress_cmd > "${TEMP_LOCATION}"/"${target_dir}".tar"${extension}"
|
||||||
target=influx_${db}_${DB_HOST#*//}_${now}.tar${extension}
|
target=influx_${db}_${DB_HOST#*//}_${now}.tar${extension}
|
||||||
target=influx_${db}_${DB_HOST#*//}
|
ltarget=influx_${db}_${DB_HOST#*//}
|
||||||
generate_checksum
|
generate_checksum
|
||||||
move_dbbackup
|
move_dbbackup
|
||||||
post_dbbackup $db
|
post_dbbackup $db
|
||||||
@@ -170,6 +184,7 @@ backup_influx() {
|
|||||||
check_exit_code $target_dir
|
check_exit_code $target_dir
|
||||||
create_archive
|
create_archive
|
||||||
target=influx2_${db}_${DB_HOST#*//}_${now}.tar${extension}
|
target=influx2_${db}_${DB_HOST#*//}_${now}.tar${extension}
|
||||||
|
ltarget=influx2_${db}_${DB_HOST#*//}
|
||||||
generate_checksum
|
generate_checksum
|
||||||
move_dbbackup
|
move_dbbackup
|
||||||
post_dbbackup $db
|
post_dbbackup $db
|
||||||
@@ -207,7 +222,7 @@ backup_mongo() {
|
|||||||
backup_mssql() {
|
backup_mssql() {
|
||||||
prepare_dbbackup
|
prepare_dbbackup
|
||||||
target=mssql_${DB_NAME,,}_${DB_HOST,,}_${now}.bak
|
target=mssql_${DB_NAME,,}_${DB_HOST,,}_${now}.bak
|
||||||
target=mssql_${DB_NAME,,}_${DB_HOST,,}
|
ltarget=mssql_${DB_NAME,,}_${DB_HOST,,}
|
||||||
compression
|
compression
|
||||||
pre_dbbackup "${DB_NAME}"
|
pre_dbbackup "${DB_NAME}"
|
||||||
print_notice "Dumping MSSQL database: '${DB_NAME}'"
|
print_notice "Dumping MSSQL database: '${DB_NAME}'"
|
||||||
@@ -369,7 +384,7 @@ backup_sqlite3() {
|
|||||||
db=$(basename "${DB_HOST}")
|
db=$(basename "${DB_HOST}")
|
||||||
db="${db%.*}"
|
db="${db%.*}"
|
||||||
target=sqlite3_${db}_${now}.sqlite3
|
target=sqlite3_${db}_${now}.sqlite3
|
||||||
target=sqlite3_${db}.sqlite3
|
ltarget=sqlite3_${db}.sqlite3
|
||||||
compression
|
compression
|
||||||
pre_dbbackup $db
|
pre_dbbackup $db
|
||||||
print_notice "Dumping sqlite3 database: '${DB_HOST}' ${compression_string}"
|
print_notice "Dumping sqlite3 database: '${DB_HOST}' ${compression_string}"
|
||||||
@@ -433,6 +448,7 @@ check_availability() {
|
|||||||
;;
|
;;
|
||||||
"mysql" )
|
"mysql" )
|
||||||
counter=0
|
counter=0
|
||||||
|
transform_file_var DB_PASS
|
||||||
export MYSQL_PWD=${DB_PASS}
|
export MYSQL_PWD=${DB_PASS}
|
||||||
while ! (mysqladmin -u"${DB_USER}" -P"${DB_PORT}" -h"${DB_HOST}" ${mysql_tls_args} status > /dev/null 2>&1) ; do
|
while ! (mysqladmin -u"${DB_USER}" -P"${DB_PORT}" -h"${DB_HOST}" ${mysql_tls_args} status > /dev/null 2>&1) ; do
|
||||||
sleep 5
|
sleep 5
|
||||||
@@ -671,8 +687,7 @@ move_dbbackup() {
|
|||||||
fi
|
fi
|
||||||
if [ -n "${DB_ARCHIVE_TIME}" ] ; then
|
if [ -n "${DB_ARCHIVE_TIME}" ] ; then
|
||||||
mkdir -p "${DB_DUMP_TARGET_ARCHIVE}"
|
mkdir -p "${DB_DUMP_TARGET_ARCHIVE}"
|
||||||
find "${DB_DUMP_TARGET}"/ -maxdepth 1 -mmin +"${DB_ARCHIVE_TIME}" -iname "*" -exec mv {} foo \;
|
find "${DB_DUMP_TARGET}"/ -maxdepth 1 -mmin +"${DB_ARCHIVE_TIME}" -iname "*" -exec mv {} "${DB_DUMP_TARGET_ARCHIVE}" \;
|
||||||
find "${DB_DUMP_TARGET}"/ -maxdepth 1 -mmin +"${DB_CLEANUP_TIME}" -iname "*" -exec rm {} \;
|
|
||||||
fi
|
fi
|
||||||
;;
|
;;
|
||||||
"s3" | "minio" )
|
"s3" | "minio" )
|
||||||
@@ -838,14 +853,6 @@ sanity_test() {
|
|||||||
sanity_var DB_NAME "Database Name to backup. Multiple seperated by commas"
|
sanity_var DB_NAME "Database Name to backup. Multiple seperated by commas"
|
||||||
;;
|
;;
|
||||||
esac
|
esac
|
||||||
|
|
||||||
if [ "${BACKUP_LOCATION,,}" = "s3" ] || [ "${BACKUP_LOCATION,,}" = "minio" ] && [ -n "${S3_KEY_ID}" ] && [ -n "${S3_KEY_SECRET}" ]; then
|
|
||||||
sanity_var S3_BUCKET "S3 Bucket"
|
|
||||||
sanity_var S3_PATH "S3 Path"
|
|
||||||
sanity_var S3_REGION "S3 Region"
|
|
||||||
file_env 'S3_KEY_ID'
|
|
||||||
file_env 'S3_KEY_SECRET'
|
|
||||||
fi
|
|
||||||
}
|
}
|
||||||
|
|
||||||
setup_mode() {
|
setup_mode() {
|
||||||
|
|||||||
@@ -37,7 +37,7 @@ else
|
|||||||
case "$1" in
|
case "$1" in
|
||||||
"-h" )
|
"-h" )
|
||||||
cat <<EOF
|
cat <<EOF
|
||||||
${IMAGE_NAME} Restore Tool
|
${IMAGE_NAME} Restore Tool ${IMAGE_VERSION}
|
||||||
(c) 2022 Dave Conroy (https://github.com/tiredofit)
|
(c) 2022 Dave Conroy (https://github.com/tiredofit)
|
||||||
|
|
||||||
This script will assist you in recovering databases taken by the Docker image.
|
This script will assist you in recovering databases taken by the Docker image.
|
||||||
|
|||||||
Reference in New Issue
Block a user