mirror of
https://github.com/tiredofit/docker-db-backup.git
synced 2025-12-22 13:44:08 +01:00
Compare commits
3 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
a5b15b4412 | ||
|
|
6692cf9834 | ||
|
|
c37de5778d |
19
CHANGELOG.md
19
CHANGELOG.md
@@ -1,3 +1,22 @@
|
|||||||
|
## 4.0.33 2023-12-18 <dave at tiredofit dot ca>
|
||||||
|
|
||||||
|
### Changed
|
||||||
|
- Allow _OPTS variabls to contain spaces
|
||||||
|
- Switch references of _DUMP_OPTS to _BACKUP_OPTS
|
||||||
|
|
||||||
|
|
||||||
|
## 4.0.32 2023-12-15 <dave at tiredofit dot ca>
|
||||||
|
|
||||||
|
### Changed
|
||||||
|
- Fix issue with directories not properly being backed up (InfluxDB)
|
||||||
|
|
||||||
|
|
||||||
|
## 4.0.31 2023-12-12 <dave at tiredofit dot ca>
|
||||||
|
|
||||||
|
### Changed
|
||||||
|
- Support backing up databases with spaces in them
|
||||||
|
|
||||||
|
|
||||||
## 4.0.30 2023-12-11 <dave at tiredofit dot ca>
|
## 4.0.30 2023-12-11 <dave at tiredofit dot ca>
|
||||||
|
|
||||||
### Added
|
### Added
|
||||||
|
|||||||
@@ -153,6 +153,11 @@ bootstrap_variables() {
|
|||||||
fi
|
fi
|
||||||
##
|
##
|
||||||
|
|
||||||
|
if grep -qo ".*_NAME='.*'" "${backup_instance_vars}"; then
|
||||||
|
print_debug "[bootstrap_variables] [backup_init] Found _NAME variable with quotes"
|
||||||
|
sed -i "s|_NAME='\(.*\)'|_NAME=\1|g" "${backup_instance_vars}"
|
||||||
|
fi
|
||||||
|
|
||||||
if grep -qo ".*_PASS='.*'" "${backup_instance_vars}"; then
|
if grep -qo ".*_PASS='.*'" "${backup_instance_vars}"; then
|
||||||
print_debug "[bootstrap_variables] [backup_init] Found _PASS variable with quotes"
|
print_debug "[bootstrap_variables] [backup_init] Found _PASS variable with quotes"
|
||||||
sed -i "s|_PASS='\(.*\)'|_PASS=\1|g" "${backup_instance_vars}"
|
sed -i "s|_PASS='\(.*\)'|_PASS=\1|g" "${backup_instance_vars}"
|
||||||
@@ -163,6 +168,11 @@ bootstrap_variables() {
|
|||||||
sed -i "s|MONGO_CUSTOM_URI='\(.*\)'|MONGO_CUSTOM_URI=\1|g" "${backup_instance_vars}"
|
sed -i "s|MONGO_CUSTOM_URI='\(.*\)'|MONGO_CUSTOM_URI=\1|g" "${backup_instance_vars}"
|
||||||
fi
|
fi
|
||||||
|
|
||||||
|
if grep -qo ".*_OPTS='.*'" "${backup_instance_vars}"; then
|
||||||
|
print_debug "[bootstrap_variables] [backup_init] Found _OPTS variable with quotes"
|
||||||
|
sed -i "s|_OPTS='\(.*\)'|_OPTS=\1|g" "${backup_instance_vars}"
|
||||||
|
fi
|
||||||
|
|
||||||
transform_backup_instance_variable() {
|
transform_backup_instance_variable() {
|
||||||
if grep -q "^DB${1}_${2}=" "${backup_instance_vars}" && [ "$(grep "^DB${1}_${2}=" "${backup_instance_vars}" | cut -d = -f2)" != "unset" ]; then
|
if grep -q "^DB${1}_${2}=" "${backup_instance_vars}" && [ "$(grep "^DB${1}_${2}=" "${backup_instance_vars}" | cut -d = -f2)" != "unset" ]; then
|
||||||
export "$3"="$(grep "^DB${1}_${2}=" "${backup_instance_vars}" | cut -d = -f2-)"
|
export "$3"="$(grep "^DB${1}_${2}=" "${backup_instance_vars}" | cut -d = -f2-)"
|
||||||
@@ -199,7 +209,7 @@ bootstrap_variables() {
|
|||||||
transform_backup_instance_variable "${backup_instance_number}" ENCRYPT_PASSPHRASE backup_job_encrypt_passphrase
|
transform_backup_instance_variable "${backup_instance_number}" ENCRYPT_PASSPHRASE backup_job_encrypt_passphrase
|
||||||
transform_backup_instance_variable "${backup_instance_number}" ENCRYPT_PRIVATE_KEY backup_job_encrypt_private_key
|
transform_backup_instance_variable "${backup_instance_number}" ENCRYPT_PRIVATE_KEY backup_job_encrypt_private_key
|
||||||
transform_backup_instance_variable "${backup_instance_number}" ENCRYPT_PUBLIC_KEY backup_job_encrypt_public_key
|
transform_backup_instance_variable "${backup_instance_number}" ENCRYPT_PUBLIC_KEY backup_job_encrypt_public_key
|
||||||
transform_backup_instance_variable "${backup_instance_number}" EXTRA_DUMP_OPTS backup_job_extra_dump_opts
|
transform_backup_instance_variable "${backup_instance_number}" EXTRA_BACKUP_OPTS backup_job_extra_backup_opts
|
||||||
transform_backup_instance_variable "${backup_instance_number}" EXTRA_ENUMERATION_OPTS backup_job_extra_enumeration_opts
|
transform_backup_instance_variable "${backup_instance_number}" EXTRA_ENUMERATION_OPTS backup_job_extra_enumeration_opts
|
||||||
transform_backup_instance_variable "${backup_instance_number}" EXTRA_OPTS backup_job_extra_opts
|
transform_backup_instance_variable "${backup_instance_number}" EXTRA_OPTS backup_job_extra_opts
|
||||||
transform_backup_instance_variable "${backup_instance_number}" FILESYSTEM_ARCHIVE_PATH backup_job_filesystem_archive_path
|
transform_backup_instance_variable "${backup_instance_number}" FILESYSTEM_ARCHIVE_PATH backup_job_filesystem_archive_path
|
||||||
@@ -505,15 +515,14 @@ backup_influx() {
|
|||||||
backup_job_filename=influx_${db}_${backup_job_db_host#*//}_${now}
|
backup_job_filename=influx_${db}_${backup_job_db_host#*//}_${now}
|
||||||
backup_job_filename_base=influx_${db}_${backup_job_db_host#*//}
|
backup_job_filename_base=influx_${db}_${backup_job_db_host#*//}
|
||||||
if var_true "${DEBUG_BACKUP_INFLUX}" ; then debug off; fi
|
if var_true "${DEBUG_BACKUP_INFLUX}" ; then debug off; fi
|
||||||
compression
|
|
||||||
pre_dbbackup "${db}"
|
pre_dbbackup "${db}"
|
||||||
write_log notice "Dumping Influx database: '${db}'"
|
write_log notice "Dumping Influx database: '${db}'"
|
||||||
if var_true "${DEBUG_BACKUP_INFLUX}" ; then debug on; fi
|
if var_true "${DEBUG_BACKUP_INFLUX}" ; then debug on; fi
|
||||||
run_as_user influxd backup ${influx_compression} ${bucket} -portable -host ${backup_job_db_host}:${backup_job_db_port} ${backup_job_extra_opts} ${backup_job_extra_dump_opts} "${temporary_directory}"/"${backup_job_filename}"
|
run_as_user influxd backup ${influx_compression} ${bucket} -portable -host ${backup_job_db_host}:${backup_job_db_port} ${backup_job_extra_opts} ${backup_job_extra_backup_opts} "${temporary_directory}"/"${backup_job_filename}"
|
||||||
exit_code=$?
|
exit_code=$?
|
||||||
check_exit_code backup "${backup_job_filename}"
|
check_exit_code backup "${backup_job_filename}"
|
||||||
|
compression
|
||||||
create_archive
|
create_archive
|
||||||
backup_job_filename=influx2_${db}_${backup_job_db_host#*//}_${now}.tar${extension}
|
|
||||||
if var_true "${DEBUG_BACKUP_INFLUX}" ; then debug off; fi
|
if var_true "${DEBUG_BACKUP_INFLUX}" ; then debug off; fi
|
||||||
timer backup finish
|
timer backup finish
|
||||||
file_encryption
|
file_encryption
|
||||||
@@ -533,15 +542,14 @@ backup_influx() {
|
|||||||
if var_true "${DEBUG_BACKUP_INFLUX}" ; then debug off; fi
|
if var_true "${DEBUG_BACKUP_INFLUX}" ; then debug off; fi
|
||||||
backup_job_filename=influx2_${db}_${backup_job_db_host#*//}_${now}
|
backup_job_filename=influx2_${db}_${backup_job_db_host#*//}_${now}
|
||||||
backup_job_filename_base=influx2_${db}_${backup_job_db_host#*//}
|
backup_job_filename_base=influx2_${db}_${backup_job_db_host#*//}
|
||||||
compression
|
|
||||||
pre_dbbackup "${db}"
|
pre_dbbackup "${db}"
|
||||||
write_log notice "Dumping Influx2 database: '${db}'"
|
write_log notice "Dumping Influx2 database: '${db}'"
|
||||||
if var_true "${DEBUG_BACKUP_INFLUX}" ; then debug on; fi
|
if var_true "${DEBUG_BACKUP_INFLUX}" ; then debug on; fi
|
||||||
run_as_user influx backup --org ${backup_job_db_user} ${bucket} --host ${backup_job_db_host}:${backup_job_db_port} --token ${backup_job_db_pass} ${backup_job_extra_opts} ${backup_job_extra_dump_opts} --compression none "${temporary_directory}"/"${backup_job_filename}"
|
run_as_user influx backup --org ${backup_job_db_user} ${bucket} --host ${backup_job_db_host}:${backup_job_db_port} --token ${backup_job_db_pass} ${backup_job_extra_opts} ${backup_job_extra_backup_opts} --compression none "${temporary_directory}"/"${backup_job_filename}"
|
||||||
exit_code=$?
|
exit_code=$?
|
||||||
check_exit_code backup "${backup_job_filename}"
|
check_exit_code backup "${backup_job_filename}"
|
||||||
|
compression
|
||||||
create_archive
|
create_archive
|
||||||
backup_job_filename=influx2_${db}_${backup_job_db_host#*//}_${now}.tar${extension}
|
|
||||||
if var_true "${DEBUG_BACKUP_INFLUX}" ; then debug off; fi
|
if var_true "${DEBUG_BACKUP_INFLUX}" ; then debug off; fi
|
||||||
timer backup finish
|
timer backup finish
|
||||||
file_encryption
|
file_encryption
|
||||||
@@ -568,9 +576,9 @@ backup_mongo() {
|
|||||||
compression_string="and compressing with gzip"
|
compression_string="and compressing with gzip"
|
||||||
fi
|
fi
|
||||||
if [ -n "${backup_job_mongo_custom_uri}" ] ; then
|
if [ -n "${backup_job_mongo_custom_uri}" ] ; then
|
||||||
mongo_backup_parameter="--uri=${backup_job_mongo_custom_uri} ${backup_job_extra_opts} ${backup_job_extra_dump_opts}"
|
mongo_backup_parameter="--uri=${backup_job_mongo_custom_uri} ${backup_job_extra_opts} ${backup_job_extra_backup_opts}"
|
||||||
else
|
else
|
||||||
mongo_backup_parameter="--host ${backup_job_db_host} --port ${backup_job_db_port} ${MONGO_USER_STR}${MONGO_PASS_STR}${MONGO_AUTH_STR}${MONGO_DB_STR} ${backup_job_extra_opts} ${backup_job_extra_dump_opts}"
|
mongo_backup_parameter="--host ${backup_job_db_host} --port ${backup_job_db_port} ${MONGO_USER_STR}${MONGO_PASS_STR}${MONGO_AUTH_STR}${MONGO_DB_STR} ${backup_job_extra_opts} ${backup_job_extra_backup_opts}"
|
||||||
fi
|
fi
|
||||||
if var_true "${DEBUG_BACKUP_MONGO}" ; then debug off; fi
|
if var_true "${DEBUG_BACKUP_MONGO}" ; then debug off; fi
|
||||||
pre_dbbackup "${backup_job_db_name}"
|
pre_dbbackup "${backup_job_db_name}"
|
||||||
@@ -599,7 +607,7 @@ backup_mssql() {
|
|||||||
pre_dbbackup "${backup_job_db_name}"
|
pre_dbbackup "${backup_job_db_name}"
|
||||||
write_log notice "Dumping MSSQL database: '${backup_job_db_name}'"
|
write_log notice "Dumping MSSQL database: '${backup_job_db_name}'"
|
||||||
if var_true "${DEBUG_BACKUP_MSSQL}" ; then debug on; fi
|
if var_true "${DEBUG_BACKUP_MSSQL}" ; then debug on; fi
|
||||||
silent run_as_user ${play_fair} /opt/mssql-tools18/bin/sqlcmd -C -S ${backup_job_db_host}\,${backup_job_db_port} -U ${backup_job_db_user} -P ${backup_job_db_pass} -Q "BACKUP DATABASE ['${backup_job_db_name}'] TO DISK = N'${temporary_directory}/${backup_job_filename}' WITH NOFORMAT, NOINIT, NAME = '${backup_job_db_name}-full', SKIP, NOREWIND, NOUNLOAD, STATS = 10"
|
silent run_as_user ${play_fair} /opt/mssql-tools18/bin/sqlcmd -C -S ${backup_job_db_host}\,${backup_job_db_port} -U ${backup_job_db_user} -P ${backup_job_db_pass} -Q "BACKUP DATABASE [${backup_job_db_name}] TO DISK = N'${temporary_directory}/${backup_job_filename}' WITH NOFORMAT, NOINIT, NAME = '${backup_job_db_name}-full', SKIP, NOREWIND, NOUNLOAD, STATS = 10"
|
||||||
exit_code=$?
|
exit_code=$?
|
||||||
if var_true "${DEBUG_BACKUP_MSSQL}" ; then debug off; fi
|
if var_true "${DEBUG_BACKUP_MSSQL}" ; then debug off; fi
|
||||||
backup_job_filename_original=${backup_job_filename}
|
backup_job_filename_original=${backup_job_filename}
|
||||||
@@ -680,7 +688,7 @@ backup_mysql() {
|
|||||||
pre_dbbackup "${db}"
|
pre_dbbackup "${db}"
|
||||||
write_log notice "Dumping MySQL/MariaDB database: '${db}' ${compression_string}"
|
write_log notice "Dumping MySQL/MariaDB database: '${db}' ${compression_string}"
|
||||||
if var_true "${DEBUG_BACKUP_MYSQL}" ; then debug on; fi
|
if var_true "${DEBUG_BACKUP_MYSQL}" ; then debug on; fi
|
||||||
run_as_user ${play_fair} mysqldump --max-allowed-packet=${backup_job_mysql_max_allowed_packet} -h ${backup_job_db_host} -P ${backup_job_db_port} -u${backup_job_db_user} ${events} ${single_transaction} ${stored_procedures} ${mysql_tls_args} ${backup_job_extra_opts} ${backup_job_extra_dump_opts} $db | ${compress_cmd} | run_as_user tee "${temporary_directory}"/"${backup_job_filename}" > /dev/null
|
run_as_user ${play_fair} mysqldump --max-allowed-packet=${backup_job_mysql_max_allowed_packet} -h ${backup_job_db_host} -P ${backup_job_db_port} -u${backup_job_db_user} ${events} ${single_transaction} ${stored_procedures} ${mysql_tls_args} ${backup_job_extra_opts} ${backup_job_extra_backup_opts} $db | ${compress_cmd} | run_as_user tee "${temporary_directory}"/"${backup_job_filename}" > /dev/null
|
||||||
exit_code=$?
|
exit_code=$?
|
||||||
if var_true "${DEBUG_BACKUP_MYSQL}" ; then debug off; fi
|
if var_true "${DEBUG_BACKUP_MYSQL}" ; then debug off; fi
|
||||||
check_exit_code backup "${backup_job_filename}"
|
check_exit_code backup "${backup_job_filename}"
|
||||||
@@ -701,7 +709,7 @@ backup_mysql() {
|
|||||||
pre_dbbackup all
|
pre_dbbackup all
|
||||||
write_log notice "Dumping all MySQL / MariaDB databases: '$(echo ${db_names} | xargs | tr ' ' ',')' ${compression_string}"
|
write_log notice "Dumping all MySQL / MariaDB databases: '$(echo ${db_names} | xargs | tr ' ' ',')' ${compression_string}"
|
||||||
if var_true "${DEBUG_BACKUP_MYSQL}" ; then debug on; fi
|
if var_true "${DEBUG_BACKUP_MYSQL}" ; then debug on; fi
|
||||||
run_as_user ${play_fair} mysqldump --max-allowed-packet=${backup_job_mysql_max_allowed_packet} -h ${backup_job_db_host} -P ${backup_job_db_port} -u${backup_job_db_user} ${events} ${single_transaction} ${stored_procedures} ${mysql_tls_args} ${backup_job_extra_opts} ${backup_job_extra_dump_opts} --databases $(echo ${db_names} | xargs) | ${compress_cmd} | run_as_user tee "${temporary_directory}"/"${backup_job_filename}" > /dev/null
|
run_as_user ${play_fair} mysqldump --max-allowed-packet=${backup_job_mysql_max_allowed_packet} -h ${backup_job_db_host} -P ${backup_job_db_port} -u${backup_job_db_user} ${events} ${single_transaction} ${stored_procedures} ${mysql_tls_args} ${backup_job_extra_opts} ${backup_job_extra_backup_opts} --databases $(echo ${db_names} | xargs) | ${compress_cmd} | run_as_user tee "${temporary_directory}"/"${backup_job_filename}" > /dev/null
|
||||||
exit_code=$?
|
exit_code=$?
|
||||||
if var_true "${DEBUG_BACKUP_MYSQL}" ; then debug off; fi
|
if var_true "${DEBUG_BACKUP_MYSQL}" ; then debug off; fi
|
||||||
check_exit_code backup "${backup_job_filename}"
|
check_exit_code backup "${backup_job_filename}"
|
||||||
@@ -723,7 +731,7 @@ backup_pgsql() {
|
|||||||
pre_dbbackup "globals"
|
pre_dbbackup "globals"
|
||||||
print_notice "Dumping PostgresSQL globals: with 'pg_dumpall -g' ${compression_string}"
|
print_notice "Dumping PostgresSQL globals: with 'pg_dumpall -g' ${compression_string}"
|
||||||
if var_true "${DEBUG_BACKUP_PGSQL}" ; then debug on; fi
|
if var_true "${DEBUG_BACKUP_PGSQL}" ; then debug on; fi
|
||||||
run_as_user ${play_fair} pg_dumpall -h "${backup_job_db_host}" -U "${backup_job_db_user}" -p "${backup_job_db_port}" -g ${backup_job_extra_opts} ${backup_job_extra_dump_opts} | ${compress_cmd} | run_as_user tee "${temporary_directory}"/"${backup_job_filename}" > /dev/null
|
run_as_user ${play_fair} pg_dumpall -h "${backup_job_db_host}" -U "${backup_job_db_user}" -p "${backup_job_db_port}" -g ${backup_job_extra_opts} ${backup_job_extra_backup_opts} | ${compress_cmd} | run_as_user tee "${temporary_directory}"/"${backup_job_filename}" > /dev/null
|
||||||
exit_code=$?
|
exit_code=$?
|
||||||
if var_true "${DEBUG_BACKUP_PGSQL}" ; then debug off; fi
|
if var_true "${DEBUG_BACKUP_PGSQL}" ; then debug off; fi
|
||||||
check_exit_code "${backup_job_filename}"
|
check_exit_code "${backup_job_filename}"
|
||||||
@@ -772,7 +780,7 @@ backup_pgsql() {
|
|||||||
pre_dbbackup "${db}"
|
pre_dbbackup "${db}"
|
||||||
write_log notice "Dumping PostgresSQL database: '${db}' ${compression_string}"
|
write_log notice "Dumping PostgresSQL database: '${db}' ${compression_string}"
|
||||||
if var_true "${DEBUG_BACKUP_PGSQL}" ; then debug on; fi
|
if var_true "${DEBUG_BACKUP_PGSQL}" ; then debug on; fi
|
||||||
run_as_user ${play_fair} pg_dump -h "${backup_job_db_host}" -p "${backup_job_db_port}" -U "${backup_job_db_user}" $db ${backup_job_extra_opts} ${backup_job_extra_dump_opts} | ${compress_cmd} | run_as_user tee "${temporary_directory}"/"${backup_job_filename}" > /dev/null
|
run_as_user ${play_fair} pg_dump -h "${backup_job_db_host}" -p "${backup_job_db_port}" -U "${backup_job_db_user}" $db ${backup_job_extra_opts} ${backup_job_extra_backup_opts} | ${compress_cmd} | run_as_user tee "${temporary_directory}"/"${backup_job_filename}" > /dev/null
|
||||||
exit_code=$?
|
exit_code=$?
|
||||||
if var_true "${DEBUG_BACKUP_PGSQL}" ; then debug off; fi
|
if var_true "${DEBUG_BACKUP_PGSQL}" ; then debug off; fi
|
||||||
check_exit_code backup "${backup_job_filename}"
|
check_exit_code backup "${backup_job_filename}"
|
||||||
@@ -802,7 +810,7 @@ backup_pgsql() {
|
|||||||
for x_db_name in ${tmp_db_names} ; do
|
for x_db_name in ${tmp_db_names} ; do
|
||||||
pgexclude_arg=$(echo ${pgexclude_arg} --exclude-database=${x_db_name})
|
pgexclude_arg=$(echo ${pgexclude_arg} --exclude-database=${x_db_name})
|
||||||
done
|
done
|
||||||
run_as_user ${play_fair} pg_dumpall -h ${backup_job_db_host} -p ${backup_job_db_port} -U ${backup_job_db_user} ${pgexclude_arg} ${backup_job_extra_opts} ${backup_job_extra_dump_opts} | ${compress_cmd} | run_as_user tee "${temporary_directory}"/"${backup_job_filename}" > /dev/null
|
run_as_user ${play_fair} pg_dumpall -h ${backup_job_db_host} -p ${backup_job_db_port} -U ${backup_job_db_user} ${pgexclude_arg} ${backup_job_extra_opts} ${backup_job_extra_backup_opts} | ${compress_cmd} | run_as_user tee "${temporary_directory}"/"${backup_job_filename}" > /dev/null
|
||||||
exit_code=$?
|
exit_code=$?
|
||||||
if var_true "${DEBUG_BACKUP_PGSQL}" ; then debug off; fi
|
if var_true "${DEBUG_BACKUP_PGSQL}" ; then debug off; fi
|
||||||
check_exit_code backup "${backup_job_filename}"
|
check_exit_code backup "${backup_job_filename}"
|
||||||
@@ -823,7 +831,7 @@ backup_redis() {
|
|||||||
backup_job_filename=redis_all_${backup_job_db_host,,}_${now}.rdb
|
backup_job_filename=redis_all_${backup_job_db_host,,}_${now}.rdb
|
||||||
backup_job_filename_base=redis_${backup_job_db_host,,}
|
backup_job_filename_base=redis_${backup_job_db_host,,}
|
||||||
if var_true "${DEBUG_BACKUP_REDIS}" ; then debug on; fi
|
if var_true "${DEBUG_BACKUP_REDIS}" ; then debug on; fi
|
||||||
echo bgsave | silent run_as_user ${play_fair} redis-cli -h ${backup_job_db_host} -p ${backup_job_db_port} ${REDIS_PASS_STR} --rdb ${temporary_directory}/${backup_job_filename} ${backup_job_extra_opts} ${backup_job_extra_dump_opts}
|
echo bgsave | silent run_as_user ${play_fair} redis-cli -h ${backup_job_db_host} -p ${backup_job_db_port} ${REDIS_PASS_STR} --rdb ${temporary_directory}/${backup_job_filename} ${backup_job_extra_opts} ${backup_job_extra_backup_opts}
|
||||||
sleep 10
|
sleep 10
|
||||||
try=5
|
try=5
|
||||||
while [ $try -gt 0 ] ; do
|
while [ $try -gt 0 ] ; do
|
||||||
@@ -1151,8 +1159,9 @@ create_archive() {
|
|||||||
if var_true "${DEBUG_CREATE_ARCHIVE}" ; then debug on; fi
|
if var_true "${DEBUG_CREATE_ARCHIVE}" ; then debug on; fi
|
||||||
if [ "${exit_code}" = "0" ] ; then
|
if [ "${exit_code}" = "0" ] ; then
|
||||||
write_log notice "Creating archive file of '${backup_job_filename}' with tar ${compression_string}"
|
write_log notice "Creating archive file of '${backup_job_filename}' with tar ${compression_string}"
|
||||||
run_as_user tar cf - "${temporary_directory}"/"${backup_job_filename}" | ${dir_compress_cmd} | run_as_user tee "${temporary_directory}"/"${backup_job_filename}".tar"${extension}" > /dev/null
|
run_as_user tar cf - "${temporary_directory}"/"${backup_job_filename_dir}" | ${dir_compress_cmd} | run_as_user tee "${temporary_directory}"/"${backup_job_filename_dir}".tar"${extension}" > /dev/null
|
||||||
rm -rf "${temporary_directory}"/"${backup_job_filename}"
|
backup_job_filename="${backup_job_filename_dir}".tar"${extension}"
|
||||||
|
rm -rf "${temporary_directory}"/"${backup_job_filename_dir}"
|
||||||
else
|
else
|
||||||
write_log error "Skipping creating archive file because backup did not complete successfully"
|
write_log error "Skipping creating archive file because backup did not complete successfully"
|
||||||
fi
|
fi
|
||||||
@@ -1167,8 +1176,8 @@ create_schedulers() {
|
|||||||
print_debug "[create_schedulers] Found '${backup_instances}' DB_HOST instances"
|
print_debug "[create_schedulers] Found '${backup_instances}' DB_HOST instances"
|
||||||
|
|
||||||
if [ -n "${DB_HOST}" ] && [ "${backup_instances}" ]; then
|
if [ -n "${DB_HOST}" ] && [ "${backup_instances}" ]; then
|
||||||
backup_instances=1;
|
backup_instances=1;
|
||||||
print_debug "[create_schedulers] Detected using old DB_ variables"
|
print_debug "[create_schedulers] Detected using old DB_ variables"
|
||||||
fi
|
fi
|
||||||
|
|
||||||
for (( instance = 01; instance <= backup_instances; )) ; do
|
for (( instance = 01; instance <= backup_instances; )) ; do
|
||||||
|
|||||||
Reference in New Issue
Block a user