|
|
|
|
@@ -5,26 +5,26 @@ bootstrap_filesystem() {
|
|
|
|
|
if [ ! -d "${backup_job_filesystem_path}" ]; then
|
|
|
|
|
mkdir -p "${backup_job_filesystem_path}"
|
|
|
|
|
fi
|
|
|
|
|
if [ "$(stat -c %U "${backup_job_filesystem_path}")" != "dbbackup" ] ; then chown -R dbbackup:dbbackup "${backup_job_filesystem_path}" ; fi
|
|
|
|
|
if [ "$(stat -c %a "${backup_job_filesystem_path}")" != "${backup_job_filesystem_permission}" ] ; then chmod -R "${backup_job_filesystem_permission}" "${backup_job_filesystem_path}" ; fi
|
|
|
|
|
if [ "$(stat -c %U "${backup_job_filesystem_path}")" != "${DBBACKUP_USER}" ] ; then chown -R "${DBBACKUP_USER}":"${DBBACKUP_GROUP}" "${backup_job_filesystem_path}" ; fi
|
|
|
|
|
if [ "$(stat -c %a "${backup_job_filesystem_path}")" != "${backup_job_filesystem_path_permission}" ] ; then chmod "${backup_job_filesystem_path_permission}" "${backup_job_filesystem_path}" ; fi
|
|
|
|
|
|
|
|
|
|
if [ -d "${backup_job_filesystem_archive_path}" ]; then
|
|
|
|
|
if [ "$(stat -c %U "${backup_job_filesystem_archive_path}")" != "dbbackup" ] ; then chown -R dbbackup:dbbackup "${backup_job_filesystem_archive_path}" ; fi
|
|
|
|
|
if [ "$(stat -c %a "${backup_job_filesystem_archive_path}")" != "${backup_job_filesystem_permission}" ] ; then chmod -R "${backup_job_filesystem_permission}" "${backup_job_filesystem_archive_path}" ; fi
|
|
|
|
|
if [ "$(stat -c %U "${backup_job_filesystem_archive_path}")" != "${DBBACKUP_USER}" ] ; then chown -R "${DBBACKUP_USER}":"${DBBACKUP_GROUP}" "${backup_job_filesystem_archive_path}" ; fi
|
|
|
|
|
if [ "$(stat -c %a "${backup_job_filesystem_archive_path}")" != "${backup_job_filesystem_path_permission}" ] ; then chmod "${backup_job_filesystem_path_permission}" "${backup_job_filesystem_archive_path}" ; fi
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
if [ ! -d "${LOG_PATH}" ]; then
|
|
|
|
|
mkdir -p "${LOG_PATH}"
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
if [ "$(stat -c %U "${LOG_PATH}")" != "dbbackup" ] ; then chown dbbackup:dbbackup "${LOG_PATH}" ; fi
|
|
|
|
|
if [ "$(stat -c %U "${LOG_PATH}")" != "${DBBACKUP_USER}" ] ; then chown -R "${DBBACKUP_USER}":"${DBBACKUP_GROUP}" "${LOG_PATH}" ; fi
|
|
|
|
|
if [ ! -d "${LOG_PATH}"/"$(date +'%Y%m%d')" ]; then run_as_user mkdir -p "${LOG_PATH}"/"$(date +'%Y%m%d')"; fi
|
|
|
|
|
if [ "$(stat -c %a "${LOG_PATH}")" != "755" ] ; then chmod -R 755 "${LOG_PATH}" ; fi
|
|
|
|
|
|
|
|
|
|
if [ ! -d "${TEMP_PATH}" ]; then
|
|
|
|
|
mkdir -p "${TEMP_PATH}"
|
|
|
|
|
fi
|
|
|
|
|
if [ "$(stat -c %U "${TEMP_PATH}")" != "dbbackup" ] ; then chown -R dbbackup:dbbackup "${TEMP_PATH}" ; fi
|
|
|
|
|
if [ "$(stat -c %U "${TEMP_PATH}")" != "${DBBACKUP_USER}" ] ; then chown -R "${DBBACKUP_USER}":"${DBBACKUP_GROUP}" "${TEMP_PATH}" ; fi
|
|
|
|
|
if var_true "${DEBUG_BOOTSTRAP_FILESYSTEM}" ; then debug off; fi
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
@@ -48,7 +48,8 @@ bootstrap_variables() {
|
|
|
|
|
DEFAULT_USER \
|
|
|
|
|
DEFAULT_PASS \
|
|
|
|
|
DEFAULT_ENCRYPT_PASSPHRASE \
|
|
|
|
|
DEFAULT_ENCRYPT_PUBKEY \
|
|
|
|
|
DEFAULT_ENCRYPT_PUBLIC_KEY \
|
|
|
|
|
DEFAULT_ENCRYPT_PRIVATE_KEY \
|
|
|
|
|
DEFAULT_MONGO_CUSTOM_URI \
|
|
|
|
|
DEFAULT_MYSQL_TLS_CA_FILE \
|
|
|
|
|
DEFAULT_MYSQL_TLS_CERT_FILE \
|
|
|
|
|
@@ -74,7 +75,8 @@ bootstrap_variables() {
|
|
|
|
|
DB"${backup_instance_number}"_USER \
|
|
|
|
|
DB"${backup_instance_number}"_PASS \
|
|
|
|
|
DB"${backup_instance_number}"_ENCRYPT_PASSPHRASE \
|
|
|
|
|
DB"${backup_instance_number}"_ENCRYPT_PUBKEY \
|
|
|
|
|
DB"${backup_instance_number}"_ENCRYPT_PUBLIC_KEY \
|
|
|
|
|
DB"${backup_instance_number}"_ENCRYPT_PRIVATE_KEY \
|
|
|
|
|
DB"${backup_instance_number}"_MONGO_CUSTOM_URI \
|
|
|
|
|
DB"${backup_instance_number}"_MYSQL_TLS_CA_FILE \
|
|
|
|
|
DB"${backup_instance_number}"_MYSQL_TLS_CERT_FILE \
|
|
|
|
|
@@ -151,18 +153,28 @@ bootstrap_variables() {
|
|
|
|
|
fi
|
|
|
|
|
##
|
|
|
|
|
|
|
|
|
|
if grep -qo ".*_PASS='.*'" "${backup_instance_vars}"; then
|
|
|
|
|
print_debug "[bootstrap_variables] [backup_init] Found _PASS variable with quotes"
|
|
|
|
|
sed -i "s|_PASS='\(.*\)'|_PASS=\1|g" "${backup_instance_vars}"
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
if grep -qo "MONGO_CUSTOM_URI='.*'" "${backup_instance_vars}"; then
|
|
|
|
|
print_debug "[bootstrap_variables] [backup_init] Found _MONGO_CUSTOM_URI variable with quotes"
|
|
|
|
|
sed -i "s|MONGO_CUSTOM_URI='\(.*\)'|MONGO_CUSTOM_URI=\1|g" "${backup_instance_vars}"
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
transform_backup_instance_variable() {
|
|
|
|
|
if grep -q "^DB${1}_${2}=" "${backup_instance_vars}" && [ "$(grep "^DB${1}_${2}=" "${backup_instance_vars}" | cut -d = -f2)" != "unset" ]; then
|
|
|
|
|
export "$3"="$(grep "^DB${1}_${2}=" "${backup_instance_vars}" | cut -d = -f2)"
|
|
|
|
|
export "$3"="$(grep "^DB${1}_${2}=" "${backup_instance_vars}" | cut -d = -f2-)"
|
|
|
|
|
elif grep -q "^DB_${2}=" "${backup_instance_vars}" && [ "$(grep "^DB_${2}=" "${backup_instance_vars}" | cut -d = -f2)" != "unset" ]; then
|
|
|
|
|
# Allow old legacy work, perhaps remove old DB_ functionality in future? This should allow for seamless upgrades
|
|
|
|
|
#print_warn "Legacy Variable 'DB_${2}'' detected - Please upgrade your variables as they will be removed in version 4.3.0"
|
|
|
|
|
export "$3"="$(grep "^DB_${2}=" "${backup_instance_vars}" | cut -d = -f2)"
|
|
|
|
|
export "$3"="$(grep "^DB_${2}=" "${backup_instance_vars}" | cut -d = -f2-)"
|
|
|
|
|
elif grep -q "^${2}=" "${backup_instance_vars}" && [ "$(grep "^${2}=" "${backup_instance_vars}" | cut -d = -f2)" != "unset" ]; then
|
|
|
|
|
print_warn "Legacy unsupported variable '${2}' detected - Please upgrade your variables as they will be removed in version 4.3.0"
|
|
|
|
|
export "$3"="$(grep "^${2}=" "${backup_instance_vars}" | cut -d = -f2)"
|
|
|
|
|
export "$3"="$(grep "^${2}=" "${backup_instance_vars}" | cut -d = -f2-)"
|
|
|
|
|
elif grep -q "^DEFAULT_${2}=" "${backup_instance_vars}" && [ "$(grep "^DEFAULT_${2}=" "${backup_instance_vars}" | cut -d = -f2)" != "unset" ]; then
|
|
|
|
|
export "$3"="$(grep "^DEFAULT_${2}=" "${backup_instance_vars}" | cut -d = -f2)"
|
|
|
|
|
export "$3"="$(grep "^DEFAULT_${2}=" "${backup_instance_vars}" | cut -d = -f2-)"
|
|
|
|
|
fi
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
@@ -185,12 +197,14 @@ bootstrap_variables() {
|
|
|
|
|
transform_backup_instance_variable "${backup_instance_number}" ENABLE_PARALLEL_COMPRESSION backup_job_parallel_compression
|
|
|
|
|
transform_backup_instance_variable "${backup_instance_number}" ENCRYPT backup_job_encrypt
|
|
|
|
|
transform_backup_instance_variable "${backup_instance_number}" ENCRYPT_PASSPHRASE backup_job_encrypt_passphrase
|
|
|
|
|
transform_backup_instance_variable "${backup_instance_number}" ENCRYPT_PUBKEY backup_job_encrypt_pubkey
|
|
|
|
|
transform_backup_instance_variable "${backup_instance_number}" ENCRYPT_PRIVATE_KEY backup_job_encrypt_private_key
|
|
|
|
|
transform_backup_instance_variable "${backup_instance_number}" ENCRYPT_PUBLIC_KEY backup_job_encrypt_public_key
|
|
|
|
|
transform_backup_instance_variable "${backup_instance_number}" EXTRA_DUMP_OPTS backup_job_extra_dump_opts
|
|
|
|
|
transform_backup_instance_variable "${backup_instance_number}" EXTRA_ENUMERATION_OPTS backup_job_extra_enumeration_opts
|
|
|
|
|
transform_backup_instance_variable "${backup_instance_number}" EXTRA_OPTS backup_job_extra_opts
|
|
|
|
|
transform_backup_instance_variable "${backup_instance_number}" FILESYSTEM_ARCHIVE_PATH backup_job_filesystem_archive_path
|
|
|
|
|
transform_backup_instance_variable "${backup_instance_number}" FILESYSTEM_PATH backup_job_filesystem_path
|
|
|
|
|
transform_backup_instance_variable "${backup_instance_number}" FILESYSTEM_PATH_PERMISSION backup_job_filesystem_path_permission
|
|
|
|
|
transform_backup_instance_variable "${backup_instance_number}" FILESYSTEM_PERMISSION backup_job_filesystem_permission
|
|
|
|
|
transform_backup_instance_variable "${backup_instance_number}" GZ_RSYNCABLE backup_job_gz_rsyncable
|
|
|
|
|
transform_backup_instance_variable "${backup_instance_number}" HOST backup_job_db_host
|
|
|
|
|
@@ -237,6 +251,14 @@ bootstrap_variables() {
|
|
|
|
|
transform_backup_instance_variable "${backup_instance_number}" USER backup_job_db_user
|
|
|
|
|
|
|
|
|
|
backup_job_backup_begin=$(echo "${backup_job_backup_begin}" | sed -e "s|'||g" -e 's|"||g')
|
|
|
|
|
if var_true "${DEBUG_BACKUP_INSTANCE_VARIABLE}" ; then cat <<EOF
|
|
|
|
|
## BEGIN Variable Dump $(TZ=${TIMEZONE} date)
|
|
|
|
|
|
|
|
|
|
$(cat ${backup_instance_vars})
|
|
|
|
|
|
|
|
|
|
## END
|
|
|
|
|
EOF
|
|
|
|
|
fi
|
|
|
|
|
rm -rf "${backup_instance_vars}"
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
@@ -315,7 +337,7 @@ bootstrap_variables() {
|
|
|
|
|
## Check is Variable is Defined
|
|
|
|
|
## Usage: check_var transformed_varname real_varname "Description"
|
|
|
|
|
output_off
|
|
|
|
|
print_debug "Looking for existence of $2 environment variable"
|
|
|
|
|
print_debug "[parse_variables] Looking for existence of $2 environment variable"
|
|
|
|
|
if [ ! -v "$1" ]; then
|
|
|
|
|
print_error "No '$3' Entered! - Set '\$$2' environment variable - Halting Backup Number ${v_instance}"
|
|
|
|
|
s6-svc -d /var/run/s6/legacy-services/dbbackup-"${v_instance}"
|
|
|
|
|
@@ -451,12 +473,7 @@ backup_couch() {
|
|
|
|
|
prepare_dbbackup
|
|
|
|
|
backup_job_filename=couch_${backup_job_db_name}_${backup_job_db_host#*//}_${now}.txt
|
|
|
|
|
backup_job_filename_base=couch_${backup_job_db_name}_${backup_job_db_host#*//}
|
|
|
|
|
compression
|
|
|
|
|
pre_dbbackup ${backup_job_db_name}
|
|
|
|
|
write_log notice "Dumping CouchDB database: '${backup_job_db_name}' ${compression_string}"
|
|
|
|
|
if var_true "${DEBUG_BACKUP_COUCH}" ; then debug on; fi
|
|
|
|
|
run_as_user curl -sSL -X GET ${backup_job_db_host}:${backup_job_db_port}/${backup_job_db_name}/_all_docs?include_docs=true | ${compress_cmd} | run_as_user tee "${TEMP_PATH}"/"${backup_job_filename}" > /dev/null
|
|
|
|
|
exit_code=$?
|
|
|
|
|
compressionzyclonite
|
|
|
|
|
if var_true "${DEBUG_BACKUP_COUCH}" ; then debug off; fi
|
|
|
|
|
check_exit_code backup "${backup_job_filename}"
|
|
|
|
|
timer backup finish
|
|
|
|
|
@@ -465,23 +482,29 @@ backup_couch() {
|
|
|
|
|
move_dbbackup
|
|
|
|
|
check_exit_code move "${backup_job_filename}"
|
|
|
|
|
post_dbbackup ${backup_job_db_name}
|
|
|
|
|
cleanup_old_data
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
backup_influx() {
|
|
|
|
|
if var_true "${DEBUG_BACKUP_INFLUX}" ; then debug on; fi
|
|
|
|
|
if [ "${backup_job_db_name,,}" = "all" ] ; then
|
|
|
|
|
write_log debug "Preparing to back up everything"
|
|
|
|
|
write_log debug "[backup_influx] Preparing to back up everything"
|
|
|
|
|
db_names=justbackupeverything
|
|
|
|
|
else
|
|
|
|
|
db_names=$(echo "${backup_job_db_name}" | tr ',' '\n')
|
|
|
|
|
fi
|
|
|
|
|
if var_true "${DEBUG_BACKUP_INFLUX}" ; then debug off; fi
|
|
|
|
|
|
|
|
|
|
case "${backup_job_db_influx_version,,}" in
|
|
|
|
|
case "${backup_job_influx_version,,}" in
|
|
|
|
|
1 )
|
|
|
|
|
print_debug "[backup_influx] Influx DB Version 1 selected"
|
|
|
|
|
for db in ${db_names}; do
|
|
|
|
|
prepare_dbbackup
|
|
|
|
|
if var_true "${DEBUG_BACKUP_INFLUX}" ; then debug on; fi
|
|
|
|
|
if [ "${db}" != "justbackupeverything" ] ; then bucket="-db ${db}" ; else db=all ; fi
|
|
|
|
|
backup_job_filename=influx_${db}_${backup_job_db_host#*//}_${now}
|
|
|
|
|
backup_job_filename_base=influx_${db}_${backup_job_db_host#*//}
|
|
|
|
|
if var_true "${DEBUG_BACKUP_INFLUX}" ; then debug off; fi
|
|
|
|
|
compression
|
|
|
|
|
pre_dbbackup "${db}"
|
|
|
|
|
write_log notice "Dumping Influx database: '${db}'"
|
|
|
|
|
@@ -500,12 +523,16 @@ backup_influx() {
|
|
|
|
|
move_dbbackup
|
|
|
|
|
check_exit_code move "${backup_job_filename_dir}"
|
|
|
|
|
post_dbbackup "${db}"
|
|
|
|
|
cleanup_old_data
|
|
|
|
|
done
|
|
|
|
|
;;
|
|
|
|
|
2 )
|
|
|
|
|
print_debug "[backup_influx] Influx DB Version 2 selected"
|
|
|
|
|
for db in ${db_names}; do
|
|
|
|
|
prepare_dbbackup
|
|
|
|
|
if var_true "${DEBUG_BACKUP_INFLUX}" ; then debug on; fi
|
|
|
|
|
if [ "${db}" != "justbackupeverything" ] ; then bucket="--bucket $db" ; else db=all ; fi
|
|
|
|
|
if var_true "${DEBUG_BACKUP_INFLUX}" ; then debug off; fi
|
|
|
|
|
backup_job_filename=influx2_${db}_${backup_job_db_host#*//}_${now}
|
|
|
|
|
backup_job_filename_base=influx2_${db}_${backup_job_db_host#*//}
|
|
|
|
|
compression
|
|
|
|
|
@@ -515,16 +542,17 @@ backup_influx() {
|
|
|
|
|
run_as_user influx backup --org ${backup_job_db_user} ${bucket} --host ${backup_job_db_host}:${backup_job_db_port} --token ${backup_job_db_pass} ${backup_job_extra_opts} ${backup_job_extra_dump_opts} --compression none "${TEMP_PATH}"/"${backup_job_filename_dir}"
|
|
|
|
|
exit_code=$?
|
|
|
|
|
check_exit_code backup "${backup_job_filename_dir}"
|
|
|
|
|
if var_true "${DEBUG_BACKUP_INFLUX}" ; then debug off; fi
|
|
|
|
|
create_archive
|
|
|
|
|
backup_job_filename=influx2_${db}_${backup_job_db_host#*//}_${now}.tar${extension}
|
|
|
|
|
backup_job_filename_base=influx2_${db}_${backup_job_db_host#*//}
|
|
|
|
|
if var_true "${DEBUG_BACKUP_INFLUX}" ; then debug off; fi
|
|
|
|
|
timer backup finish
|
|
|
|
|
file_encryption
|
|
|
|
|
generate_checksum
|
|
|
|
|
move_dbbackup
|
|
|
|
|
check_exit_code move "${backup_job_filename_dir}"
|
|
|
|
|
post_dbbackup "${db}"
|
|
|
|
|
cleanup_old_data
|
|
|
|
|
done
|
|
|
|
|
;;
|
|
|
|
|
esac
|
|
|
|
|
@@ -532,6 +560,7 @@ backup_influx() {
|
|
|
|
|
|
|
|
|
|
backup_mongo() {
|
|
|
|
|
prepare_dbbackup
|
|
|
|
|
if var_true "${DEBUG_BACKUP_MONGO}" ; then debug on; fi
|
|
|
|
|
if [ "${backup_job_compression,,}" = "none" ] ; then
|
|
|
|
|
backup_job_filename=${dbtype}_${backup_job_db_name,,}_${backup_job_db_host,,}_${now}.archive
|
|
|
|
|
backup_job_filename_base=${dbtype}_${backup_job_db_name,,}_${backup_job_db_host,,}
|
|
|
|
|
@@ -546,6 +575,7 @@ backup_mongo() {
|
|
|
|
|
else
|
|
|
|
|
mongo_backup_parameter="--host ${backup_job_db_host} --port ${backup_job_db_port} ${MONGO_USER_STR}${MONGO_PASS_STR}${MONGO_AUTH_STR}${MONGO_DB_STR} ${backup_job_extra_opts} ${backup_job_extra_dump_opts}"
|
|
|
|
|
fi
|
|
|
|
|
if var_true "${DEBUG_BACKUP_MONGO}" ; then debug off; fi
|
|
|
|
|
pre_dbbackup "${backup_job_db_name}"
|
|
|
|
|
write_log notice "Dumping MongoDB database: '${DB_NAME}' ${compression_string}"
|
|
|
|
|
if var_true "${DEBUG_BACKUP_MONGO}" ; then debug on; fi
|
|
|
|
|
@@ -559,6 +589,7 @@ backup_mongo() {
|
|
|
|
|
move_dbbackup
|
|
|
|
|
check_exit_code move "${backup_job_filename}"
|
|
|
|
|
post_dbbackup "${backup_job_db_name}"
|
|
|
|
|
cleanup_old_data
|
|
|
|
|
if var_true "${DEBUG_BACKUP_MONGO}" ; then debug off; fi
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
@@ -577,14 +608,17 @@ backup_mssql() {
|
|
|
|
|
backup_job_filename_original=${backup_job_filename}
|
|
|
|
|
compression
|
|
|
|
|
pre_dbbackup all
|
|
|
|
|
if var_true "${DEBUG_BACKUP_MSSQL}" ; then debug on; fi
|
|
|
|
|
run_as_user ${compress_cmd} "${TEMP_PATH}/${backup_job_filename_original}"
|
|
|
|
|
check_exit_code backup "${backup_job_filename}"
|
|
|
|
|
if var_true "${DEBUG_BACKUP_MSSQL}" ; then debug off; fi
|
|
|
|
|
timer backup finish
|
|
|
|
|
file_encryption
|
|
|
|
|
generate_checksum
|
|
|
|
|
move_dbbackup
|
|
|
|
|
check_exit_code move "${backup_job_filename}"
|
|
|
|
|
post_dbbackup "${backup_job_db_name}"
|
|
|
|
|
cleanup_old_data
|
|
|
|
|
;;
|
|
|
|
|
trn|transaction )
|
|
|
|
|
prepare_dbbackup
|
|
|
|
|
@@ -607,12 +641,13 @@ backup_mssql() {
|
|
|
|
|
move_dbbackup
|
|
|
|
|
check_exit_code move "${backup_job_filename}"
|
|
|
|
|
post_dbbackup "${backup_job_db_name}"
|
|
|
|
|
cleanup_old_data
|
|
|
|
|
;;
|
|
|
|
|
esac
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
backup_mysql() {
|
|
|
|
|
|
|
|
|
|
if var_true "${DEBUG_BACKUP_MYSQL}" ; then debug on; fi
|
|
|
|
|
if var_true "${backup_job_mysql_events}" ; then
|
|
|
|
|
events="--events"
|
|
|
|
|
fi
|
|
|
|
|
@@ -636,7 +671,7 @@ backup_mysql() {
|
|
|
|
|
else
|
|
|
|
|
db_names=$(echo "${backup_job_db_name}" | tr ',' '\n')
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
if var_true "${DEBUG_BACKUP_MYSQL}" ; then debug off; fi
|
|
|
|
|
write_log debug "Databases Found: $(echo ${db_names} | xargs | tr ' ' ',')"
|
|
|
|
|
|
|
|
|
|
if var_true "${backup_job_split_db}" ; then
|
|
|
|
|
@@ -658,6 +693,7 @@ backup_mysql() {
|
|
|
|
|
move_dbbackup
|
|
|
|
|
check_exit_code move "${backup_job_filename}"
|
|
|
|
|
post_dbbackup "${db}"
|
|
|
|
|
cleanup_old_data
|
|
|
|
|
done
|
|
|
|
|
else
|
|
|
|
|
write_log debug "Not splitting database dumps into their own files"
|
|
|
|
|
@@ -678,6 +714,7 @@ backup_mysql() {
|
|
|
|
|
move_dbbackup
|
|
|
|
|
check_exit_code move "${backup_job_filename}"
|
|
|
|
|
post_dbbackup all
|
|
|
|
|
cleanup_old_data
|
|
|
|
|
fi
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
@@ -699,8 +736,10 @@ backup_pgsql() {
|
|
|
|
|
move_dbbackup
|
|
|
|
|
check_exit_code move "${backup_job_filename}"
|
|
|
|
|
post_dbbackup "globals"
|
|
|
|
|
cleanup_old_data
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
if var_true "${DEBUG_BACKUP_PGSQL}" ; then debug on; fi
|
|
|
|
|
export PGPASSWORD=${backup_job_db_pass}
|
|
|
|
|
if [ -n "${backup_job_db_auth}" ] ; then
|
|
|
|
|
authdb=${backup_job_db_auth}
|
|
|
|
|
@@ -709,7 +748,7 @@ backup_pgsql() {
|
|
|
|
|
fi
|
|
|
|
|
if [ "${backup_job_db_name,,}" = "all" ] ; then
|
|
|
|
|
write_log debug "Preparing to back up all databases"
|
|
|
|
|
db_names=$(run_as_user psql -h ${backup_job_db_host} -U ${backup_job_db_user} -p ${backup_job_db_port} -d ${authdb} -c 'COPY (SELECT datname FROM pg_database WHERE datistemplate = false) TO STDOUT;' )
|
|
|
|
|
db_names=$(psql -h ${backup_job_db_host} -U ${backup_job_db_user} -p ${backup_job_db_port} -d ${authdb} -c 'COPY (SELECT datname FROM pg_database WHERE datistemplate = false) TO STDOUT;' )
|
|
|
|
|
if [ -n "${backup_job_db_name_exclude}" ] ; then
|
|
|
|
|
db_names_exclusions=$(echo "${backup_job_db_name_exclude}" | tr ',' '\n')
|
|
|
|
|
for db_exclude in ${db_names_exclusions} ; do
|
|
|
|
|
@@ -724,7 +763,7 @@ backup_pgsql() {
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
if var_false "${_postgres_backup_globals}" && var_true "${backup_job_backup_pgsql_globals}" ; then _postgres_backup_globals=true; fi
|
|
|
|
|
|
|
|
|
|
if var_true "${DEBUG_BACKUP_PGSQL}" ; then debug off; fi
|
|
|
|
|
write_log debug "Databases Found: $(echo ${db_names} | xargs | tr ' ' ',')"
|
|
|
|
|
|
|
|
|
|
if var_true "${backup_job_split_db}" ; then
|
|
|
|
|
@@ -746,6 +785,7 @@ backup_pgsql() {
|
|
|
|
|
move_dbbackup
|
|
|
|
|
check_exit_code move "${backup_job_filename}"
|
|
|
|
|
post_dbbackup "${db}"
|
|
|
|
|
cleanup_old_data
|
|
|
|
|
done
|
|
|
|
|
if var_true "${_postgres_backup_globals}" ; then backup_pgsql_globals; fi
|
|
|
|
|
else
|
|
|
|
|
@@ -756,7 +796,8 @@ backup_pgsql() {
|
|
|
|
|
compression
|
|
|
|
|
pre_dbbackup all
|
|
|
|
|
write_log notice "Dumping all PostgreSQL databases: '$(echo ${db_names} | xargs | tr ' ' ',')' ${compression_string}"
|
|
|
|
|
tmp_db_names=$(run_as_user psql -h ${backup_job_db_host} -p ${backup_job_db_port} -U ${backup_job_db_user} -d ${authdb} -c 'COPY (SELECT datname FROM pg_database WHERE datistemplate = false) TO STDOUT;' )
|
|
|
|
|
if var_true "${DEBUG_BACKUP_PGSQL}" ; then debug on; fi
|
|
|
|
|
tmp_db_names=$(psql -h ${backup_job_db_host} -p ${backup_job_db_port} -U ${backup_job_db_user} -d ${authdb} -c 'COPY (SELECT datname FROM pg_database WHERE datistemplate = false) TO STDOUT;' )
|
|
|
|
|
for r_db_name in $(echo $db_names | xargs); do
|
|
|
|
|
tmp_db_names=$(echo "$tmp_db_names" | xargs | sed "s|${r_db_name}||g" )
|
|
|
|
|
done
|
|
|
|
|
@@ -764,7 +805,6 @@ backup_pgsql() {
|
|
|
|
|
for x_db_name in ${tmp_db_names} ; do
|
|
|
|
|
pgexclude_arg=$(echo ${pgexclude_arg} --exclude-database=${x_db_name})
|
|
|
|
|
done
|
|
|
|
|
if var_true "${DEBUG_BACKUP_PGSQL}" ; then debug on; fi
|
|
|
|
|
run_as_user ${play_fair} pg_dumpall -h ${backup_job_db_host} -p ${backup_job_db_port} -U ${backup_job_db_user} ${pgexclude_arg} ${backup_job_extra_opts} ${backup_job_extra_dump_opts} | ${compress_cmd} | run_as_user tee "${TEMP_PATH}"/"${backup_job_filename}" > /dev/null
|
|
|
|
|
exit_code=$?
|
|
|
|
|
if var_true "${DEBUG_BACKUP_PGSQL}" ; then debug off; fi
|
|
|
|
|
@@ -775,6 +815,7 @@ backup_pgsql() {
|
|
|
|
|
move_dbbackup
|
|
|
|
|
check_exit_code move "${backup_job_filename}"
|
|
|
|
|
post_dbbackup all
|
|
|
|
|
cleanup_old_data
|
|
|
|
|
if var_true "${_postgres_backup_globals}" ; then backup_pgsql_globals; fi
|
|
|
|
|
fi
|
|
|
|
|
}
|
|
|
|
|
@@ -801,17 +842,20 @@ backup_redis() {
|
|
|
|
|
sleep 5
|
|
|
|
|
done
|
|
|
|
|
backup_job_filename_original=${backup_job_filename}
|
|
|
|
|
if var_true "${DEBUG_BACKUP_REDIS}" ; then debug off; fi
|
|
|
|
|
compression
|
|
|
|
|
pre_dbbackup all
|
|
|
|
|
run_as_user ${compress_cmd} "${TEMP_PATH}/${backup_job_filename_original}"
|
|
|
|
|
timer backup finish
|
|
|
|
|
if var_true "${DEBUG_BACKUP_REDIS}" ; then debug on; fi
|
|
|
|
|
run_as_user ${compress_cmd} "${TEMP_PATH}/${backup_job_filename_original}"
|
|
|
|
|
if var_true "${DEBUG_BACKUP_REDIS}" ; then debug off; fi
|
|
|
|
|
timer backup finish
|
|
|
|
|
check_exit_code backup "${backup_job_filename}"
|
|
|
|
|
file_encryption
|
|
|
|
|
generate_checksum
|
|
|
|
|
move_dbbackup
|
|
|
|
|
check_exit_code move "${backup_job_filename}"
|
|
|
|
|
post_dbbackup all
|
|
|
|
|
cleanup_old_data
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
backup_sqlite3() {
|
|
|
|
|
@@ -819,22 +863,28 @@ backup_sqlite3() {
|
|
|
|
|
db=$(basename "${backup_job_db_host}")
|
|
|
|
|
db="${db%.*}"
|
|
|
|
|
backup_job_filename=sqlite3_${db}_${now}.sqlite3
|
|
|
|
|
backup_job_filename_base=sqlite3_${db}.sqlite3
|
|
|
|
|
compression
|
|
|
|
|
backup_job_filename_base=sqlite3_${db}
|
|
|
|
|
pre_dbbackup "${db}"
|
|
|
|
|
write_log notice "Dumping sqlite3 database: '${backup_job_db_host}' ${compression_string}"
|
|
|
|
|
if var_true "${DEBUG_BACKUP_SQLITE3}" ; then debug on; fi
|
|
|
|
|
silent run_as_user ${play_fair} sqlite3 "${backup_job_db_host}" ".backup '${TEMP_PATH}/backup.sqlite3'"
|
|
|
|
|
silent ${play_fair} sqlite3 "${backup_job_db_host}" ".backup '${TEMP_PATH}/backup_${now}.sqlite3'"
|
|
|
|
|
exit_code=$?
|
|
|
|
|
check_exit_code backup "${backup_job_filename}"
|
|
|
|
|
run_as_user ${play_fair} cat "${TEMP_PATH}"/backup.sqlite3 | ${dir_compress_cmd} | run_as_user tee "${TEMP_PATH}/${backup_job_filename}" > /dev/null
|
|
|
|
|
timer backup finish
|
|
|
|
|
if [ ! -f "${TEMP_PATH}"/backup_${now}.sqlite3 ] ; then
|
|
|
|
|
print_error "SQLite3 backup failed! Exitting"
|
|
|
|
|
return 1
|
|
|
|
|
fi
|
|
|
|
|
compression
|
|
|
|
|
run_as_user ${play_fair} cat "${TEMP_PATH}"/backup_${now}.sqlite3 | ${dir_compress_cmd} | run_as_user tee "${TEMP_PATH}/${backup_job_filename}" > /dev/null
|
|
|
|
|
rm -rf "${TEMP_PATH}"/backup_${now}.sqlite3
|
|
|
|
|
if var_true "${DEBUG_BACKUP_SQLITE3}" ; then debug off; fi
|
|
|
|
|
timer backup finish
|
|
|
|
|
file_encryption
|
|
|
|
|
generate_checksum
|
|
|
|
|
move_dbbackup
|
|
|
|
|
check_exit_code move "${backup_job_filename}"
|
|
|
|
|
post_dbbackup "${db}"
|
|
|
|
|
cleanup_old_data
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
check_availability() {
|
|
|
|
|
@@ -1036,6 +1086,7 @@ compression() {
|
|
|
|
|
|
|
|
|
|
case "${backup_job_compression,,}" in
|
|
|
|
|
bz* )
|
|
|
|
|
print_debug "[compression] Selected BZIP"
|
|
|
|
|
compress_cmd="${play_fair} pbzip2 -q -${backup_job_compression_level} -p${backup_job_parallel_compression_threads} "
|
|
|
|
|
compression_type="bzip2"
|
|
|
|
|
dir_compress_cmd=${compress_cmd}
|
|
|
|
|
@@ -1044,6 +1095,7 @@ compression() {
|
|
|
|
|
backup_job_filename=${backup_job_filename}.bz2
|
|
|
|
|
;;
|
|
|
|
|
gz* )
|
|
|
|
|
print_debug "[compression] Selected GZIP"
|
|
|
|
|
compress_cmd="${play_fair} pigz -q -${backup_job_compression_level} -p ${backup_job_parallel_compression_threads} ${gz_rsyncable}"
|
|
|
|
|
compression_type="gzip"
|
|
|
|
|
extension=".gz"
|
|
|
|
|
@@ -1052,6 +1104,7 @@ compression() {
|
|
|
|
|
backup_job_filename=${backup_job_filename}.gz
|
|
|
|
|
;;
|
|
|
|
|
xz* )
|
|
|
|
|
print_debug "[compression] Selected XZIP"
|
|
|
|
|
compress_cmd="${play_fair} pixz -${backup_job_compression_level} -p ${backup_job_parallel_compression_threads} "
|
|
|
|
|
compression_type="xzip"
|
|
|
|
|
dir_compress_cmd=${compress_cmd}
|
|
|
|
|
@@ -1060,6 +1113,7 @@ compression() {
|
|
|
|
|
backup_job_filename=${backup_job_filename}.xz
|
|
|
|
|
;;
|
|
|
|
|
zst* )
|
|
|
|
|
print_debug "[compression] Selected ZSTD"
|
|
|
|
|
compress_cmd="${play_fair} zstd -q -q --rm -${backup_job_compression_level} -T${backup_job_parallel_compression_threads} ${gz_rsyncable}"
|
|
|
|
|
compression_type="zstd"
|
|
|
|
|
dir_compress_cmd=${compress_cmd}
|
|
|
|
|
@@ -1105,25 +1159,28 @@ create_archive() {
|
|
|
|
|
|
|
|
|
|
create_schedulers() {
|
|
|
|
|
if var_true "${DEBUG_CREATE_SCHEDULERS}" ; then debug on; fi
|
|
|
|
|
|
|
|
|
|
backup() {
|
|
|
|
|
bootstrap_variables upgrade BACKUP
|
|
|
|
|
local backup_instances=$(printenv | sort | grep -c "^DB[0-9]._HOST")
|
|
|
|
|
local backup_instances=$(set -o posix ; set | grep -Pc "^(DB[0-9]._HOST=|.*MONGO_CUSTOM_URI=)")
|
|
|
|
|
print_debug "[create_schedulers] Found '${backup_instances}' DB_HOST instances"
|
|
|
|
|
|
|
|
|
|
if [ -n "${DB_HOST}" ] && [ "${backup_instances}" ]; then
|
|
|
|
|
backup_instances=1;
|
|
|
|
|
print_debug "Detected using old DB_ variables"
|
|
|
|
|
print_debug "[create_schedulers] Detected using old DB_ variables"
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
for (( instance = 01; instance <= backup_instances; )) ; do
|
|
|
|
|
instance=$(printf "%02d" $instance)
|
|
|
|
|
cp -R /assets/dbbackup/template-dbbackup /etc/services.available/dbbackup-"${instance}"
|
|
|
|
|
sed -i "s|{{BACKUP_NUMBER}}|${instance}|g" /etc/services.available/dbbackup-"${instance}"/run
|
|
|
|
|
|
|
|
|
|
if [ "${MODE,,}" = "manual" ] ; then service_stop dbbackup-"${instance}" ; fi
|
|
|
|
|
cat <<EOF >> /usr/bin/backup"${instance}"-now
|
|
|
|
|
#!/bin/bash
|
|
|
|
|
source /assets/functions/00-container
|
|
|
|
|
PROCESS_NAME=db-backup${instance}
|
|
|
|
|
print_info "Starting Manual Backup for db-backup${instance}"
|
|
|
|
|
/var/run/s6/legacy-services/dbbackup-${instance}/run now
|
|
|
|
|
#/var/run/s6/legacy-services/dbbackup-${instance}/run now
|
|
|
|
|
/etc/services.available/dbbackup-${instance}/run now
|
|
|
|
|
|
|
|
|
|
EOF
|
|
|
|
|
chmod +x /usr/bin/backup"${instance}"-now
|
|
|
|
|
@@ -1133,11 +1190,11 @@ EOF
|
|
|
|
|
cat <<EOF > /usr/bin/backup-now
|
|
|
|
|
#!/bin/bash
|
|
|
|
|
|
|
|
|
|
/usr/bin/backup${instance}-now
|
|
|
|
|
/usr/bin/backup${instance}-now now
|
|
|
|
|
|
|
|
|
|
EOF
|
|
|
|
|
else
|
|
|
|
|
echo "/usr/bin/backup${instance}-now" >> /usr/bin/backup-now
|
|
|
|
|
echo "/usr/bin/backup${instance}-now now" >> /usr/bin/backup-now
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
instance=$(echo "${instance} +1" | bc)
|
|
|
|
|
@@ -1151,7 +1208,7 @@ EOF
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
ctrl_c() {
|
|
|
|
|
sed -i "/^{{BACKUP_NUMBER}}/d" /tmp/.container/db-backup-backups
|
|
|
|
|
sed -i "/^${backup_instance_number}/d" /tmp/.container/db-backup-backups
|
|
|
|
|
symlink_log
|
|
|
|
|
print_warn "User aborted"
|
|
|
|
|
exit
|
|
|
|
|
@@ -1166,7 +1223,11 @@ db_backup_container_init() {
|
|
|
|
|
debug() {
|
|
|
|
|
case "${1}" in
|
|
|
|
|
off)
|
|
|
|
|
backup_job_log_level=$_original_job_log_level}
|
|
|
|
|
CONTAINER_LOG_LEVEL=${_original_container_log_level}
|
|
|
|
|
DEBUG_MODE=${_original_debug_mode}
|
|
|
|
|
SHOW_OUTPUT=${_original_show_output}
|
|
|
|
|
|
|
|
|
|
if var_true "${DEBUG_MODE}" ; then
|
|
|
|
|
set -x
|
|
|
|
|
else
|
|
|
|
|
@@ -1174,9 +1235,25 @@ debug() {
|
|
|
|
|
fi
|
|
|
|
|
;;
|
|
|
|
|
on)
|
|
|
|
|
if [ -z "${_original_container_log_level}" ]; then
|
|
|
|
|
_original_container_log_level="${CONTAINER_LOG_LEVEL}"
|
|
|
|
|
fi
|
|
|
|
|
if [ -z "${_original_job_log_level}" ]; then
|
|
|
|
|
_original_job_log_level="${backup_job_log_level}"
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
if [ -z "${_original_debug_mode}" ]; then
|
|
|
|
|
_original_debug_mode="${DEBUG_MODE}"
|
|
|
|
|
fi
|
|
|
|
|
if [ -z "${_original_show_output}" ]; then
|
|
|
|
|
_original_show_output="${SHOW_OUTPUT}"
|
|
|
|
|
if ! [[ "${_original_show_output,,}" =~ true|false ]]; then
|
|
|
|
|
__original_show_output="FALSE"
|
|
|
|
|
fi
|
|
|
|
|
fi
|
|
|
|
|
backup_job_log_level=DEBUG
|
|
|
|
|
CONTAINER_LOG_LEVEL=DEBUG
|
|
|
|
|
SHOW_OUTPUT=TRUE
|
|
|
|
|
set -x
|
|
|
|
|
;;
|
|
|
|
|
esac
|
|
|
|
|
@@ -1186,27 +1263,33 @@ file_encryption() {
|
|
|
|
|
if var_true "${DEBUG_FILE_ENCRYPTION}" ; then debug on; fi
|
|
|
|
|
if var_true "${backup_job_encrypt}" ; then
|
|
|
|
|
if [ "${exit_code}" = "0" ] ; then
|
|
|
|
|
print_debug "Encrypting"
|
|
|
|
|
print_debug "[file_encryption] Encrypting"
|
|
|
|
|
output_off
|
|
|
|
|
if [ -n "${backup_job_encrypt_passphrase}" ] && [ -n "${backup_job_encrypt_pubkey}" ]; then
|
|
|
|
|
if [ -n "${backup_job_encrypt_passphrase}" ] && [ -n "${backup_job_encrypt_public_key}" ]; then
|
|
|
|
|
print_error "Can't encrypt as both ENCRYPT_PASSPHRASE and ENCRYPT_PUBKEY exist!"
|
|
|
|
|
return
|
|
|
|
|
elif [ -n "${backup_job_encrypt_passphrase}" ] && [ -z "${backup_job_encrypt_pubkey}" ]; then
|
|
|
|
|
elif [ -n "${backup_job_encrypt_passphrase}" ] && [ -z "${backup_job_encrypt_public_key}" ]; then
|
|
|
|
|
print_notice "Encrypting with GPG Passphrase"
|
|
|
|
|
encrypt_routines_start_time=$(date +'%s')
|
|
|
|
|
encrypt_tmp_dir=$(run_as_user mktemp -d)
|
|
|
|
|
echo "${backup_job_encrypt_passphrase}" | silent run_as_user ${play_fair} gpg --batch --home ${encrypt_tmp_dir} --yes --passphrase-fd 0 -c "${TEMP_PATH}"/"${backup_job_filename}"
|
|
|
|
|
rm -rf "${encrypt_tmp_dir}"
|
|
|
|
|
elif [ -z "${backup_job_encrypt_passphrase}" ] && [ -n "${backup_job_encrypt_pubkey}" ]; then
|
|
|
|
|
if [ -f "${backup_job_encrypt_pubkey}" ]; then
|
|
|
|
|
elif [ -z "${backup_job_encrypt_passphrase}" ] && [ -n "${backup_job_encrypt_public_key}" ] && [ -n "${backup_job_encrypt_private_key}" ]; then
|
|
|
|
|
if [ -f "${backup_job_encrypt_private_key}" ]; then
|
|
|
|
|
encrypt_routines_start_time=$(date +'%s')
|
|
|
|
|
print_notice "Encrypting with GPG Public Key"
|
|
|
|
|
print_notice "Encrypting with GPG Private Key"
|
|
|
|
|
encrypt_tmp_dir=$(run_as_user mktemp -d)
|
|
|
|
|
silent run_as_user ${play_fair} gpg --batch --yes --home "${encrypt_tmp_dir}" --recipient-file "${backup_job_encrypt_pubkey}" -c "${TEMP_PATH}"/"${backup_job_filename}"
|
|
|
|
|
cat "${backup_job_encrypt_private_key}" | run_as_user tee "${encrypt_tmp_dir}"/private_key.asc > /dev/null
|
|
|
|
|
print_debug "[file_encryption] [key] Importing Private Key"
|
|
|
|
|
silent run_as_user gpg --home ${encrypt_tmp_dir} --batch --import "${encrypt_tmp_dir}"/private_key.asc
|
|
|
|
|
print_debug "[file_encryption] [key] Encrypting to Public Key"
|
|
|
|
|
cat "${backup_job_encrypt_public_key}" | run_as_user tee "${encrypt_tmp_dir}"/public_key.asc > /dev/null
|
|
|
|
|
silent run_as_user ${play_fair} gpg --batch --yes --home "${encrypt_tmp_dir}" --encrypt --recipient-file "${encrypt_tmp_dir}"/public_key.asc "${TEMP_PATH}"/"${backup_job_filename}"
|
|
|
|
|
rm -rf "${encrypt_tmp_dir}"
|
|
|
|
|
fi
|
|
|
|
|
fi
|
|
|
|
|
if [ -f "${TEMP_PATH}"/"${backup_job_filename}".gpg ]; then
|
|
|
|
|
print_debug "[file_encryption] Deleting original file"
|
|
|
|
|
rm -rf "${TEMP_PATH:?}"/"${backup_job_filename:?}"
|
|
|
|
|
backup_job_filename="${backup_job_filename}.gpg"
|
|
|
|
|
|
|
|
|
|
@@ -1216,6 +1299,9 @@ file_encryption() {
|
|
|
|
|
- dbbackup.backup.encrypt.duration.[${backup_job_db_host}.${backup_job_db_name}] ${encrypt_routines_total_time}
|
|
|
|
|
EOF
|
|
|
|
|
)
|
|
|
|
|
else
|
|
|
|
|
print_error "Encryption failed! Could not detect encrypted file"
|
|
|
|
|
return 99
|
|
|
|
|
fi
|
|
|
|
|
else
|
|
|
|
|
write_log error "Skipping encryption because backup did not complete successfully"
|
|
|
|
|
@@ -1265,20 +1351,20 @@ EOF
|
|
|
|
|
notify() {
|
|
|
|
|
if var_true "${DEBUG_NOTIFY}" ; then debug on; fi
|
|
|
|
|
notification_custom() {
|
|
|
|
|
if [ -n "${NOTIFICATION_SCRIPT}" ] ; then
|
|
|
|
|
if var_true "${NOTIFICATION_SCRIPT_SKIP_X_VERIFY}" ; then
|
|
|
|
|
eval "${NOTIFICATION_SCRIPT}" "${1}" "${2}" "${3}" "${4}" "${5}"
|
|
|
|
|
if [ -n "${NOTIFICATION_CUSTOM_SCRIPT}" ] ; then
|
|
|
|
|
if var_true "${NOTIFICATION_CUSTOM_SCRIPT_SKIP_X_VERIFY}" ; then
|
|
|
|
|
eval "${NOTIFICATION_CUSTOM_SCRIPT}" "${1}" "${2}" "${3}" "${4}" "${5}"
|
|
|
|
|
else
|
|
|
|
|
if [ -x "${NOTIFICATION_SCRIPT}" ] ; then
|
|
|
|
|
write_log notice "Found NOTIFICATION_SCRIPT environment variable. Executing '${NOTIFICATION_SCRIPT}"
|
|
|
|
|
if [ -x "${NOTIFICATION_CUSTOM_SCRIPT}" ] ; then
|
|
|
|
|
write_log notice "Found NOTIFICATION_CUSTOM_SCRIPT environment variable. Executing '${NOTIFICATION_CUSTOM_SCRIPT}"
|
|
|
|
|
# script timestamp logfile errorcode subject body
|
|
|
|
|
eval "${NOTIFICATION_SCRIPT}" "${1}" "${2}" "${3}" "${4}" "${5}"
|
|
|
|
|
eval "${NOTIFICATION_CUSTOM_SCRIPT}" "${1}" "${2}" "${3}" "${4}" "${5}"
|
|
|
|
|
else
|
|
|
|
|
write_log error "Can't execute NOTIFICATION_SCRIPT environment variable '${NOTIFICATION_SCRIPT}' as its filesystem bit is not executible!"
|
|
|
|
|
write_log error "Can't execute NOTIFICATION_CUSTOM_SCRIPT environment variable '${NOTIFICATION_CUSTOM_SCRIPT}' as its filesystem bit is not executible!"
|
|
|
|
|
fi
|
|
|
|
|
fi
|
|
|
|
|
else
|
|
|
|
|
print_error "[notifications] No NOTIFICATION_SCRIPT variable set - Skipping sending Custom notifications"
|
|
|
|
|
print_error "[notifications] No NOTIFICATION_CUSTOM_SCRIPT variable set - Skipping sending Custom notifications"
|
|
|
|
|
fi
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
@@ -1291,18 +1377,20 @@ notify() {
|
|
|
|
|
if [ -z "${SMTP_HOST}" ] ; then write_log error "[notifications] No SMTP_HOST variable set - Skipping sending Email notifications" ; skip_mail=true ; fi
|
|
|
|
|
if [ -z "${SMTP_PORT}" ] ; then write_log error "[notifications] No SMTP_PORT variable set - Skipping sending Email notifications" ; skip_mail=true ; fi
|
|
|
|
|
if var_nottrue "${skip_mail}" ; then
|
|
|
|
|
if ! grep -q ^from /etc/msmptrc ; then
|
|
|
|
|
echo "from ${MAIL_FROM}" >> /etc/msmtprc
|
|
|
|
|
fi
|
|
|
|
|
mail_recipients=$(echo "${MAIL_TO}" | tr "," "\n")
|
|
|
|
|
for mail_recipient in $mail_recipients ; do
|
|
|
|
|
cat <<EOF | s-nail -v \
|
|
|
|
|
-r "${MAIL_FROM}" \
|
|
|
|
|
-s "[db-backup] [${DOMAIN}] ${3}" \
|
|
|
|
|
-S smtp="${SMTP_HOST}":"${SMTP_PORT}" \
|
|
|
|
|
"${mail_recipient}"
|
|
|
|
|
cat <<EOF | msmtp -t "${mail_recipient}" -C /etc/msmtprc
|
|
|
|
|
To: ${mail_recipient}
|
|
|
|
|
Subject: [db-backup] ${4}
|
|
|
|
|
|
|
|
|
|
Time: ${1}
|
|
|
|
|
Log File: {2}
|
|
|
|
|
Error Code: ${3}
|
|
|
|
|
|
|
|
|
|
${4}
|
|
|
|
|
${5}
|
|
|
|
|
EOF
|
|
|
|
|
done
|
|
|
|
|
fi
|
|
|
|
|
@@ -1318,7 +1406,7 @@ EOF
|
|
|
|
|
if [ -z "${MATTERMOST_WEBHOOK_URL}" ] ; then write_log error "[notifications] No MATTERMOST_WEBHOOK_URL variable set - Skipping sending Mattermost notifications" ; skip_mattermost=true ; fi
|
|
|
|
|
if var_nottrue "${skip_mattermost}" ; then
|
|
|
|
|
emoji=":bomb:"
|
|
|
|
|
message="*[db-backup] ${3}*\n${4}\n*Timestamp:* ${1}\n*Logfile:* ${2}\n*Error Code: ${3}"
|
|
|
|
|
message="*[db-backup] ${4}*\n${5}\n*Timestamp:* ${1}\n*Logfile:* ${2}\n*Error Code: ${3}"
|
|
|
|
|
mattermost_recipients=$(echo "${MATTERMOST_RECIPIENT}" | tr "," "\n")
|
|
|
|
|
for mattermost_recipient in $mattermost_recipients ; do
|
|
|
|
|
payload="payload={\"channel\": \"${mattermost_recipient//\"/\\\"}\", \"username\": \"${MATTERMOST_USERNAME//\"/\\\"}\", \"text\": \"${message//\"/\\\"}\", \"icon_emoji\": \"${emoji}\"}"
|
|
|
|
|
@@ -1343,7 +1431,7 @@ EOF
|
|
|
|
|
for matrix_room in $matrix_rooms ; do
|
|
|
|
|
curl \
|
|
|
|
|
-XPOST \
|
|
|
|
|
-d "{\"msgtype\":\"m.text\", \"body\":\"*[db-backup] ${3}*\n${4}\n*Timestamp:* ${1}\n*Logfile:* ${2}\n*Error Code: ${3}*\"}" \
|
|
|
|
|
-d "{\"msgtype\":\"m.text\", \"body\":\"*[db-backup] ${4}*\n${5}\n*Timestamp:* ${1}\n*Logfile:* ${2}\n*Error Code: ${3}*\"}" \
|
|
|
|
|
"${MATRIX_HOST}/_matrix/client/r0/rooms/${matrix_room}/send/m.room.message?access_token=${MATRIX_ACCESS_TOKEN}"
|
|
|
|
|
done
|
|
|
|
|
fi
|
|
|
|
|
@@ -1359,7 +1447,7 @@ EOF
|
|
|
|
|
if [ -z "${ROCKETCHAT_WEBHOOK_URL}" ] ; then write_log error "[notifications] No ROCKETCHAT_WEBHOOK_URL variable set - Skipping sending Rocket.Chat notifications" ; skip_rocketchat=true ; fi
|
|
|
|
|
if var_nottrue "${skip_rocketchat}" ; then
|
|
|
|
|
emoji=":bomb:"
|
|
|
|
|
message="*[db-backup] ${3}*\n${4}\n*Timestamp:* ${1}\n*Logfile:* ${2}\n*Error Code: ${3}"
|
|
|
|
|
message="*[db-backup] ${4}*\n${5}\n*Timestamp:* ${1}\n*Logfile:* ${2}\n*Error Code: ${3}"
|
|
|
|
|
rocketchat_recipients=$(echo "${ROCKETCHAT_RECIPIENT}" | tr "," "\n")
|
|
|
|
|
for rocketchat_recipient in $rocketchat_recipients ; do
|
|
|
|
|
payload="payload={\"channel\": \"${rocketchat_recipient//\"/\\\"}\", \"username\": \"${ROCKETCHAT_USERNAME//\"/\\\"}\", \"text\": \"${message//\"/\\\"}\", \"icon_emoji\": \"${emoji}\"}"
|
|
|
|
|
@@ -1378,27 +1466,27 @@ EOF
|
|
|
|
|
# $4 body
|
|
|
|
|
|
|
|
|
|
if var_true "${ENABLE_NOTIFICATIONS}" ; then
|
|
|
|
|
notification_types=$(echo "${NOTIIFICATION_TYPE}" | tr "," "\n")
|
|
|
|
|
notification_types=$(echo "${NOTIFICATION_TYPE}" | tr "," "\n")
|
|
|
|
|
for notification_type in $notification_types ; do
|
|
|
|
|
case "${notification_type,,}" in
|
|
|
|
|
"custom" )
|
|
|
|
|
print_debug "Sending Notification via custom"
|
|
|
|
|
print_debug "[notify] Sending Notification via custom"
|
|
|
|
|
notification_custom "${1}" "${2}" "${3}" "${4}" "${5}"
|
|
|
|
|
;;
|
|
|
|
|
"email" | "mail" )
|
|
|
|
|
print_debug "Sending Notification via email"
|
|
|
|
|
print_debug "[notify] Sending Notification via email"
|
|
|
|
|
notification_email "${1}" "${2}" "${3}" "${4}" "${5}"
|
|
|
|
|
;;
|
|
|
|
|
"matrix" )
|
|
|
|
|
print_debug "Sending Notification via Matrix"
|
|
|
|
|
print_debug "[notify] Sending Notification via Matrix"
|
|
|
|
|
notification_matrix "${1}" "${2}" "${3}" "${4}" "${5}"
|
|
|
|
|
;;
|
|
|
|
|
"mattermost" )
|
|
|
|
|
print_debug "Sending Notification via Mattermost"
|
|
|
|
|
print_debug "[notify] Sending Notification via Mattermost"
|
|
|
|
|
notification_mattermost "${1}" "${2}" "${3}" "${4}" "${5}"
|
|
|
|
|
;;
|
|
|
|
|
"rocketchat" )
|
|
|
|
|
print_debug "Sending Notification via Rocketchat"
|
|
|
|
|
print_debug "[notify] Sending Notification via Rocketchat"
|
|
|
|
|
notification_rocketchat "${1}" "${2}" "${3}" "${4}" "${5}"
|
|
|
|
|
;;
|
|
|
|
|
* )
|
|
|
|
|
@@ -1441,8 +1529,37 @@ move_dbbackup() {
|
|
|
|
|
write_log debug "Moving backup to filesystem"
|
|
|
|
|
run_as_user mkdir -p "${backup_job_filesystem_path}"
|
|
|
|
|
if [ "${backup_job_checksum,,}" != "none" ] ; then run_as_user mv "${TEMP_PATH}"/*."${checksum_extension}" "${backup_job_filesystem_path}"/ ; fi
|
|
|
|
|
if var_true "${DEBUG_MOVE_DBBACKUP}"; then
|
|
|
|
|
cat <<EOF
|
|
|
|
|
## BEGIN Before Moving file from TEMP_PATH $(TZ=${TIMEZONE} date)
|
|
|
|
|
##
|
|
|
|
|
|
|
|
|
|
$(ls -l "${TEMP_PATH}"/*)
|
|
|
|
|
|
|
|
|
|
## END
|
|
|
|
|
EOF
|
|
|
|
|
fi
|
|
|
|
|
run_as_user mv "${TEMP_PATH}"/"${backup_job_filename}" "${backup_job_filesystem_path}"/"${backup_job_filename}"
|
|
|
|
|
move_exit_code=$?
|
|
|
|
|
if var_true "${DEBUG_MOVE_DBBACKUP}"; then
|
|
|
|
|
cat <<EOF
|
|
|
|
|
## BEGIN After Moving file from TEMP_PATH $(TZ=${TIMEZONE} date)
|
|
|
|
|
##
|
|
|
|
|
|
|
|
|
|
$(ls -l "${TEMP_PATH}"/*)
|
|
|
|
|
|
|
|
|
|
## END
|
|
|
|
|
|
|
|
|
|
## BEGIN After Moving file to _FILESYSTEM_PATH $(TZ=${TIMEZONE} date)
|
|
|
|
|
##
|
|
|
|
|
|
|
|
|
|
$(ls -l "${backup_job_filesystem_path}"/*)
|
|
|
|
|
|
|
|
|
|
## END
|
|
|
|
|
|
|
|
|
|
EOF
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
if var_true "${backup_job_create_latest_symlink}" ; then
|
|
|
|
|
run_as_user ln -sfr "${backup_job_filesystem_path}"/"${backup_job_filename}" "${backup_job_filesystem_path}"/latest-"${backup_job_filename_base}"
|
|
|
|
|
fi
|
|
|
|
|
@@ -1530,20 +1647,20 @@ pre_dbbackup() {
|
|
|
|
|
|
|
|
|
|
### Pre Backup Custom Script Support
|
|
|
|
|
if [ -d "/assets/custom-scripts/pre" ] && dir_notempty "/assets/custom-scripts/pre" ; then
|
|
|
|
|
write_log warning "Found Custom Post Scripts in /assets/custom-scripts/pre - Automatically moving them to '${backup_job_script_location_pre}'"
|
|
|
|
|
run_as_user mkdir -p "${backup_job_script_location_pre}"
|
|
|
|
|
silent run_as_user cp /assets/custom-scripts/pre/* "${backup_job_script_location_pre}"
|
|
|
|
|
write_log warn "Found Custom Post Scripts in /assets/custom-scripts/pre - Automatically moving them to '${backup_job_script_location_pre}'"
|
|
|
|
|
mkdir -p "${backup_job_script_location_pre}"
|
|
|
|
|
silent cp -aR /assets/custom-scripts/pre/* "${backup_job_script_location_pre}"
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
if [ -d "${backup_job_script_location_pre}" ] && dir_notempty "${backup_job_script_location_pre}" ; then
|
|
|
|
|
for f in $(find ${backup_job_script_location_pre} -name \*.sh -type f); do
|
|
|
|
|
if var_true "${backup_job_pre_script_x_verify}" ; then
|
|
|
|
|
run_as_user ${f} "${dbtype}" "${backup_job_db_host}" "${1}" "${dbbackup_start_time}" "${backup_job_filename}"
|
|
|
|
|
${f} "${dbtype}" "${backup_job_db_host}" "${1}" "${dbbackup_start_time}" "${backup_job_filename}"
|
|
|
|
|
else
|
|
|
|
|
if [ -x "${f}" ] ; then
|
|
|
|
|
write_log notice "Executing pre backup custom script : '${f}'"
|
|
|
|
|
## script DB_TYPE DB_HOST DB_NAME STARTEPOCH BACKUP_FILENAME
|
|
|
|
|
run_as_user ${f} "${dbtype}" "${backup_job_db_host}" "${1}" "${dbbackup_start_time}" "${backup_job_filename}"
|
|
|
|
|
${f} "${dbtype}" "${backup_job_db_host}" "${1}" "${dbbackup_start_time}" "${backup_job_filename}"
|
|
|
|
|
else
|
|
|
|
|
write_log error "Can't run pre backup custom script: '${f}' as its filesystem bit is not executible!"
|
|
|
|
|
fi
|
|
|
|
|
@@ -1586,11 +1703,11 @@ EOZP
|
|
|
|
|
### Post Script Support
|
|
|
|
|
if [ -n "${backup_job_post_script}" ] ; then
|
|
|
|
|
if var_true "${backup_job_post_script_x_verify}" ; then
|
|
|
|
|
run_as_user eval "${backup_job_post_script}" "${exit_code}" "${dbtype}" "${backup_job_db_host}" "${1}" "${dbbackup_start_time}" "${dbbackup_finish_time}" "${dbbackup_total_time}" "${backup_job_filename}" "${filesize}" "${checksum_value}" "${move_exit_code}"
|
|
|
|
|
eval "${backup_job_post_script}" "${exit_code}" "${dbtype}" "${backup_job_db_host}" "${1}" "${dbbackup_start_time}" "${dbbackup_finish_time}" "${dbbackup_total_time}" "${backup_job_filename}" "${filesize}" "${checksum_value}" "${move_exit_code}"
|
|
|
|
|
else
|
|
|
|
|
if [ -x "${backup_job_post_script}" ] ; then
|
|
|
|
|
write_log notice "Found POST_SCRIPT environment variable. Executing '${backup_job_post_script}"
|
|
|
|
|
run_as_user eval "${backup_job_post_script}" "${exit_code}" "${dbtype}" "${backup_job_db_host}" "${1}" "${dbbackup_start_time}" "${dbbackup_finish_time}" "${dbbackup_total_time}" "${backup_job_filename}" "${filesize}" "${checksum_value}" "${move_exit_code}"
|
|
|
|
|
eval "${backup_job_post_script}" "${exit_code}" "${dbtype}" "${backup_job_db_host}" "${1}" "${dbbackup_start_time}" "${dbbackup_finish_time}" "${dbbackup_total_time}" "${backup_job_filename}" "${filesize}" "${checksum_value}" "${move_exit_code}"
|
|
|
|
|
else
|
|
|
|
|
write_log error "Can't execute POST_SCRIPT environment variable '${backup_job_post_script}' as its filesystem bit is not executible!"
|
|
|
|
|
fi
|
|
|
|
|
@@ -1599,20 +1716,20 @@ EOZP
|
|
|
|
|
|
|
|
|
|
### Post Backup Custom Script Support
|
|
|
|
|
if [ -d "/assets/custom-scripts/" ] && dir_notempty "/assets/custom-scripts" ; then
|
|
|
|
|
write_log warning "Found Custom Post Scripts in /assets/custom-scripts/ - Automatically moving them to '${backup_job_script_location_post}'"
|
|
|
|
|
run_as_user mkdir -p "${backup_job_script_location_post}"
|
|
|
|
|
silent run_as_user cp /assets/custom-scripts/* "${backup_job_script_location_post}"
|
|
|
|
|
write_log warn "Found Custom Post Scripts in /assets/custom-scripts/ - Automatically moving them to '${backup_job_script_location_post}'"
|
|
|
|
|
mkdir -p "${backup_job_script_location_post}"
|
|
|
|
|
cp -aR /assets/custom-scripts/* "${backup_job_script_location_post}"
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
if [ -d "${backup_job_script_location_post}" ] && dir_notempty "${backup_job_script_location_post}" ; then
|
|
|
|
|
for f in $(run_as_user find "${backup_job_script_location_post}" -name \*.sh -type f); do
|
|
|
|
|
if var_true "${backup_job_post_script_x_verify}" ; then
|
|
|
|
|
run_as_user ${f} "${exit_code}" "${dbtype}" "${backup_job_db_host}" "${1}" "${dbbackup_start_time}" "${dbbackup_finish_time}" "${dbbackup_total_time}" "${backup_job_filename}" "${filesize}" "${checksum_value}" "${move_exit_code}"
|
|
|
|
|
${f} "${exit_code}" "${dbtype}" "${backup_job_db_host}" "${1}" "${dbbackup_start_time}" "${dbbackup_finish_time}" "${dbbackup_total_time}" "${backup_job_filename}" "${filesize}" "${checksum_value}" "${move_exit_code}"
|
|
|
|
|
else
|
|
|
|
|
if [ -x "${f}" ] ; then
|
|
|
|
|
write_log notice "Executing post backup custom script : '${f}'"
|
|
|
|
|
## script EXIT_CODE DB_TYPE DB_HOST DB_NAME STARTEPOCH FINISHEPOCH DURATIONEPOCH BACKUP_FILENAME FILESIZE CHECKSUMVALUE
|
|
|
|
|
run_as_user ${f} "${exit_code}" "${dbtype}" "${backup_job_db_host}" "${1}" "${dbbackup_start_time}" "${dbbackup_finish_time}" "${dbbackup_total_time}" "${backup_job_filename}" "${filesize}" "${checksum_value}" "${move_exit_code}"
|
|
|
|
|
${f} "${exit_code}" "${dbtype}" "${backup_job_db_host}" "${1}" "${dbbackup_start_time}" "${dbbackup_finish_time}" "${dbbackup_total_time}" "${backup_job_filename}" "${filesize}" "${checksum_value}" "${move_exit_code}"
|
|
|
|
|
else
|
|
|
|
|
write_log error "Can't run post backup custom script: '${f}' as its filesystem bit is not executible!"
|
|
|
|
|
fi
|
|
|
|
|
@@ -1648,7 +1765,7 @@ process_limiter() {
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
run_as_user() {
|
|
|
|
|
s6-setuidgid dbbackup $@
|
|
|
|
|
sudo -Eu "${DBBACKUP_USER}" "$@"
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
setup_mode() {
|
|
|
|
|
@@ -1740,7 +1857,7 @@ timer() {
|
|
|
|
|
if [ "${expression_step}" != "${expression}" ]; then
|
|
|
|
|
for step in ${validate_temp}; do
|
|
|
|
|
if [ $(( ( step - expression_start ) % expression_step )) -eq 0 ]; then
|
|
|
|
|
validate_all="$validate_all ${step}"
|
|
|
|
|
validate_all="${validate_all} ${step}"
|
|
|
|
|
fi
|
|
|
|
|
done
|
|
|
|
|
else
|
|
|
|
|
@@ -1748,15 +1865,16 @@ timer() {
|
|
|
|
|
fi
|
|
|
|
|
done
|
|
|
|
|
|
|
|
|
|
validate_all=$(echo "${validate_all}" | tr ' ' '\n' | sort -n -u | tr '\n' ' ')
|
|
|
|
|
validate_all=$(echo "${validate_all}" | tr ' ' '\n' | sort -g -u | tr '\n' ' ')
|
|
|
|
|
for entry in $validate_all; do
|
|
|
|
|
if [ "${entry}" -ge "${3}" ]; then
|
|
|
|
|
if [ ${entry} -ge ${3} ]; then
|
|
|
|
|
echo "${entry}"
|
|
|
|
|
return 0
|
|
|
|
|
fi
|
|
|
|
|
done
|
|
|
|
|
|
|
|
|
|
echo "${validate_all%% *}"
|
|
|
|
|
echo "${validate_all// /}"
|
|
|
|
|
#echo "${validate_all%% *}"
|
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
local cron_compare="${3}"
|
|
|
|
|
@@ -1775,7 +1893,10 @@ timer() {
|
|
|
|
|
local cron_minute="$(echo -n "${2}" | awk '{print $1}')"
|
|
|
|
|
local cron_hour="$(echo -n "${2}" | awk '{print $2}')"
|
|
|
|
|
local cron_day_of_month="$(echo -n "${2}" | awk '{print $3}')"
|
|
|
|
|
local cron_month="$(echo -n "${2}" | awk '{print $4}')"Generating
|
|
|
|
|
local cron_month="$(echo -n "${2}" | awk '{print $4}')"
|
|
|
|
|
local cron_day_of_week="$(echo -n "${2}" | awk '{print $5}')"
|
|
|
|
|
|
|
|
|
|
local cron_next_minute="$(date --date=@"${cron_compare}" +"%-M")"
|
|
|
|
|
local cron_next_hour="$(date --date=@"${cron_compare}" +"%-H")"
|
|
|
|
|
local cron_next_day_of_month="$(date --date=@"${cron_compare}" +"%-d")"
|
|
|
|
|
local cron_next_month="$(date --date=@"${cron_compare}" +"%-m")"
|
|
|
|
|
@@ -1784,8 +1905,10 @@ timer() {
|
|
|
|
|
local cron_next_year="$(date --date=@"${cron_compare}" +"%-Y")"
|
|
|
|
|
|
|
|
|
|
local cron_next=
|
|
|
|
|
local cron_parsed=1
|
|
|
|
|
|
|
|
|
|
while [ "$cron_parsed" != "0" ]; do
|
|
|
|
|
while [ "${cron_parsed}" != "0" ]; do
|
|
|
|
|
print_debug "[timer] [cron] Parse Minute"
|
|
|
|
|
cron_next=$(parse_expression "${cron_minute}" 59 "${cron_next_minute}")
|
|
|
|
|
if [ "${cron_next}" != "${cron_next_minute}" ]; then
|
|
|
|
|
if [ "${cron_next_minute}" -gt "${cron_next}" ]; then
|
|
|
|
|
@@ -1795,20 +1918,22 @@ timer() {
|
|
|
|
|
cron_next_minute="${cron_next}"
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
print_debug "[timer] [cron] Parse Hour"
|
|
|
|
|
cron_next=$(parse_expression "${cron_hour}" 23 "${cron_next_hour}")
|
|
|
|
|
if [ "${cron_next}" != "${cron_next_hour}" ]; then
|
|
|
|
|
if [ "${cron_next_hour}" -gt "${cron_next}" ]; then
|
|
|
|
|
cron_next_day_of_month=$(( cron_next_day_of_month + 1 ))
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
cron_next_hour="${cron_next}"
|
|
|
|
|
#cron_next_minute=0
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
print_debug "[timer] [cron] Parse Day of Week"
|
|
|
|
|
cron_next=$(parse_expression "${cron_day_of_week}" 6 "${cron_next_day_of_week}")
|
|
|
|
|
if [ "${cron_next}" != "${cron_next_day_of_week}" ]; then
|
|
|
|
|
day_of_week_difference=$(( ${cron_next} - ${cron_next_day_of_week} ))
|
|
|
|
|
day_of_week_difference=$(( cron_next - cron_next_day_of_week ))
|
|
|
|
|
|
|
|
|
|
if [ "${day_of_week_difference}" -lt "0" ]; then
|
|
|
|
|
if [ "${day_of_week_difference}" -lt 0 ]; then
|
|
|
|
|
day_of_week_difference=$(( day_of_week_difference + 7 ))
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
@@ -1817,6 +1942,7 @@ timer() {
|
|
|
|
|
cron_next_minute=0
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
print_debug "[timer] [cron] Parse day of month"
|
|
|
|
|
case "${cron_next_month}" in
|
|
|
|
|
1|3|5|7|8|10|12)
|
|
|
|
|
last_day_of_month="31"
|
|
|
|
|
@@ -1856,6 +1982,7 @@ timer() {
|
|
|
|
|
cron_next_day_of_month=$cron_next
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
print_debug "[timer] [cron] Parse Next Month"
|
|
|
|
|
cron_next=$(parse_expression "${cron_month}" 12 "${cron_next_month}")
|
|
|
|
|
if [ "${cron_next}" != "${cron_next_month}" ]; then
|
|
|
|
|
if [ "${cron_next}" -gt "12" ]; then
|
|
|
|
|
@@ -1881,18 +2008,18 @@ timer() {
|
|
|
|
|
;;
|
|
|
|
|
datetime)
|
|
|
|
|
time_begin=$(date -d "${backup_job_backup_begin}" +%s)
|
|
|
|
|
print_debug "BACKUP_BEGIN time = ${time_begin}"
|
|
|
|
|
print_debug "[timer] [datetime] BACKUP_BEGIN time = ${time_begin}"
|
|
|
|
|
time_wait=$(( time_begin - time_current ))
|
|
|
|
|
print_debug "Difference in seconds: ${time_wait}"
|
|
|
|
|
print_debug "[timer] [datetime] Difference in seconds: ${time_wait}"
|
|
|
|
|
|
|
|
|
|
if (( ${time_wait} < 0 )); then
|
|
|
|
|
time_wait=$(( (${time_wait} + (${backup_job_backup_interval} - 1)) / (${backup_job_backup_interval} * 60) ))
|
|
|
|
|
time_wait=$(( ${time_wait} * -1 ))
|
|
|
|
|
print_debug "Difference in seconds (rounded) time_wait is in the past : ${time_wait}"
|
|
|
|
|
print_debug "[timer] [datetime] Difference in seconds (rounded) time_wait is in the past : ${time_wait}"
|
|
|
|
|
fi
|
|
|
|
|
|
|
|
|
|
time_future=$(( time_current + time_wait ))
|
|
|
|
|
print_debug "Future execution time = ${time_future}"
|
|
|
|
|
print_debug "[timer] [datetime] Future execution time = ${time_future}"
|
|
|
|
|
;;
|
|
|
|
|
job)
|
|
|
|
|
case "${2}" in
|
|
|
|
|
@@ -1967,4 +2094,4 @@ write_log() {
|
|
|
|
|
print_${_arg_log_level} "${_arg_log_message}"
|
|
|
|
|
output_on
|
|
|
|
|
if var_true "${DEBUG_WRITE_LOG}" ; then debug off; fi
|
|
|
|
|
}
|
|
|
|
|
}
|
|
|
|
|
|