参考这篇文章,当设置到 Use auto config? 是否使用自动配置,选 n 不自动配置,然后根据提示用浏览器打开 rclone 给出的 URL,点击接受(Accept),然后将浏览器上显示出来的字符串粘贴回命令行里,完成授权,然后退出即可。参考文章里有挂载的操作,记得这里不需要挂载 Google Drive。
#!/usr/bin/env bash # Copyright (C) 2013 - 2020 Teddysun <i@teddysun.com> # # This file is part of the LAMP script. # # LAMP is a powerful bash script for the installation of # Apache + PHP + MySQL/MariaDB and so on. # You can install Apache + PHP + MySQL/MariaDB in an very easy way. # Just need to input numbers to choose what you want to install before installation. # And all things will be done in a few minutes. # # Description: Auto backup shell script # Description URL: https://teddysun.com/469.html # # Website: https://lamp.sh # Github: https://github.com/teddysun/lamp # # You must to modify the config before run it!!! # Backup MySQL/MariaDB datebases, files and directories # Backup file is encrypted with AES256-cbc with SHA1 message-digest (option) # Auto transfer backup file to Google Drive (need install rclone command) (option) # Auto transfer backup file to FTP server (option) # Auto delete Google Drive's or FTP server's remote file (option)
[[ $EUID -ne 0 ]] && echo"Error: This script must be run as root!" && exit 1
########## START OF CONFIG ##########
# Encrypt flag (true: encrypt, false: not encrypt) ENCRYPTFLG=true
# WARNING: KEEP THE PASSWORD SAFE!!! # The password used to encrypt the backup # To decrypt backups made by this script, run the following command: # openssl enc -aes256 -in [encrypted backup] -out decrypted_backup.tgz -pass pass:[backup password] -d -md sha1 BACKUPPASS=""
# Directory to store backups LOCALDIR=""
# Temporary directory used during backup creation TEMPDIR="/tmp/backups/temp/"
# File to log the outcome of backups LOGFILE=""
# OPTIONAL: # If you want to backup the MySQL database, enter the MySQL root password below, otherwise leave it blank MYSQL_ROOT_PASSWORD=""
# Below is a list of MySQL database name that will be backed up # If you want backup ALL databases, leave it blank. MYSQL_DATABASE_NAME[0]=""
# Below is a list of files and directories that will be backed up in the tar backup # For example: # File: /data/www/default/test.tgz # Directory: /data/www/default/test BACKUP[0]=""
# Number of days to store daily local backups (default 7 days) LOCALAGEDAILIES="7"
# Delete remote file from Googole Drive or FTP server flag (true: delete, false: not delete) DELETE_REMOTE_FILE_FLG=true
# Rclone remote name RCLONE_NAME=""
# Rclone remote folder name (default "") RCLONE_FOLDER=""
# Cos remote folder name (default "") COS_FOLDER=""
# AliyunDrive remote folder name (default "") ALI_FOLDER=""
# Upload local file to FTP server flag (true: upload, false: not upload) FTP_FLG=false
# Upload local file to Google Drive flag (true: upload, false: not upload) RCLONE_FLG=false
# Upload local file to Cos flag (true: upload, false: not upload) COS_FLG=false
# Upload local file to AliyunDrive flag (true: upload, false: not upload) ALI_FLG=true ALI_PY_FILE="/usr/local/bin/aliyunpan-cli"
# FTP server # OPTIONAL: If you want to upload to FTP server, enter the Hostname or IP address below FTP_HOST=""
# FTP username # OPTIONAL: If you want to upload to FTP server, enter the FTP username below FTP_USER=""
# FTP password # OPTIONAL: If you want to upload to FTP server, enter the username's password below FTP_PASS=""
# FTP server remote folder # OPTIONAL: If you want to upload to FTP server, enter the FTP remote folder below # For example: public_html FTP_DIR=""
########## END OF CONFIG ##########
# Date & Time DAY=$(date +%d) MONTH=$(date +%m) YEAR=$(date +%C%y) BACKUPDATE=$(date +%Y%m%d%H%M%S) # Backup file name TARFILE="${LOCALDIR}""$(hostname)"_"${BACKUPDATE}".tgz # Encrypted backup file name ENC_TARFILE="${TARFILE}.enc" # Backup MySQL dump file name SQLFILE="${TEMPDIR}mysql_${BACKUPDATE}.sql"
# Check for list of mandatory binaries check_commands() { # This section checks for all of the binaries used in the backup # Do not check mysql command if you do not want to backup the MySQL database if [ -z "${MYSQL_ROOT_PASSWORD}" ]; then BINARIES=( catcddudatedirnameecho openssl pwdrm tar ) else BINARIES=( catcddudatedirnameecho openssl mysql mysqldump pwdrm tar ) fi
# Iterate over the list of binaries, and if one isn't found, abort for BINARY in"${BINARIES[@]}"; do if [ ! "$(command -v "$BINARY")" ]; then log"$BINARY is not installed. Install it and try again" exit 1 fi done
# check rclone command RCLONE_COMMAND=false if [ "$(command -v "rclone")" ]; then RCLONE_COMMAND=true fi
# check COS command COS_COMMAND=false if [ "$(command -v "coscmd")" ]; then COS_COMMAND=true fi
# check AliyunDrive command ALI_COMMAND=false if [ -f "${ALI_PY_FILE}" ]; then ALI_COMMAND=true fi
# check ftp command if${FTP_FLG}; then if [ ! "$(command -v "ftp")" ]; then log"ftp is not installed. Install it and try again" exit 1 fi fi }
calculate_size() { local file_name=$1 local file_size=$(du -h $file_name 2>/dev/null | awk '{print $1}') if [ "x${file_size}" = "x" ]; then echo"unknown" else echo"${file_size}" fi }
# Backup MySQL databases mysql_backup() { if [ -z "${MYSQL_ROOT_PASSWORD}" ]; then log"MySQL root password not set, MySQL backup skipped" else log"MySQL dump start" mysql -u root -p"${MYSQL_ROOT_PASSWORD}" 2>/dev/null <<EOF exit EOF if [ $? -ne 0 ]; then log"MySQL root password is incorrect. Please check it and try again" exit 1 fi if [ "${MYSQL_DATABASE_NAME[@]}" == "" ]; then mysqldump -u root -p"${MYSQL_ROOT_PASSWORD}" --all-databases > "${SQLFILE}" 2>/dev/null if [ $? -ne 0 ]; then log"MySQL all databases backup failed" exit 1 fi log"MySQL all databases dump file name: ${SQLFILE}" #Add MySQL backup dump file to BACKUP list BACKUP=(${BACKUP[@]}${SQLFILE}) else for db in${MYSQL_DATABASE_NAME[@]}; do unset DBFILE DBFILE="${TEMPDIR}${db}_${BACKUPDATE}.sql" mysqldump -u root -p"${MYSQL_ROOT_PASSWORD}"${db} > "${DBFILE}" 2>/dev/null if [ $? -ne 0 ]; then log"MySQL database name [${db}] backup failed, please check database name is correct and try again" exit 1 fi log"MySQL database name [${db}] dump file name: ${DBFILE}" #Add MySQL backup dump file to BACKUP list BACKUP=(${BACKUP[@]}${DBFILE}) done fi log"MySQL dump completed" fi }
start_backup() { [ "${#BACKUP[@]}" -eq 0 ] && echo"Error: You must to modify the [$(basename $0)] config before run it!" && exit 1
log"Tar backup file start" tar -zcPf ${TARFILE}${BACKUP[@]} if [ $? -gt 1 ]; then log"Tar backup file failed" exit 1 fi log"Tar backup file completed"
# Delete unencrypted tar log"Delete unencrypted tar file: ${TARFILE}" rm -f ${TARFILE} fi
# Delete MySQL temporary dump file for sql in $(ls${TEMPDIR}*.sql); do log"Delete MySQL temporary dump file: ${sql}" rm -f ${sql} done
if${ENCRYPTFLG}; then OUT_FILE="${ENC_TARFILE}" else OUT_FILE="${TARFILE}" fi log"File name: ${OUT_FILE}, File size: $(calculate_size ${OUT_FILE})" }
# Transfer backup file to Google Drive # If you want to install rclone command, please visit website: # https://rclone.org/downloads/ rclone_upload() { if${RCLONE_FLG} && ${RCLONE_COMMAND}; then [ -z "${RCLONE_NAME}" ] && log"Error: RCLONE_NAME can not be empty!" && return 1 if [ -n "${RCLONE_FOLDER}" ]; then rclone ls${RCLONE_NAME}:${RCLONE_FOLDER} 2>&1 > /dev/null if [ $? -ne 0 ]; then log"Create the path ${RCLONE_NAME}:${RCLONE_FOLDER}" rclone mkdir${RCLONE_NAME}:${RCLONE_FOLDER} fi fi log"Tranferring backup file: ${OUT_FILE} to Google Drive" rclone copy ${OUT_FILE}${RCLONE_NAME}:${RCLONE_FOLDER} >> ${LOGFILE} if [ $? -ne 0 ]; then log"Error: Tranferring backup file: ${OUT_FILE} to Google Drive failed" return 1 fi log"Tranferring backup file: ${OUT_FILE} to Google Drive completed" fi }
# Tranferring backup file to COS cos_upload() { if${COS_FLG} && ${COS_COMMAND}; then [ -z "${COS_FOLDER}" ] && log"Error: COS_FOLDER can not be empty!" && return 1 log"Tranferring backup file: ${OUT_FILE} to COS" coscmd upload ${OUT_FILE}${COS_FOLDER}/ >> ${LOGFILE} if [ $? -ne 0 ]; then log"Error: Tranferring backup file: ${OUT_FILE} to COS" return 1 fi log"Tranferring backup file: ${OUT_FILE} to COS completed" fi }
# Tranferring backup file to AliyunDrive ali_upload() { if${ALI_FLG} && ${ALI_COMMAND}; then [ -z "${ALI_FOLDER}" ] && log"Error: ALI_FOLDER can not be empty!" && return 1 log"Tranferring backup file: ${OUT_FILE} to AliyunDrive" ${ALI_PY_FILE} upload ${OUT_FILE}${ALI_FOLDER}# >> ${LOGFILE} if [ $? -ne 0 ]; then log"Error: Tranferring backup file: ${OUT_FILE} to AliyunDrive" return 1 fi log"Tranferring backup file: ${OUT_FILE} to AliyunDrive completed" fi }
# Tranferring backup file to FTP server ftp_upload() { if${FTP_FLG}; then [ -z "${FTP_HOST}" ] && log"Error: FTP_HOST can not be empty!" && return 1 [ -z "${FTP_USER}" ] && log"Error: FTP_USER can not be empty!" && return 1 [ -z "${FTP_PASS}" ] && log"Error: FTP_PASS can not be empty!" && return 1 [ -z "${FTP_DIR}" ] && log"Error: FTP_DIR can not be empty!" && return 1 local FTP_OUT_FILE=$(basename${OUT_FILE}) log"Tranferring backup file: ${FTP_OUT_FILE} to FTP server" ftp -in${FTP_HOST} 2>&1 >> ${LOGFILE} <<EOF user $FTP_USER $FTP_PASS binary lcd $LOCALDIR cd $FTP_DIR put $FTP_OUT_FILE quit EOF if [ $? -ne 0 ]; then log"Error: Tranferring backup file: ${FTP_OUT_FILE} to FTP server failed" return 1 fi log"Tranferring backup file: ${FTP_OUT_FILE} to FTP server completed" fi }
# Get file date get_file_date() { #Approximate a 30-day month and 365-day year DAYS=$(( $((10#${YEAR}*365)) + $((10#${MONTH}*30)) + $((10#${DAY})) )) unset FILEYEAR FILEMONTH FILEDAY FILEDAYS FILEAGE FILEYEAR=$(echo"$1" | cut -d_ -f2 | cut -c 1-4) FILEMONTH=$(echo"$1" | cut -d_ -f2 | cut -c 5-6) FILEDAY=$(echo"$1" | cut -d_ -f2 | cut -c 7-8) if [[ "${FILEYEAR}" && "${FILEMONTH}" && "${FILEDAY}" ]]; then #Approximate a 30-day month and 365-day year FILEDAYS=$(( $((10#${FILEYEAR}*365)) + $((10#${FILEMONTH}*30)) + $((10#${FILEDAY})) )) FILEAGE=$(( 10#${DAYS} - 10#${FILEDAYS} )) return 0 fi return 1 }
# Delete Google Drive's old backup file delete_gdrive_file() { local FILENAME=$1 if${DELETE_REMOTE_FILE_FLG} && ${RCLONE_COMMAND}; then rclone ls${RCLONE_NAME}:${RCLONE_FOLDER}/${FILENAME} 2>&1 > /dev/null if [ $? -eq 0 ]; then rclone delete ${RCLONE_NAME}:${RCLONE_FOLDER}/${FILENAME} >> ${LOGFILE} if [ $? -eq 0 ]; then log"Google Drive's old backup file: ${FILENAME} has been deleted" else log"Failed to delete Google Drive's old backup file: ${FILENAME}" fi else log"Google Drive's old backup file: ${FILENAME} is not exist" fi fi }
# Delete COS's old backup file delete_cos_file() { local FILENAME=$1 if${DELETE_REMOTE_FILE_FLG} && ${COS_COMMAND}; then cos delete ${COS_FOLDER}/${FILENAME} >> ${LOGFILE} if [ $? -eq 0 ]; then log"COS's old backup file: ${FILENAME} has been deleted" else log"Failed to delete COS's old backup file: ${FILENAME}" fi fi }
# Delete AliyunDrive's old backup file delete_ali_file() { local FILENAME=$1 if${DELETE_REMOTE_FILE_FLG} && ${ALI_COMMAND}; then ${ALI_PY_FILE} delete ${ALI_FOLDER}${FILENAME}# >> ${LOGFILE} if [ $? -eq 0 ]; then log"AliyunDrive's old backup file: ${FILENAME} has been deleted" else log"Failed to delete AliyunDrive's old backup file: ${FILENAME}" fi fi }
# Delete FTP server's old backup file delete_ftp_file() { local FILENAME=$1 if${DELETE_REMOTE_FILE_FLG} && ${FTP_FLG}; then ftp -in${FTP_HOST} 2>&1 >> ${LOGFILE} <<EOF user $FTP_USER $FTP_PASS cd $FTP_DIR del $FILENAME quit EOF if [ $? -eq 0 ]; then log"FTP server's old backup file: ${FILENAME} has been deleted" else log"Failed to delete FTP server's old backup file: ${FILENAME}" fi fi }
# Clean up old file clean_up_files() { cd${LOCALDIR} || exit if${ENCRYPTFLG}; then LS=($(ls *.enc)) else LS=($(ls *.tgz)) fi for f in${LS[@]}; do get_file_date ${f} if [ $? -eq 0 ]; then if [[ ${FILEAGE} -gt ${LOCALAGEDAILIES} ]]; then rm -f ${f} log"Old backup file name: ${f} has been deleted" delete_gdrive_file ${f} delete_ftp_file ${f} delete_cos_file ${f} delete_ali_file ${f} fi fi done }
# Main progress STARTTIME=$(date +%s)
# Check if the backup folders exist and are writeable [ ! -d "${LOCALDIR}" ] && mkdir -p ${LOCALDIR} [ ! -d "${TEMPDIR}" ] && mkdir -p ${TEMPDIR}