- weekly-backup.sh: replace hardcoded BACKUP_DIRS with glob auto-discovery (catches nextcloud-backup, council-complaints-backup, future dirs) - weekly-backup.sh: add auto SQLite backup from PVC snapshots (magic number check, ?mode=ro URI, fallback to raw copy) - offsite-sync-backup.sh: add NFS media direct-to-Synology sync (Immich, calibre, audiobookshelf — reuses existing TrueNAS Cloud Sync paths) - Cleaned up 9 orphaned LVs + 38 snapshots on PVE host (101GB reclaimed) Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
133 lines
5.1 KiB
Bash
133 lines
5.1 KiB
Bash
#!/usr/bin/env bash
|
|
# offsite-sync-backup — Sync /mnt/backup to Synology NAS using changed-files manifest
|
|
# Deploy to PVE host at /usr/local/bin/offsite-sync-backup
|
|
# Schedule: Weekly Sunday 08:00 via systemd timer (After=weekly-backup.service)
|
|
set -euo pipefail
|
|
|
|
# --- Configuration ---
|
|
BACKUP_ROOT="/mnt/backup"
|
|
DEST="Administrator@192.168.1.13:/volume1/Backup/Viki/pve-backup"
|
|
MANIFEST="${BACKUP_ROOT}/.changed-files"
|
|
PUSHGATEWAY="${OFFSITE_SYNC_PUSHGATEWAY:-http://10.0.20.100:30091}"
|
|
PUSHGATEWAY_JOB="offsite-backup-sync"
|
|
LOCKFILE="/run/offsite-sync-backup.lock"
|
|
|
|
# NFS media — synced directly to Synology (bypasses sda, too large to fit)
|
|
NFS_BASE="/srv/nfs"
|
|
NFS_SSD_BASE="/srv/nfs-ssd"
|
|
SYNOLOGY_NFS_DEST="Administrator@192.168.1.13:/volume1/Backup/Viki/truenas"
|
|
|
|
# --- Logging ---
|
|
log() { echo "[$(date '+%Y-%m-%d %H:%M:%S')] $*"; }
|
|
warn() { log "WARN: $*" >&2; }
|
|
|
|
# --- Locking ---
|
|
cleanup() { rm -f "${LOCKFILE}"; }
|
|
trap cleanup EXIT
|
|
if ! ( set -o noclobber; echo $$ > "${LOCKFILE}" ) 2>/dev/null; then
|
|
log "FATAL: Another instance running"; exit 1
|
|
fi
|
|
|
|
# --- Main ---
|
|
log "=== Offsite sync starting ==="
|
|
STATUS=0
|
|
|
|
if ! mountpoint -q "${BACKUP_ROOT}"; then
|
|
log "FATAL: ${BACKUP_ROOT} is not mounted"; exit 1
|
|
fi
|
|
|
|
# Test SSH connectivity first
|
|
if ! timeout 10 ssh -o BatchMode=yes -o ConnectTimeout=5 Administrator@192.168.1.13 true 2>/dev/null; then
|
|
log "FATAL: Cannot SSH to Synology (192.168.1.13)"
|
|
echo "backup_last_success_timestamp 0" | \
|
|
curl -s --connect-timeout 5 --max-time 10 --data-binary @- \
|
|
"${PUSHGATEWAY}/metrics/job/${PUSHGATEWAY_JOB}" 2>/dev/null || true
|
|
exit 1
|
|
fi
|
|
|
|
DAY_OF_MONTH=$(date +%d)
|
|
|
|
if [ "${DAY_OF_MONTH}" -le 7 ]; then
|
|
# First Sunday of month: full sync with --delete to clean orphans on Synology
|
|
log "Monthly full sync (1st Sunday)..."
|
|
rsync -rltz --delete --chmod=Du=rwx,Dgo=rx,Fu=rw,Fog=r \
|
|
--exclude='.changed-files' \
|
|
--exclude='.last-offsite-sync' \
|
|
--exclude='.lv-pvc-mapping.json' \
|
|
"${BACKUP_ROOT}/" "${DEST}/" 2>&1 || STATUS=1
|
|
elif [ -s "${MANIFEST}" ]; then
|
|
# Incremental: only send files listed in manifest (no remote dir walk)
|
|
MANIFEST_LINES=$(wc -l < "${MANIFEST}")
|
|
log "Incremental sync (${MANIFEST_LINES} files from manifest)..."
|
|
rsync -rltz --chmod=Du=rwx,Dgo=rx,Fu=rw,Fog=r --files-from="${MANIFEST}" --no-traverse \
|
|
"${BACKUP_ROOT}/" "${DEST}/" 2>&1 || STATUS=1
|
|
else
|
|
log "No changed files in manifest, nothing to sync"
|
|
fi
|
|
|
|
# ============================================================
|
|
# STEP 2: NFS media direct to Synology (bypasses sda — too large)
|
|
# Reuses existing TrueNAS Cloud Sync paths on Synology
|
|
# ============================================================
|
|
log "--- Step 2: NFS media direct to Synology ---"
|
|
|
|
# Immich (map Proxmox paths to existing Synology layout)
|
|
for subdir in backups encoded-video library profile upload; do
|
|
if [ -d "${NFS_BASE}/immich/${subdir}" ]; then
|
|
rsync -rltz --delete \
|
|
"${NFS_BASE}/immich/${subdir}/" \
|
|
"${SYNOLOGY_NFS_DEST}/immich/immich/${subdir}/" 2>&1 \
|
|
&& log " OK: immich/${subdir}" \
|
|
|| { warn "Failed: immich/${subdir}"; STATUS=1; }
|
|
fi
|
|
done
|
|
# Immich PG data + dumps
|
|
if [ -d "${NFS_BASE}/immich/postgresql" ]; then
|
|
rsync -rltz --delete "${NFS_BASE}/immich/postgresql/" \
|
|
"${SYNOLOGY_NFS_DEST}/immich/data-immich-postgresql/" 2>&1 \
|
|
&& log " OK: immich/postgresql" \
|
|
|| { warn "Failed: immich/postgresql"; STATUS=1; }
|
|
fi
|
|
# Immich SSD (thumbs, ML cache)
|
|
if [ -d "${NFS_SSD_BASE}/immich/thumbs" ]; then
|
|
rsync -rltz --delete "${NFS_SSD_BASE}/immich/thumbs/" \
|
|
"${SYNOLOGY_NFS_DEST}/immich/immich/thumbs/" 2>&1 \
|
|
&& log " OK: immich/thumbs" \
|
|
|| { warn "Failed: immich/thumbs"; STATUS=1; }
|
|
fi
|
|
if [ -d "${NFS_SSD_BASE}/immich/machine-learning" ]; then
|
|
rsync -rltz --delete "${NFS_SSD_BASE}/immich/machine-learning/" \
|
|
"${SYNOLOGY_NFS_DEST}/immich/machine-learning/" 2>&1 \
|
|
&& log " OK: immich/machine-learning" \
|
|
|| { warn "Failed: immich/machine-learning"; STATUS=1; }
|
|
fi
|
|
# Calibre + Audiobookshelf
|
|
for media_dir in calibre audiobookshelf; do
|
|
if [ -d "${NFS_BASE}/${media_dir}" ]; then
|
|
rsync -rltz --delete "${NFS_BASE}/${media_dir}/" \
|
|
"${SYNOLOGY_NFS_DEST}/${media_dir}/" 2>&1 \
|
|
&& log " OK: ${media_dir}" \
|
|
|| { warn "Failed: ${media_dir}"; STATUS=1; }
|
|
fi
|
|
done
|
|
|
|
# ============================================================
|
|
# Finish
|
|
# ============================================================
|
|
if [ "${STATUS}" -eq 0 ]; then
|
|
# Only clear manifest + update timestamp on SUCCESS
|
|
touch "${BACKUP_ROOT}/.last-offsite-sync"
|
|
> "${MANIFEST}"
|
|
log "=== Offsite sync complete (success) ==="
|
|
else
|
|
# Keep manifest for retry next week
|
|
warn "Offsite sync had errors — manifest preserved for retry"
|
|
log "=== Offsite sync complete (with errors) ==="
|
|
fi
|
|
|
|
cat <<EOF | curl -s --connect-timeout 5 --max-time 10 --data-binary @- "${PUSHGATEWAY}/metrics/job/${PUSHGATEWAY_JOB}" 2>/dev/null || true
|
|
backup_last_success_timestamp $(date +%s)
|
|
offsite_sync_last_status ${STATUS}
|
|
EOF
|
|
|
|
exit "${STATUS}"
|