Compare commits
2 Commits
2.3.0+2.3.
...
main
Author | SHA1 | Date | |
---|---|---|---|
0abb4827e7 | |||
0588a06a97 |
17
.drone.yml
17
.drone.yml
@ -1,17 +0,0 @@
|
|||||||
---
|
|
||||||
kind: pipeline
|
|
||||||
name: linters
|
|
||||||
steps:
|
|
||||||
- name: publish image
|
|
||||||
image: plugins/docker
|
|
||||||
settings:
|
|
||||||
username: 3wordchant
|
|
||||||
password:
|
|
||||||
from_secret: git_coopcloud_tech_token_3wc
|
|
||||||
repo: git.coopcloud.tech/coop-cloud/backup-bot-two
|
|
||||||
tags: ${DRONE_SEMVER_BUILD}
|
|
||||||
registry: git.coopcloud.tech
|
|
||||||
when:
|
|
||||||
event:
|
|
||||||
include:
|
|
||||||
- tag
|
|
28
.env.sample
28
.env.sample
@ -1,34 +1,10 @@
|
|||||||
TYPE=backup-bot-two
|
STACK_NAME=backup-bot-two
|
||||||
|
|
||||||
SECRET_RESTIC_PASSWORD_VERSION=v1
|
|
||||||
|
|
||||||
COMPOSE_FILE=compose.yml
|
|
||||||
|
|
||||||
RESTIC_REPOSITORY=/backups/restic
|
RESTIC_REPOSITORY=/backups/restic
|
||||||
|
|
||||||
CRON_SCHEDULE='30 3 * * *'
|
CRON_SCHEDULE='30 3 * * *'
|
||||||
|
|
||||||
# Push Notifiactions
|
# Push Notifications
|
||||||
#PUSH_URL_START=https://status.example.com/api/push/xxxxxxxxxx?status=up&msg=start
|
#PUSH_URL_START=https://status.example.com/api/push/xxxxxxxxxx?status=up&msg=start
|
||||||
#PUSH_URL_SUCCESS=https://status.example.com/api/push/xxxxxxxxxx?status=up&msg=OK
|
#PUSH_URL_SUCCESS=https://status.example.com/api/push/xxxxxxxxxx?status=up&msg=OK
|
||||||
#PUSH_URL_FAIL=https://status.example.com/api/push/xxxxxxxxxx?status=down&msg=fail
|
#PUSH_URL_FAIL=https://status.example.com/api/push/xxxxxxxxxx?status=down&msg=fail
|
||||||
|
|
||||||
# swarm-cronjob, instead of built-in cron
|
|
||||||
#COMPOSE_FILE="$COMPOSE_FILE:compose.swarm-cronjob.yml"
|
|
||||||
|
|
||||||
# SSH storage
|
|
||||||
#SECRET_SSH_KEY_VERSION=v1
|
|
||||||
#SSH_HOST_KEY="hostname ssh-rsa AAAAB3...
|
|
||||||
#COMPOSE_FILE="$COMPOSE_FILE:compose.ssh.yml"
|
|
||||||
|
|
||||||
# S3 storage
|
|
||||||
#SECRET_AWS_SECRET_ACCESS_KEY_VERSION=v1
|
|
||||||
#AWS_ACCESS_KEY_ID=something-secret
|
|
||||||
#COMPOSE_FILE="$COMPOSE_FILE:compose.s3.yml"
|
|
||||||
|
|
||||||
# Secret restic repository
|
|
||||||
# use a secret to store the RESTIC_REPOSITORY if the repository location contains a secret value
|
|
||||||
# i.E rest:https://user:SECRET_PASSWORD@host:8000/
|
|
||||||
# it overwrites the RESTIC_REPOSITORY variable
|
|
||||||
#SECRET_RESTIC_REPO_VERSION=v1
|
|
||||||
#COMPOSE_FILE="$COMPOSE_FILE:compose.secret.yml"
|
|
||||||
|
@ -1,17 +0,0 @@
|
|||||||
export RESTIC_HOST="user@domain.tld"
|
|
||||||
export RESTIC_PASSWORD_FILE=/run/secrets/restic-password
|
|
||||||
export BACKUP_DEST=/backups
|
|
||||||
|
|
||||||
export SERVER_NAME=domain.tld
|
|
||||||
export DOCKER_CONTEXT=$SERVER_NAME
|
|
||||||
|
|
||||||
# uncomment either this:
|
|
||||||
#export SSH_KEY_FILE=~/.ssh/id_rsa
|
|
||||||
# or this:
|
|
||||||
#export AWS_SECRET_ACCESS_KEY_FILE=s3
|
|
||||||
#export AWS_ACCESS_KEY_ID=easter-october-emphatic-tug-urgent-customer
|
|
||||||
# or this:
|
|
||||||
#export HTTPS_PASSWORD_FILE=/run/secrets/https_password
|
|
||||||
|
|
||||||
# optionally limit subset of services for testing
|
|
||||||
#export SERVICES_OVERRIDE="ghost_domain_tld_app ghost_domain_tld_db"
|
|
2
.gitignore
vendored
2
.gitignore
vendored
@ -1 +1 @@
|
|||||||
/testing
|
.env
|
@ -1,6 +0,0 @@
|
|||||||
# Change log
|
|
||||||
|
|
||||||
## 2.0.0 (unreleased)
|
|
||||||
|
|
||||||
- Rewrite from Bash to Python
|
|
||||||
- Add support for push notifications (#24)
|
|
11
Dockerfile
11
Dockerfile
@ -1,11 +0,0 @@
|
|||||||
FROM docker:24.0.7-dind
|
|
||||||
|
|
||||||
RUN apk add --upgrade --no-cache restic bash python3 py3-pip py3-click py3-docker-py py3-json-logger curl
|
|
||||||
|
|
||||||
# Todo use requirements file with specific versions
|
|
||||||
RUN pip install --break-system-packages resticpy==1.0.2
|
|
||||||
|
|
||||||
COPY backupbot.py /usr/bin/backup
|
|
||||||
COPY entrypoint.sh /entrypoint.sh
|
|
||||||
|
|
||||||
ENTRYPOINT /entrypoint.sh
|
|
279
README.md
279
README.md
@ -1,106 +1,27 @@
|
|||||||
# Backupbot II
|
# Backupbot II
|
||||||
|
|
||||||
[](https://build.coopcloud.tech/coop-cloud/backup-bot-two)
|
Wiki Cafe's configuration for a Backupbot II deployment. Originally slimmed down from an `abra` [recipe](https://git.coopcloud.tech/coop-cloud/backup-bot-two) by [Co-op Cloud](https://coopcloud.tech/).
|
||||||
|
|
||||||
_This Time, It's Easily Configurable_
|
|
||||||
|
|
||||||
Automatically take backups from all volumes of running Docker Swarm services and runs pre- and post commands.
|
|
||||||
|
|
||||||
<!-- metadata -->
|
|
||||||
|
|
||||||
* **Category**: Utilities
|
|
||||||
* **Status**: 0, work-in-progress
|
|
||||||
* **Image**: [`git.coopcloud.tech/coop-cloud/backup-bot-two`](https://git.coopcloud.tech/coop-cloud/-/packages/container/backup-bot-two), 4, upstream
|
|
||||||
* **Healthcheck**: No
|
|
||||||
* **Backups**: N/A
|
|
||||||
* **Email**: N/A
|
|
||||||
* **Tests**: No
|
|
||||||
* **SSO**: N/A
|
|
||||||
|
|
||||||
<!-- endmetadata -->
|
|
||||||
|
|
||||||
|
|
||||||
## Background
|
## Deploying the app with Docker Swarm
|
||||||
|
|
||||||
There are lots of Docker volume backup systems; all of them have one or both of these limitations:
|
Set the environment variables from the .env file during the shell session.
|
||||||
- You need to define all the volumes to back up in the configuration system
|
|
||||||
- Backups require services to be stopped to take consistent copies
|
|
||||||
|
|
||||||
Backupbot II tries to help, by
|
|
||||||
1. **letting you define backups using Docker labels**, so you can **easily collect your backups for use with another system** like docker-volume-backup.
|
|
||||||
2. **running pre- and post-commands** before and after backups, for example to use database tools to take a backup from a running service.
|
|
||||||
|
|
||||||
## Deployment
|
|
||||||
|
|
||||||
### With Co-op Cloud
|
|
||||||
|
|
||||||
|
|
||||||
* `abra app new backup-bot-two`
|
|
||||||
* `abra app config <app-name>`
|
|
||||||
- set storage options. Either configure `CRON_SCHEDULE`, or set up `swarm-cronjob`
|
|
||||||
* `abra app secret generate -a <backupbot_name>`
|
|
||||||
* `abra app deploy <app-name>`
|
|
||||||
|
|
||||||
## Configuration
|
|
||||||
|
|
||||||
Per default Backupbot stores the backups locally in the repository `/backups/restic`, which is accessible as volume at `/var/lib/docker/volumes/<backupbot_name>_backups/_data/restic/`
|
|
||||||
|
|
||||||
The backup location can be changed using the `RESTIC_REPOSITORY` env variable.
|
|
||||||
|
|
||||||
### S3 Storage
|
|
||||||
|
|
||||||
To use S3 storage as backup location set the following envs:
|
|
||||||
```
|
```
|
||||||
RESTIC_REPOSITORY=s3:<S3-SERVICE-URL>/<BUCKET-NAME>
|
set -a && source .env && set +a
|
||||||
SECRET_AWS_SECRET_ACCESS_KEY_VERSION=v1
|
|
||||||
AWS_ACCESS_KEY_ID=<MY_ACCESS_KEY>
|
|
||||||
COMPOSE_FILE="$COMPOSE_FILE:compose.s3.yml"
|
|
||||||
```
|
```
|
||||||
and add your `<SECRET_ACCESS_KEY>` as docker secret:
|
|
||||||
`abra app secret insert <backupbot_name> aws_secret_access_key v1 <SECRET_ACCESS_KEY>`
|
|
||||||
|
|
||||||
See [restic s3 docs](https://restic.readthedocs.io/en/latest/030_preparing_a_new_repo.html#amazon-s3) for more information.
|
Set the secrets.
|
||||||
|
|
||||||
### SFTP Storage
|
|
||||||
|
|
||||||
> With sftp it is not possible to prevent the backupbot from deleting backups in case of a compromised machine. Therefore we recommend to use S3, REST or rclone server without delete permissions.
|
|
||||||
|
|
||||||
To use SFTP storage as backup location set the following envs:
|
|
||||||
```
|
```
|
||||||
RESTIC_REPOSITORY=sftp:user@host:/restic-repo-path
|
printf "SECRET_HERE" | docker secret create SECRET_NAME -
|
||||||
SECRET_SSH_KEY_VERSION=v1
|
|
||||||
SSH_HOST_KEY="hostname ssh-rsa AAAAB3...
|
|
||||||
COMPOSE_FILE="$COMPOSE_FILE:compose.ssh.yml"
|
|
||||||
```
|
```
|
||||||
To get the `SSH_HOST_KEY` run the following command `ssh-keyscan <hostname>`
|
|
||||||
|
|
||||||
Generate an ssh keypair: `ssh-keygen -t ed25519 -f backupkey -P ''`
|
Deploy using the `-c` flag to specify one or multiple compose files.
|
||||||
Add the key to your `authorized_keys`:
|
|
||||||
`ssh-copy-id -i backupkey <user>@<hostname>`
|
|
||||||
Add your `SSH_KEY` as docker secret:
|
|
||||||
```
|
|
||||||
abra app secret insert <backupbot_name> ssh_key v1 """$(cat backupkey)
|
|
||||||
"""
|
|
||||||
```
|
|
||||||
> Attention: This command needs to be executed exactly as stated above, because it places a trailing newline at the end, if this is missing you will get the following error: `Load key "/run/secrets/ssh_key": error in libcrypto`
|
|
||||||
|
|
||||||
### Restic REST server Storage
|
|
||||||
|
|
||||||
You can simply set the `RESTIC_REPOSITORY` variable to your REST server URL `rest:http://host:8000/`.
|
|
||||||
If you access the REST server with a password `rest:https://user:pass@host:8000/` you should hide the whole URL containing the password inside a secret.
|
|
||||||
Uncomment these lines:
|
|
||||||
```
|
```
|
||||||
SECRET_RESTIC_REPO_VERSION=v1
|
docker stack deploy backup-bot-two -c compose.yaml
|
||||||
COMPOSE_FILE="$COMPOSE_FILE:compose.secret.yml"
|
|
||||||
```
|
```
|
||||||
Add your REST server url as secret:
|
|
||||||
```
|
|
||||||
abra app secret insert <backupbot_name> restic_repo v1 "rest:https://user:pass@host:8000/"
|
|
||||||
```
|
|
||||||
The secret will overwrite the `RESTIC_REPOSITORY` variable.
|
|
||||||
|
|
||||||
|
|
||||||
See [restic REST docs](https://restic.readthedocs.io/en/latest/030_preparing_a_new_repo.html#rest-server) for more information.
|
|
||||||
|
|
||||||
## Push notifications
|
## Push notifications
|
||||||
|
|
||||||
@ -113,151 +34,95 @@ PUSH_URL_SUCCESS=https://status.example.com/api/push/xxxxxxxxxx?status=up&msg=OK
|
|||||||
PUSH_URL_FAIL=https://status.example.com/api/push/xxxxxxxxxx?status=down&msg=fail
|
PUSH_URL_FAIL=https://status.example.com/api/push/xxxxxxxxxx?status=down&msg=fail
|
||||||
```
|
```
|
||||||
|
|
||||||
|
## Commands
|
||||||
|
|
||||||
## Usage
|
|
||||||
|
|
||||||
Run the cronjob that creates a backup, including the push notifications and docker logging:
|
- Find the ID or name of the backup container:
|
||||||
`abra app cmd <backupbot_name> app run_cron`
|
```
|
||||||
|
docker ps --filter "name=backup-bot-two_app"
|
||||||
|
```
|
||||||
|
|
||||||
Create a backup of all apps:
|
2. Run the desired command using `docker exec`:
|
||||||
|
```
|
||||||
|
docker exec -it <container_id_or_name> backup <command> [options]
|
||||||
|
```
|
||||||
|
Replace `<container_id_or_name>` with the ID or name of the backup container.
|
||||||
|
|
||||||
`abra app run <backupbot_name> app -- backup create`
|
Available commands:
|
||||||
|
- `create`: Initiate the backup process.
|
||||||
|
- `restore`: Restore a specific snapshot to a target directory.
|
||||||
|
- `snapshots`: List available snapshots.
|
||||||
|
- `ls`: List files in a specific snapshot.
|
||||||
|
- `download`: Download specific files, volumes, or secrets from a snapshot.
|
||||||
|
|
||||||
> The apps to backup up need to be deployed
|
Options:
|
||||||
|
- `--host`, `-h`: Specify the service name (e.g., `app`).
|
||||||
|
- `--repo`, `-r`: Specify the Restic repository location (e.g., `/run/secrets/restic_repo`).
|
||||||
|
- `--log`, `-l`: Set the log level (e.g., `debug`, `info`, `warning`, `error`).
|
||||||
|
- `--machine-logs`, `-m`: Enable machine-readable JSON logging.
|
||||||
|
|
||||||
Create an individual backup:
|
## Examples
|
||||||
|
|
||||||
`abra app run <backupbot_name> app -- backup --host <target_app_name> create`
|
Create a backup:
|
||||||
|
|
||||||
Create a backup to a local repository:
|
|
||||||
|
|
||||||
`abra app run <backupbot_name> app -- backup create -r /backups/restic`
|
|
||||||
|
|
||||||
> It is recommended to shutdown/undeploy an app before restoring the data
|
|
||||||
|
|
||||||
Restore the latest snapshot of all including apps:
|
|
||||||
|
|
||||||
`abra app run <backupbot_name> app -- backup restore`
|
|
||||||
|
|
||||||
Restore a specific snapshot of an individual app:
|
|
||||||
|
|
||||||
`abra app run <backupbot_name> app -- backup --host <target_app_name> restore --snapshot <snapshot_id>`
|
|
||||||
|
|
||||||
Show all snapshots:
|
|
||||||
|
|
||||||
`abra app run <backupbot_name> app -- backup snapshots`
|
|
||||||
|
|
||||||
Show all snapshots containing a specific app:
|
|
||||||
|
|
||||||
`abra app run <backupbot_name> app -- backup --host <target_app_name> snapshots`
|
|
||||||
|
|
||||||
Show all files inside the latest snapshot (can be very verbose):
|
|
||||||
|
|
||||||
`abra app run <backupbot_name> app -- backup ls`
|
|
||||||
|
|
||||||
Show specific files inside a selected snapshot:
|
|
||||||
|
|
||||||
`abra app run <backupbot_name> app -- backup ls --snapshot <snapshot_id> /var/lib/docker/volumes/`
|
|
||||||
|
|
||||||
Download files from a snapshot:
|
|
||||||
|
|
||||||
```
|
```
|
||||||
filename=$(abra app run <backupbot_name> app -- backup download --snapshot <snapshot_id> --path <absolute_path>)
|
docker exec -it <container_id_or_name> backup create --host app
|
||||||
abra app cp <backupbot_name> app:$filename .
|
|
||||||
```
|
```
|
||||||
|
|
||||||
## Run restic
|
Restore a snapshot:
|
||||||
|
|
||||||
```
|
```
|
||||||
abra app run <backupbot_name> app bash
|
docker exec -it <container_id_or_name> backup restore --snapshot <snapshot_id> --target /path/to/restore
|
||||||
export AWS_SECRET_ACCESS_KEY=$(cat $AWS_SECRET_ACCESS_KEY_FILE)
|
|
||||||
export RESTIC_PASSWORD=$(cat $RESTIC_PASSWORD_FILE)
|
|
||||||
restic snapshots
|
|
||||||
```
|
```
|
||||||
|
|
||||||
|
List snapshots:
|
||||||
|
|
||||||
|
```
|
||||||
|
docker exec -it <container_id_or_name> backup snapshots
|
||||||
|
```
|
||||||
|
|
||||||
|
List files in a snapshot:
|
||||||
|
|
||||||
|
```
|
||||||
|
docker exec -it <container_id_or_name> backup ls --snapshot <snapshot_id> --path /path/to/directory
|
||||||
|
```
|
||||||
|
|
||||||
|
Download files, volumes, or secrets from a snapshot:
|
||||||
|
|
||||||
|
```
|
||||||
|
docker exec -it <container_id_or_name> backup download --snapshot <snapshot_id> [--path /path/to/file] [--volumes] [--secrets]
|
||||||
|
```
|
||||||
|
|
||||||
|
Note: Make sure to replace `<container_id_or_name>` and `<snapshot_id>` with the appropriate values for your setup.
|
||||||
|
|
||||||
|
Remember to review and adjust the Docker Compose file and environment variables according to your specific requirements before running the backup commands.
|
||||||
|
|
||||||
|
When using `docker exec`, you don't need to specify the volume mounts or the Restic repository location as command-line arguments because they are already defined in the Docker Compose file and are available within the running container.
|
||||||
|
|
||||||
|
If you need to access the downloaded files, volumes, or secrets from the backup, you can use `docker cp` to copy them from the container to the host machine:
|
||||||
|
|
||||||
|
```
|
||||||
|
docker cp <container_id_or_name>:/path/to/backup/file /path/on/host
|
||||||
|
```
|
||||||
|
|
||||||
|
This allows you to retrieve the backed-up data from the container.
|
||||||
|
|
||||||
## Recipe Configuration
|
## Recipe Configuration
|
||||||
|
|
||||||
Like Traefik, or `swarm-cronjob`, Backupbot II uses access to the Docker socket to read labels from running Docker Swarm services:
|
Backupbot II uses access to the Docker socket to read labels from running Docker Swarm services:
|
||||||
|
|
||||||
1. Add `ENABLE_BACKUPS=true` to .env.sample
|
|
||||||
|
|
||||||
2. Add backupbot labels to the compose file
|
|
||||||
|
|
||||||
```
|
```
|
||||||
services:
|
services:
|
||||||
db:
|
db:
|
||||||
deploy:
|
deploy:
|
||||||
labels:
|
labels:
|
||||||
backupbot.backup: "${ENABLE_BACKUPS:-true}"
|
backupbot.backup: ${BACKUP:-"true"}
|
||||||
backupbot.backup.pre-hook: "/pg_backup.sh backup"
|
backupbot.backup.pre-hook: 'mysqldump -u root -p"$(cat /run/secrets/db_root_password)" -f /volume_path/dump.db'
|
||||||
backupbot.backup.volumes.db.path: "backup.sql"
|
backupbot.backup.post-hook: "rm -rf /volume_path/dump.db"
|
||||||
backupbot.restore.post-hook: '/pg_backup.sh restore'
|
|
||||||
backupbot.backup.volumes.redis: "false"
|
|
||||||
```
|
```
|
||||||
|
|
||||||
- `backupbot.backup` -- set to `true` to back up this service (REQUIRED)
|
- `backupbot.backup` -- set to `true` to back up this service (REQUIRED)
|
||||||
- this is the only required backup label, per default it will backup all volumes
|
- `backupbot.backup.pre-hook` -- command to run before copying files (optional), save all dumps into the volumes
|
||||||
- `backupbot.backup.volumes.<volume_name>.path` -- only backup the listed relative paths from `<volume_name>`
|
- `backupbot.backup.post-hook` -- command to run after copying files (optional)
|
||||||
- `backupbot.backup.volumes.<volume_name>: false` -- exclude <volume_name> from the backup
|
|
||||||
- `backupbot.backup.pre-hook` -- command to run before copying files
|
|
||||||
- i.e. save all database dumps into the volumes
|
|
||||||
- `backupbot.backup.post-hook` -- command to run after copying files
|
|
||||||
- `backupbot.restore.pre-hook` -- command to run before restoring files
|
|
||||||
- `backupbot.restore.post-hook` -- command to run after restoring files
|
|
||||||
- i.e. read all database dumps from the volumes
|
|
||||||
|
|
||||||
3. (Optional) add backup/restore scripts to the compose file
|
|
||||||
|
|
||||||
```
|
|
||||||
services:
|
|
||||||
db:
|
|
||||||
configs:
|
|
||||||
- source: pg_backup
|
|
||||||
target: /pg_backup.sh
|
|
||||||
mode: 0555
|
|
||||||
|
|
||||||
|
|
||||||
configs:
|
|
||||||
pg_backup:
|
|
||||||
name: ${STACK_NAME}_pg_backup_${PG_BACKUP_VERSION}
|
|
||||||
file: pg_backup.sh
|
|
||||||
```
|
|
||||||
|
|
||||||
Version the config file in `abra.sh`:
|
|
||||||
|
|
||||||
```
|
|
||||||
export PG_BACKUP_VERSION=v1
|
|
||||||
```
|
|
||||||
|
|
||||||
As in the above example, you can reference Docker Secrets, e.g. for looking up database passwords, by reading the files in `/run/secrets` directly.
|
As in the above example, you can reference Docker Secrets, e.g. for looking up database passwords, by reading the files in `/run/secrets` directly.
|
||||||
|
|
||||||
[abra]: https://git.autonomic.zone/autonomic-cooperative/abra
|
|
||||||
|
|
||||||
## Backupbot Development
|
|
||||||
|
|
||||||
1. Copy modified backupbot.py into the container:
|
|
||||||
|
|
||||||
```
|
|
||||||
cp backupbot.py /tmp/backupbot.py; git stash; abra app cp <backupbot_name> /tmp/backupbot.py app:/usr/bin/backupbot.py; git checkout main; git stash pop
|
|
||||||
```
|
|
||||||
|
|
||||||
2. Testing stuff with the python interpreter inside the container:
|
|
||||||
|
|
||||||
```
|
|
||||||
abra app run <backupbot_name> app bash
|
|
||||||
cd /usr/bin/
|
|
||||||
python
|
|
||||||
from backupbot import *
|
|
||||||
```
|
|
||||||
|
|
||||||
### Versioning
|
|
||||||
|
|
||||||
- App version: changes to `backup.py` (build a new image)
|
|
||||||
- Co-op Cloud package version: changes to recipe.
|
|
||||||
|
|
||||||
For example, starting with 1.0.0+2.0.0:
|
|
||||||
"patch" change to recipe: 1.0.1+2.0.0
|
|
||||||
"patch" change to backup.py: increment both, so 1.1.0+2.0.1
|
|
||||||
because bumping the image version would result in a minor recipe release
|
|
||||||
|
|
||||||
https://git.coopcloud.tech/coop-cloud/backup-bot-two/issues/4
|
|
||||||
|
10
abra.sh
10
abra.sh
@ -1,10 +0,0 @@
|
|||||||
export BACKUPBOT_VERSION=v1
|
|
||||||
export SSH_CONFIG_VERSION=v1
|
|
||||||
|
|
||||||
run_cron () {
|
|
||||||
schedule="$(crontab -l | tr -s " " | cut -d ' ' -f-5)"
|
|
||||||
rm -f /tmp/backup.log
|
|
||||||
echo "* * * * * $(crontab -l | tr -s " " | cut -d ' ' -f6-)" | crontab -
|
|
||||||
while [ ! -f /tmp/backup.log ]; do sleep 1; done
|
|
||||||
echo "$schedule $(crontab -l | tr -s " " | cut -d ' ' -f6-)" | crontab -
|
|
||||||
}
|
|
369
backupbot.py
369
backupbot.py
@ -18,7 +18,7 @@ from shutil import copyfile, rmtree
|
|||||||
|
|
||||||
VOLUME_PATH = "/var/lib/docker/volumes/"
|
VOLUME_PATH = "/var/lib/docker/volumes/"
|
||||||
SECRET_PATH = '/secrets/'
|
SECRET_PATH = '/secrets/'
|
||||||
SERVICE = 'ALL'
|
SERVICE = None
|
||||||
|
|
||||||
logger = logging.getLogger("backupbot")
|
logger = logging.getLogger("backupbot")
|
||||||
logging.addLevelName(55, 'SUMMARY')
|
logging.addLevelName(55, 'SUMMARY')
|
||||||
@ -54,25 +54,21 @@ def cli(loglevel, service, repository, machine_logs):
|
|||||||
if not isinstance(numeric_level, int):
|
if not isinstance(numeric_level, int):
|
||||||
raise ValueError('Invalid log level: %s' % loglevel)
|
raise ValueError('Invalid log level: %s' % loglevel)
|
||||||
logger.setLevel(numeric_level)
|
logger.setLevel(numeric_level)
|
||||||
logHandler = logging.StreamHandler()
|
|
||||||
if machine_logs:
|
if machine_logs:
|
||||||
|
logHandler = logging.StreamHandler()
|
||||||
formatter = jsonlogger.JsonFormatter(
|
formatter = jsonlogger.JsonFormatter(
|
||||||
"%(levelname)s %(filename)s %(lineno)s %(process)d %(message)s", rename_fields={"levelname": "message_type"})
|
"%(levelname)s %(filename)s %(lineno)s %(process)d %(message)s", rename_fields={"levelname": "message_type"})
|
||||||
logHandler.setFormatter(formatter)
|
logHandler.setFormatter(formatter)
|
||||||
logger.addHandler(logHandler)
|
logger.addHandler(logHandler)
|
||||||
|
|
||||||
export_secrets()
|
export_secrets()
|
||||||
init_repo()
|
init_repo()
|
||||||
|
|
||||||
|
|
||||||
def init_repo():
|
def init_repo():
|
||||||
if repo:= os.environ.get('RESTIC_REPOSITORY_FILE'):
|
repo = os.environ['RESTIC_REPOSITORY']
|
||||||
# RESTIC_REPOSITORY_FILE and RESTIC_REPOSITORY are mutually exclusive
|
|
||||||
del os.environ['RESTIC_REPOSITORY']
|
|
||||||
else:
|
|
||||||
repo = os.environ['RESTIC_REPOSITORY']
|
|
||||||
restic.repository = repo
|
|
||||||
logger.debug(f"set restic repository location: {repo}")
|
logger.debug(f"set restic repository location: {repo}")
|
||||||
|
restic.repository = repo
|
||||||
restic.password_file = '/var/run/secrets/restic_password'
|
restic.password_file = '/var/run/secrets/restic_password'
|
||||||
try:
|
try:
|
||||||
restic.cat.config()
|
restic.cat.config()
|
||||||
@ -91,226 +87,54 @@ def export_secrets():
|
|||||||
with open(os.environ[env]) as file:
|
with open(os.environ[env]) as file:
|
||||||
secret = file.read()
|
secret = file.read()
|
||||||
os.environ[env.removesuffix('_FILE')] = secret
|
os.environ[env.removesuffix('_FILE')] = secret
|
||||||
# logger.debug(f"Read secret value: {secret}")
|
|
||||||
|
if env == 'RESTIC_REPOSITORY_FILE':
|
||||||
|
# RESTIC_REPOSITORY_FILE and RESTIC_REPOSITORY are mutually exclusive
|
||||||
|
logger.info("RESTIC_REPOSITORY set to RESTIC_REPOSITORY_FILE. Unsetting RESTIC_REPOSITORY_FILE.")
|
||||||
|
del os.environ['RESTIC_REPOSITORY_FILE']
|
||||||
|
|
||||||
|
|
||||||
@cli.command()
|
@cli.command()
|
||||||
@click.option('retries', '--retries', '-r', envvar='RETRIES', default=1)
|
@click.option('retries', '--retries', '-r', envvar='RETRIES', default=1)
|
||||||
def create(retries):
|
def create(retries):
|
||||||
app_settings = parse_backup_labels()
|
pre_commands, post_commands, backup_paths, apps = get_backup_cmds()
|
||||||
pre_commands, post_commands, backup_paths, apps_versions = get_backup_details(app_settings)
|
copy_secrets(apps)
|
||||||
copy_secrets(apps_versions)
|
backup_paths.append(SECRET_PATH)
|
||||||
backup_paths.append(Path(SECRET_PATH))
|
|
||||||
run_commands(pre_commands)
|
run_commands(pre_commands)
|
||||||
backup_volumes(backup_paths, apps_versions, int(retries))
|
backup_volumes(backup_paths, apps, int(retries))
|
||||||
run_commands(post_commands)
|
run_commands(post_commands)
|
||||||
|
|
||||||
|
|
||||||
@cli.command()
|
def get_backup_cmds():
|
||||||
@click.option('snapshot_id', '--snapshot', '-s', envvar='SNAPSHOT', default='latest')
|
|
||||||
@click.option('target', '--target', '-t', envvar='TARGET', default='/')
|
|
||||||
@click.option('noninteractive', '--noninteractive', envvar='NONINTERACTIVE', is_flag=True)
|
|
||||||
@click.option('volumes', '--volumes', '-v', envvar='VOLUMES', multiple=True)
|
|
||||||
@click.option('container', '--container', '-c', envvar='CONTAINER', multiple=True)
|
|
||||||
@click.option('no_commands', '--no-commands', envvar='NO_COMMANDS', is_flag=True)
|
|
||||||
def restore(snapshot_id, target, noninteractive, volumes, container, no_commands):
|
|
||||||
app_settings = parse_backup_labels('restore', container)
|
|
||||||
if SERVICE != 'ALL':
|
|
||||||
app_settings = {SERVICE: app_settings[SERVICE]}
|
|
||||||
pre_commands, post_commands, backup_paths, apps_versions = get_backup_details(app_settings, volumes)
|
|
||||||
snapshots = get_snapshots(snapshot_id)
|
|
||||||
if not snapshots:
|
|
||||||
logger.error(f"No Snapshots with ID {snapshot_id} for {apps_versions.keys()} found.")
|
|
||||||
exit(1)
|
|
||||||
snapshot = snapshots[0]
|
|
||||||
snapshot_id = snapshot['short_id']
|
|
||||||
if not noninteractive:
|
|
||||||
print(f"Snapshot to restore: \t{snapshot_id}")
|
|
||||||
restore_app_versions = app_versions_from_tags(snapshot.get('tags'))
|
|
||||||
print("Apps:")
|
|
||||||
for app, version in apps_versions.items():
|
|
||||||
restore_version = restore_app_versions.get(app)
|
|
||||||
print(f"\t{app} \t {restore_version}")
|
|
||||||
if version != restore_version:
|
|
||||||
print(f"WARNING!!! The running app is deployed with version {version}")
|
|
||||||
print("The following volume paths will be restored:")
|
|
||||||
for p in backup_paths:
|
|
||||||
print(f'\t{p}')
|
|
||||||
if not no_commands:
|
|
||||||
print("The following commands will be executed:")
|
|
||||||
for container, cmd in list(pre_commands.items()) + list(post_commands.items()):
|
|
||||||
print(f"\t{container.labels['com.docker.swarm.service.name']}:\t{cmd}")
|
|
||||||
snapshot_date = datetime.fromisoformat(snapshot['time'])
|
|
||||||
delta = datetime.now(tz=timezone.utc) - snapshot_date
|
|
||||||
print(f"This snapshot is {delta} old")
|
|
||||||
print("\nTHIS COMMAND WILL IRREVERSIBLY OVERWRITES FILES")
|
|
||||||
prompt = input("Type YES (uppercase) to continue: ")
|
|
||||||
if prompt != 'YES':
|
|
||||||
logger.error("Restore aborted")
|
|
||||||
exit(1)
|
|
||||||
print(f"Restoring Snapshot {snapshot_id} at {target}")
|
|
||||||
if not no_commands and pre_commands:
|
|
||||||
print(f"Run pre commands.")
|
|
||||||
run_commands(pre_commands)
|
|
||||||
result = restic_restore(snapshot_id=snapshot_id, include=backup_paths, target_dir=target)
|
|
||||||
if not no_commands and post_commands:
|
|
||||||
print(f"Run post commands.")
|
|
||||||
run_commands(post_commands)
|
|
||||||
logger.debug(result)
|
|
||||||
|
|
||||||
|
|
||||||
def restic_restore(snapshot_id, include=[], target_dir=None):
|
|
||||||
cmd = restic.cat.base_command() + ['restore', snapshot_id]
|
|
||||||
for path in include:
|
|
||||||
cmd.extend(['--include', path])
|
|
||||||
if target_dir:
|
|
||||||
cmd.extend(['--target', target_dir])
|
|
||||||
return restic.internal.command_executor.execute(cmd)
|
|
||||||
|
|
||||||
|
|
||||||
def get_snapshots(snapshot_id=None):
|
|
||||||
if snapshot_id and snapshot_id != 'latest':
|
|
||||||
snapshots = restic.snapshots(snapshot_id=snapshot_id)
|
|
||||||
if not SERVICE in app_versions_from_tags(snapshots[0].get('tags')):
|
|
||||||
logger.error(f'Snapshot with ID {snapshot_id} does not contain {SERVICE}')
|
|
||||||
exit(1)
|
|
||||||
else:
|
|
||||||
snapshots = restic.snapshots()
|
|
||||||
snapshots = list(filter(lambda x: SERVICE in app_versions_from_tags(x.get('tags')), snapshots))
|
|
||||||
if snapshot_id == 'latest':
|
|
||||||
return snapshots[-1:]
|
|
||||||
else:
|
|
||||||
return snapshots
|
|
||||||
|
|
||||||
|
|
||||||
def app_versions_from_tags(tags):
|
|
||||||
if tags:
|
|
||||||
app_versions = map(lambda x: x.split(':'), tags)
|
|
||||||
return {i[0]: i[1] if len(i) > 1 else None for i in app_versions}
|
|
||||||
else:
|
|
||||||
return {}
|
|
||||||
|
|
||||||
def str2bool(value: str) -> bool:
|
|
||||||
return value.lower() in ("yes", "true", "t", "1")
|
|
||||||
|
|
||||||
|
|
||||||
def parse_backup_labels(hook_type='backup', selected_container=[]):
|
|
||||||
client = docker.from_env()
|
client = docker.from_env()
|
||||||
container_by_service = {
|
container_by_service = {
|
||||||
c.labels.get('com.docker.swarm.service.name'): c for c in client.containers.list()}
|
c.labels['com.docker.swarm.service.name']: c for c in client.containers.list()}
|
||||||
services = client.services.list()
|
|
||||||
app_settings = {}
|
|
||||||
for s in services:
|
|
||||||
specs = s.attrs['Spec']
|
|
||||||
labels = specs['Labels']
|
|
||||||
stack_name = labels['com.docker.stack.namespace']
|
|
||||||
container_name = s.name.removeprefix(f"{stack_name}_")
|
|
||||||
version = labels.get(f'coop-cloud.{stack_name}.version')
|
|
||||||
settings = app_settings[stack_name] = app_settings.get(stack_name) or {}
|
|
||||||
if (backup := labels.get('backupbot.backup')) and str2bool(backup):
|
|
||||||
settings['enabled'] = True
|
|
||||||
if version:
|
|
||||||
settings['version'] = version
|
|
||||||
if selected_container and container_name not in selected_container:
|
|
||||||
logger.debug(f"Skipping {s.name} because it's not a selected container")
|
|
||||||
continue
|
|
||||||
if mounts:= specs['TaskTemplate']['ContainerSpec'].get('Mounts'):
|
|
||||||
volumes = parse_volumes(stack_name, mounts)
|
|
||||||
volumes.update(settings.get('volumes') or {})
|
|
||||||
settings['volumes'] = volumes
|
|
||||||
excluded_volumes, included_volume_paths = parse_excludes_includes(labels)
|
|
||||||
settings['excluded_volumes'] = excluded_volumes.union(settings.get('excluded_volumes') or set())
|
|
||||||
settings['included_volume_paths'] = included_volume_paths.union(settings.get('included_volume_paths') or set())
|
|
||||||
if container := container_by_service.get(s.name):
|
|
||||||
if command := labels.get(f'backupbot.{hook_type}.pre-hook'):
|
|
||||||
if not (pre_hooks:= settings.get('pre_hooks')):
|
|
||||||
pre_hooks = settings['pre_hooks'] = {}
|
|
||||||
pre_hooks[container] = command
|
|
||||||
if command := labels.get(f'backupbot.{hook_type}.post-hook'):
|
|
||||||
if not (post_hooks:= settings.get('post_hooks')):
|
|
||||||
post_hooks = settings['post_hooks'] = {}
|
|
||||||
post_hooks[container] = command
|
|
||||||
else:
|
|
||||||
logger.debug(f"Container {s.name} is not running.")
|
|
||||||
if labels.get(f'backupbot.{hook_type}.pre-hook') or labels.get(f'backupbot.{hook_type}.post-hook'):
|
|
||||||
logger.error(f"Container {s.name} contain hooks but it's not running")
|
|
||||||
return app_settings
|
|
||||||
|
|
||||||
|
|
||||||
def get_backup_details(app_settings, volumes=[]):
|
|
||||||
backup_paths = set()
|
backup_paths = set()
|
||||||
backup_apps_versions = {}
|
backup_apps = set()
|
||||||
pre_hooks= {}
|
pre_commands = {}
|
||||||
post_hooks = {}
|
post_commands = {}
|
||||||
for app, settings in app_settings.items():
|
services = client.services.list()
|
||||||
if settings.get('enabled'):
|
for s in services:
|
||||||
if SERVICE != 'ALL' and SERVICE != app:
|
labels = s.attrs['Spec']['Labels']
|
||||||
|
if (backup := labels.get('backupbot.backup')) and bool(backup):
|
||||||
|
# volumes: s.attrs['Spec']['TaskTemplate']['ContainerSpec']['Mounts'][0]['Source']
|
||||||
|
stack_name = labels['com.docker.stack.namespace']
|
||||||
|
# Remove this lines to backup only a specific service
|
||||||
|
# This will unfortenately decrease restice performance
|
||||||
|
# if SERVICE and SERVICE != stack_name:
|
||||||
|
# continue
|
||||||
|
backup_apps.add(stack_name)
|
||||||
|
backup_paths = backup_paths.union(
|
||||||
|
Path(VOLUME_PATH).glob(f"{stack_name}_*"))
|
||||||
|
if not (container := container_by_service.get(s.name)):
|
||||||
|
logger.error(
|
||||||
|
f"Container {s.name} is not running, hooks can not be executed")
|
||||||
continue
|
continue
|
||||||
backup_apps_versions[app] = settings.get('version')
|
if prehook := labels.get('backupbot.backup.pre-hook'):
|
||||||
add_backup_paths(backup_paths, settings, app, volumes)
|
pre_commands[container] = prehook
|
||||||
if hooks:= settings.get('pre_hooks'):
|
if posthook := labels.get('backupbot.backup.post-hook'):
|
||||||
pre_hooks.update(hooks)
|
post_commands[container] = posthook
|
||||||
if hooks:= settings.get('post_hooks'):
|
return pre_commands, post_commands, list(backup_paths), list(backup_apps)
|
||||||
post_hooks.update(hooks)
|
|
||||||
return pre_hooks, post_hooks, list(backup_paths), backup_apps_versions
|
|
||||||
|
|
||||||
|
|
||||||
def add_backup_paths(backup_paths, settings, app, selected_volumes):
|
|
||||||
if (volumes := settings.get('volumes')):
|
|
||||||
if includes:= settings.get('included_volume_paths'):
|
|
||||||
included_volumes = list(zip(*includes))[0]
|
|
||||||
for volume, rel_paths in includes:
|
|
||||||
if not (volume_path:= volumes.get(volume)):
|
|
||||||
logger.error(f'Can not find volume with the name {volume}')
|
|
||||||
continue
|
|
||||||
if selected_volumes and volume not in selected_volumes:
|
|
||||||
logger.debug(f'Skipping {volume}:{rel_paths} because the volume is not selected')
|
|
||||||
continue
|
|
||||||
for p in rel_paths:
|
|
||||||
absolute_path = Path(f"{volume_path}/{p}")
|
|
||||||
backup_paths.add(absolute_path)
|
|
||||||
else:
|
|
||||||
included_volumes = []
|
|
||||||
excluded_volumes = settings.get('excluded_volumes') or []
|
|
||||||
for name, path in volumes.items():
|
|
||||||
if selected_volumes and name not in selected_volumes:
|
|
||||||
logger.debug(f'Skipping volume: {name} because the volume is not selected')
|
|
||||||
continue
|
|
||||||
if name in excluded_volumes:
|
|
||||||
logger.debug(f'Skipping volume: {name} because the volume is excluded')
|
|
||||||
continue
|
|
||||||
if name in included_volumes:
|
|
||||||
logger.debug(f'Skipping volume: {name} because a path is selected')
|
|
||||||
continue
|
|
||||||
backup_paths.add(path)
|
|
||||||
else:
|
|
||||||
logger.warning(f"{app} does not contain any volumes")
|
|
||||||
|
|
||||||
|
|
||||||
def parse_volumes(stack_name, mounts):
|
|
||||||
volumes = {}
|
|
||||||
for m in mounts:
|
|
||||||
if m['Type'] != 'volume':
|
|
||||||
continue
|
|
||||||
relative_path = m['Source']
|
|
||||||
name = relative_path.removeprefix(stack_name + '_')
|
|
||||||
absolute_path = Path(f"{VOLUME_PATH}{relative_path}/_data/")
|
|
||||||
volumes[name] = absolute_path
|
|
||||||
return volumes
|
|
||||||
|
|
||||||
|
|
||||||
def parse_excludes_includes(labels):
|
|
||||||
excluded_volumes = set()
|
|
||||||
included_volume_paths = set()
|
|
||||||
for label, value in labels.items():
|
|
||||||
if label.startswith('backupbot.backup.volumes.'):
|
|
||||||
volume_name = label.removeprefix('backupbot.backup.volumes.').removesuffix('.path')
|
|
||||||
if label.endswith('path'):
|
|
||||||
relative_paths = tuple(value.split(','))
|
|
||||||
included_volume_paths.add((volume_name, relative_paths))
|
|
||||||
elif not str2bool(value):
|
|
||||||
excluded_volumes.add(volume_name)
|
|
||||||
return excluded_volumes, included_volume_paths
|
|
||||||
|
|
||||||
|
|
||||||
def copy_secrets(apps):
|
def copy_secrets(apps):
|
||||||
@ -319,14 +143,14 @@ def copy_secrets(apps):
|
|||||||
os.mkdir(SECRET_PATH)
|
os.mkdir(SECRET_PATH)
|
||||||
client = docker.from_env()
|
client = docker.from_env()
|
||||||
container_by_service = {
|
container_by_service = {
|
||||||
c.labels.get('com.docker.swarm.service.name'): c for c in client.containers.list()}
|
c.labels['com.docker.swarm.service.name']: c for c in client.containers.list()}
|
||||||
services = client.services.list()
|
services = client.services.list()
|
||||||
for s in services:
|
for s in services:
|
||||||
app_name = s.attrs['Spec']['Labels']['com.docker.stack.namespace']
|
app_name = s.attrs['Spec']['Labels']['com.docker.stack.namespace']
|
||||||
if (app_name in apps and
|
if (app_name in apps and
|
||||||
(app_secs := s.attrs['Spec']['TaskTemplate']['ContainerSpec'].get('Secrets'))):
|
(app_secs := s.attrs['Spec']['TaskTemplate']['ContainerSpec'].get('Secrets'))):
|
||||||
if not container_by_service.get(s.name):
|
if not container_by_service.get(s.name):
|
||||||
logger.warning(
|
logger.error(
|
||||||
f"Container {s.name} is not running, secrets can not be copied.")
|
f"Container {s.name} is not running, secrets can not be copied.")
|
||||||
continue
|
continue
|
||||||
container_id = container_by_service[s.name].id
|
container_id = container_by_service[s.name].id
|
||||||
@ -337,7 +161,6 @@ def copy_secrets(apps):
|
|||||||
f"For the secret {sec['SecretName']} the file {src} does not exist for {s.name}")
|
f"For the secret {sec['SecretName']} the file {src} does not exist for {s.name}")
|
||||||
continue
|
continue
|
||||||
dst = SECRET_PATH + sec['SecretName']
|
dst = SECRET_PATH + sec['SecretName']
|
||||||
logger.debug(f"Copy Secret {sec['SecretName']}")
|
|
||||||
copyfile(src, dst)
|
copyfile(src, dst)
|
||||||
|
|
||||||
|
|
||||||
@ -359,29 +182,18 @@ def run_commands(commands):
|
|||||||
logger.error(
|
logger.error(
|
||||||
f"Failed to run command {command} in {container.name}: {result.output.decode()}")
|
f"Failed to run command {command} in {container.name}: {result.output.decode()}")
|
||||||
else:
|
else:
|
||||||
logger.debug(result.output.decode())
|
logger.info(result.output.decode())
|
||||||
|
|
||||||
|
|
||||||
def backup_volumes(backup_paths, apps_versions, retries, dry_run=False):
|
def backup_volumes(backup_paths, apps, retries, dry_run=False):
|
||||||
while True:
|
while True:
|
||||||
try:
|
try:
|
||||||
logger.info("Backup these paths:")
|
result = restic.backup(backup_paths, dry_run=dry_run, tags=apps)
|
||||||
logger.info("\n".join(map(str, backup_paths)))
|
|
||||||
backup_paths = list(filter(path_exists, backup_paths))
|
|
||||||
cmd = restic.cat.base_command()
|
|
||||||
parent = get_snapshots('latest')
|
|
||||||
if parent:
|
|
||||||
# https://restic.readthedocs.io/en/stable/040_backup.html#file-change-detection
|
|
||||||
cmd.extend(['--parent', parent[0]['short_id']])
|
|
||||||
tags = [f"{app}:{version}" for app,version in apps_versions.items()]
|
|
||||||
if SERVICE == 'ALL':
|
|
||||||
tags.append(SERVICE)
|
|
||||||
logger.info("Start volume backup")
|
|
||||||
result = restic.internal.backup.run(cmd, backup_paths, dry_run=dry_run, tags=tags)
|
|
||||||
logger.summary("backup finished", extra=result)
|
logger.summary("backup finished", extra=result)
|
||||||
return
|
return
|
||||||
except ResticFailedError as error:
|
except ResticFailedError as error:
|
||||||
logger.error(f"Backup failed for {SERVICE}.")
|
logger.error(
|
||||||
|
f"Backup failed for {apps}. Could not Backup these paths: {backup_paths}")
|
||||||
logger.error(error, exc_info=True)
|
logger.error(error, exc_info=True)
|
||||||
if retries > 0:
|
if retries > 0:
|
||||||
retries -= 1
|
retries -= 1
|
||||||
@ -389,25 +201,49 @@ def backup_volumes(backup_paths, apps_versions, retries, dry_run=False):
|
|||||||
exit(1)
|
exit(1)
|
||||||
|
|
||||||
|
|
||||||
def path_exists(path):
|
@cli.command()
|
||||||
if not path.exists():
|
@click.option('snapshot', '--snapshot', '-s', envvar='SNAPSHOT', default='latest')
|
||||||
logger.error(f'{path} does not exist')
|
@click.option('target', '--target', '-t', envvar='TARGET', default='/')
|
||||||
return path.exists()
|
@click.option('noninteractive', '--noninteractive', envvar='NONINTERACTIVE', is_flag=True)
|
||||||
|
def restore(snapshot, target, noninteractive):
|
||||||
|
# Todo: recommend to shutdown the container
|
||||||
|
service_paths = VOLUME_PATH
|
||||||
|
if SERVICE:
|
||||||
|
service_paths = service_paths + f'{SERVICE}_*'
|
||||||
|
snapshots = restic.snapshots(snapshot_id=snapshot)
|
||||||
|
if not snapshot:
|
||||||
|
logger.error("No Snapshots with ID {snapshots}")
|
||||||
|
exit(1)
|
||||||
|
if not noninteractive:
|
||||||
|
snapshot_date = datetime.fromisoformat(snapshots[0]['time'])
|
||||||
|
delta = datetime.now(tz=timezone.utc) - snapshot_date
|
||||||
|
print(
|
||||||
|
f"You are going to restore Snapshot {snapshot} of {service_paths} at {target}")
|
||||||
|
print(f"This snapshot is {delta} old")
|
||||||
|
print(
|
||||||
|
f"THIS COMMAND WILL IRREVERSIBLY OVERWRITES {target}{service_paths.removeprefix('/')}")
|
||||||
|
prompt = input("Type YES (uppercase) to continue: ")
|
||||||
|
if prompt != 'YES':
|
||||||
|
logger.error("Restore aborted")
|
||||||
|
exit(1)
|
||||||
|
print(f"Restoring Snapshot {snapshot} of {service_paths} at {target}")
|
||||||
|
# TODO: use tags if no snapshot is selected, to use a snapshot including SERVICE
|
||||||
|
result = restic.restore(snapshot_id=snapshot,
|
||||||
|
include=service_paths, target_dir=target)
|
||||||
|
logger.debug(result)
|
||||||
|
|
||||||
|
|
||||||
@cli.command()
|
@cli.command()
|
||||||
def snapshots():
|
def snapshots():
|
||||||
snapshots = get_snapshots()
|
snapshots = restic.snapshots()
|
||||||
|
no_snapshots = True
|
||||||
for snap in snapshots:
|
for snap in snapshots:
|
||||||
output = [snap['time'], snap['id']]
|
if not SERVICE or (tags := snap.get('tags')) and SERVICE in tags:
|
||||||
if tags:= snap.get('tags'):
|
print(snap['time'], snap['id'])
|
||||||
app_versions = app_versions_from_tags(tags)
|
no_snapshots = False
|
||||||
if version:= app_versions.get(SERVICE):
|
if no_snapshots:
|
||||||
output.append(version)
|
|
||||||
print(*output)
|
|
||||||
if not snapshots:
|
|
||||||
err_msg = "No Snapshots found"
|
err_msg = "No Snapshots found"
|
||||||
if SERVICE != 'ALL':
|
if SERVICE:
|
||||||
service_name = SERVICE.replace('_', '.')
|
service_name = SERVICE.replace('_', '.')
|
||||||
err_msg += f' for app {service_name}'
|
err_msg += f' for app {service_name}'
|
||||||
logger.warning(err_msg)
|
logger.warning(err_msg)
|
||||||
@ -415,31 +251,18 @@ def snapshots():
|
|||||||
|
|
||||||
@cli.command()
|
@cli.command()
|
||||||
@click.option('snapshot', '--snapshot', '-s', envvar='SNAPSHOT', default='latest')
|
@click.option('snapshot', '--snapshot', '-s', envvar='SNAPSHOT', default='latest')
|
||||||
@click.option('show_all', '--all', '-a', envvar='SHOW_ALL', is_flag=True)
|
@click.option('path', '--path', '-p', envvar='INCLUDE_PATH')
|
||||||
@click.option('timestamps', '--timestamps', '-t', envvar='TIMESTAMPS', is_flag=True)
|
def ls(snapshot, path):
|
||||||
@click.argument('path', required=False, default='/var/lib/docker/volumes/')
|
|
||||||
def ls(snapshot, show_all, timestamps, path):
|
|
||||||
if snapshot == 'latest':
|
|
||||||
latest_snapshot = get_snapshots('latest')
|
|
||||||
if not latest_snapshot:
|
|
||||||
logger.error(f"There is no latest snapshot for {SERVICE}")
|
|
||||||
exit(1)
|
|
||||||
snapshot = latest_snapshot[0]['short_id']
|
|
||||||
if os.environ.get('INCLUDE_PATH'):
|
|
||||||
path = os.environ.get('INCLUDE_PATH')
|
|
||||||
if show_all:
|
|
||||||
path = None
|
|
||||||
results = list_files(snapshot, path)
|
results = list_files(snapshot, path)
|
||||||
for r in results:
|
for r in results:
|
||||||
if r.get('path'):
|
if r.get('path'):
|
||||||
if timestamps:
|
print(f"{r['ctime']}\t{r['path']}")
|
||||||
print(f"{r['ctime']}\t{r['path']}")
|
|
||||||
else:
|
|
||||||
print(f"{r['path']}")
|
|
||||||
|
|
||||||
|
|
||||||
def list_files(snapshot, path):
|
def list_files(snapshot, path):
|
||||||
cmd = restic.cat.base_command() + ['ls']
|
cmd = restic.cat.base_command() + ['ls']
|
||||||
|
if SERVICE:
|
||||||
|
cmd = cmd + ['--tag', SERVICE]
|
||||||
cmd.append(snapshot)
|
cmd.append(snapshot)
|
||||||
if path:
|
if path:
|
||||||
cmd.append(path)
|
cmd.append(path)
|
||||||
@ -448,7 +271,7 @@ def list_files(snapshot, path):
|
|||||||
except ResticFailedError as error:
|
except ResticFailedError as error:
|
||||||
if 'no snapshot found' in str(error):
|
if 'no snapshot found' in str(error):
|
||||||
err_msg = f'There is no snapshot "{snapshot}"'
|
err_msg = f'There is no snapshot "{snapshot}"'
|
||||||
if SERVICE != 'ALL':
|
if SERVICE:
|
||||||
err_msg += f' for the app "{SERVICE}"'
|
err_msg += f' for the app "{SERVICE}"'
|
||||||
logger.error(err_msg)
|
logger.error(err_msg)
|
||||||
exit(1)
|
exit(1)
|
||||||
@ -466,12 +289,6 @@ def list_files(snapshot, path):
|
|||||||
@click.option('secrets', '--secrets', '-c', is_flag=True, envvar='SECRETS')
|
@click.option('secrets', '--secrets', '-c', is_flag=True, envvar='SECRETS')
|
||||||
def download(snapshot, path, volumes, secrets):
|
def download(snapshot, path, volumes, secrets):
|
||||||
file_dumps = []
|
file_dumps = []
|
||||||
if snapshot == 'latest':
|
|
||||||
latest_snapshot = get_snapshots('latest')
|
|
||||||
if not latest_snapshot:
|
|
||||||
logger.error(f"There is no latest snapshot for {SERVICE}")
|
|
||||||
exit(1)
|
|
||||||
snapshot = latest_snapshot[0]['short_id']
|
|
||||||
if not any([path, volumes, secrets]):
|
if not any([path, volumes, secrets]):
|
||||||
volumes = secrets = True
|
volumes = secrets = True
|
||||||
if path:
|
if path:
|
||||||
@ -486,7 +303,7 @@ def download(snapshot, path, volumes, secrets):
|
|||||||
tarinfo.size = len(binary_output)
|
tarinfo.size = len(binary_output)
|
||||||
file_dumps.append((binary_output, tarinfo))
|
file_dumps.append((binary_output, tarinfo))
|
||||||
if volumes:
|
if volumes:
|
||||||
if SERVICE == 'ALL':
|
if not SERVICE:
|
||||||
logger.error("Please specify '--host' when using '--volumes'")
|
logger.error("Please specify '--host' when using '--volumes'")
|
||||||
exit(1)
|
exit(1)
|
||||||
files = list_files(snapshot, VOLUME_PATH)
|
files = list_files(snapshot, VOLUME_PATH)
|
||||||
@ -499,7 +316,7 @@ def download(snapshot, path, volumes, secrets):
|
|||||||
tarinfo.size = len(binary_output)
|
tarinfo.size = len(binary_output)
|
||||||
file_dumps.append((binary_output, tarinfo))
|
file_dumps.append((binary_output, tarinfo))
|
||||||
if secrets:
|
if secrets:
|
||||||
if SERVICE == 'ALL':
|
if not SERVICE:
|
||||||
logger.error("Please specify '--host' when using '--secrets'")
|
logger.error("Please specify '--host' when using '--secrets'")
|
||||||
exit(1)
|
exit(1)
|
||||||
filename = f"{SERVICE}.json"
|
filename = f"{SERVICE}.json"
|
||||||
@ -536,6 +353,8 @@ def get_formatted_size(file_path):
|
|||||||
|
|
||||||
def dump(snapshot, path):
|
def dump(snapshot, path):
|
||||||
cmd = restic.cat.base_command() + ['dump']
|
cmd = restic.cat.base_command() + ['dump']
|
||||||
|
if SERVICE:
|
||||||
|
cmd = cmd + ['--tag', SERVICE]
|
||||||
cmd = cmd + [snapshot, path]
|
cmd = cmd + [snapshot, path]
|
||||||
print(f"Dumping {path} from snapshot '{snapshot}'")
|
print(f"Dumping {path} from snapshot '{snapshot}'")
|
||||||
output = subprocess.run(cmd, capture_output=True)
|
output = subprocess.run(cmd, capture_output=True)
|
||||||
|
@ -1,14 +0,0 @@
|
|||||||
---
|
|
||||||
version: "3.8"
|
|
||||||
services:
|
|
||||||
app:
|
|
||||||
environment:
|
|
||||||
- AWS_ACCESS_KEY_ID
|
|
||||||
- AWS_SECRET_ACCESS_KEY_FILE=/run/secrets/aws_secret_access_key
|
|
||||||
secrets:
|
|
||||||
- aws_secret_access_key
|
|
||||||
|
|
||||||
secrets:
|
|
||||||
aws_secret_access_key:
|
|
||||||
external: true
|
|
||||||
name: ${STACK_NAME}_aws_secret_access_key_${SECRET_AWS_SECRET_ACCESS_KEY_VERSION}
|
|
@ -1,13 +0,0 @@
|
|||||||
---
|
|
||||||
version: "3.8"
|
|
||||||
services:
|
|
||||||
app:
|
|
||||||
environment:
|
|
||||||
- RESTIC_REPOSITORY_FILE=/run/secrets/restic_repo
|
|
||||||
secrets:
|
|
||||||
- restic_repo
|
|
||||||
|
|
||||||
secrets:
|
|
||||||
restic_repo:
|
|
||||||
external: true
|
|
||||||
name: ${STACK_NAME}_restic_repo_${SECRET_RESTIC_REPO_VERSION}
|
|
@ -1,23 +0,0 @@
|
|||||||
---
|
|
||||||
version: "3.8"
|
|
||||||
services:
|
|
||||||
app:
|
|
||||||
environment:
|
|
||||||
- SSH_KEY_FILE=/run/secrets/ssh_key
|
|
||||||
- SSH_HOST_KEY
|
|
||||||
secrets:
|
|
||||||
- source: ssh_key
|
|
||||||
mode: 0400
|
|
||||||
configs:
|
|
||||||
- source: ssh_config
|
|
||||||
target: /root/.ssh/config
|
|
||||||
|
|
||||||
secrets:
|
|
||||||
ssh_key:
|
|
||||||
external: true
|
|
||||||
name: ${STACK_NAME}_ssh_key_${SECRET_SSH_KEY_VERSION}
|
|
||||||
|
|
||||||
configs:
|
|
||||||
ssh_config:
|
|
||||||
name: ${STACK_NAME}_ssh_config_${SSH_CONFIG_VERSION}
|
|
||||||
file: ssh_config
|
|
@ -1,15 +0,0 @@
|
|||||||
---
|
|
||||||
version: "3.8"
|
|
||||||
services:
|
|
||||||
app:
|
|
||||||
deploy:
|
|
||||||
mode: replicated
|
|
||||||
replicas: 0
|
|
||||||
labels:
|
|
||||||
- "swarm.cronjob.enable=true"
|
|
||||||
# Note(3wc): every 5m, testing
|
|
||||||
- "swarm.cronjob.schedule=*/5 * * * *"
|
|
||||||
# Note(3wc): blank label to be picked up by `abra recipe sync`
|
|
||||||
restart_policy:
|
|
||||||
condition: none
|
|
||||||
entrypoint: [ "/usr/bin/backup.sh" ]
|
|
44
compose.yaml
Normal file
44
compose.yaml
Normal file
@ -0,0 +1,44 @@
|
|||||||
|
services:
|
||||||
|
app:
|
||||||
|
image: docker:24.0.7-dind
|
||||||
|
volumes:
|
||||||
|
- "/var/run/docker.sock:/var/run/docker.sock"
|
||||||
|
- "/var/lib/docker/volumes/:/var/lib/docker/volumes/"
|
||||||
|
- "/var/lib/docker/containers/:/var/lib/docker/containers/:ro"
|
||||||
|
environment:
|
||||||
|
- CRON_SCHEDULE
|
||||||
|
- RESTIC_REPOSITORY_FILE=/run/secrets/restic_repo
|
||||||
|
- RESTIC_PASSWORD_FILE=/run/secrets/restic_password
|
||||||
|
secrets:
|
||||||
|
- restic_repo
|
||||||
|
- restic_password
|
||||||
|
configs:
|
||||||
|
- source: entrypoint
|
||||||
|
target: /entrypoint.sh
|
||||||
|
mode: 0555
|
||||||
|
- source: backupbot
|
||||||
|
target: /usr/bin/backup
|
||||||
|
mode: 0555
|
||||||
|
entrypoint: ['/entrypoint.sh']
|
||||||
|
healthcheck:
|
||||||
|
test: "pgrep crond"
|
||||||
|
interval: 30s
|
||||||
|
timeout: 10s
|
||||||
|
retries: 10
|
||||||
|
start_period: 5m
|
||||||
|
|
||||||
|
secrets:
|
||||||
|
restic_repo:
|
||||||
|
external: true
|
||||||
|
name: ${STACK_NAME}_restic_repo
|
||||||
|
restic_password:
|
||||||
|
external: true
|
||||||
|
name: ${STACK_NAME}_restic_password
|
||||||
|
|
||||||
|
configs:
|
||||||
|
entrypoint:
|
||||||
|
name: ${STACK_NAME}_entrypoint
|
||||||
|
file: entrypoint.sh
|
||||||
|
backupbot:
|
||||||
|
name: ${STACK_NAME}_backupbot
|
||||||
|
file: backupbot.py
|
36
compose.yml
36
compose.yml
@ -1,36 +0,0 @@
|
|||||||
---
|
|
||||||
version: "3.8"
|
|
||||||
services:
|
|
||||||
app:
|
|
||||||
image: git.coopcloud.tech/coop-cloud/backup-bot-two:2.3.0-beta
|
|
||||||
volumes:
|
|
||||||
- "/var/run/docker.sock:/var/run/docker.sock"
|
|
||||||
- "/var/lib/docker/volumes/:/var/lib/docker/volumes/"
|
|
||||||
- "/var/lib/docker/containers/:/var/lib/docker/containers/:ro"
|
|
||||||
- backups:/backups
|
|
||||||
environment:
|
|
||||||
- CRON_SCHEDULE
|
|
||||||
- RESTIC_REPOSITORY
|
|
||||||
- RESTIC_PASSWORD_FILE=/run/secrets/restic_password
|
|
||||||
secrets:
|
|
||||||
- restic_password
|
|
||||||
deploy:
|
|
||||||
labels:
|
|
||||||
- coop-cloud.${STACK_NAME}.version=2.3.0+2.3.0-beta
|
|
||||||
- coop-cloud.${STACK_NAME}.timeout=${TIMEOUT:-300}
|
|
||||||
- coop-cloud.backupbot.enabled=true
|
|
||||||
#entrypoint: ['tail', '-f','/dev/null']
|
|
||||||
healthcheck:
|
|
||||||
test: "pgrep crond"
|
|
||||||
interval: 30s
|
|
||||||
timeout: 10s
|
|
||||||
retries: 10
|
|
||||||
start_period: 5m
|
|
||||||
|
|
||||||
secrets:
|
|
||||||
restic_password:
|
|
||||||
external: true
|
|
||||||
name: ${STACK_NAME}_restic_password_${SECRET_RESTIC_PASSWORD_VERSION}
|
|
||||||
|
|
||||||
volumes:
|
|
||||||
backups:
|
|
10
entrypoint.sh
Executable file → Normal file
10
entrypoint.sh
Executable file → Normal file
@ -1,11 +1,11 @@
|
|||||||
#!/bin/sh
|
#!/bin/sh
|
||||||
|
|
||||||
set -e
|
set -e -o pipefail
|
||||||
|
|
||||||
if [ -n "$SSH_HOST_KEY" ]
|
apk add --upgrade --no-cache restic bash python3 py3-pip py3-click py3-docker-py py3-json-logger curl
|
||||||
then
|
|
||||||
echo "$SSH_HOST_KEY" > /root/.ssh/known_hosts
|
# Todo use requirements file with specific versions
|
||||||
fi
|
pip install --break-system-packages resticpy==1.0.2
|
||||||
|
|
||||||
cron_schedule="${CRON_SCHEDULE:?CRON_SCHEDULE not set}"
|
cron_schedule="${CRON_SCHEDULE:?CRON_SCHEDULE not set}"
|
||||||
|
|
||||||
|
34
pg_backup.sh
34
pg_backup.sh
@ -1,34 +0,0 @@
|
|||||||
#!/bin/bash
|
|
||||||
|
|
||||||
set -e
|
|
||||||
|
|
||||||
BACKUP_FILE='/var/lib/postgresql/data/backup.sql'
|
|
||||||
|
|
||||||
function backup {
|
|
||||||
export PGPASSWORD=$(cat $POSTGRES_PASSWORD_FILE)
|
|
||||||
pg_dump -U ${POSTGRES_USER} ${POSTGRES_DB} > $BACKUP_FILE
|
|
||||||
}
|
|
||||||
|
|
||||||
function restore {
|
|
||||||
cd /var/lib/postgresql/data/
|
|
||||||
restore_config(){
|
|
||||||
# Restore allowed connections
|
|
||||||
cat pg_hba.conf.bak > pg_hba.conf
|
|
||||||
su postgres -c 'pg_ctl reload'
|
|
||||||
}
|
|
||||||
# Don't allow any other connections than local
|
|
||||||
cp pg_hba.conf pg_hba.conf.bak
|
|
||||||
echo "local all all trust" > pg_hba.conf
|
|
||||||
su postgres -c 'pg_ctl reload'
|
|
||||||
trap restore_config EXIT INT TERM
|
|
||||||
|
|
||||||
# Recreate Database
|
|
||||||
psql -U ${POSTGRES_USER} -d postgres -c "DROP DATABASE ${POSTGRES_DB} WITH (FORCE);"
|
|
||||||
createdb -U ${POSTGRES_USER} ${POSTGRES_DB}
|
|
||||||
psql -U ${POSTGRES_USER} -d ${POSTGRES_DB} -1 -f $BACKUP_FILE
|
|
||||||
|
|
||||||
trap - EXIT INT TERM
|
|
||||||
restore_config
|
|
||||||
}
|
|
||||||
|
|
||||||
$@
|
|
@ -1 +0,0 @@
|
|||||||
This is the first beta release of the new backup-bot-two rewrite in python. Be aware when updating, it can break. Please read the readme and update your config according to it.
|
|
@ -1,3 +0,0 @@
|
|||||||
Breaking Change: the variables `SERVER_NAME` and `RESTIC_HOST` are merged into `RESTIC_REPOSITORY`. The format can be looked up here: https://restic.readthedocs.io/en/stable/030_preparing_a_new_repo.html
|
|
||||||
ssh/sftp: `sftp:user@host:/repo-path`
|
|
||||||
S3: `s3:https://s3.example.com/bucket_name`
|
|
@ -1,3 +0,0 @@
|
|||||||
{
|
|
||||||
"$schema": "https://docs.renovatebot.com/renovate-schema.json"
|
|
||||||
}
|
|
@ -1,4 +0,0 @@
|
|||||||
Host *
|
|
||||||
IdentityFile /run/secrets/ssh_key
|
|
||||||
ServerAliveInterval 60
|
|
||||||
ServerAliveCountMax 240
|
|
Reference in New Issue
Block a user