Compare commits
1 Commits
2.0.1+2.1.
...
prom-mon
Author | SHA1 | Date | |
---|---|---|---|
c1902b2dbc |
@ -9,9 +9,9 @@ steps:
|
|||||||
password:
|
password:
|
||||||
from_secret: git_coopcloud_tech_token_3wc
|
from_secret: git_coopcloud_tech_token_3wc
|
||||||
repo: git.coopcloud.tech/coop-cloud/backup-bot-two
|
repo: git.coopcloud.tech/coop-cloud/backup-bot-two
|
||||||
tags: ${DRONE_SEMVER_BUILD}
|
tags: 2.0.0
|
||||||
registry: git.coopcloud.tech
|
registry: git.coopcloud.tech
|
||||||
when:
|
when:
|
||||||
event:
|
event:
|
||||||
include:
|
exclude:
|
||||||
- tag
|
- pull_request
|
||||||
|
@ -9,10 +9,16 @@ RESTIC_REPOSITORY=/backups/restic
|
|||||||
CRON_SCHEDULE='30 3 * * *'
|
CRON_SCHEDULE='30 3 * * *'
|
||||||
|
|
||||||
# Push Notifiactions
|
# Push Notifiactions
|
||||||
|
#PUSH_PROMETHEUS_URL=https://pushgateway.example.com/metrics/job/backup
|
||||||
|
# or
|
||||||
#PUSH_URL_START=https://status.example.com/api/push/xxxxxxxxxx?status=up&msg=start
|
#PUSH_URL_START=https://status.example.com/api/push/xxxxxxxxxx?status=up&msg=start
|
||||||
#PUSH_URL_SUCCESS=https://status.example.com/api/push/xxxxxxxxxx?status=up&msg=OK
|
#PUSH_URL_SUCCESS=https://status.example.com/api/push/xxxxxxxxxx?status=up&msg=OK
|
||||||
#PUSH_URL_FAIL=https://status.example.com/api/push/xxxxxxxxxx?status=down&msg=fail
|
#PUSH_URL_FAIL=https://status.example.com/api/push/xxxxxxxxxx?status=down&msg=fail
|
||||||
|
|
||||||
|
# Push Basic Auth
|
||||||
|
#COMPOSE_FILE="$COMPOSE_FILE:compose.pushbasicauth.yml"
|
||||||
|
#SECRET_PUSH_BASICAUTH=v1
|
||||||
|
|
||||||
# swarm-cronjob, instead of built-in cron
|
# swarm-cronjob, instead of built-in cron
|
||||||
#COMPOSE_FILE="$COMPOSE_FILE:compose.swarm-cronjob.yml"
|
#COMPOSE_FILE="$COMPOSE_FILE:compose.swarm-cronjob.yml"
|
||||||
|
|
||||||
|
27
README.md
27
README.md
@ -95,7 +95,7 @@ COMPOSE_FILE="$COMPOSE_FILE:compose.secret.yml"
|
|||||||
```
|
```
|
||||||
Add your REST server url as secret:
|
Add your REST server url as secret:
|
||||||
```
|
```
|
||||||
abra app secret insert <backupbot_name> restic_repo v1 "rest:https://user:pass@host:8000/"
|
`abra app secret insert <backupbot_name> restic_repo v1 "rest:https://user:pass@host:8000/"`
|
||||||
```
|
```
|
||||||
The secret will overwrite the `RESTIC_REPOSITORY` variable.
|
The secret will overwrite the `RESTIC_REPOSITORY` variable.
|
||||||
|
|
||||||
@ -104,15 +104,38 @@ See [restic REST docs](https://restic.readthedocs.io/en/latest/030_preparing_a_n
|
|||||||
|
|
||||||
## Push notifications
|
## Push notifications
|
||||||
|
|
||||||
|
It is possible to configure three push events, that may trigger on the backup cronjob. Those can be used to detect failures from mointoring systems.
|
||||||
|
The events are:
|
||||||
|
- start
|
||||||
|
- success
|
||||||
|
- fail
|
||||||
|
|
||||||
|
### Using a Prometheus Push Gateway
|
||||||
|
|
||||||
|
[A prometheus push gateway](https://git.coopcloud.tech/coop-cloud/monitoring-ng#setup-push-gateway) can be used by setting the following env variables:
|
||||||
|
- `PUSH_PROMETHEUS_URL=pushgateway.example.com/metrics/job/backup`
|
||||||
|
|
||||||
|
### Using custom URLs
|
||||||
|
|
||||||
The following env variables can be used to setup push notifications for backups. `PUSH_URL_START` is requested just before the backups starts, `PUSH_URL_SUCCESS` is only requested if the backup was successful and if the backup fails `PUSH_URL_FAIL` will be requested.
|
The following env variables can be used to setup push notifications for backups. `PUSH_URL_START` is requested just before the backups starts, `PUSH_URL_SUCCESS` is only requested if the backup was successful and if the backup fails `PUSH_URL_FAIL` will be requested.
|
||||||
Each variable is optional and independent of the other.
|
Each variable is optional and independent of the other.
|
||||||
```
|
|
||||||
|
|
||||||
|
```
|
||||||
PUSH_URL_START=https://status.example.com/api/push/xxxxxxxxxx?status=up&msg=start
|
PUSH_URL_START=https://status.example.com/api/push/xxxxxxxxxx?status=up&msg=start
|
||||||
PUSH_URL_SUCCESS=https://status.example.com/api/push/xxxxxxxxxx?status=up&msg=OK
|
PUSH_URL_SUCCESS=https://status.example.com/api/push/xxxxxxxxxx?status=up&msg=OK
|
||||||
PUSH_URL_FAIL=https://status.example.com/api/push/xxxxxxxxxx?status=down&msg=fail
|
PUSH_URL_FAIL=https://status.example.com/api/push/xxxxxxxxxx?status=down&msg=fail
|
||||||
```
|
```
|
||||||
|
|
||||||
|
### Push endpoint behind basic auth
|
||||||
|
|
||||||
|
Insert the basic auth secret
|
||||||
|
`abra app secret insert <backupbot_name> push_basicauth v1 "user:password"`
|
||||||
|
|
||||||
|
Enable basic auth in the env file, by uncommenting the following line:
|
||||||
|
```
|
||||||
|
#COMPOSE_FILE="$COMPOSE_FILE:compose.pushbasicauth.yml"
|
||||||
|
#SECRET_PUSH_BASICAUTH=v1
|
||||||
|
```
|
||||||
|
|
||||||
## Usage
|
## Usage
|
||||||
|
|
||||||
|
2
abra.sh
2
abra.sh
@ -1,5 +1,7 @@
|
|||||||
export BACKUPBOT_VERSION=v1
|
export BACKUPBOT_VERSION=v1
|
||||||
export SSH_CONFIG_VERSION=v1
|
export SSH_CONFIG_VERSION=v1
|
||||||
|
export ENTRYPOINT_VERSION=v17
|
||||||
|
export CRONJOB_VERSION=v2
|
||||||
|
|
||||||
run_cron () {
|
run_cron () {
|
||||||
schedule="$(crontab -l | tr -s " " | cut -d ' ' -f-5)"
|
schedule="$(crontab -l | tr -s " " | cut -d ' ' -f-5)"
|
||||||
|
283
backupbot.py
283
backupbot.py
@ -18,7 +18,7 @@ from shutil import copyfile, rmtree
|
|||||||
|
|
||||||
VOLUME_PATH = "/var/lib/docker/volumes/"
|
VOLUME_PATH = "/var/lib/docker/volumes/"
|
||||||
SECRET_PATH = '/secrets/'
|
SECRET_PATH = '/secrets/'
|
||||||
SERVICE = 'ALL'
|
SERVICE = None
|
||||||
|
|
||||||
logger = logging.getLogger("backupbot")
|
logger = logging.getLogger("backupbot")
|
||||||
logging.addLevelName(55, 'SUMMARY')
|
logging.addLevelName(55, 'SUMMARY')
|
||||||
@ -97,197 +97,44 @@ def export_secrets():
|
|||||||
@cli.command()
|
@cli.command()
|
||||||
@click.option('retries', '--retries', '-r', envvar='RETRIES', default=1)
|
@click.option('retries', '--retries', '-r', envvar='RETRIES', default=1)
|
||||||
def create(retries):
|
def create(retries):
|
||||||
app_settings = parse_backup_labels()
|
pre_commands, post_commands, backup_paths, apps = get_backup_cmds()
|
||||||
pre_commands, post_commands, backup_paths, apps = get_backup_details(app_settings)
|
|
||||||
copy_secrets(apps)
|
copy_secrets(apps)
|
||||||
backup_paths.append(Path(SECRET_PATH))
|
backup_paths.append(SECRET_PATH)
|
||||||
run_commands(pre_commands)
|
run_commands(pre_commands)
|
||||||
backup_volumes(backup_paths, apps, int(retries))
|
backup_volumes(backup_paths, apps, int(retries))
|
||||||
run_commands(post_commands)
|
run_commands(post_commands)
|
||||||
|
|
||||||
|
|
||||||
@cli.command()
|
def get_backup_cmds():
|
||||||
@click.option('snapshot', '--snapshot', '-s', envvar='SNAPSHOT', default='latest')
|
|
||||||
@click.option('target', '--target', '-t', envvar='TARGET', default='/')
|
|
||||||
@click.option('noninteractive', '--noninteractive', envvar='NONINTERACTIVE', is_flag=True)
|
|
||||||
@click.option('volumes', '--volumes', '-v', envvar='VOLUMES', multiple=True)
|
|
||||||
@click.option('container', '--container', '-c', envvar='CONTAINER', multiple=True)
|
|
||||||
@click.option('no_commands', '--no-commands', envvar='NO_COMMANDS', is_flag=True)
|
|
||||||
def restore(snapshot, target, noninteractive, volumes, container, no_commands):
|
|
||||||
app_settings = parse_backup_labels('restore', container)
|
|
||||||
if SERVICE != 'ALL':
|
|
||||||
app_settings = {SERVICE: app_settings[SERVICE]}
|
|
||||||
pre_commands, post_commands, backup_paths, apps = get_backup_details(app_settings, volumes)
|
|
||||||
snapshots = get_snapshots(snapshot_id=snapshot)
|
|
||||||
if not snapshot:
|
|
||||||
logger.error("No Snapshots with ID {snapshots} for {apps} found.")
|
|
||||||
exit(1)
|
|
||||||
if not noninteractive:
|
|
||||||
snapshot_date = datetime.fromisoformat(snapshots[0]['time'])
|
|
||||||
delta = datetime.now(tz=timezone.utc) - snapshot_date
|
|
||||||
print(f"You are going to restore Snapshot {snapshot} of {apps} at {target}")
|
|
||||||
print("The following volume paths will be restored:")
|
|
||||||
for p in backup_paths:
|
|
||||||
print(f'\t{p}')
|
|
||||||
if not no_commands:
|
|
||||||
print("The following commands will be executed:")
|
|
||||||
for container, cmd in list(pre_commands.items()) + list(post_commands.items()):
|
|
||||||
print(f"\t{container.labels['com.docker.swarm.service.name']}:\t{cmd}")
|
|
||||||
print(f"This snapshot is {delta} old")
|
|
||||||
print("\nTHIS COMMAND WILL IRREVERSIBLY OVERWRITES FILES")
|
|
||||||
prompt = input("Type YES (uppercase) to continue: ")
|
|
||||||
if prompt != 'YES':
|
|
||||||
logger.error("Restore aborted")
|
|
||||||
exit(1)
|
|
||||||
print(f"Restoring Snapshot {snapshot} at {target}")
|
|
||||||
if not no_commands and pre_commands:
|
|
||||||
print(f"Run pre commands.")
|
|
||||||
run_commands(pre_commands)
|
|
||||||
result = restic_restore(snapshot_id=snapshot, include=backup_paths, target_dir=target)
|
|
||||||
if not no_commands and post_commands:
|
|
||||||
print(f"Run post commands.")
|
|
||||||
run_commands(post_commands)
|
|
||||||
logger.debug(result)
|
|
||||||
|
|
||||||
|
|
||||||
def restic_restore(snapshot_id='latest', include=[], target_dir=None):
|
|
||||||
cmd = restic.cat.base_command() + ['restore', snapshot_id]
|
|
||||||
for path in include:
|
|
||||||
cmd.extend(['--include', path])
|
|
||||||
if target_dir:
|
|
||||||
cmd.extend(['--target', target_dir])
|
|
||||||
return restic.internal.command_executor.execute(cmd)
|
|
||||||
|
|
||||||
|
|
||||||
def get_snapshots(snapshot_id=None):
|
|
||||||
if snapshot_id and snapshot_id != 'latest':
|
|
||||||
snapshots = restic.snapshots(snapshot_id=snapshot_id)
|
|
||||||
if SERVICE not in snapshots[0]['tags']:
|
|
||||||
logger.error(f'Snapshot with ID {snapshot_id} does not contain {SERVICE}')
|
|
||||||
exit(1)
|
|
||||||
else:
|
|
||||||
snapshots = restic.snapshots()
|
|
||||||
snapshots = list(filter(lambda x: x.get('tags') and SERVICE in x.get('tags'), snapshots))
|
|
||||||
if snapshot_id == 'latest':
|
|
||||||
return snapshots[-1:]
|
|
||||||
else:
|
|
||||||
return snapshots
|
|
||||||
|
|
||||||
|
|
||||||
def parse_backup_labels(hook_type='backup', selected_container=[]):
|
|
||||||
client = docker.from_env()
|
client = docker.from_env()
|
||||||
container_by_service = {
|
container_by_service = {
|
||||||
c.labels.get('com.docker.swarm.service.name'): c for c in client.containers.list()}
|
c.labels.get('com.docker.swarm.service.name'): c for c in client.containers.list()}
|
||||||
services = client.services.list()
|
|
||||||
app_settings = {}
|
|
||||||
for s in services:
|
|
||||||
specs = s.attrs['Spec']
|
|
||||||
labels = specs['Labels']
|
|
||||||
stack_name = labels['com.docker.stack.namespace']
|
|
||||||
container_name = s.name.removeprefix(f"{stack_name}_")
|
|
||||||
settings = app_settings[stack_name] = app_settings.get(stack_name) or {}
|
|
||||||
if (backup := labels.get('backupbot.backup')) and bool(backup):
|
|
||||||
settings['enabled'] = True
|
|
||||||
if selected_container and container_name not in selected_container:
|
|
||||||
logger.debug(f"Skipping {s.name} because it's not a selected container")
|
|
||||||
continue
|
|
||||||
if mounts:= specs['TaskTemplate']['ContainerSpec'].get('Mounts'):
|
|
||||||
volumes = parse_volumes(stack_name, mounts)
|
|
||||||
volumes.update(settings.get('volumes') or {})
|
|
||||||
settings['volumes'] = volumes
|
|
||||||
excluded_volumes, included_volume_paths = parse_excludes_includes(labels)
|
|
||||||
settings['excluded_volumes'] = excluded_volumes.union(settings.get('excluded_volumes') or set())
|
|
||||||
settings['included_volume_paths'] = included_volume_paths.union(settings.get('included_volume_paths') or set())
|
|
||||||
if container := container_by_service.get(s.name):
|
|
||||||
if command := labels.get(f'backupbot.{hook_type}.pre-hook'):
|
|
||||||
if not (pre_hooks:= settings.get('pre_hooks')):
|
|
||||||
pre_hooks = settings['pre_hooks'] = {}
|
|
||||||
pre_hooks[container] = command
|
|
||||||
if command := labels.get(f'backupbot.{hook_type}.post-hook'):
|
|
||||||
if not (post_hooks:= settings.get('post_hooks')):
|
|
||||||
post_hooks = settings['post_hooks'] = {}
|
|
||||||
post_hooks[container] = command
|
|
||||||
else:
|
|
||||||
logger.debug(f"Container {s.name} is not running.")
|
|
||||||
if labels.get(f'backupbot.{hook_type}.pre-hook') or labels.get(f'backupbot.{hook_type}.post-hook'):
|
|
||||||
logger.error(f"Container {s.name} contain hooks but it's not running")
|
|
||||||
return app_settings
|
|
||||||
|
|
||||||
|
|
||||||
def get_backup_details(app_settings, volumes=[]):
|
|
||||||
backup_paths = set()
|
backup_paths = set()
|
||||||
backup_apps = []
|
backup_apps = set()
|
||||||
pre_hooks= {}
|
pre_commands = {}
|
||||||
post_hooks = {}
|
post_commands = {}
|
||||||
for app, settings in app_settings.items():
|
services = client.services.list()
|
||||||
if settings.get('enabled'):
|
for s in services:
|
||||||
if SERVICE != 'ALL' and SERVICE != app:
|
labels = s.attrs['Spec']['Labels']
|
||||||
|
if (backup := labels.get('backupbot.backup')) and bool(backup):
|
||||||
|
# volumes: s.attrs['Spec']['TaskTemplate']['ContainerSpec']['Mounts'][0]['Source']
|
||||||
|
stack_name = labels['com.docker.stack.namespace']
|
||||||
|
# Remove this lines to backup only a specific service
|
||||||
|
# This will unfortenately decrease restice performance
|
||||||
|
# if SERVICE and SERVICE != stack_name:
|
||||||
|
# continue
|
||||||
|
backup_apps.add(stack_name)
|
||||||
|
backup_paths = backup_paths.union(
|
||||||
|
Path(VOLUME_PATH).glob(f"{stack_name}_*"))
|
||||||
|
if not (container := container_by_service.get(s.name)):
|
||||||
|
logger.error(
|
||||||
|
f"Container {s.name} is not running, hooks can not be executed")
|
||||||
continue
|
continue
|
||||||
backup_apps.append(app)
|
if prehook := labels.get('backupbot.backup.pre-hook'):
|
||||||
add_backup_paths(backup_paths, settings, app, volumes)
|
pre_commands[container] = prehook
|
||||||
if hooks:= settings.get('pre_hooks'):
|
if posthook := labels.get('backupbot.backup.post-hook'):
|
||||||
pre_hooks.update(hooks)
|
post_commands[container] = posthook
|
||||||
if hooks:= settings.get('post_hooks'):
|
return pre_commands, post_commands, list(backup_paths), list(backup_apps)
|
||||||
post_hooks.update(hooks)
|
|
||||||
return pre_hooks, post_hooks, list(backup_paths), backup_apps
|
|
||||||
|
|
||||||
|
|
||||||
def add_backup_paths(backup_paths, settings, app, selected_volumes):
|
|
||||||
if (volumes := settings.get('volumes')):
|
|
||||||
if includes:= settings.get('included_volume_paths'):
|
|
||||||
included_volumes = list(zip(*includes))[0]
|
|
||||||
for volume, rel_paths in includes:
|
|
||||||
if not (volume_path:= volumes.get(volume)):
|
|
||||||
logger.error(f'Can not find volume with the name {volume}')
|
|
||||||
continue
|
|
||||||
if selected_volumes and volume not in selected_volumes:
|
|
||||||
logger.debug(f'Skipping {volume}:{rel_paths} because the volume is not selected')
|
|
||||||
continue
|
|
||||||
for p in rel_paths:
|
|
||||||
absolute_path = Path(f"{volume_path}/{p}")
|
|
||||||
backup_paths.add(absolute_path)
|
|
||||||
else:
|
|
||||||
included_volumes = []
|
|
||||||
excluded_volumes = settings.get('excluded_volumes') or []
|
|
||||||
for name, path in volumes.items():
|
|
||||||
if selected_volumes and name not in selected_volumes:
|
|
||||||
logger.debug(f'Skipping volume: {name} because the volume is not selected')
|
|
||||||
continue
|
|
||||||
if name in excluded_volumes:
|
|
||||||
logger.debug(f'Skipping volume: {name} because the volume is excluded')
|
|
||||||
continue
|
|
||||||
if name in included_volumes:
|
|
||||||
logger.debug(f'Skipping volume: {name} because a path is selected')
|
|
||||||
continue
|
|
||||||
backup_paths.add(path)
|
|
||||||
else:
|
|
||||||
logger.warning(f"{app} does not contain any volumes")
|
|
||||||
|
|
||||||
|
|
||||||
def parse_volumes(stack_name, mounts):
|
|
||||||
volumes = {}
|
|
||||||
for m in mounts:
|
|
||||||
if m['Type'] != 'volume':
|
|
||||||
continue
|
|
||||||
relative_path = m['Source']
|
|
||||||
name = relative_path.removeprefix(stack_name + '_')
|
|
||||||
absolute_path = Path(f"{VOLUME_PATH}{relative_path}/_data/")
|
|
||||||
volumes[name] = absolute_path
|
|
||||||
return volumes
|
|
||||||
|
|
||||||
|
|
||||||
def parse_excludes_includes(labels):
|
|
||||||
excluded_volumes = set()
|
|
||||||
included_volume_paths = set()
|
|
||||||
for label, value in labels.items():
|
|
||||||
if label.startswith('backupbot.backup.volumes.'):
|
|
||||||
volume_name = label.removeprefix('backupbot.backup.volumes.').removesuffix('.path')
|
|
||||||
if label.endswith('path'):
|
|
||||||
relative_paths = tuple(value.split(','))
|
|
||||||
included_volume_paths.add((volume_name, relative_paths))
|
|
||||||
elif bool(value):
|
|
||||||
excluded_volumes.add(volume_name)
|
|
||||||
return excluded_volumes, included_volume_paths
|
|
||||||
|
|
||||||
|
|
||||||
def copy_secrets(apps):
|
def copy_secrets(apps):
|
||||||
@ -314,7 +161,7 @@ def copy_secrets(apps):
|
|||||||
f"For the secret {sec['SecretName']} the file {src} does not exist for {s.name}")
|
f"For the secret {sec['SecretName']} the file {src} does not exist for {s.name}")
|
||||||
continue
|
continue
|
||||||
dst = SECRET_PATH + sec['SecretName']
|
dst = SECRET_PATH + sec['SecretName']
|
||||||
logger.debug(f"Copy Secret {sec['SecretName']}")
|
logger.debug("Copy Secret {sec['SecretName']}")
|
||||||
copyfile(src, dst)
|
copyfile(src, dst)
|
||||||
|
|
||||||
|
|
||||||
@ -342,21 +189,14 @@ def run_commands(commands):
|
|||||||
def backup_volumes(backup_paths, apps, retries, dry_run=False):
|
def backup_volumes(backup_paths, apps, retries, dry_run=False):
|
||||||
while True:
|
while True:
|
||||||
try:
|
try:
|
||||||
logger.info("Backup these paths:")
|
|
||||||
logger.debug("\n".join(map(str, backup_paths)))
|
|
||||||
backup_paths = list(filter(path_exists, backup_paths))
|
|
||||||
cmd = restic.cat.base_command()
|
|
||||||
parent = get_snapshots('latest')
|
|
||||||
if parent:
|
|
||||||
# https://restic.readthedocs.io/en/stable/040_backup.html#file-change-detection
|
|
||||||
cmd.extend(['--parent', parent[0]['short_id']])
|
|
||||||
tags = set(apps + [SERVICE])
|
|
||||||
logger.info("Start volume backup")
|
logger.info("Start volume backup")
|
||||||
result = restic.internal.backup.run(cmd, backup_paths, dry_run=dry_run, tags=tags)
|
logger.debug(backup_paths)
|
||||||
|
result = restic.backup(backup_paths, dry_run=dry_run, tags=apps)
|
||||||
logger.summary("backup finished", extra=result)
|
logger.summary("backup finished", extra=result)
|
||||||
return
|
return
|
||||||
except ResticFailedError as error:
|
except ResticFailedError as error:
|
||||||
logger.error(f"Backup failed for {SERVICE}.")
|
logger.error(
|
||||||
|
f"Backup failed for {apps}. Could not Backup these paths: {backup_paths}")
|
||||||
logger.error(error, exc_info=True)
|
logger.error(error, exc_info=True)
|
||||||
if retries > 0:
|
if retries > 0:
|
||||||
retries -= 1
|
retries -= 1
|
||||||
@ -364,20 +204,49 @@ def backup_volumes(backup_paths, apps, retries, dry_run=False):
|
|||||||
exit(1)
|
exit(1)
|
||||||
|
|
||||||
|
|
||||||
def path_exists(path):
|
@cli.command()
|
||||||
if not path.exists():
|
@click.option('snapshot', '--snapshot', '-s', envvar='SNAPSHOT', default='latest')
|
||||||
logger.error(f'{path} does not exist')
|
@click.option('target', '--target', '-t', envvar='TARGET', default='/')
|
||||||
return path.exists()
|
@click.option('noninteractive', '--noninteractive', envvar='NONINTERACTIVE', is_flag=True)
|
||||||
|
def restore(snapshot, target, noninteractive):
|
||||||
|
# Todo: recommend to shutdown the container
|
||||||
|
service_paths = VOLUME_PATH
|
||||||
|
if SERVICE:
|
||||||
|
service_paths = service_paths + f'{SERVICE}_*'
|
||||||
|
snapshots = restic.snapshots(snapshot_id=snapshot)
|
||||||
|
if not snapshot:
|
||||||
|
logger.error("No Snapshots with ID {snapshots}")
|
||||||
|
exit(1)
|
||||||
|
if not noninteractive:
|
||||||
|
snapshot_date = datetime.fromisoformat(snapshots[0]['time'])
|
||||||
|
delta = datetime.now(tz=timezone.utc) - snapshot_date
|
||||||
|
print(
|
||||||
|
f"You are going to restore Snapshot {snapshot} of {service_paths} at {target}")
|
||||||
|
print(f"This snapshot is {delta} old")
|
||||||
|
print(
|
||||||
|
f"THIS COMMAND WILL IRREVERSIBLY OVERWRITES {target}{service_paths.removeprefix('/')}")
|
||||||
|
prompt = input("Type YES (uppercase) to continue: ")
|
||||||
|
if prompt != 'YES':
|
||||||
|
logger.error("Restore aborted")
|
||||||
|
exit(1)
|
||||||
|
print(f"Restoring Snapshot {snapshot} of {service_paths} at {target}")
|
||||||
|
# TODO: use tags if no snapshot is selected, to use a snapshot including SERVICE
|
||||||
|
result = restic.restore(snapshot_id=snapshot,
|
||||||
|
include=service_paths, target_dir=target)
|
||||||
|
logger.debug(result)
|
||||||
|
|
||||||
|
|
||||||
@cli.command()
|
@cli.command()
|
||||||
def snapshots():
|
def snapshots():
|
||||||
snapshots = get_snapshots()
|
snapshots = restic.snapshots()
|
||||||
|
no_snapshots = True
|
||||||
for snap in snapshots:
|
for snap in snapshots:
|
||||||
print(snap['time'], snap['id'])
|
if not SERVICE or (tags := snap.get('tags')) and SERVICE in tags:
|
||||||
if not snapshots:
|
print(snap['time'], snap['id'])
|
||||||
|
no_snapshots = False
|
||||||
|
if no_snapshots:
|
||||||
err_msg = "No Snapshots found"
|
err_msg = "No Snapshots found"
|
||||||
if SERVICE != 'ALL':
|
if SERVICE:
|
||||||
service_name = SERVICE.replace('_', '.')
|
service_name = SERVICE.replace('_', '.')
|
||||||
err_msg += f' for app {service_name}'
|
err_msg += f' for app {service_name}'
|
||||||
logger.warning(err_msg)
|
logger.warning(err_msg)
|
||||||
@ -395,7 +264,8 @@ def ls(snapshot, path):
|
|||||||
|
|
||||||
def list_files(snapshot, path):
|
def list_files(snapshot, path):
|
||||||
cmd = restic.cat.base_command() + ['ls']
|
cmd = restic.cat.base_command() + ['ls']
|
||||||
cmd = cmd + ['--tag', SERVICE]
|
if SERVICE:
|
||||||
|
cmd = cmd + ['--tag', SERVICE]
|
||||||
cmd.append(snapshot)
|
cmd.append(snapshot)
|
||||||
if path:
|
if path:
|
||||||
cmd.append(path)
|
cmd.append(path)
|
||||||
@ -404,7 +274,7 @@ def list_files(snapshot, path):
|
|||||||
except ResticFailedError as error:
|
except ResticFailedError as error:
|
||||||
if 'no snapshot found' in str(error):
|
if 'no snapshot found' in str(error):
|
||||||
err_msg = f'There is no snapshot "{snapshot}"'
|
err_msg = f'There is no snapshot "{snapshot}"'
|
||||||
if SERVICE != 'ALL':
|
if SERVICE:
|
||||||
err_msg += f' for the app "{SERVICE}"'
|
err_msg += f' for the app "{SERVICE}"'
|
||||||
logger.error(err_msg)
|
logger.error(err_msg)
|
||||||
exit(1)
|
exit(1)
|
||||||
@ -436,7 +306,7 @@ def download(snapshot, path, volumes, secrets):
|
|||||||
tarinfo.size = len(binary_output)
|
tarinfo.size = len(binary_output)
|
||||||
file_dumps.append((binary_output, tarinfo))
|
file_dumps.append((binary_output, tarinfo))
|
||||||
if volumes:
|
if volumes:
|
||||||
if SERVICE == 'ALL':
|
if not SERVICE:
|
||||||
logger.error("Please specify '--host' when using '--volumes'")
|
logger.error("Please specify '--host' when using '--volumes'")
|
||||||
exit(1)
|
exit(1)
|
||||||
files = list_files(snapshot, VOLUME_PATH)
|
files = list_files(snapshot, VOLUME_PATH)
|
||||||
@ -449,7 +319,7 @@ def download(snapshot, path, volumes, secrets):
|
|||||||
tarinfo.size = len(binary_output)
|
tarinfo.size = len(binary_output)
|
||||||
file_dumps.append((binary_output, tarinfo))
|
file_dumps.append((binary_output, tarinfo))
|
||||||
if secrets:
|
if secrets:
|
||||||
if SERVICE == 'ALL':
|
if not SERVICE:
|
||||||
logger.error("Please specify '--host' when using '--secrets'")
|
logger.error("Please specify '--host' when using '--secrets'")
|
||||||
exit(1)
|
exit(1)
|
||||||
filename = f"{SERVICE}.json"
|
filename = f"{SERVICE}.json"
|
||||||
@ -486,7 +356,8 @@ def get_formatted_size(file_path):
|
|||||||
|
|
||||||
def dump(snapshot, path):
|
def dump(snapshot, path):
|
||||||
cmd = restic.cat.base_command() + ['dump']
|
cmd = restic.cat.base_command() + ['dump']
|
||||||
cmd = cmd + ['--tag', SERVICE]
|
if SERVICE:
|
||||||
|
cmd = cmd + ['--tag', SERVICE]
|
||||||
cmd = cmd + [snapshot, path]
|
cmd = cmd + [snapshot, path]
|
||||||
print(f"Dumping {path} from snapshot '{snapshot}'")
|
print(f"Dumping {path} from snapshot '{snapshot}'")
|
||||||
output = subprocess.run(cmd, capture_output=True)
|
output = subprocess.run(cmd, capture_output=True)
|
||||||
|
11
compose.pushbasicauth.yml
Normal file
11
compose.pushbasicauth.yml
Normal file
@ -0,0 +1,11 @@
|
|||||||
|
---
|
||||||
|
version: "3.8"
|
||||||
|
services:
|
||||||
|
app:
|
||||||
|
secrets:
|
||||||
|
- push_basicauth
|
||||||
|
|
||||||
|
secrets:
|
||||||
|
push_basicauth:
|
||||||
|
external: true
|
||||||
|
name: ${STACK_NAME}_push_basicauth_${SECRET_PUSH_BASICAUTH}
|
19
compose.yml
19
compose.yml
@ -2,7 +2,7 @@
|
|||||||
version: "3.8"
|
version: "3.8"
|
||||||
services:
|
services:
|
||||||
app:
|
app:
|
||||||
image: git.coopcloud.tech/coop-cloud/backup-bot-two:2.1.1-beta
|
image: git.coopcloud.tech/coop-cloud/backup-bot-two:2.0.0
|
||||||
volumes:
|
volumes:
|
||||||
- "/var/run/docker.sock:/var/run/docker.sock"
|
- "/var/run/docker.sock:/var/run/docker.sock"
|
||||||
- "/var/lib/docker/volumes/:/var/lib/docker/volumes/"
|
- "/var/lib/docker/volumes/:/var/lib/docker/volumes/"
|
||||||
@ -14,9 +14,16 @@ services:
|
|||||||
- RESTIC_PASSWORD_FILE=/run/secrets/restic_password
|
- RESTIC_PASSWORD_FILE=/run/secrets/restic_password
|
||||||
secrets:
|
secrets:
|
||||||
- restic_password
|
- restic_password
|
||||||
|
configs:
|
||||||
|
- source: entrypoint
|
||||||
|
target: /entrypoint.sh
|
||||||
|
mode: 666
|
||||||
|
- source: cronjob
|
||||||
|
target: /cronjob.sh
|
||||||
|
mode: 666
|
||||||
deploy:
|
deploy:
|
||||||
labels:
|
labels:
|
||||||
- coop-cloud.${STACK_NAME}.version=2.0.1+2.1.1-beta
|
- coop-cloud.${STACK_NAME}.version=0.1.0+latest
|
||||||
- coop-cloud.${STACK_NAME}.timeout=${TIMEOUT:-300}
|
- coop-cloud.${STACK_NAME}.timeout=${TIMEOUT:-300}
|
||||||
- coop-cloud.backupbot.enabled=true
|
- coop-cloud.backupbot.enabled=true
|
||||||
#entrypoint: ['tail', '-f','/dev/null']
|
#entrypoint: ['tail', '-f','/dev/null']
|
||||||
@ -32,5 +39,13 @@ secrets:
|
|||||||
external: true
|
external: true
|
||||||
name: ${STACK_NAME}_restic_password_${SECRET_RESTIC_PASSWORD_VERSION}
|
name: ${STACK_NAME}_restic_password_${SECRET_RESTIC_PASSWORD_VERSION}
|
||||||
|
|
||||||
|
configs:
|
||||||
|
entrypoint:
|
||||||
|
name: ${STACK_NAME}_entrypoint_${ENTRYPOINT_VERSION}
|
||||||
|
file: entrypoint.sh
|
||||||
|
cronjob:
|
||||||
|
name: ${STACK_NAME}_cronjob_${CRONJOB_VERSION}
|
||||||
|
file: cronjob.sh
|
||||||
|
|
||||||
volumes:
|
volumes:
|
||||||
backups:
|
backups:
|
||||||
|
40
cronjob.sh
Executable file
40
cronjob.sh
Executable file
@ -0,0 +1,40 @@
|
|||||||
|
#!/bin/sh
|
||||||
|
|
||||||
|
set -e
|
||||||
|
|
||||||
|
CURL_OPTS="-s"
|
||||||
|
# Check for basic auth
|
||||||
|
if [ -n "$(cat /run/secrets/push_basicauth)" ]
|
||||||
|
then
|
||||||
|
CURL_OPTS="$CURL_OPTS -u $(cat /run/secrets/push_basicauth)"
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [ -n "$PUSH_PROMETHEUS_URL" ]
|
||||||
|
then
|
||||||
|
push_start_notification="(echo 'backup 1' | curl $CURL_OPTS --data-binary @- $PUSH_PROMETHEUS_URL)"
|
||||||
|
push_success_notification="(echo 'backup 0' | curl $CURL_OPTS --data-binary @- $PUSH_PROMETHEUS_URL)"
|
||||||
|
push_fail_notification="(echo 'backup -1' | curl $CURL_OPTS --data-binary @- $PUSH_PROMETHEUS_URL)"
|
||||||
|
else
|
||||||
|
if [ -n "$PUSH_URL_START" ]
|
||||||
|
then
|
||||||
|
push_start_notification="curl $CURL_OPTS '$PUSH_URL_START'"
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [ -n "$PUSH_URL_FAIL" ]
|
||||||
|
then
|
||||||
|
push_fail_notification="curl $CURL_OPTS '$PUSH_URL_FAIL'"
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [ -n "$PUSH_URL_SUCCESS" ]
|
||||||
|
then
|
||||||
|
push_success_notification="curl $CURL_OPTS '$PUSH_URL_SUCCESS'"
|
||||||
|
fi
|
||||||
|
fi
|
||||||
|
|
||||||
|
eval "$push_start_notification"
|
||||||
|
if [ "$(backup --machine-logs create 2>&1 | tee /tmp/backup.log && (grep -q 'backup finished' /tmp/backup.log))" ]
|
||||||
|
then
|
||||||
|
eval "$push_success_notification"
|
||||||
|
else
|
||||||
|
eval "$push_fail_notification"
|
||||||
|
fi
|
@ -9,22 +9,7 @@ fi
|
|||||||
|
|
||||||
cron_schedule="${CRON_SCHEDULE:?CRON_SCHEDULE not set}"
|
cron_schedule="${CRON_SCHEDULE:?CRON_SCHEDULE not set}"
|
||||||
|
|
||||||
if [ -n "$PUSH_URL_START" ]
|
echo "$cron_schedule /cronjob.sh" | crontab -
|
||||||
then
|
|
||||||
push_start_notification="curl -s '$PUSH_URL_START' &&"
|
|
||||||
fi
|
|
||||||
|
|
||||||
if [ -n "$PUSH_URL_FAIL" ]
|
|
||||||
then
|
|
||||||
push_fail_notification="|| curl -s '$PUSH_URL_FAIL'"
|
|
||||||
fi
|
|
||||||
|
|
||||||
if [ -n "$PUSH_URL_SUCCESS" ]
|
|
||||||
then
|
|
||||||
push_notification=" && (grep -q 'backup finished' /tmp/backup.log && curl -s '$PUSH_URL_SUCCESS' $push_fail_notification)"
|
|
||||||
fi
|
|
||||||
|
|
||||||
echo "$cron_schedule $push_start_notification backup --machine-logs create 2>&1 | tee /tmp/backup.log $push_notification" | crontab -
|
|
||||||
crontab -l
|
crontab -l
|
||||||
|
|
||||||
crond -f -d8 -L /dev/stdout
|
crond -f -d8 -L /dev/stdout
|
||||||
|
@ -1 +0,0 @@
|
|||||||
This is the first beta release of the new backup-bot-two rewrite in python. Be aware when updating, it can break. Please read the readme and update your config according to it.
|
|
Reference in New Issue
Block a user