generated from coop-cloud/example
Compare commits
18 Commits
0.13.0+0.7
...
main
Author | SHA1 | Date |
---|---|---|
Moritz | c1d6826d82 | |
Moritz | ac7fb7c3dd | |
Moritz | ba63176598 | |
iexos | 6eee864ba1 | |
iexos | f3c8e08441 | |
iexos | 6da688ad1b | |
3wc | ea012f2628 | |
iexos | 1705383533 | |
iexos | 6575668586 | |
iexos | 39f3a61ce0 | |
iexos | 158a3c8b1c | |
iexos | 03b7d984f0 | |
Simon Thiessen | 907597aab3 | |
Simon Thiessen | d9ce8fb168 | |
Simon Thiessen | 73de12d12f | |
iexos | 8c3521b87f | |
iexos | 4273faad76 | |
iexos | 0e84bbc2ee |
|
@ -21,7 +21,6 @@ steps:
|
|||
SECRET_DB_PASSWORD_VERSION: v1
|
||||
SECRET_SECRET_KEY_VERSION: v1 # length=64
|
||||
SECRET_UTILS_SECRET_VERSION: v1 # length=64
|
||||
SECRET_AWS_SECRET_KEY_VERSION: v1
|
||||
trigger:
|
||||
branch:
|
||||
- main
|
||||
|
|
23
.env.sample
23
.env.sample
|
@ -15,15 +15,9 @@ COMPOSE_FILE="compose.yml"
|
|||
SECRET_DB_PASSWORD_VERSION=v1
|
||||
SECRET_SECRET_KEY_VERSION=v1 # length=64
|
||||
SECRET_UTILS_SECRET_VERSION=v1 # length=64
|
||||
SECRET_AWS_SECRET_KEY_VERSION=v1
|
||||
|
||||
AWS_ACCESS_KEY_ID=
|
||||
AWS_REGION=
|
||||
AWS_S3_UPLOAD_BUCKET_URL=
|
||||
AWS_S3_UPLOAD_BUCKET_NAME=
|
||||
AWS_S3_UPLOAD_MAX_SIZE=26214400
|
||||
AWS_S3_FORCE_PATH_STYLE=true
|
||||
AWS_S3_ACL=private
|
||||
# Set to s3 to use AWS S3 bucket
|
||||
FILE_STORAGE=local
|
||||
|
||||
# –––––––––––––––– OPTIONAL ––––––––––––––––
|
||||
|
||||
|
@ -85,3 +79,16 @@ ALLOWED_DOMAINS=
|
|||
#GOOGLE_ENABLED=1
|
||||
#GOOGLE_CLIENT_ID=
|
||||
#SECRET_GOOGLE_CLIENT_SECRET_VERSION=v1
|
||||
|
||||
COMPOSE_FILE="$COMPOSE_FILE:compose.local.yml"
|
||||
FILE_STORAGE_UPLOAD_MAX_SIZE=26214400
|
||||
|
||||
#COMPOSE_FILE="$COMPOSE_FILE:compose.aws.yml"
|
||||
#AWS_ACCESS_KEY_ID=
|
||||
#AWS_REGION=
|
||||
#AWS_S3_UPLOAD_BUCKET_URL=
|
||||
#AWS_S3_UPLOAD_BUCKET_NAME=
|
||||
#AWS_S3_UPLOAD_MAX_SIZE=26214400
|
||||
#AWS_S3_FORCE_PATH_STYLE=true
|
||||
#AWS_S3_ACL=private
|
||||
#SECRET_AWS_SECRET_KEY_VERSION=v1
|
||||
|
|
40
README.md
40
README.md
|
@ -22,13 +22,12 @@ Wiki and knowledge base for growing teams
|
|||
3. `abra app new ${REPO_NAME}`
|
||||
- **WARNING**: Choose "n" when `abra` asks if you'd like to generate secrets
|
||||
4. `abra app config YOURAPPNAME` - be sure to change `$DOMAIN` to something that resolves to
|
||||
your Docker swarm box. For Minio, you'll want:
|
||||
- `AWS_ACCESS_KEY_ID=<minio username>`
|
||||
- `AWS_REGION="us-east-1"`
|
||||
- `AWS_S3_UPLOAD_BUCKET_URL=https://minio.example.com`
|
||||
- `AWS_S3_UPLOAD_BUCKET_NAME=
|
||||
5. `abra app deploy YOURAPPNAME`
|
||||
7. Open the configured domain in your browser to finish set-up
|
||||
your Docker swarm box
|
||||
5. Insert secrets:
|
||||
- `abra app secret insert YOURAPPNAME secret_key v1 $(openssl rand -hex 32)` #12
|
||||
- `abra app secret generate -a YOURAPPNAME`
|
||||
6. `abra app deploy YOURAPPNAME`
|
||||
8. Open the configured domain in your browser to finish set-up
|
||||
|
||||
[`abra`]: https://git.coopcloud.tech/coop-cloud/abra
|
||||
[`coop-cloud/traefik`]: https://git.coopcloud.tech/coop-cloud/traefik
|
||||
|
@ -41,14 +40,6 @@ Wiki and knowledge base for growing teams
|
|||
abra app cmd YOURAPPNAME app create_email_user test@example.com
|
||||
```
|
||||
|
||||
### Post-deploy migration
|
||||
|
||||
```
|
||||
abra app cmd YOURAPPNAME app migrate
|
||||
```
|
||||
|
||||
_As of 2022-03-30, this requires `abra` RC version, run `abra upgrade --rc`._
|
||||
|
||||
### Setting up your `.env` config
|
||||
|
||||
Avoid the use of quotes (`"..."`) as much as possible, the NodeJS scripts flip out for some reason on some vars.
|
||||
|
@ -61,7 +52,24 @@ Where `<username-to-delete>` is the username of the user to be removed, and
|
|||
`<username-to-replace>` is the username of another user, to assign documents and
|
||||
revisions to (instead of deleting them).
|
||||
|
||||
_As of 2022-03-30, this requires `abra` RC version, run `abra upgrade --rc`._
|
||||
### Migrate from S3 to local storage
|
||||
|
||||
- `abra app config <domain>`, add
|
||||
- `COMPOSE_FILE="$COMPOSE_FILE:compose.local.yml"`
|
||||
- `FILE_STORAGE_UPLOAD_MAX_SIZE=26214400`
|
||||
- `abra app deploy <domain> -f`
|
||||
- compose.aws.yml should still be deployed!
|
||||
- `abra app undeploy <domain>`
|
||||
- on the docker host, find mountpoint of newly created volume via `docker volume ls` and `docker volume inspect`
|
||||
- volume name is smth like `<domain>_storage-data`
|
||||
- take note which linux user owns `<storage_mountpoint>` (likely `1001`)
|
||||
- use s3cmd/rclone/... to sync your bucket to `<storage_mountpoint>`
|
||||
- `chown -R <storage_user>:<storage_user> <storage_mountpoint>`
|
||||
- `abra app config <domain>`, switch storage backend
|
||||
- remove `AWS_*` vars, `SECRET_AWS_SECRET_KEY_VERSION` and `COMPOSE_FILE="$COMPOSE_FILE:compose.aws.yml"`
|
||||
- set `FILE_STORAGE=local`
|
||||
- `abra app deploy <domain> -f`
|
||||
- enjoy getting rid of S3 🥳
|
||||
|
||||
## Single Sign On with Keycloak
|
||||
|
||||
|
|
2
abra.sh
2
abra.sh
|
@ -1,4 +1,4 @@
|
|||
export APP_ENTRYPOINT_VERSION=v8
|
||||
export APP_ENTRYPOINT_VERSION=v9
|
||||
export DB_ENTRYPOINT_VERSION=v2
|
||||
|
||||
create_email_user() {
|
||||
|
|
|
@ -0,0 +1,15 @@
|
|||
authentik:
|
||||
env:
|
||||
OIDC_CLIENT_ID: outline
|
||||
OIDC_AUTH_URI: https://authentik.example.com/application/o/authorize/
|
||||
OIDC_TOKEN_URI: https://authentik.example.com/application/o/token/
|
||||
OIDC_USERINFO_URI: https://authentik.example.com/application/o/userinfo/
|
||||
OIDC_DISPLAY_NAME: "Authentik"
|
||||
uncomment:
|
||||
- compose.oidc.yml
|
||||
- OIDC_ENABLED
|
||||
- OIDC_USERNAME_CLAIM
|
||||
- OIDC_SCOPES
|
||||
- SECRET_OIDC_CLIENT_SECRET_VERSION
|
||||
shared_secrets:
|
||||
outline_secret: oidc_client_secret
|
|
@ -0,0 +1,22 @@
|
|||
---
|
||||
version: "3.8"
|
||||
|
||||
services:
|
||||
app:
|
||||
secrets:
|
||||
- aws_secret_key
|
||||
environment:
|
||||
- AWS_ACCESS_KEY_ID
|
||||
- AWS_REGION
|
||||
- AWS_S3_ACL
|
||||
- AWS_S3_FORCE_PATH_STYLE
|
||||
- AWS_S3_UPLOAD_BUCKET_NAME
|
||||
- AWS_S3_UPLOAD_BUCKET_URL
|
||||
- AWS_S3_UPLOAD_MAX_SIZE
|
||||
- AWS_SDK_LOAD_CONFIG=0
|
||||
- AWS_SECRET_KEY_FILE=/run/secrets/aws_secret_key
|
||||
|
||||
secrets:
|
||||
aws_secret_key:
|
||||
name: ${STACK_NAME}_aws_secret_key_${SECRET_AWS_SECRET_KEY_VERSION}
|
||||
external: true
|
|
@ -0,0 +1,13 @@
|
|||
---
|
||||
version: "3.8"
|
||||
|
||||
services:
|
||||
app:
|
||||
volumes:
|
||||
- storage-data:/var/lib/outline/data
|
||||
environment:
|
||||
- FILE_STORAGE
|
||||
- FILE_STORAGE_UPLOAD_MAX_SIZE
|
||||
|
||||
volumes:
|
||||
storage-data:
|
28
compose.yml
28
compose.yml
|
@ -6,9 +6,8 @@ services:
|
|||
networks:
|
||||
- backend
|
||||
- proxy
|
||||
image: outlinewiki/outline:0.71.0
|
||||
image: outlinewiki/outline:0.75.2
|
||||
secrets:
|
||||
- aws_secret_key
|
||||
- db_password
|
||||
- secret_key
|
||||
- utils_secret
|
||||
|
@ -17,15 +16,7 @@ services:
|
|||
target: /docker-entrypoint.sh
|
||||
mode: 0555
|
||||
environment:
|
||||
- AWS_ACCESS_KEY_ID
|
||||
- AWS_REGION
|
||||
- AWS_S3_ACL
|
||||
- AWS_S3_FORCE_PATH_STYLE
|
||||
- AWS_S3_UPLOAD_BUCKET_NAME
|
||||
- AWS_S3_UPLOAD_BUCKET_URL
|
||||
- AWS_S3_UPLOAD_MAX_SIZE
|
||||
- AWS_SDK_LOAD_CONFIG=0
|
||||
- AWS_SECRET_KEY_FILE=/run/secrets/aws_secret_key
|
||||
- FILE_STORAGE
|
||||
- DATABASE_PASSWORD_FILE=/run/secrets/db_password
|
||||
- FORCE_HTTPS=true
|
||||
- PGSSLMODE=disable
|
||||
|
@ -43,19 +34,19 @@ services:
|
|||
- "traefik.http.routers.${STACK_NAME}.rule=Host(`${DOMAIN}`${EXTRA_DOMAINS})"
|
||||
- "traefik.http.routers.${STACK_NAME}.entrypoints=web-secure"
|
||||
- "traefik.http.routers.${STACK_NAME}.tls.certresolver=${LETS_ENCRYPT_ENV}"
|
||||
- "coop-cloud.${STACK_NAME}.version=0.13.0+0.71.0"
|
||||
- "coop-cloud.${STACK_NAME}.version=2.2.0+0.75.2"
|
||||
## Redirect from EXTRA_DOMAINS to DOMAIN
|
||||
#- "traefik.http.routers.${STACK_NAME}.middlewares=${STACK_NAME}-redirect"
|
||||
#- "traefik.http.middlewares.${STACK_NAME}-redirect.headers.SSLForceHost=true"
|
||||
#- "traefik.http.middlewares.${STACK_NAME}-redirect.headers.SSLHost=${DOMAIN}"
|
||||
|
||||
cache:
|
||||
image: redis:7.2.0
|
||||
image: redis:7.2.4
|
||||
networks:
|
||||
- backend
|
||||
|
||||
db:
|
||||
image: postgres:15.4
|
||||
image: postgres:16.2
|
||||
networks:
|
||||
- backend
|
||||
secrets:
|
||||
|
@ -74,9 +65,9 @@ services:
|
|||
deploy:
|
||||
labels:
|
||||
backupbot.backup: "true"
|
||||
backupbot.backup.path: "/tmp/dump.sql.gz"
|
||||
backupbot.backup.post-hook: "rm -f /tmp/dump.sql.gz"
|
||||
backupbot.backup.pre-hook: "sh -c 'PGPASSWORD=$$(cat $${POSTGRES_PASSWORD_FILE}) pg_dump -U outline outline | gzip > /tmp/dump.sql.gz'"
|
||||
backupbot.backup.path: "/var/lib/postgresql/data/dump.sql.gz"
|
||||
backupbot.backup.post-hook: "rm -f /var/lib/postgresql/data/dump.sql.gz"
|
||||
backupbot.backup.pre-hook: "sh -c 'PGPASSWORD=$$(cat $${POSTGRES_PASSWORD_FILE}) pg_dump -U outline outline | gzip > /var/lib/postgresql/data/dump.sql.gz'"
|
||||
|
||||
secrets:
|
||||
secret_key:
|
||||
|
@ -85,9 +76,6 @@ secrets:
|
|||
utils_secret:
|
||||
name: ${STACK_NAME}_utils_secret_${SECRET_UTILS_SECRET_VERSION}
|
||||
external: true
|
||||
aws_secret_key:
|
||||
name: ${STACK_NAME}_aws_secret_key_${SECRET_AWS_SECRET_KEY_VERSION}
|
||||
external: true
|
||||
db_password:
|
||||
name: ${STACK_NAME}_db_password_${SECRET_DB_PASSWORD_VERSION}
|
||||
external: true
|
||||
|
|
|
@ -1,6 +1,8 @@
|
|||
#!/bin/sh
|
||||
|
||||
{{ if eq (env "FILE_STORAGE") "s3" }}
|
||||
export AWS_SECRET_ACCESS_KEY=$(cat /run/secrets/aws_secret_key)
|
||||
{{ end }}
|
||||
|
||||
{{ if eq (env "SMTP_ENABLED") "1" }}
|
||||
export SMTP_PASSWORD=$(cat /run/secrets/smtp_password)
|
||||
|
|
|
@ -0,0 +1,4 @@
|
|||
Due to the introduction of local storage, you need to adapt your config to continue using S3 storage. Just add the following lines to your config:
|
||||
|
||||
FILE_STORAGE=s3
|
||||
COMPOSE_FILE="$COMPOSE_FILE:compose.aws.yml"
|
Loading…
Reference in New Issue