Compare commits

..

3 commits

8 changed files with 94 additions and 27 deletions

View file

@ -6,6 +6,7 @@ services:
networks:
- web
- grafana
- monitoring
labels:
- "traefik.enable=true"
- "traefik.http.routers.grafana.rule=Host(`monitoring.${DOMAIN}`)"
@ -29,7 +30,7 @@ services:
- GF_DATABASE_NAME=grafana
- GF_DATABASE_USER=${POSTGRES_USER}
- GF_DATABASE_PASSWORD=${POSTGRES_PASSWORD}
- GF_INSTALL_PLUGINS=flant-statusmap-panel
- GF_INSTALL_PLUGINS=flant-statusmap-panel,redis-datasource
depends_on:
- influxdb
- grafanadb
@ -55,6 +56,8 @@ services:
restart: unless-stopped
networks:
- web
- monitoring
- grafana
environment:
- INFLUXDB_MONITOR_STORE_ENABLED=false
volumes:
@ -72,6 +75,19 @@ services:
- "docker.group=monitoring"
prometheus:
image: prom/prometheus
restart: unless-stopped
networks:
- grafana
- monitoring
- web # also used to get traefik metrics
volumes:
- ./prometheus.yml:/etc/prometheus/prometheus.yml
- ${VOLUMES_PATH}/prometheus:/prometheus
labels:
- "docker.group=monitoring"
##################################################################
# here starts data collection
@ -98,11 +114,10 @@ services:
depends_on:
- influxdb
telegraf_net:
image: telegraf:${TELEGRAF_VERSION}
restart: unless-stopped
env_file:
- ./.env # set environments into container
volumes:
- ./telegraf_net.conf:/etc/telegraf/telegraf.conf:ro
networks:
@ -113,6 +128,7 @@ services:
depends_on:
- influxdb
networks:
grafana:
monitoring:

37
monitoring/prometheus.yml Normal file
View file

@ -0,0 +1,37 @@
# my global config
global:
scrape_interval: 15s # Set the scrape interval to every 15 seconds. Default is every 1 minute.
evaluation_interval: 15s # Evaluate rules every 15 seconds. The default is every 1 minute.
# scrape_timeout is set to the global default (10s).
# Alertmanager configuration
alerting:
alertmanagers:
- static_configs:
- targets:
# - alertmanager:9093
# Load rules once and periodically evaluate them according to the global 'evaluation_interval'.
rule_files:
# - "first_rules.yml"
# - "second_rules.yml"
# A scrape configuration containing exactly one endpoint to scrape:
# Here it's Prometheus itself.
scrape_configs:
# The job name is added as a label `job=<job_name>` to any timeseries scraped from this config.
- job_name: "prometheus"
# metrics_path defaults to '/metrics'
# scheme defaults to 'http'.
static_configs:
- targets: ["localhost:9090"]
- job_name: "traefik"
scrape_interval: 5s
static_configs:
- targets: ["traefik:8080"]
- job_name: "grafana"
static_configs:
- targets: ["grafana:3000"]

View file

@ -82,10 +82,7 @@
# urls = ["udp://127.0.0.1:8089"]
# urls = ["http://127.0.0.1:8086"]
## HTTP Basic Auth
username = "${INFLUXDB_HTTP_BASIC_AUTH_USER}"
password = "${INFLUXDB_HTTP_BASIC_AUTH_PASSWORD}"
urls = ["https://influxdb.florianzirker.de"] # required
urls = ["http://influxdb:8086"]
###############################################################################

5
nextcloud/crontab Normal file → Executable file
View file

@ -1,4 +1,3 @@
*/5 * * * * php -f /var/www/html/cron.php
1,11,21,31,41,51 * * * * php -f /var/www/html/occ files:cleanup
2,12,22,32,42,52 * * * * php -f /var/www/html/occ files:scan --all
#3,13,23,33,43,53 * * * * php -f /var/www/html/occ preview:pre-generate

View file

@ -57,7 +57,12 @@ services:
volumes:
- ${VOLUMES_PATH}/nextcloud_html:/var/www/html
- ${VOLUMES_PATH}/nextcloud_data:/var/www/html/data
- $PWD/crontab:/var/spool/cron/crontabs/www-data:ro
# If I mount my crontab into the container crond is not working any more :(
# docker log should print 11110001 lines
# https://github.com/nextcloud/docker/issues/1775
# https://github.com/nextcloud/docker/issues/1695
# build own cron image?
# - $PWD/crontab:/var/spool/cron/crontabs/www-data
entrypoint: /cron.sh
depends_on:
- db
@ -115,11 +120,12 @@ services:
redis:
image: redis:alpine
image: redis
restart: unless-stopped
command: redis-server --requirepass ${REDIS_HOST_PASSWORD}
networks:
- nextcloud
- monitoring
volumes:
- ${VOLUMES_PATH}/nextcloud_redis:/data
labels:
@ -177,6 +183,12 @@ services:
- LETS_ENCRYPT_ENABLED=false
- EXPORT_URL=http://drawio-export:8000/
restart: unless-stopped
healthcheck:
test: curl -f http://localhost:8080 || exit 1
interval: 1m30s
timeout: 10s
retries: 5
start_period: 20s
labels:
- "traefik.enable=true"
- "traefik.http.routers.drawio.rule=Host(`drawio.${DOMAIN}`)"
@ -191,3 +203,5 @@ networks:
external: true
nextcloud:
mariadb:
monitoring:
external: true

View file

@ -29,6 +29,10 @@ services:
- "--certificatesresolvers.netcup.acme.dnschallenge.delayBeforeCheck=900"
- "--certificatesresolvers.netcup.acme.email=${LETSENCRYPT_MAIL}"
- "--certificatesresolvers.netcup.acme.storage=/letsencrypt/acme.json"
- "--metrics.prometheus=true"
- "--metrics.prometheus.addEntryPointsLabels=true"
- "--metrics.prometheus.addRoutersLabels=true"
- "--metrics.prometheus.addServicesLabels=true"
ports:
- "80:80"
- "443:443"
@ -65,17 +69,17 @@ services:
privileged: true
# whoami:
# image: containous/whoami
# networks:
# - web
# labels:
# - "traefik.enable=true"
# - "traefik.http.routers.whoami.rule=Host(`whoami.${DOMAIN}`)"
# - "traefik.http.routers.whoami.entrypoints=websecure"
# - "traefik.http.routers.whoami.tls.certresolver=netcup"
# - "docker.group=proxy"
# restart: unless-stopped
whoami:
image: containous/whoami
networks:
- web
labels:
- "traefik.enable=true"
- "traefik.http.routers.whoami.rule=Host(`whoami.${DOMAIN}`)"
- "traefik.http.routers.whoami.entrypoints=websecure"
- "traefik.http.routers.whoami.tls.certresolver=netcup"
- "docker.group=proxy"
restart: unless-stopped
networks:

View file

@ -1,9 +1,9 @@
#/bin/bash/
for dir in ./*/ ; do
(cd "$dir" && echo "[$dir]" && docker compose build);
( cd "$dir" && echo "[Building $dir]" && docker compose build --pull );
done
for dir in ./*/ ; do
(cd "$dir" && echo "[$dir]" && docker compose pull);
( cd "$dir" && echo "[Pulling $dir]" && docker compose pull --ignore-buildable );
done

View file

@ -5,9 +5,9 @@ function up {
(cd "$1" && echo "[$1]" && docker compose up -d "${@:2}");
}
up proxy;
up proxy --scale whoami=3;
up monitoring;
up nextcloud --scale app=3 --scale web=2;
up nextcloud --scale app=5 --scale web=3;
up gitea;
up wallabag;
up www;