moved unused to archive
This commit is contained in:
parent
5b03dbf90f
commit
275b8989ec
31 changed files with 0 additions and 918 deletions
|
|
@ -1,75 +0,0 @@
|
|||
version: "3.8"
|
||||
|
||||
services:
|
||||
authelia:
|
||||
image: authelia/authelia:latest
|
||||
container_name: authelia
|
||||
hostname: authelia
|
||||
# user: "1001:998"
|
||||
networks:
|
||||
- netgrimoire
|
||||
environment:
|
||||
- PUID=1001
|
||||
- PGID=998
|
||||
- TZ=America/Chicago
|
||||
volumes:
|
||||
- /DockerVol/authelia/config:/config
|
||||
- /DockerVol/authelia/secrets:/secrets
|
||||
ports:
|
||||
# Optional: usually you do NOT need to publish this if only Caddy talks to it
|
||||
# - "9091:9091"
|
||||
- target: 9091
|
||||
published: 9091
|
||||
protocol: tcp
|
||||
mode: ingress
|
||||
deploy:
|
||||
replicas: 1
|
||||
placement:
|
||||
constraints:
|
||||
- node.hostname == nas
|
||||
labels:
|
||||
# Caddy - Authelia portal URL
|
||||
- caddy=login.wasted-bandwidth.net
|
||||
- caddy.reverse_proxy={{upstreams 9091}}
|
||||
|
||||
# Homepage
|
||||
- homepage.group=Management
|
||||
- homepage.name=Authelia
|
||||
- homepage.icon=authelia.png
|
||||
- homepage.href=https://login.wasted-bandwidth.net
|
||||
- homepage.description=SSO / Forward-Auth
|
||||
|
||||
# Uptime Kuma
|
||||
- kuma.authelia.http.name="Authelia"
|
||||
- kuma.authelia.http.url=http://authelia:9091
|
||||
|
||||
restart: unless-stopped
|
||||
depends_on:
|
||||
- redis
|
||||
|
||||
redis:
|
||||
image: redis:7-alpine
|
||||
container_name: authelia-redis
|
||||
hostname: authelia-redis
|
||||
#user: "1001:998"
|
||||
networks:
|
||||
- netgrimoire
|
||||
environment:
|
||||
- TZ=America/Chicago
|
||||
- PUID=1001
|
||||
- PGID=998
|
||||
command: >
|
||||
sh -c "redis-server --appendonly yes --dir /data"
|
||||
volumes:
|
||||
- /DockerVol/authelia/redis:/data
|
||||
deploy:
|
||||
replicas: 1
|
||||
placement:
|
||||
constraints:
|
||||
- node.hostname == nas
|
||||
|
||||
restart: unless-stopped
|
||||
|
||||
networks:
|
||||
netgrimoire:
|
||||
external: true
|
||||
|
|
@ -1,51 +0,0 @@
|
|||
services:
|
||||
bazarr:
|
||||
image: lscr.io/linuxserver/bazarr:latest
|
||||
networks:
|
||||
- netgrimoire
|
||||
|
||||
environment:
|
||||
TZ: America/Chicago
|
||||
PUID: "1964"
|
||||
PGID: "1964"
|
||||
# Match your template preference (run container as 1001:998);l'k'k'klo[oghhghg[]]
|
||||
# user: "1001:998"
|
||||
|
||||
volumes:
|
||||
- /DockerVol/bazarr/config:/config
|
||||
- /data/nfs/Baxter/Data/:/data:shared
|
||||
|
||||
ports:
|
||||
- target: 6767
|
||||
published: 6767
|
||||
protocol: tcp
|
||||
mode: ingress
|
||||
|
||||
deploy:
|
||||
placement:
|
||||
constraints:
|
||||
- node.hostname == docker4
|
||||
|
||||
labels:
|
||||
# Homepage
|
||||
homepage.group: "Jolly Roger"
|
||||
homepage.name: "Bazarr"
|
||||
homepage.icon: "bazarr.png"
|
||||
homepage.href: "http://bazarr.netgrimoire.com"
|
||||
homepage.description: "Subtitle Search"
|
||||
|
||||
# Kuma
|
||||
kuma.bazarr.http.name: "Bazarr"
|
||||
kuma.bazarr.http.url: "http://bazarr:6767"
|
||||
|
||||
# Caddy
|
||||
caddy: "bazarr.netgrimoire.com"
|
||||
caddy.import: "authentik"
|
||||
caddy.reverse_proxy: "{{upstreams 6767}}"
|
||||
|
||||
# Diun (image update monitoring)
|
||||
diun.enable: "true"
|
||||
|
||||
networks:
|
||||
netgrimoire:
|
||||
external: true
|
||||
|
|
@ -1,53 +0,0 @@
|
|||
services:
|
||||
beets:
|
||||
image: lscr.io/linuxserver/beets:latest
|
||||
networks:
|
||||
- netgrimoire
|
||||
|
||||
environment:
|
||||
TZ: America/Chicago
|
||||
PUID: "1964"
|
||||
PGID: "1964"
|
||||
|
||||
# Match your standard UID:GID execution model
|
||||
#user: "1001:998"
|
||||
|
||||
volumes:
|
||||
- /DockerVol/beets/config:/config
|
||||
- /data/nfs/Baxter/Data/media/music/Collection:/music
|
||||
- /data/nfs/Baxter/Data/media/music/ingest:/downloads
|
||||
|
||||
ports:
|
||||
- target: 8337
|
||||
published: 8337
|
||||
protocol: tcp
|
||||
mode: ingress
|
||||
|
||||
deploy:
|
||||
placement:
|
||||
constraints:
|
||||
- node.hostname == znas
|
||||
|
||||
labels:
|
||||
# Homepage
|
||||
homepage.group: "Media Management"
|
||||
homepage.name: "Beets"
|
||||
homepage.icon: "beets.png"
|
||||
homepage.href: "https://beets.netgrimoire.com"
|
||||
homepage.description: "A Music Manager"
|
||||
|
||||
# Kuma
|
||||
kuma.beets.http.name: "Beets"
|
||||
kuma.beets.http.url: "http://beets:8337"
|
||||
|
||||
# Caddy
|
||||
caddy: "beets.netgrimoire.com"
|
||||
caddy.import: "authentik"
|
||||
caddy.reverse_proxy: "{{upstreams 8337}}"
|
||||
|
||||
# Diun (image update monitoring)
|
||||
diun.enable: "true"
|
||||
|
||||
networks:
|
||||
netgrimoire:
|
||||
external: true
|
||||
|
|
@ -1,41 +0,0 @@
|
|||
services:
|
||||
|
||||
cloudcmd:
|
||||
image: coderaiser/cloudcmd
|
||||
container_name: cloudcmd
|
||||
hostname: cloudcmd
|
||||
networks:
|
||||
- netgrimoire
|
||||
ports:
|
||||
- 8000:8000
|
||||
environment:
|
||||
- PUID=1001
|
||||
- PGID=998
|
||||
- TZ=America/Chicago
|
||||
volumes:
|
||||
- ~:/root
|
||||
- /:/mnt/fs
|
||||
deploy:
|
||||
labels:
|
||||
- homepage.group=Application
|
||||
- homepage.name=Cloud Commander
|
||||
- homepage.icon=cloudcmd.png
|
||||
- homepage.href=http://commander.netgrimoire.com
|
||||
- homepage.description=Cloud Commander
|
||||
- kuma.cloud.http.name="Cloudcmd
|
||||
- kuma.cloud.http.url=http://cloudcmd:8000
|
||||
- caddy=commander.netgrimoire.com
|
||||
- caddy.import=authentik
|
||||
- caddy.reverse_proxy="{{upstreams 8000}}"
|
||||
placement:
|
||||
constraints:
|
||||
- node.hostname == nas
|
||||
|
||||
|
||||
networks:
|
||||
netgrimoire:
|
||||
external: true
|
||||
|
||||
|
||||
|
||||
|
||||
|
|
@ -1,48 +0,0 @@
|
|||
# /config is where plugins, extensions, and configuration files are placed.
|
||||
# /imports is where comics to be imported are placed
|
||||
# /library is where the comics library is maintaned
|
||||
# /more
|
||||
version: "3.6"
|
||||
services:
|
||||
comixed:
|
||||
image: comixed/comixed
|
||||
container_name: comixed
|
||||
environment:
|
||||
- PUID=1001
|
||||
- PGID=998
|
||||
- TZ=America/Chicago
|
||||
volumes:
|
||||
- /DockerVol/comixed/config:/config
|
||||
- /data/nfs/Baxter/Data/media/comics/library:/library:shared
|
||||
- /data/nfs/Baxter/Data/media/comics/downloads:/imports:shared
|
||||
ports:
|
||||
- 7171:7171
|
||||
entrypoint: ["bash", "/app/comixed-release/bin/docker-run.sh", "-L", "/library/comixed.log", "-c", "/config"]
|
||||
restart: unless-stopped
|
||||
networks:
|
||||
- netgrimoire
|
||||
|
||||
deploy:
|
||||
labels:
|
||||
homepage.group: "Application"
|
||||
homepage.name: "Comixed"
|
||||
homepage.icon: "comixed.png"
|
||||
homepage.href: "https://comics.netgrimoire.com"
|
||||
homepage.description: "Task Manager"
|
||||
kuma.cxd.http.name: "Vikunja"
|
||||
kuma.cxd.http.url: "http://comixed:7171"
|
||||
caddy: "comics.netgrimoire.com"
|
||||
caddy.import: "authentik"
|
||||
caddy.reverse_proxy: "{{upstreams 7171}}"
|
||||
placement:
|
||||
constraints:
|
||||
- node.hostname == nas
|
||||
|
||||
|
||||
networks:
|
||||
netgrimoire:
|
||||
external: true
|
||||
|
||||
|
||||
|
||||
|
||||
|
|
@ -1,41 +0,0 @@
|
|||
services:
|
||||
|
||||
cloudcmd:
|
||||
image: coderaiser/cloudcmd
|
||||
container_name: cloudcmd
|
||||
hostname: cloudcmd
|
||||
networks:
|
||||
- netgrimoire
|
||||
ports:
|
||||
- 8024:8000
|
||||
environment:
|
||||
- TZ=America/Chicago
|
||||
user: "1001:998"
|
||||
volumes:
|
||||
- ~:/root
|
||||
- /:/mnt/fs
|
||||
deploy:
|
||||
placement:
|
||||
constraints:
|
||||
- node.hostname == nas
|
||||
labels:
|
||||
# Diun
|
||||
- "diun.enable=true"
|
||||
|
||||
# Homepage
|
||||
- "homepage.group=Applications"
|
||||
- "homepage.name=Cloud Commander"
|
||||
- "homepage.icon=mdi-cloud"
|
||||
- "homepage.href=https://commander.netgrimoire.com"
|
||||
- "homepage.description=Web File Manager"
|
||||
|
||||
|
||||
# Caddy
|
||||
- "caddy=commander.netgrimoire.com"
|
||||
- "caddy.reverse_proxy={{upstreams 8000}}"
|
||||
|
||||
restart: unless-stopped
|
||||
|
||||
networks:
|
||||
netgrimoire:
|
||||
external: true
|
||||
|
|
@ -1,60 +0,0 @@
|
|||
services:
|
||||
configarr:
|
||||
image: ghcr.io/raydak-labs/configarr:latest
|
||||
user: "1001:998"
|
||||
environment:
|
||||
PUID: "1001"
|
||||
PGID: "998"
|
||||
TZ: America/Chicago
|
||||
|
||||
volumes:
|
||||
- /data/nfs/Baxter/Docker/configarr/config:/app/config
|
||||
- /data/nfs/Baxter/Docker/configarr/repos:/app/repos
|
||||
- /data/nfs/Baxter/Docker/configarr/cfs:/app/cfs
|
||||
- /data/nfs/Baxter/Docker/configarr/templates:/app/templates
|
||||
|
||||
networks:
|
||||
- netgrimoire
|
||||
|
||||
deploy:
|
||||
mode: replicated
|
||||
replicas: 1
|
||||
restart_policy:
|
||||
condition: any
|
||||
delay: 5s
|
||||
max_attempts: 3
|
||||
window: 30s
|
||||
|
||||
labels:
|
||||
# -------------------------
|
||||
# Homepage Dashboard Labels
|
||||
# -------------------------
|
||||
homepage.group: "Jolly Roger"
|
||||
homepage.name: "Configarr"
|
||||
homepage.icon: "si-config"
|
||||
homepage.href: "https://configarr.netgrimoire.com"
|
||||
homepage.description: "Automatically sync TRaSH formats & configs"
|
||||
|
||||
# -------------------------
|
||||
# Kuma Monitoring Labels
|
||||
# -------------------------
|
||||
kuma.configarr.http.name: "Configarr"
|
||||
kuma.configarr.http.url: "https://configarr.netgrimoire.com"
|
||||
|
||||
# -------------------------
|
||||
# Caddy Reverse Proxy
|
||||
# (Swarm label syntax works the same)
|
||||
# -------------------------
|
||||
caddy: configarr.netgrimoire.com
|
||||
caddy.import: "authentik"
|
||||
|
||||
caddy.reverse_proxy: "{{upstreams 8000}}"
|
||||
|
||||
# -------------------------
|
||||
# Diun Image Monitoring
|
||||
# -------------------------
|
||||
diun.enable: "true"
|
||||
|
||||
networks:
|
||||
netgrimoire:
|
||||
external: true
|
||||
|
|
@ -1,34 +0,0 @@
|
|||
services:
|
||||
dailytxt:
|
||||
# choose the correct image tag
|
||||
image: phitux/dailytxt:2.x.x
|
||||
container_name: dailytxt
|
||||
restart: unless-stopped
|
||||
volumes:
|
||||
# Change the left path to your needs
|
||||
- ./data:/data
|
||||
environment:
|
||||
# Create a secret token by running: openssl rand -base64 32
|
||||
- SECRET_TOKEN=...
|
||||
|
||||
# If you want to have the json-files pretty-printed, set some indent.
|
||||
# (Otherwise just remove the line)
|
||||
- INDENT=4
|
||||
|
||||
# Allow new user registrations.
|
||||
# I strongly recommend to keep this disabled except for the first user.
|
||||
# You can later temporarily enable it again in the admin panel.
|
||||
- ALLOW_REGISTRATION=true
|
||||
|
||||
# Set the Admin-Password (for the admin-panel).
|
||||
- ADMIN_PASSWORD=your_admin_password
|
||||
|
||||
# After how many days shall the login-cookie expire?
|
||||
- LOGOUT_AFTER_DAYS=40
|
||||
|
||||
# Set the BASE_PATH if you are running DailyTxT under a subpath (e.g. /dailytxt).
|
||||
# - BASE_PATH=/dailytxt
|
||||
ports:
|
||||
# Change the left port to your needs.
|
||||
# You often would only see 8000:80. But this way, port 8000 is publicly accessible (without TLS!).
|
||||
- 127.0.0.1:8000:80
|
||||
|
|
@ -1,47 +0,0 @@
|
|||
services:
|
||||
dumbterm:
|
||||
container_name: dumbterm
|
||||
image: dumbwareio/dumbterm:latest
|
||||
ports:
|
||||
- 8062:3000
|
||||
networks:
|
||||
- netgrimoire
|
||||
# user: 1001:998
|
||||
environment:
|
||||
# Container timezone
|
||||
TZ: America/Chicago
|
||||
# The title shown in the web interface
|
||||
SITE_TITLE: https://cli.netgrimoire.com
|
||||
# Optional PIN protection (leave empty to disable)
|
||||
DUMBTERM_PIN: 2810
|
||||
# The base URL for the application
|
||||
BASE_URL: http://localhost:3000 # Use ALLOWED_ORIGINS below to restrict cors to specific origins
|
||||
ENABLE_STARSHIP: "true" # Enable starship prompt
|
||||
LOCKOUT_TIME: 15 # Minutes
|
||||
MAX_SESSION_AGE: 24 # Hours
|
||||
|
||||
|
||||
deploy:
|
||||
labels:
|
||||
- homepage.group=Remote Access
|
||||
- homepage.name=Dumbterm
|
||||
- homepage.icon=dumbterm.png
|
||||
- homepage.href=https://cli.netgrimoire.com
|
||||
- homepage.description=Terminal
|
||||
- kuma.cli.http.name="dumbterm"
|
||||
- kuma.cli.http.url=http://dumbterm:3000
|
||||
- caddy=cli.netgrimoire.com
|
||||
- caddy.import=authentik
|
||||
- caddy.reverse_proxy="{{upstreams 3000}}"
|
||||
|
||||
volumes:
|
||||
#- /data/nfs/Baxter/Docker/dumbterm/config:/root/.config
|
||||
#- /data/nfs/Baxter/Docker/dumbterm/:/root/data
|
||||
- /data/nfs/Baxter/Docker/dumbterm/root:/root
|
||||
|
||||
|
||||
networks:
|
||||
netgrimoire:
|
||||
external: true
|
||||
|
||||
|
||||
|
|
@ -1,42 +0,0 @@
|
|||
services:
|
||||
freshrss:
|
||||
image: lscr.io/linuxserver/freshrss:latest
|
||||
container_name: freshrss
|
||||
user: "1001:998"
|
||||
# user: "0:0"
|
||||
environment:
|
||||
TZ: "America/Chicago"
|
||||
PUID: "1001"
|
||||
PGID: "998"
|
||||
|
||||
networks:
|
||||
- netgrimoire
|
||||
|
||||
volumes:
|
||||
- /data/nfs/Baxter/Docker/freshrss:/config
|
||||
|
||||
deploy:
|
||||
endpoint_mode: dnsrr
|
||||
labels:
|
||||
# Homepage
|
||||
homepage.group: "Services"
|
||||
homepage.name: "FreshRSS"
|
||||
homepage.icon: "rss"
|
||||
homepage.href: "https://rss.netgrimoire.com"
|
||||
|
||||
# Kuma integration
|
||||
kuma.freshrss.http.name: "FreshRSS"
|
||||
kuma.freshrss.http.url: "https://rss.netgrimoire.com"
|
||||
|
||||
# Caddy (replace with your domain)
|
||||
caddy: "https://rss.netgrimoire.com"
|
||||
caddy.reverse_proxy: "{{upstreams 80}}"
|
||||
|
||||
# Diun
|
||||
diun.enable: "true"
|
||||
|
||||
restart: unless-stopped
|
||||
|
||||
networks:
|
||||
netgrimoire:
|
||||
external: true
|
||||
|
|
@ -1,29 +0,0 @@
|
|||
version: "3.9"
|
||||
|
||||
services:
|
||||
forgejo-runner:
|
||||
image: data.forgejo.org/forgejo/runner:4.0.0
|
||||
user: "1001:998"
|
||||
environment:
|
||||
PUID: "1001"
|
||||
PGID: "998"
|
||||
TZ: "America/Chicago"
|
||||
volumes:
|
||||
- type: bind
|
||||
source: /DockerVol/forgejo-runner
|
||||
target: /data
|
||||
- type: bind
|
||||
source: /var/run/docker.sock
|
||||
target: /var/run/docker.sock
|
||||
networks:
|
||||
- netgrimoire
|
||||
command: ["forgejo-runner", "daemon"]
|
||||
deploy:
|
||||
mode: global
|
||||
placement:
|
||||
constraints:
|
||||
- node.role == manager
|
||||
|
||||
networks:
|
||||
netgrimoire:
|
||||
external: true
|
||||
|
|
@ -1,51 +0,0 @@
|
|||
services:
|
||||
kavita:
|
||||
image: jvmilazz0/kavita:latest # Change latest to nightly for latest develop builds (can't go back to stable)
|
||||
networks:
|
||||
- netgrimoire
|
||||
|
||||
volumes:
|
||||
- /data/nfs/Baxter/Data/media/comics:/comics:shared # Use as many as you want
|
||||
# - ./books:/books #
|
||||
- /data/nfs/Baxter/Docker/Kavita/config:/kavita/config # Change './data if you want to have the config files in a different place.
|
||||
# /kavita/config must not be changed
|
||||
|
||||
environment:
|
||||
TZ: America/Chicago
|
||||
PUID: "1001"
|
||||
PGID: "998"
|
||||
|
||||
# Match your standard UID:GID execution model
|
||||
#user: "1001:998"
|
||||
|
||||
ports:
|
||||
# Change the public port (the first 5000) if you have conflicts with other services
|
||||
- target: 5000
|
||||
published: 8054
|
||||
protocol: tcp
|
||||
mode: ingress
|
||||
|
||||
deploy:
|
||||
labels:
|
||||
# Homepage
|
||||
homepage.group: "PNCHarris Apps"
|
||||
homepage.name: "Kavita"
|
||||
homepage.icon: "kavita.png"
|
||||
homepage.href: "https://kavita.netgrimoire.com"
|
||||
homepage.description: "Comic Book Reader"
|
||||
|
||||
# Kuma
|
||||
kuma.kavita.http.name: "Kavita"
|
||||
kuma.kavita.http.url: "http://kavita:5000"
|
||||
|
||||
# Caddy
|
||||
caddy: "kavita.netgrimoire.com"
|
||||
caddy.import: "authentik"
|
||||
caddy.reverse_proxy: "Kavita:5000"
|
||||
|
||||
# Diun (image update monitoring)
|
||||
diun.enable: "true"
|
||||
|
||||
networks:
|
||||
netgrimoire:
|
||||
external: true
|
||||
|
|
@ -1,53 +0,0 @@
|
|||
services:
|
||||
pinchflat:
|
||||
image: ghcr.io/kieraneglin/pinchflat:latest
|
||||
networks:
|
||||
- netgrimoire
|
||||
|
||||
environment:
|
||||
# Set the timezone to your local timezone
|
||||
TZ: America/Chicago
|
||||
PUID: "1001"
|
||||
PGID: "998"
|
||||
|
||||
# Match your standard UID:GID execution model
|
||||
#user: "1001:998"
|
||||
|
||||
ports:
|
||||
- target: 8945
|
||||
published: 8945
|
||||
protocol: tcp
|
||||
mode: ingress
|
||||
|
||||
volumes:
|
||||
- /DockerVol/pinchflat/config:/config
|
||||
- /data/nfs/Baxter/Data/media/other/pinchflat:/downloads
|
||||
|
||||
deploy:
|
||||
placement:
|
||||
constraints:
|
||||
- node.hostname == nas
|
||||
|
||||
labels:
|
||||
# Homepage
|
||||
homepage.group: "Downloaders"
|
||||
homepage.name: "PinchFlat"
|
||||
homepage.icon: "pinchflat.png"
|
||||
homepage.href: "https://pinchflat.netgrimoire.com"
|
||||
homepage.description: "YouTube Library"
|
||||
|
||||
# Kuma
|
||||
kuma.pf.http.name: "PinchFlat"
|
||||
kuma.pf.http.url: "http://pinchflat:8945"
|
||||
|
||||
# Caddy
|
||||
caddy: "pinchflat.netgrimoire.com"
|
||||
caddy.import: "authentik"
|
||||
caddy.reverse_proxy: "pinchflat:8945"
|
||||
|
||||
# Diun (image update monitoring)
|
||||
diun.enable: "true"
|
||||
|
||||
networks:
|
||||
netgrimoire:
|
||||
external: true
|
||||
|
|
@ -1,59 +0,0 @@
|
|||
services:
|
||||
profilarr:
|
||||
image: santiagosayshey/profilarr:latest
|
||||
container_name: profilarr
|
||||
|
||||
#user: "1001:998"
|
||||
environment:
|
||||
- PUID=1001
|
||||
- PGID=998
|
||||
- TZ=America/Chicago
|
||||
|
||||
ports:
|
||||
- "6868:6868"
|
||||
|
||||
volumes:
|
||||
- /DockerVol/profilarr:/config
|
||||
|
||||
restart: unless-stopped
|
||||
|
||||
networks:
|
||||
- netgrimoire
|
||||
|
||||
deploy:
|
||||
placement:
|
||||
constraints:
|
||||
- node.hostname == docker4
|
||||
|
||||
labels:
|
||||
# -------------------------
|
||||
# Diun (image update notify)
|
||||
# -------------------------
|
||||
diun.enable: "true"
|
||||
|
||||
# -------------------------
|
||||
# Uptime Kuma (your standard labels)
|
||||
# -------------------------
|
||||
kuma.profilarr.http.name: "profilarr"
|
||||
kuma.profilarr.http.url: "http://profilarr.netgrimoire.com"
|
||||
|
||||
# -------------------------
|
||||
# Homepage (no homepage.weight per your preference)
|
||||
# -------------------------
|
||||
homepage.group: "Jolly Roger"
|
||||
homepage.name: "Profilarr"
|
||||
homepage.description: "Profilarr"
|
||||
homepage.href: "https://profilarr.netgrimoire.com"
|
||||
homepage.icon: "profilarr"
|
||||
|
||||
|
||||
# -------------------------
|
||||
# Caddy (your domain + authentik import)
|
||||
# -------------------------
|
||||
caddy: "profilarr.netgrimoire.com"
|
||||
caddy.reverse_proxy: "{{upstreams 6868}}"
|
||||
caddy.import: "authentik"
|
||||
|
||||
networks:
|
||||
netgrimoire:
|
||||
external: true
|
||||
|
|
@ -1,40 +0,0 @@
|
|||
services:
|
||||
readarr:
|
||||
image: blampe/rreading-glasses:latest
|
||||
container_name: readarr
|
||||
hostname: readarr
|
||||
environment:
|
||||
- PUID=1001
|
||||
- PGID=998
|
||||
- TZ=America/Chicago
|
||||
networks:
|
||||
- netgrimoire
|
||||
volumes:
|
||||
- /DockerVol/readarr/config:/config
|
||||
- /data/nfs/Baxter/Data:/data:shared
|
||||
ports:
|
||||
- 8787:8787
|
||||
restart: unless-stopped
|
||||
deploy:
|
||||
placement:
|
||||
constraints:
|
||||
- node.hostname == docker4
|
||||
labels:
|
||||
- homepage.group=Jolly Roger
|
||||
- homepage.name=Readarr
|
||||
- homepage.icon=readarr.png
|
||||
- homepage.href=http://readarr.netgrimoire.com
|
||||
- homepage.description=Ebook Library
|
||||
- homepage.widget.type=readarr
|
||||
- homepage.widget.url=http://readarr:8787
|
||||
- homepage.widget.key=78954fcf696e4da9b2e9391a54e87478
|
||||
- kuma.readarr.http.name="Readarr"
|
||||
- kuma.readarr.http.url=http://readarr:8787
|
||||
- caddy=readarr.netgrimoire.com
|
||||
- caddy.import=authentik
|
||||
- caddy.reverse_proxy="{{upstreams 8787}}"
|
||||
|
||||
|
||||
networks:
|
||||
netgrimoire:
|
||||
external: true
|
||||
|
|
@ -1,53 +0,0 @@
|
|||
version: "3.8"
|
||||
|
||||
services:
|
||||
recyclarr:
|
||||
image: ghcr.io/recyclarr/recyclarr:latest
|
||||
user: "1001:998"
|
||||
environment:
|
||||
- PUID=1001
|
||||
- PGID=998
|
||||
- TZ=America/Chicago
|
||||
|
||||
# Run on an interval (simple + reliable)
|
||||
# Adjust to taste: 6h, 12h, 24h, etc.
|
||||
- RECYCLARR_RUN_MODE=interval
|
||||
- RECYCLARR_INTERVAL=12h
|
||||
|
||||
# Optional: if you want extra logging
|
||||
# - RECYCLARR_LOG_LEVEL=Information
|
||||
|
||||
volumes:
|
||||
- /data/nfs/Baxter/Docker/recyclarr:/config
|
||||
|
||||
networks:
|
||||
- netgrimoire
|
||||
|
||||
deploy:
|
||||
placement:
|
||||
constraints:
|
||||
- node.hostname == docker4
|
||||
mode: replicated
|
||||
replicas: 1
|
||||
|
||||
restart_policy:
|
||||
condition: any
|
||||
delay: 10s
|
||||
max_attempts: 0
|
||||
window: 30s
|
||||
|
||||
update_config:
|
||||
order: start-first
|
||||
parallelism: 1
|
||||
delay: 10s
|
||||
|
||||
labels:
|
||||
# -------------------------
|
||||
# Diun (image update notify)
|
||||
# -------------------------
|
||||
diun.enable: "true"
|
||||
|
||||
|
||||
networks:
|
||||
netgrimoire:
|
||||
external: true
|
||||
|
|
@ -1,41 +0,0 @@
|
|||
services:
|
||||
tunarr:
|
||||
image: chrisbenincasa/tunarr
|
||||
container_name: tunarr
|
||||
hostname: tunarr
|
||||
networks:
|
||||
- netgrimoire
|
||||
ports:
|
||||
- "8051:8000" # Web UI
|
||||
- "8052:8001" # Streaming endpoint
|
||||
volumes:
|
||||
- /data/nfs/Baxter/Docker/tunarr/config:/config
|
||||
- /data/nfs/Baxter/Data/:/media # Adjust to your media directory
|
||||
environment:
|
||||
- TZ=America/Chicago
|
||||
- PGID=998
|
||||
- PUID=1001
|
||||
user: "1001:998"
|
||||
restart: unless-stopped
|
||||
deploy:
|
||||
labels:
|
||||
- homepage.group=Jolly Roger
|
||||
- homepage.name=Tunarr
|
||||
- homepage.icon=tunarr.png
|
||||
- homepage.href=https://tunarr.netgrimoire.com
|
||||
- homepage.description=Virtual TV Channel Creator
|
||||
- kuma.tunarr.http.name="Tunarr"
|
||||
- kuma.tunarr.http.url=http://tunarr:8000
|
||||
- caddy=tunarr.netgrimoire.com
|
||||
- caddy.import=authentik
|
||||
- caddy.reverse_proxy="{{upstreams 8000}}"
|
||||
placement:
|
||||
constraints:
|
||||
- node.labels.general == true
|
||||
|
||||
networks:
|
||||
netgrimoire:
|
||||
external: true
|
||||
|
||||
|
||||
|
||||
Loading…
Add table
Add a link
Reference in a new issue