Compare commits
1 commit
main
...
renovate/n
Author | SHA1 | Date | |
---|---|---|---|
bc4a1edb5f |
58 changed files with 977 additions and 523 deletions
|
@ -1,22 +0,0 @@
|
||||||
on:
|
|
||||||
push:
|
|
||||||
paths:
|
|
||||||
- containers/deployer/**
|
|
||||||
- .forgejo/workflows/build-deployer.yaml
|
|
||||||
jobs:
|
|
||||||
build-deployer:
|
|
||||||
runs-on: docker
|
|
||||||
container:
|
|
||||||
image: library/docker:dind
|
|
||||||
steps:
|
|
||||||
- run: apk add --no-cache nodejs git
|
|
||||||
- name: login to container registry
|
|
||||||
run: echo "${{ secrets.DEPLOY_TOKEN }}" | docker login --username ${{ secrets.DEPLOY_USER }} --password-stdin git.janky.solutions
|
|
||||||
- name: build container image
|
|
||||||
uses: docker/build-push-action@v6
|
|
||||||
with:
|
|
||||||
file: Containerfile
|
|
||||||
context: "{{defaultContext}}:containers/deployer"
|
|
||||||
tags: git.janky.solutions/jankysolutions/infra/deployer:latest
|
|
||||||
platforms: linux/amd64
|
|
||||||
push: ${{ github.ref == 'refs/heads/main' }}
|
|
|
@ -18,4 +18,4 @@ jobs:
|
||||||
context: FreeTakServer
|
context: FreeTakServer
|
||||||
tags: git.janky.solutions/jankysolutions/infra/freetakserver:latest
|
tags: git.janky.solutions/jankysolutions/infra/freetakserver:latest
|
||||||
platforms: linux/amd64
|
platforms: linux/amd64
|
||||||
push: ${{ github.ref == 'refs/heads/main' }}
|
push: true
|
||||||
|
|
|
@ -19,4 +19,4 @@ jobs:
|
||||||
context: "{{defaultContext}}:containers/keycloak"
|
context: "{{defaultContext}}:containers/keycloak"
|
||||||
tags: git.janky.solutions/jankysolutions/infra/keycloak:latest
|
tags: git.janky.solutions/jankysolutions/infra/keycloak:latest
|
||||||
platforms: linux/amd64
|
platforms: linux/amd64
|
||||||
push: ${{ github.ref == 'refs/heads/main' }}
|
push: true
|
||||||
|
|
|
@ -19,4 +19,4 @@ jobs:
|
||||||
context: "{{defaultContext}}:containers/openbao-csi-provider"
|
context: "{{defaultContext}}:containers/openbao-csi-provider"
|
||||||
tags: git.janky.solutions/jankysolutions/infra/openbao-csi-provider:latest
|
tags: git.janky.solutions/jankysolutions/infra/openbao-csi-provider:latest
|
||||||
platforms: linux/amd64
|
platforms: linux/amd64
|
||||||
push: ${{ github.ref == 'refs/heads/main' }}
|
push: true
|
||||||
|
|
|
@ -19,4 +19,4 @@ jobs:
|
||||||
context: "{{defaultContext}}:containers/openbao"
|
context: "{{defaultContext}}:containers/openbao"
|
||||||
tags: git.janky.solutions/jankysolutions/infra/openbao:latest
|
tags: git.janky.solutions/jankysolutions/infra/openbao:latest
|
||||||
platforms: linux/amd64
|
platforms: linux/amd64
|
||||||
push: ${{ github.ref == 'refs/heads/main' }}
|
push: true
|
||||||
|
|
|
@ -18,4 +18,4 @@ jobs:
|
||||||
context: pethublocal
|
context: pethublocal
|
||||||
tags: git.janky.solutions/jankysolutions/infra/pethublocal:latest
|
tags: git.janky.solutions/jankysolutions/infra/pethublocal:latest
|
||||||
platforms: linux/amd64
|
platforms: linux/amd64
|
||||||
push: ${{ github.ref == 'refs/heads/main' }}
|
push: true
|
||||||
|
|
|
@ -18,4 +18,4 @@ jobs:
|
||||||
file: containers/synapse/Containerfile
|
file: containers/synapse/Containerfile
|
||||||
tags: git.janky.solutions/jankysolutions/infra/synapse:latest
|
tags: git.janky.solutions/jankysolutions/infra/synapse:latest
|
||||||
platforms: linux/amd64
|
platforms: linux/amd64
|
||||||
push: ${{ github.ref == 'refs/heads/main' }}
|
push: true
|
||||||
|
|
|
@ -19,4 +19,4 @@ jobs:
|
||||||
context: "{{defaultContext}}:containers/traefik-forward-auth"
|
context: "{{defaultContext}}:containers/traefik-forward-auth"
|
||||||
tags: git.janky.solutions/jankysolutions/infra/traefik-forward-auth:latest
|
tags: git.janky.solutions/jankysolutions/infra/traefik-forward-auth:latest
|
||||||
platforms: linux/amd64
|
platforms: linux/amd64
|
||||||
push: ${{ github.ref == 'refs/heads/main' }}
|
push: true
|
||||||
|
|
|
@ -1,37 +0,0 @@
|
||||||
on:
|
|
||||||
push:
|
|
||||||
paths:
|
|
||||||
- k8s/**
|
|
||||||
- .forgejo/workflows/k8s-diff-and-deploy.yaml
|
|
||||||
jobs:
|
|
||||||
diff-and-deploy:
|
|
||||||
runs-on: ubuntu-latest
|
|
||||||
container:
|
|
||||||
image: git.janky.solutions/jankysolutions/infra/deployer:latest
|
|
||||||
steps:
|
|
||||||
- uses: actions/checkout@v4
|
|
||||||
- name: kubectl diff and deploy
|
|
||||||
run: |
|
|
||||||
set -euo pipefail
|
|
||||||
echo "${{ secrets.KUBERNETES_CLIENT_CONFIG }}" > ~/.kube/config
|
|
||||||
|
|
||||||
for component in k8s/*; do
|
|
||||||
if [ ! -d "${component}" ]; then
|
|
||||||
continue
|
|
||||||
fi
|
|
||||||
|
|
||||||
touch "${component}/secrets.yaml"
|
|
||||||
|
|
||||||
echo "👀 $ kubectl diff -k ${component}"
|
|
||||||
kubectl diff -k "${component}" || echo
|
|
||||||
|
|
||||||
if [[ "${GITHUB_REF_NAME}" == "main" ]]; then
|
|
||||||
echo "🚀 $ kubectl apply -k ${component}"
|
|
||||||
if [[ "${component}" == "k8s/operators" ]]; then
|
|
||||||
kubectl apply -k "${component}" --server-side
|
|
||||||
else
|
|
||||||
kubectl apply -k "${component}"
|
|
||||||
fi
|
|
||||||
echo
|
|
||||||
fi
|
|
||||||
done
|
|
|
@ -1,3 +0,0 @@
|
||||||
FROM library/alpine:3.20
|
|
||||||
RUN apk add --no-cache nodejs git bash helm kubectl
|
|
||||||
RUN mkdir -p ~/.kube
|
|
|
@ -1,4 +1,4 @@
|
||||||
FROM node:22 AS web-build
|
FROM node:23 AS web-build
|
||||||
RUN git clone -b v2.0.0 https://github.com/openbao/openbao /usr/src/openbao
|
RUN git clone -b v2.0.0 https://github.com/openbao/openbao /usr/src/openbao
|
||||||
WORKDIR /usr/src/openbao
|
WORKDIR /usr/src/openbao
|
||||||
RUN make ember-dist
|
RUN make ember-dist
|
||||||
|
|
|
@ -1,6 +1,4 @@
|
||||||
FROM matrixdotorg/synapse:v1.119.0
|
FROM matrixdotorg/synapse:v1.117.0
|
||||||
RUN pip install boto3 humanize tqdm
|
RUN pip install boto3 humanize tqdm
|
||||||
# there is probably a better way to figure out where the site packages are
|
RUN curl -Lo /usr/local/lib/python3.11/site-packages/s3_storage_provider.py https://github.com/matrix-org/synapse-s3-storage-provider/raw/v1.4.0/s3_storage_provider.py
|
||||||
# this used to be hard coded to /usr/local/lib/python3.11/site-packages but then synapse updated it's minor python version and it broke
|
|
||||||
RUN curl -Lo $(python -c 'import sys; print([x for x in sys.path if "site-packages" in x][0])')/s3_storage_provider.py https://github.com/matrix-org/synapse-s3-storage-provider/raw/v1.4.0/s3_storage_provider.py
|
|
||||||
RUN curl -L https://github.com/matrix-org/synapse-s3-storage-provider/raw/main/scripts/s3_media_upload | sed "s#/usr/bin/env python#/usr/local/bin/python#" > /usr/local/bin/s3_media_upload && chmod +x /usr/local/bin/s3_media_upload
|
RUN curl -L https://github.com/matrix-org/synapse-s3-storage-provider/raw/main/scripts/s3_media_upload | sed "s#/usr/bin/env python#/usr/local/bin/python#" > /usr/local/bin/s3_media_upload && chmod +x /usr/local/bin/s3_media_upload
|
||||||
|
|
|
@ -7,5 +7,5 @@ helmCharts:
|
||||||
enabled: false # default, bitwarden-sdk-server doesn't work with vaultwarden (https://github.com/external-secrets/bitwarden-sdk-server/issues/18)
|
enabled: false # default, bitwarden-sdk-server doesn't work with vaultwarden (https://github.com/external-secrets/bitwarden-sdk-server/issues/18)
|
||||||
namespace: external-secrets
|
namespace: external-secrets
|
||||||
releaseName: external-secrets
|
releaseName: external-secrets
|
||||||
version: 0.10.5
|
version: 0.10.4
|
||||||
repo: https://charts.external-secrets.io
|
repo: https://charts.external-secrets.io
|
||||||
|
|
|
@ -26,6 +26,13 @@ monitoring:
|
||||||
ansible_host: 10.5.1.251
|
ansible_host: 10.5.1.251
|
||||||
home_network: true
|
home_network: true
|
||||||
|
|
||||||
|
matrix.home.finn.io:
|
||||||
|
ansible_host: 10.5.1.34
|
||||||
|
home_network: true
|
||||||
|
logs:
|
||||||
|
jobs:
|
||||||
|
synapse: /var/log/matrix-synapse/homeserver.log
|
||||||
|
|
||||||
minio.home.finn.io:
|
minio.home.finn.io:
|
||||||
ansible_host: 10.5.1.250
|
ansible_host: 10.5.1.250
|
||||||
home_network: true
|
home_network: true
|
||||||
|
@ -48,15 +55,6 @@ monitoring:
|
||||||
jobs:
|
jobs:
|
||||||
minecraft: /var/minecraft/logs/*.log
|
minecraft: /var/minecraft/logs/*.log
|
||||||
|
|
||||||
freepbx:
|
|
||||||
ansible_host: 10.5.1.169
|
|
||||||
home_network: true
|
|
||||||
logs:
|
|
||||||
jobs:
|
|
||||||
apache2: /var/log/apache2/*.log
|
|
||||||
redis: /var/log/redis/*.log
|
|
||||||
asterisk: /var/log/asterisk/*.log
|
|
||||||
|
|
||||||
authentik:
|
authentik:
|
||||||
hosts:
|
hosts:
|
||||||
authentik.home.finn.io:
|
authentik.home.finn.io:
|
||||||
|
|
|
@ -1,20 +0,0 @@
|
||||||
apiVersion: rbac.authorization.k8s.io/v1
|
|
||||||
kind: ClusterRole
|
|
||||||
metadata:
|
|
||||||
name: infra-deployer
|
|
||||||
rules:
|
|
||||||
- apiGroups: ["*"]
|
|
||||||
resources: ["*"]
|
|
||||||
verbs: ["*"]
|
|
||||||
---
|
|
||||||
apiVersion: rbac.authorization.k8s.io/v1
|
|
||||||
kind: ClusterRoleBinding
|
|
||||||
metadata:
|
|
||||||
name: infra-deployer
|
|
||||||
roleRef:
|
|
||||||
apiGroup: rbac.authorization.k8s.io
|
|
||||||
kind: ClusterRole
|
|
||||||
name: infra-deployer
|
|
||||||
subjects:
|
|
||||||
- kind: User
|
|
||||||
name: infra-deployer
|
|
|
@ -1,5 +0,0 @@
|
||||||
apiVersion: kustomize.config.k8s.io/v1beta1
|
|
||||||
kind: Kustomization
|
|
||||||
resources:
|
|
||||||
- infra-deployer.yaml
|
|
||||||
- matrix-bridge-meshtastic-deployer.yaml
|
|
|
@ -1,22 +0,0 @@
|
||||||
apiVersion: rbac.authorization.k8s.io/v1
|
|
||||||
kind: Role
|
|
||||||
metadata:
|
|
||||||
name: matrix-bridge-meshtastic-deployer
|
|
||||||
namespace: meshtastic
|
|
||||||
rules:
|
|
||||||
- apiGroups: ["apps"]
|
|
||||||
resources: ["deployments"]
|
|
||||||
verbs: ["get", "patch"]
|
|
||||||
---
|
|
||||||
apiVersion: rbac.authorization.k8s.io/v1
|
|
||||||
kind: RoleBinding
|
|
||||||
metadata:
|
|
||||||
name: matrix-bridge-meshtastic-deployer
|
|
||||||
namespace: meshtastic
|
|
||||||
roleRef:
|
|
||||||
apiGroup: rbac.authorization.k8s.io
|
|
||||||
kind: Role
|
|
||||||
name: matrix-bridge-meshtastic-deployer
|
|
||||||
subjects:
|
|
||||||
- kind: User
|
|
||||||
name: matrix-bridge-meshtastic-deployer
|
|
|
@ -1,62 +0,0 @@
|
||||||
apiVersion: batch/v1
|
|
||||||
kind: CronJob
|
|
||||||
metadata:
|
|
||||||
name: forgejo-secret-sync
|
|
||||||
spec:
|
|
||||||
schedule: "0 0 * * *"
|
|
||||||
jobTemplate:
|
|
||||||
spec:
|
|
||||||
template:
|
|
||||||
spec:
|
|
||||||
containers:
|
|
||||||
- name: secret-sync
|
|
||||||
image: library/python:3
|
|
||||||
command:
|
|
||||||
- bash
|
|
||||||
- -c
|
|
||||||
- pip install requests && python /code/forgejo-secret-sync.py
|
|
||||||
env:
|
|
||||||
- name: REPO_MAPPINGS
|
|
||||||
value: |
|
|
||||||
[
|
|
||||||
{"k8s_name": "infra-deployer", "owner": "JankySolutions", "repo": "infra"}
|
|
||||||
]
|
|
||||||
envFrom:
|
|
||||||
- secretRef:
|
|
||||||
name: forgejo-secret-sync
|
|
||||||
volumeMounts:
|
|
||||||
- name: code
|
|
||||||
mountPath: /code
|
|
||||||
- name: host-tls
|
|
||||||
mountPath: /var/lib/rancher/k3s/server/tls
|
|
||||||
restartPolicy: OnFailure
|
|
||||||
affinity:
|
|
||||||
nodeAffinity:
|
|
||||||
requiredDuringSchedulingIgnoredDuringExecution:
|
|
||||||
nodeSelectorTerms:
|
|
||||||
- matchExpressions:
|
|
||||||
- key: node-role.kubernetes.io/control-plane
|
|
||||||
operator: In
|
|
||||||
values: ["true"]
|
|
||||||
volumes:
|
|
||||||
- name: code
|
|
||||||
configMap:
|
|
||||||
name: forgejo-secret-sync
|
|
||||||
- name: host-tls
|
|
||||||
hostPath:
|
|
||||||
path: /var/lib/rancher/k3s/server/tls
|
|
||||||
---
|
|
||||||
apiVersion: external-secrets.io/v1beta1
|
|
||||||
kind: ExternalSecret
|
|
||||||
metadata:
|
|
||||||
name: forgejo-secret-sync
|
|
||||||
spec:
|
|
||||||
secretStoreRef:
|
|
||||||
kind: SecretStore
|
|
||||||
name: openbao
|
|
||||||
target:
|
|
||||||
name: forgejo-secret-sync
|
|
||||||
creationPolicy: Owner
|
|
||||||
dataFrom:
|
|
||||||
- extract:
|
|
||||||
key: forgejo/default/secret-sync
|
|
|
@ -1,86 +0,0 @@
|
||||||
#!/usr/bin/env python3
|
|
||||||
import subprocess
|
|
||||||
import logging
|
|
||||||
import base64
|
|
||||||
import os
|
|
||||||
import requests
|
|
||||||
import json
|
|
||||||
|
|
||||||
logging.basicConfig(level=logging.DEBUG)
|
|
||||||
|
|
||||||
with open("/var/lib/rancher/k3s/server/tls/server-ca.crt") as f:
|
|
||||||
ca = base64.b64encode(f.read().encode()).decode()
|
|
||||||
|
|
||||||
forgejo_token = os.getenv("FORGEJO_TOKEN")
|
|
||||||
|
|
||||||
|
|
||||||
def run(cmd: list[str], stdin=None) -> str:
|
|
||||||
logging.debug("executing %s", cmd)
|
|
||||||
p = subprocess.Popen(cmd, stdout=subprocess.PIPE, stdin=subprocess.PIPE)
|
|
||||||
out = p.communicate(stdin)
|
|
||||||
if p.returncode != 0:
|
|
||||||
logging.critical("{} exited with code {}", cmd, p.returncode)
|
|
||||||
os.exit(1)
|
|
||||||
return out[0]
|
|
||||||
|
|
||||||
|
|
||||||
def update_cert(k8s_name: str, owner: str, repo: str):
|
|
||||||
key = run(["openssl", "genrsa", "4096"])
|
|
||||||
req = run(
|
|
||||||
["openssl", "req", "-key", "/dev/stdin", "-new", "-nodes", "-subj", f"/CN={k8s_name}"], stdin=key
|
|
||||||
)
|
|
||||||
cert = run(
|
|
||||||
[
|
|
||||||
"openssl",
|
|
||||||
"x509",
|
|
||||||
"-req",
|
|
||||||
"-CA",
|
|
||||||
"/var/lib/rancher/k3s/server/tls/client-ca.nochain.crt",
|
|
||||||
"-CAkey",
|
|
||||||
"/var/lib/rancher/k3s/server/tls/client-ca.key",
|
|
||||||
"-CAcreateserial",
|
|
||||||
"-days",
|
|
||||||
"10",
|
|
||||||
],
|
|
||||||
stdin=req,
|
|
||||||
)
|
|
||||||
|
|
||||||
keyb64 = base64.b64encode(key).decode()
|
|
||||||
certb64 = base64.b64encode(cert).decode()
|
|
||||||
|
|
||||||
kubeconfig = f"""
|
|
||||||
apiVersion: v1
|
|
||||||
clusters:
|
|
||||||
- cluster:
|
|
||||||
certificate-authority-data: {ca}
|
|
||||||
server: https://10.5.1.110:6443
|
|
||||||
name: default
|
|
||||||
contexts:
|
|
||||||
- context:
|
|
||||||
cluster: default
|
|
||||||
user: default
|
|
||||||
name: default
|
|
||||||
current-context: default
|
|
||||||
kind: Config
|
|
||||||
preferences: {"{}"}
|
|
||||||
users:
|
|
||||||
- name: default
|
|
||||||
user:
|
|
||||||
client-certificate-data: {certb64}
|
|
||||||
client-key-data: {keyb64}
|
|
||||||
"""
|
|
||||||
logging.info(f"updating secret for {owner}/{repo}")
|
|
||||||
requests.put(
|
|
||||||
f"https://git.janky.solutions/api/v1/repos/{owner}/{repo}/actions/secrets/KUBERNETES_CLIENT_CONFIG",
|
|
||||||
data=json.dumps(
|
|
||||||
{"data": kubeconfig},
|
|
||||||
),
|
|
||||||
headers={
|
|
||||||
"Authorization": f"token {forgejo_token}",
|
|
||||||
"Content-Type": "application/json",
|
|
||||||
},
|
|
||||||
).raise_for_status()
|
|
||||||
|
|
||||||
|
|
||||||
for entry in json.loads(os.getenv("REPO_MAPPINGS")):
|
|
||||||
update_cert(**entry)
|
|
|
@ -5,10 +5,8 @@ resources:
|
||||||
- namespace.yaml
|
- namespace.yaml
|
||||||
- config.yaml
|
- config.yaml
|
||||||
- ingress.yaml
|
- ingress.yaml
|
||||||
- forgejo-secret-sync.yaml
|
|
||||||
- services.yaml
|
- services.yaml
|
||||||
- statefulset.yaml
|
- statefulset.yaml
|
||||||
- secret-store.yaml
|
|
||||||
- secrets.yaml
|
- secrets.yaml
|
||||||
- renovatebot.yaml
|
- renovatebot.yaml
|
||||||
configMapGenerator:
|
configMapGenerator:
|
||||||
|
@ -18,6 +16,3 @@ configMapGenerator:
|
||||||
- name: renovate-config
|
- name: renovate-config
|
||||||
files:
|
files:
|
||||||
- renovate/config.js
|
- renovate/config.js
|
||||||
- name: forgejo-secret-sync
|
|
||||||
files:
|
|
||||||
- forgejo-secret-sync/forgejo-secret-sync.py
|
|
||||||
|
|
|
@ -11,7 +11,7 @@ spec:
|
||||||
spec:
|
spec:
|
||||||
containers:
|
containers:
|
||||||
- name: renovate
|
- name: renovate
|
||||||
image: ghcr.io/renovatebot/renovate:39
|
image: ghcr.io/renovatebot/renovate:38.124-full
|
||||||
env:
|
env:
|
||||||
- name: RENOVATE_CONFIG_FILE
|
- name: RENOVATE_CONFIG_FILE
|
||||||
value: /etc/renovate/config.js
|
value: /etc/renovate/config.js
|
||||||
|
|
|
@ -1,16 +0,0 @@
|
||||||
apiVersion: external-secrets.io/v1beta1
|
|
||||||
kind: SecretStore
|
|
||||||
metadata:
|
|
||||||
name: openbao
|
|
||||||
spec:
|
|
||||||
provider:
|
|
||||||
vault:
|
|
||||||
server: http://openbao.openbao:8200
|
|
||||||
path: static-secrets
|
|
||||||
version: v2
|
|
||||||
auth:
|
|
||||||
kubernetes:
|
|
||||||
mountPath: kubernetes
|
|
||||||
role: kubernetes-default
|
|
||||||
serviceAccountRef:
|
|
||||||
name: default
|
|
|
@ -14,7 +14,7 @@ spec:
|
||||||
app: forgejo
|
app: forgejo
|
||||||
spec:
|
spec:
|
||||||
containers:
|
containers:
|
||||||
- image: codeberg.org/forgejo/forgejo:9.0.2
|
- image: codeberg.org/forgejo/forgejo:9.0.0
|
||||||
imagePullPolicy: Always
|
imagePullPolicy: Always
|
||||||
name: forgejo
|
name: forgejo
|
||||||
resources: {}
|
resources: {}
|
||||||
|
|
|
@ -46,8 +46,8 @@ spec:
|
||||||
groups:
|
groups:
|
||||||
- count: 1
|
- count: 1
|
||||||
usb:
|
usb:
|
||||||
- vendor: "10C4"
|
- vendor: "1A86"
|
||||||
product: "EA60"
|
product: "55D4"
|
||||||
name: generic-device-plugin
|
name: generic-device-plugin
|
||||||
resources:
|
resources:
|
||||||
requests:
|
requests:
|
||||||
|
|
|
@ -30,7 +30,7 @@ spec:
|
||||||
spec:
|
spec:
|
||||||
containers:
|
containers:
|
||||||
- name: mysql
|
- name: mysql
|
||||||
image: docker.io/library/mysql:8.4.3
|
image: mysql:8
|
||||||
envFrom:
|
envFrom:
|
||||||
- secretRef:
|
- secretRef:
|
||||||
name: mysql
|
name: mysql
|
||||||
|
|
|
@ -6,7 +6,7 @@ spec:
|
||||||
spiloFSGroup: 103 # https://github.com/zalando/postgres-operator/issues/988
|
spiloFSGroup: 103 # https://github.com/zalando/postgres-operator/issues/988
|
||||||
teamId: keycloak
|
teamId: keycloak
|
||||||
volume:
|
volume:
|
||||||
size: 20Gi
|
size: 10Gi
|
||||||
numberOfInstances: 2
|
numberOfInstances: 2
|
||||||
users:
|
users:
|
||||||
superuser:
|
superuser:
|
||||||
|
|
17
k8s/matrix/appservices-janky.solutions/telegram.yaml
Normal file
17
k8s/matrix/appservices-janky.solutions/telegram.yaml
Normal file
|
@ -0,0 +1,17 @@
|
||||||
|
id: telegram
|
||||||
|
as_token: SECRET_TELEGRAM_AS_TOKEN
|
||||||
|
hs_token: SECRET_TELEGRAM_HS_TOKEN
|
||||||
|
namespaces:
|
||||||
|
users:
|
||||||
|
- exclusive: true
|
||||||
|
regex: '@telegram_.*:janky\.solutions'
|
||||||
|
- exclusive: true
|
||||||
|
regex: '@telegrambot:janky\.solutions'
|
||||||
|
aliases:
|
||||||
|
- exclusive: true
|
||||||
|
regex: \#telegram_.*:janky\.solutions
|
||||||
|
url: http://bridge-telegram:29317
|
||||||
|
sender_localpart: SECRET_TELEGRAM_SENDER_LOCALPART
|
||||||
|
rate_limited: false
|
||||||
|
de.sorunome.msc2409.push_ephemeral: true
|
||||||
|
push_ephemeral: true
|
|
@ -60,7 +60,7 @@ spec:
|
||||||
- secretRef:
|
- secretRef:
|
||||||
name: bridge-facebook
|
name: bridge-facebook
|
||||||
containers:
|
containers:
|
||||||
- image: dock.mau.dev/mautrix/meta:v0.4.2
|
- image: dock.mau.dev/mautrix/meta:v0.4.1
|
||||||
name: bridge-facebook
|
name: bridge-facebook
|
||||||
resources: {}
|
resources: {}
|
||||||
command: ["/usr/bin/mautrix-meta", "-c", "/data/config.yaml", "--no-update"]
|
command: ["/usr/bin/mautrix-meta", "-c", "/data/config.yaml", "--no-update"]
|
||||||
|
@ -70,6 +70,8 @@ spec:
|
||||||
volumeMounts:
|
volumeMounts:
|
||||||
- name: storage
|
- name: storage
|
||||||
mountPath: /data
|
mountPath: /data
|
||||||
|
- name: config
|
||||||
|
mountPath: /config
|
||||||
volumes:
|
volumes:
|
||||||
- name: config
|
- name: config
|
||||||
configMap:
|
configMap:
|
||||||
|
@ -168,8 +170,8 @@ data:
|
||||||
async_transactions: false
|
async_transactions: false
|
||||||
|
|
||||||
# Authentication tokens for AS <-> HS communication. Autogenerated; do not modify.
|
# Authentication tokens for AS <-> HS communication. Autogenerated; do not modify.
|
||||||
as_token: "SECRET_FACEBOOK_AS_TOKEN"
|
as_token: "AS_TOKEN"
|
||||||
hs_token: "SECRET_FACEBOOK_HS_TOKEN"
|
hs_token: "HS_TOKEN"
|
||||||
|
|
||||||
meta:
|
meta:
|
||||||
# Which service is this bridge for? Available options:
|
# Which service is this bridge for? Available options:
|
||||||
|
|
|
@ -57,7 +57,7 @@ spec:
|
||||||
- secretRef:
|
- secretRef:
|
||||||
name: bridge-signal
|
name: bridge-signal
|
||||||
containers:
|
containers:
|
||||||
- image: dock.mau.dev/mautrix/signal:v0.7.3
|
- image: dock.mau.dev/mautrix/signal:v0.7.2
|
||||||
name: bridge-signal
|
name: bridge-signal
|
||||||
resources: {}
|
resources: {}
|
||||||
ports:
|
ports:
|
||||||
|
@ -162,8 +162,8 @@ data:
|
||||||
async_transactions: false
|
async_transactions: false
|
||||||
|
|
||||||
# Authentication tokens for AS <-> HS communication. Autogenerated; do not modify.
|
# Authentication tokens for AS <-> HS communication. Autogenerated; do not modify.
|
||||||
as_token: "SECRET_SIGNAL_AS_TOKEN"
|
as_token: "AS_TOKEN"
|
||||||
hs_token: "SECRET_SIGNAL_HS_TOKEN"
|
hs_token: "HS_TOKEN"
|
||||||
|
|
||||||
# Prometheus config.
|
# Prometheus config.
|
||||||
metrics:
|
metrics:
|
||||||
|
|
752
k8s/matrix/bridge-telegram.yaml
Normal file
752
k8s/matrix/bridge-telegram.yaml
Normal file
|
@ -0,0 +1,752 @@
|
||||||
|
apiVersion: networking.k8s.io/v1
|
||||||
|
kind: Ingress
|
||||||
|
metadata:
|
||||||
|
name: bridge-telegram
|
||||||
|
namespace: matrix
|
||||||
|
spec:
|
||||||
|
rules:
|
||||||
|
- host: bridge-telegram.matrix.k8s
|
||||||
|
http:
|
||||||
|
paths:
|
||||||
|
- path: /
|
||||||
|
pathType: Prefix
|
||||||
|
backend:
|
||||||
|
service:
|
||||||
|
name: bridge-telegram
|
||||||
|
port:
|
||||||
|
name: bridge
|
||||||
|
---
|
||||||
|
apiVersion: networking.k8s.io/v1
|
||||||
|
kind: Ingress
|
||||||
|
metadata:
|
||||||
|
name: bridge-telegram-public
|
||||||
|
namespace: matrix
|
||||||
|
spec:
|
||||||
|
rules:
|
||||||
|
- host: telegram-bridge.k8s.home.finn.io
|
||||||
|
http:
|
||||||
|
paths:
|
||||||
|
- path: /public
|
||||||
|
pathType: Prefix
|
||||||
|
backend:
|
||||||
|
service:
|
||||||
|
name: bridge-telegram
|
||||||
|
port:
|
||||||
|
name: bridge
|
||||||
|
---
|
||||||
|
apiVersion: v1
|
||||||
|
kind: Service
|
||||||
|
metadata:
|
||||||
|
name: bridge-telegram
|
||||||
|
namespace: matrix
|
||||||
|
spec:
|
||||||
|
publishNotReadyAddresses: true
|
||||||
|
ports:
|
||||||
|
- name: bridge
|
||||||
|
port: 29317
|
||||||
|
selector:
|
||||||
|
app: bridge-telegram
|
||||||
|
---
|
||||||
|
apiVersion: apps/v1
|
||||||
|
kind: StatefulSet
|
||||||
|
metadata:
|
||||||
|
name: bridge-telegram
|
||||||
|
namespace: matrix
|
||||||
|
spec:
|
||||||
|
selector:
|
||||||
|
matchLabels:
|
||||||
|
app: bridge-telegram
|
||||||
|
serviceName: bridge-telegram
|
||||||
|
replicas: 1
|
||||||
|
template:
|
||||||
|
metadata:
|
||||||
|
labels:
|
||||||
|
app: bridge-telegram
|
||||||
|
spec:
|
||||||
|
initContainers:
|
||||||
|
- name: initialize-secrets
|
||||||
|
image: docker.io/library/python
|
||||||
|
command: ["python", "/init/initialize-secrets.py", "config.yaml"]
|
||||||
|
volumeMounts:
|
||||||
|
- name: init
|
||||||
|
mountPath: /init
|
||||||
|
- name: storage
|
||||||
|
mountPath: /data
|
||||||
|
- name: config
|
||||||
|
mountPath: /config
|
||||||
|
envFrom:
|
||||||
|
- secretRef:
|
||||||
|
name: bridge-telegram
|
||||||
|
containers:
|
||||||
|
- image: dock.mau.dev/mautrix/telegram:v0.15.2
|
||||||
|
name: bridge-telegram
|
||||||
|
resources: {}
|
||||||
|
command: ["python3", "-m", "mautrix_telegram", "-c", "/data/config.yaml"]
|
||||||
|
ports:
|
||||||
|
- name: bridge
|
||||||
|
containerPort: 29317
|
||||||
|
volumeMounts:
|
||||||
|
- name: storage
|
||||||
|
mountPath: /data
|
||||||
|
- name: config
|
||||||
|
mountPath: /config
|
||||||
|
volumes:
|
||||||
|
- name: config
|
||||||
|
configMap:
|
||||||
|
name: bridge-telegram
|
||||||
|
- name: init
|
||||||
|
configMap:
|
||||||
|
name: secrets-init
|
||||||
|
volumeClaimTemplates:
|
||||||
|
- metadata:
|
||||||
|
name: storage
|
||||||
|
spec:
|
||||||
|
accessModes: ["ReadWriteOnce"]
|
||||||
|
resources:
|
||||||
|
requests:
|
||||||
|
storage: 1Gi
|
||||||
|
---
|
||||||
|
apiVersion: v1
|
||||||
|
kind: ConfigMap
|
||||||
|
metadata:
|
||||||
|
name: bridge-telegram
|
||||||
|
namespace: matrix
|
||||||
|
data:
|
||||||
|
config.yaml: |
|
||||||
|
# Homeserver details
|
||||||
|
homeserver:
|
||||||
|
# The address that this appservice can use to connect to the homeserver.
|
||||||
|
address: https://matrix.janky.solutions
|
||||||
|
# The domain of the homeserver (for MXIDs, etc).
|
||||||
|
domain: janky.solutions
|
||||||
|
# Whether or not to verify the SSL certificate of the homeserver.
|
||||||
|
# Only applies if address starts with https://
|
||||||
|
verify_ssl: true
|
||||||
|
# What software is the homeserver running?
|
||||||
|
# Standard Matrix homeservers like Synapse, Dendrite and Conduit should just use "standard" here.
|
||||||
|
software: standard
|
||||||
|
# Number of retries for all HTTP requests if the homeserver isn't reachable.
|
||||||
|
http_retry_count: 4
|
||||||
|
# The URL to push real-time bridge status to.
|
||||||
|
# If set, the bridge will make POST requests to this URL whenever a user's Telegram connection state changes.
|
||||||
|
# The bridge will use the appservice as_token to authorize requests.
|
||||||
|
status_endpoint: null
|
||||||
|
# Endpoint for reporting per-message status.
|
||||||
|
message_send_checkpoint_endpoint: null
|
||||||
|
# Whether asynchronous uploads via MSC2246 should be enabled for media.
|
||||||
|
# Requires a media repo that supports MSC2246.
|
||||||
|
async_media: false
|
||||||
|
# Application service host/registration related details
|
||||||
|
# Changing these values requires regeneration of the registration.
|
||||||
|
appservice:
|
||||||
|
# The address that the homeserver can use to connect to this appservice.
|
||||||
|
address: http://bridge-telegram.matrix.k8s:80
|
||||||
|
# When using https:// the TLS certificate and key files for the address.
|
||||||
|
tls_cert: false
|
||||||
|
tls_key: false
|
||||||
|
# The hostname and port where this appservice should listen.
|
||||||
|
hostname: 0.0.0.0
|
||||||
|
port: 29317
|
||||||
|
# The maximum body size of appservice API requests (from the homeserver) in mebibytes
|
||||||
|
# Usually 1 is enough, but on high-traffic bridges you might need to increase this to avoid 413s
|
||||||
|
max_body_size: 1
|
||||||
|
# The full URI to the database. SQLite and Postgres are supported.
|
||||||
|
# Format examples:
|
||||||
|
# SQLite: sqlite:filename.db
|
||||||
|
# Postgres: postgres://username:password@hostname/dbname
|
||||||
|
database: sqlite:/data/telegram.db
|
||||||
|
# Additional arguments for asyncpg.create_pool() or sqlite3.connect()
|
||||||
|
# https://magicstack.github.io/asyncpg/current/api/index.html#asyncpg.pool.create_pool
|
||||||
|
# https://docs.python.org/3/library/sqlite3.html#sqlite3.connect
|
||||||
|
# For sqlite, min_size is used as the connection thread pool size and max_size is ignored.
|
||||||
|
# Additionally, SQLite supports init_commands as an array of SQL queries to run on connect (e.g. to set PRAGMAs).
|
||||||
|
database_opts:
|
||||||
|
min_size: 1
|
||||||
|
max_size: 10
|
||||||
|
# Public part of web server for out-of-Matrix interaction with the bridge.
|
||||||
|
# Used for things like login if the user wants to make sure the 2FA password isn't stored in
|
||||||
|
# the HS database.
|
||||||
|
public:
|
||||||
|
# Whether or not the public-facing endpoints should be enabled.
|
||||||
|
enabled: true
|
||||||
|
# The prefix to use in the public-facing endpoints.
|
||||||
|
prefix: /public
|
||||||
|
# The base URL where the public-facing endpoints are available. The prefix is not added
|
||||||
|
# implicitly.
|
||||||
|
external: https://telegram-bridge.k8s.home.finn.io/public
|
||||||
|
# Provisioning API part of the web server for automated portal creation and fetching information.
|
||||||
|
# Used by things like mautrix-manager (https://github.com/tulir/mautrix-manager).
|
||||||
|
provisioning:
|
||||||
|
# Whether or not the provisioning API should be enabled.
|
||||||
|
enabled: true
|
||||||
|
# The prefix to use in the provisioning API endpoints.
|
||||||
|
prefix: /_matrix/provision
|
||||||
|
# The shared secret to authorize users of the API.
|
||||||
|
# Set to "generate" to generate and save a new token.
|
||||||
|
shared_secret: generate
|
||||||
|
# The unique ID of this appservice.
|
||||||
|
id: telegram
|
||||||
|
# Username of the appservice bot.
|
||||||
|
bot_username: telegrambot
|
||||||
|
# Display name and avatar for bot. Set to "remove" to remove display name/avatar, leave empty
|
||||||
|
# to leave display name/avatar as-is.
|
||||||
|
bot_displayname: Telegram bridge bot
|
||||||
|
bot_avatar: mxc://maunium.net/tJCRmUyJDsgRNgqhOgoiHWbX
|
||||||
|
# Whether or not to receive ephemeral events via appservice transactions.
|
||||||
|
# Requires MSC2409 support (i.e. Synapse 1.22+).
|
||||||
|
# You should disable bridge -> sync_with_custom_puppets when this is enabled.
|
||||||
|
ephemeral_events: true
|
||||||
|
# Authentication tokens for AS <-> HS communication. Autogenerated; do not modify.
|
||||||
|
as_token: "AS_TOKEN"
|
||||||
|
hs_token: "HS_TOKEN"
|
||||||
|
# Prometheus telemetry config. Requires prometheus-client to be installed.
|
||||||
|
metrics:
|
||||||
|
enabled: true
|
||||||
|
listen_port: 8000
|
||||||
|
# Manhole config.
|
||||||
|
manhole:
|
||||||
|
# Whether or not opening the manhole is allowed.
|
||||||
|
enabled: false
|
||||||
|
# The path for the unix socket.
|
||||||
|
path: /var/tmp/mautrix-telegram.manhole
|
||||||
|
# The list of UIDs who can be added to the whitelist.
|
||||||
|
# If empty, any UIDs can be specified in the open-manhole command.
|
||||||
|
whitelist:
|
||||||
|
- 0
|
||||||
|
# Bridge config
|
||||||
|
bridge:
|
||||||
|
# Localpart template of MXIDs for Telegram users.
|
||||||
|
# {userid} is replaced with the user ID of the Telegram user.
|
||||||
|
username_template: "telegram_{userid}"
|
||||||
|
# Localpart template of room aliases for Telegram portal rooms.
|
||||||
|
# {groupname} is replaced with the name part of the public channel/group invite link ( https://t.me/{} )
|
||||||
|
alias_template: "telegram_{groupname}"
|
||||||
|
# Displayname template for Telegram users.
|
||||||
|
# {displayname} is replaced with the display name of the Telegram user.
|
||||||
|
displayname_template: "{displayname} (Telegram)"
|
||||||
|
# Set the preferred order of user identifiers which to use in the Matrix puppet display name.
|
||||||
|
# In the (hopefully unlikely) scenario that none of the given keys are found, the numeric user
|
||||||
|
# ID is used.
|
||||||
|
#
|
||||||
|
# If the bridge is working properly, a phone number or an username should always be known, but
|
||||||
|
# the other one can very well be empty.
|
||||||
|
#
|
||||||
|
# Valid keys:
|
||||||
|
# "full name" (First and/or last name)
|
||||||
|
# "full name reversed" (Last and/or first name)
|
||||||
|
# "first name"
|
||||||
|
# "last name"
|
||||||
|
# "username"
|
||||||
|
# "phone number"
|
||||||
|
displayname_preference:
|
||||||
|
- full name
|
||||||
|
- username
|
||||||
|
- phone number
|
||||||
|
# Maximum length of displayname
|
||||||
|
displayname_max_length: 100
|
||||||
|
# Remove avatars from Telegram ghost users when removed on Telegram. This is disabled by default
|
||||||
|
# as there's no way to determine whether an avatar is removed or just hidden from some users. If
|
||||||
|
# you're on a single-user instance, this should be safe to enable.
|
||||||
|
allow_avatar_remove: false
|
||||||
|
# Should contact names and profile pictures be allowed?
|
||||||
|
# This is only safe to enable on single-user instances.
|
||||||
|
allow_contact_info: false
|
||||||
|
# Maximum number of members to sync per portal when starting up. Other members will be
|
||||||
|
# synced when they send messages. The maximum is 10000, after which the Telegram server
|
||||||
|
# will not send any more members.
|
||||||
|
# -1 means no limit (which means it's limited to 10000 by the server)
|
||||||
|
max_initial_member_sync: 100
|
||||||
|
# Maximum number of participants in chats to bridge. Only applies when the portal is being created.
|
||||||
|
# If there are more members when trying to create a room, the room creation will be cancelled.
|
||||||
|
# -1 means no limit (which means all chats can be bridged)
|
||||||
|
max_member_count: -1
|
||||||
|
# Whether or not to sync the member list in channels.
|
||||||
|
# If no channel admins have logged into the bridge, the bridge won't be able to sync the member
|
||||||
|
# list regardless of this setting.
|
||||||
|
sync_channel_members: false
|
||||||
|
# Whether or not to skip deleted members when syncing members.
|
||||||
|
skip_deleted_members: true
|
||||||
|
# Whether or not to automatically synchronize contacts and chats of Matrix users logged into
|
||||||
|
# their Telegram account at startup.
|
||||||
|
startup_sync: false
|
||||||
|
# Number of most recently active dialogs to check when syncing chats.
|
||||||
|
# Set to 0 to remove limit.
|
||||||
|
sync_update_limit: 0
|
||||||
|
# Number of most recently active dialogs to create portals for when syncing chats.
|
||||||
|
# Set to 0 to remove limit.
|
||||||
|
sync_create_limit: 15
|
||||||
|
# Should all chats be scheduled to be created later?
|
||||||
|
# This is best used in combination with MSC2716 infinite backfill.
|
||||||
|
sync_deferred_create_all: false
|
||||||
|
# Whether or not to sync and create portals for direct chats at startup.
|
||||||
|
sync_direct_chats: false
|
||||||
|
# The maximum number of simultaneous Telegram deletions to handle.
|
||||||
|
# A large number of simultaneous redactions could put strain on your homeserver.
|
||||||
|
max_telegram_delete: 10
|
||||||
|
# Whether or not to automatically sync the Matrix room state (mostly unpuppeted displaynames)
|
||||||
|
# at startup and when creating a bridge.
|
||||||
|
sync_matrix_state: true
|
||||||
|
# Allow logging in within Matrix. If false, users can only log in using login-qr or the
|
||||||
|
# out-of-Matrix login website (see appservice.public config section)
|
||||||
|
allow_matrix_login: true
|
||||||
|
# Whether or not to make portals of publicly joinable channels/supergroups publicly joinable on Matrix.
|
||||||
|
public_portals: false
|
||||||
|
# Whether or not to use /sync to get presence, read receipts and typing notifications
|
||||||
|
# when double puppeting is enabled
|
||||||
|
sync_with_custom_puppets: false
|
||||||
|
# Whether or not to update the m.direct account data event when double puppeting is enabled.
|
||||||
|
# Note that updating the m.direct event is not atomic (except with mautrix-asmux)
|
||||||
|
# and is therefore prone to race conditions.
|
||||||
|
sync_direct_chat_list: false
|
||||||
|
# Servers to always allow double puppeting from
|
||||||
|
double_puppet_server_map:
|
||||||
|
example.com: https://example.com
|
||||||
|
# Allow using double puppeting from any server with a valid client .well-known file.
|
||||||
|
double_puppet_allow_discovery: false
|
||||||
|
# Shared secrets for https://github.com/devture/matrix-synapse-shared-secret-auth
|
||||||
|
#
|
||||||
|
# If set, custom puppets will be enabled automatically for local users
|
||||||
|
# instead of users having to find an access token and run `login-matrix`
|
||||||
|
# manually.
|
||||||
|
# If using this for other servers than the bridge's server,
|
||||||
|
# you must also set the URL in the double_puppet_server_map.
|
||||||
|
login_shared_secret_map:
|
||||||
|
example.com: foobar
|
||||||
|
# Set to false to disable link previews in messages sent to Telegram.
|
||||||
|
telegram_link_preview: true
|
||||||
|
# Whether or not the !tg join command should do a HTTP request
|
||||||
|
# to resolve redirects in invite links.
|
||||||
|
invite_link_resolve: false
|
||||||
|
# Send captions in the same message as images. This will send data compatible with both MSC2530 and MSC3552.
|
||||||
|
# This is currently not supported in most clients.
|
||||||
|
caption_in_message: false
|
||||||
|
# Maximum size of image in megabytes before sending to Telegram as a document.
|
||||||
|
image_as_file_size: 10
|
||||||
|
# Maximum number of pixels in an image before sending to Telegram as a document. Defaults to 4096x4096 = 16777216.
|
||||||
|
image_as_file_pixels: 16777216
|
||||||
|
# Maximum size of Telegram documents before linking to Telegrm instead of bridge
|
||||||
|
# to Matrix media.
|
||||||
|
document_as_link_size:
|
||||||
|
channel:
|
||||||
|
bot:
|
||||||
|
# Enable experimental parallel file transfer, which makes uploads/downloads much faster by
|
||||||
|
# streaming from/to Matrix and using many connections for Telegram.
|
||||||
|
# Note that generating HQ thumbnails for videos is not possible with streamed transfers.
|
||||||
|
# This option uses internal Telethon implementation details and may break with minor updates.
|
||||||
|
parallel_file_transfer: false
|
||||||
|
# Whether or not created rooms should have federation enabled.
|
||||||
|
# If false, created portal rooms will never be federated.
|
||||||
|
federate_rooms: true
|
||||||
|
# Should the bridge send all unicode reactions as custom emoji reactions to Telegram?
|
||||||
|
# By default, the bridge only uses custom emojis for unicode emojis that aren't allowed in reactions.
|
||||||
|
always_custom_emoji_reaction: false
|
||||||
|
# Settings for converting animated stickers.
|
||||||
|
animated_sticker:
|
||||||
|
# Format to which animated stickers should be converted.
|
||||||
|
# disable - No conversion, send as-is (gzipped lottie)
|
||||||
|
# png - converts to non-animated png (fastest),
|
||||||
|
# gif - converts to animated gif
|
||||||
|
# webm - converts to webm video, requires ffmpeg executable with vp9 codec and webm container support
|
||||||
|
# webp - converts to animated webp, requires ffmpeg executable with webp codec/container support
|
||||||
|
target: gif
|
||||||
|
# Should video stickers be converted to the specified format as well?
|
||||||
|
convert_from_webm: false
|
||||||
|
# Arguments for converter. All converters take width and height.
|
||||||
|
args:
|
||||||
|
width: 256
|
||||||
|
height: 256
|
||||||
|
fps: 25 # only for webm, webp and gif (2, 5, 10, 20 or 25 recommended)
|
||||||
|
# Settings for converting animated emoji.
|
||||||
|
# Same as animated_sticker, but webm is not supported as the target
|
||||||
|
# (because inline images can only contain images, not videos).
|
||||||
|
animated_emoji:
|
||||||
|
target: webp
|
||||||
|
args:
|
||||||
|
width: 64
|
||||||
|
height: 64
|
||||||
|
fps: 25
|
||||||
|
# End-to-bridge encryption support options.
|
||||||
|
#
|
||||||
|
# See https://docs.mau.fi/bridges/general/end-to-bridge-encryption.html for more info.
|
||||||
|
encryption:
|
||||||
|
# Allow encryption, work in group chat rooms with e2ee enabled
|
||||||
|
allow: false
|
||||||
|
# Default to encryption, force-enable encryption in all portals the bridge creates
|
||||||
|
# This will cause the bridge bot to be in private chats for the encryption to work properly.
|
||||||
|
default: false
|
||||||
|
# Whether to use MSC2409/MSC3202 instead of /sync long polling for receiving encryption-related data.
|
||||||
|
appservice: false
|
||||||
|
# Require encryption, drop any unencrypted messages.
|
||||||
|
require: false
|
||||||
|
# Enable key sharing? If enabled, key requests for rooms where users are in will be fulfilled.
|
||||||
|
# You must use a client that supports requesting keys from other users to use this feature.
|
||||||
|
allow_key_sharing: false
|
||||||
|
# Options for deleting megolm sessions from the bridge.
|
||||||
|
delete_keys:
|
||||||
|
# Beeper-specific: delete outbound sessions when hungryserv confirms
|
||||||
|
# that the user has uploaded the key to key backup.
|
||||||
|
delete_outbound_on_ack: false
|
||||||
|
# Don't store outbound sessions in the inbound table.
|
||||||
|
dont_store_outbound: false
|
||||||
|
# Ratchet megolm sessions forward after decrypting messages.
|
||||||
|
ratchet_on_decrypt: false
|
||||||
|
# Delete fully used keys (index >= max_messages) after decrypting messages.
|
||||||
|
delete_fully_used_on_decrypt: false
|
||||||
|
# Delete previous megolm sessions from same device when receiving a new one.
|
||||||
|
delete_prev_on_new_session: false
|
||||||
|
# Delete megolm sessions received from a device when the device is deleted.
|
||||||
|
delete_on_device_delete: false
|
||||||
|
# Periodically delete megolm sessions when 2x max_age has passed since receiving the session.
|
||||||
|
periodically_delete_expired: false
|
||||||
|
# Delete inbound megolm sessions that don't have the received_at field used for
|
||||||
|
# automatic ratcheting and expired session deletion. This is meant as a migration
|
||||||
|
# to delete old keys prior to the bridge update.
|
||||||
|
delete_outdated_inbound: false
|
||||||
|
# What level of device verification should be required from users?
|
||||||
|
#
|
||||||
|
# Valid levels:
|
||||||
|
# unverified - Send keys to all device in the room.
|
||||||
|
# cross-signed-untrusted - Require valid cross-signing, but trust all cross-signing keys.
|
||||||
|
# cross-signed-tofu - Require valid cross-signing, trust cross-signing keys on first use (and reject changes).
|
||||||
|
# cross-signed-verified - Require valid cross-signing, plus a valid user signature from the bridge bot.
|
||||||
|
# Note that creating user signatures from the bridge bot is not currently possible.
|
||||||
|
# verified - Require manual per-device verification
|
||||||
|
# (currently only possible by modifying the `trust` column in the `crypto_device` database table).
|
||||||
|
verification_levels:
|
||||||
|
# Minimum level for which the bridge should send keys to when bridging messages from Telegram to Matrix.
|
||||||
|
receive: unverified
|
||||||
|
# Minimum level that the bridge should accept for incoming Matrix messages.
|
||||||
|
send: unverified
|
||||||
|
# Minimum level that the bridge should require for accepting key requests.
|
||||||
|
share: cross-signed-tofu
|
||||||
|
# Options for Megolm room key rotation. These options allow you to
|
||||||
|
# configure the m.room.encryption event content. See:
|
||||||
|
# https://spec.matrix.org/v1.3/client-server-api/#mroomencryption for
|
||||||
|
# more information about that event.
|
||||||
|
rotation:
|
||||||
|
# Enable custom Megolm room key rotation settings. Note that these
|
||||||
|
# settings will only apply to rooms created after this option is
|
||||||
|
# set.
|
||||||
|
enable_custom: false
|
||||||
|
# The maximum number of milliseconds a session should be used
|
||||||
|
# before changing it. The Matrix spec recommends 604800000 (a week)
|
||||||
|
# as the default.
|
||||||
|
milliseconds: 604800000
|
||||||
|
# The maximum number of messages that should be sent with a given a
|
||||||
|
# session before changing it. The Matrix spec recommends 100 as the
|
||||||
|
# default.
|
||||||
|
messages: 100
|
||||||
|
# Disable rotating keys when a user's devices change?
|
||||||
|
# You should not enable this option unless you understand all the implications.
|
||||||
|
disable_device_change_key_rotation: false
|
||||||
|
# Whether to explicitly set the avatar and room name for private chat portal rooms.
|
||||||
|
# If set to `default`, this will be enabled in encrypted rooms and disabled in unencrypted rooms.
|
||||||
|
# If set to `always`, all DM rooms will have explicit names and avatars set.
|
||||||
|
# If set to `never`, DM rooms will never have names and avatars set.
|
||||||
|
private_chat_portal_meta: default
|
||||||
|
# Disable generating reply fallbacks? Some extremely bad clients still rely on them,
|
||||||
|
# but they're being phased out and will be completely removed in the future.
|
||||||
|
disable_reply_fallbacks: false
|
||||||
|
# Should cross-chat replies from Telegram be bridged? Most servers and clients don't support this.
|
||||||
|
cross_room_replies: false
|
||||||
|
# Whether or not the bridge should send a read receipt from the bridge bot when a message has
|
||||||
|
# been sent to Telegram.
|
||||||
|
delivery_receipts: false
|
||||||
|
# Whether or not delivery errors should be reported as messages in the Matrix room.
|
||||||
|
delivery_error_reports: false
|
||||||
|
# Should errors in incoming message handling send a message to the Matrix room?
|
||||||
|
incoming_bridge_error_reports: false
|
||||||
|
# Whether the bridge should send the message status as a custom com.beeper.message_send_status event.
|
||||||
|
message_status_events: false
|
||||||
|
# Set this to true to tell the bridge to re-send m.bridge events to all rooms on the next run.
|
||||||
|
# This field will automatically be changed back to false after it,
|
||||||
|
# except if the config file is not writable.
|
||||||
|
resend_bridge_info: false
|
||||||
|
# When using double puppeting, should muted chats be muted in Matrix?
|
||||||
|
mute_bridging: false
|
||||||
|
# When using double puppeting, should pinned chats be moved to a specific tag in Matrix?
|
||||||
|
# The favorites tag is `m.favourite`.
|
||||||
|
pinned_tag: null
|
||||||
|
# Same as above for archived chats, the low priority tag is `m.lowpriority`.
|
||||||
|
archive_tag: null
|
||||||
|
# Whether or not mute status and tags should only be bridged when the portal room is created.
|
||||||
|
tag_only_on_create: true
|
||||||
|
# Should leaving the room on Matrix make the user leave on Telegram?
|
||||||
|
bridge_matrix_leave: true
|
||||||
|
# Should the user be kicked out of all portals when logging out of the bridge?
|
||||||
|
kick_on_logout: true
|
||||||
|
# Should the "* user joined Telegram" notice always be marked as read automatically?
|
||||||
|
always_read_joined_telegram_notice: true
|
||||||
|
# Should the bridge auto-create a group chat on Telegram when a ghost is invited to a room?
|
||||||
|
# Requires the user to have sufficient power level and double puppeting enabled.
|
||||||
|
create_group_on_invite: true
|
||||||
|
# Settings for backfilling messages from Telegram.
|
||||||
|
backfill:
|
||||||
|
# Allow backfilling at all?
|
||||||
|
enable: true
|
||||||
|
# Whether or not to enable backfilling in normal groups.
|
||||||
|
# Normal groups have numerous technical problems in Telegram, and backfilling normal groups
|
||||||
|
# will likely cause problems if there are multiple Matrix users in the group.
|
||||||
|
normal_groups: false
|
||||||
|
# If a backfilled chat is older than this number of hours, mark it as read even if it's unread on Telegram.
|
||||||
|
# Set to -1 to let any chat be unread.
|
||||||
|
unread_hours_threshold: 720
|
||||||
|
# Forward backfilling limits.
|
||||||
|
#
|
||||||
|
# Using a negative initial limit is not recommended, as it would try to backfill everything in a single batch.
|
||||||
|
forward_limits:
|
||||||
|
# Number of messages to backfill immediately after creating a portal.
|
||||||
|
initial:
|
||||||
|
user: 50
|
||||||
|
normal_group: 100
|
||||||
|
supergroup: 10
|
||||||
|
channel: 10
|
||||||
|
# Number of messages to backfill when syncing chats.
|
||||||
|
sync:
|
||||||
|
user: 100
|
||||||
|
normal_group: 100
|
||||||
|
supergroup: 100
|
||||||
|
channel: 100
|
||||||
|
# Timeout for forward backfills in seconds. If you have a high limit, you'll have to increase this too.
|
||||||
|
forward_timeout: 900
|
||||||
|
# Settings for incremental backfill of history. These only apply to Beeper, as upstream abandoned MSC2716.
|
||||||
|
incremental:
|
||||||
|
# Maximum number of messages to backfill per batch.
|
||||||
|
messages_per_batch: 100
|
||||||
|
# The number of seconds to wait after backfilling the batch of messages.
|
||||||
|
post_batch_delay: 20
|
||||||
|
# The maximum number of batches to backfill per portal, split by the chat type.
|
||||||
|
# If set to -1, all messages in the chat will eventually be backfilled.
|
||||||
|
max_batches:
|
||||||
|
# Direct chats
|
||||||
|
user: -1
|
||||||
|
# Normal groups. Note that the normal_groups option above must be enabled
|
||||||
|
# for these to be backfilled.
|
||||||
|
normal_group: -1
|
||||||
|
# Supergroups
|
||||||
|
supergroup: 10
|
||||||
|
# Broadcast channels
|
||||||
|
channel: -1
|
||||||
|
# Overrides for base power levels.
|
||||||
|
initial_power_level_overrides:
|
||||||
|
user: {}
|
||||||
|
group: {}
|
||||||
|
# Whether to bridge Telegram bot messages as m.notices or m.texts.
|
||||||
|
bot_messages_as_notices: true
|
||||||
|
bridge_notices:
|
||||||
|
# Whether or not Matrix bot messages (type m.notice) should be bridged.
|
||||||
|
default: false
|
||||||
|
# List of user IDs for whom the previous flag is flipped.
|
||||||
|
# e.g. if bridge_notices.default is false, notices from other users will not be bridged, but
|
||||||
|
# notices from users listed here will be bridged.
|
||||||
|
exceptions: []
|
||||||
|
# An array of possible values for the $distinguisher variable in message formats.
|
||||||
|
# Each user gets one of the values here, based on a hash of their user ID.
|
||||||
|
# If the array is empty, the $distinguisher variable will also be empty.
|
||||||
|
relay_user_distinguishers: ["\U0001F7E6", "\U0001F7E3", "\U0001F7E9", "⭕️", "\U0001F536", "⬛️", "\U0001F535", "\U0001F7E2"]
|
||||||
|
# The formats to use when sending messages to Telegram via the relay bot.
|
||||||
|
# Text msgtypes (m.text, m.notice and m.emote) support HTML, media msgtypes don't.
|
||||||
|
#
|
||||||
|
# Available variables:
|
||||||
|
# $sender_displayname - The display name of the sender (e.g. Example User)
|
||||||
|
# $sender_username - The username (Matrix ID localpart) of the sender (e.g. exampleuser)
|
||||||
|
# $sender_mxid - The Matrix ID of the sender (e.g. @exampleuser:example.com)
|
||||||
|
# $distinguisher - A random string from the options in the relay_user_distinguishers array.
|
||||||
|
# $message - The message content
|
||||||
|
message_formats:
|
||||||
|
m.text: "$distinguisher <b>$sender_displayname</b>: $message"
|
||||||
|
m.notice: "$distinguisher <b>$sender_displayname</b>: $message"
|
||||||
|
m.emote: "* $distinguisher <b>$sender_displayname</b> $message"
|
||||||
|
m.file: "$distinguisher <b>$sender_displayname</b> sent a file: $message"
|
||||||
|
m.image: "$distinguisher <b>$sender_displayname</b> sent an image: $message"
|
||||||
|
m.audio: "$distinguisher <b>$sender_displayname</b> sent an audio file: $message"
|
||||||
|
m.video: "$distinguisher <b>$sender_displayname</b> sent a video: $message"
|
||||||
|
m.location: "$distinguisher <b>$sender_displayname</b> sent a location: $message"
|
||||||
|
# Telegram doesn't have built-in emotes, this field specifies how m.emote's from authenticated
|
||||||
|
# users are sent to telegram. All fields in message_formats are supported. Additionally, the
|
||||||
|
# Telegram user info is available in the following variables:
|
||||||
|
# $displayname - Telegram displayname
|
||||||
|
# $username - Telegram username (may not exist)
|
||||||
|
# $mention - Telegram @username or displayname mention (depending on which exists)
|
||||||
|
emote_format: "* $mention $formatted_body"
|
||||||
|
# The formats to use when sending state events to Telegram via the relay bot.
|
||||||
|
#
|
||||||
|
# Variables from `message_formats` that have the `sender_` prefix are available without the prefix.
|
||||||
|
# In name_change events, `$prev_displayname` is the previous displayname.
|
||||||
|
#
|
||||||
|
# Set format to an empty string to disable the messages for that event.
|
||||||
|
state_event_formats:
|
||||||
|
join: "$distinguisher <b>$displayname</b> joined the room."
|
||||||
|
leave: "$distinguisher <b>$displayname</b> left the room."
|
||||||
|
name_change: "$distinguisher <b>$prev_displayname</b> changed their name to $distinguisher <b>$displayname</b>"
|
||||||
|
# Filter rooms that can/can't be bridged. Can also be managed using the `filter` and
|
||||||
|
# `filter-mode` management commands.
|
||||||
|
#
|
||||||
|
# An empty blacklist will essentially disable the filter.
|
||||||
|
filter:
|
||||||
|
# Filter mode to use. Either "blacklist" or "whitelist".
|
||||||
|
# If the mode is "blacklist", the listed chats will never be bridged.
|
||||||
|
# If the mode is "whitelist", only the listed chats can be bridged.
|
||||||
|
mode: blacklist
|
||||||
|
# The list of group/channel IDs to filter.
|
||||||
|
list: []
|
||||||
|
# How to handle direct chats:
|
||||||
|
# If users is "null", direct chats will follow the previous settings.
|
||||||
|
# If users is "true", direct chats will always be bridged.
|
||||||
|
# If users is "false", direct chats will never be bridged.
|
||||||
|
users: true
|
||||||
|
# The prefix for commands. Only required in non-management rooms.
|
||||||
|
command_prefix: "!tg"
|
||||||
|
# Messages sent upon joining a management room.
|
||||||
|
# Markdown is supported. The defaults are listed below.
|
||||||
|
management_room_text:
|
||||||
|
# Sent when joining a room.
|
||||||
|
welcome: "Hello, I'm a Telegram bridge bot."
|
||||||
|
# Sent when joining a management room and the user is already logged in.
|
||||||
|
welcome_connected: "Use `help` for help."
|
||||||
|
# Sent when joining a management room and the user is not logged in.
|
||||||
|
welcome_unconnected: "Use `help` for help or `login` to log in."
|
||||||
|
# Optional extra text sent when joining a management room.
|
||||||
|
additional_help: ""
|
||||||
|
# Send each message separately (for readability in some clients)
|
||||||
|
management_room_multiple_messages: false
|
||||||
|
# Permissions for using the bridge.
|
||||||
|
# Permitted values:
|
||||||
|
# relaybot - Only use the bridge via the relaybot, no access to commands.
|
||||||
|
# user - Relaybot level + access to commands to create bridges.
|
||||||
|
# puppeting - User level + logging in with a Telegram account.
|
||||||
|
# full - Full access to use the bridge, i.e. previous levels + Matrix login.
|
||||||
|
# admin - Full access to use the bridge and some extra administration commands.
|
||||||
|
# Permitted keys:
|
||||||
|
# * - All Matrix users
|
||||||
|
# domain - All users on that homeserver
|
||||||
|
# mxid - Specific user
|
||||||
|
permissions:
|
||||||
|
"*": "relaybot"
|
||||||
|
"janky.solutions": "full"
|
||||||
|
"@finn:janky.solutions": "admin"
|
||||||
|
# Options related to the message relay Telegram bot.
|
||||||
|
relaybot:
|
||||||
|
private_chat:
|
||||||
|
# List of users to invite to the portal when someone starts a private chat with the bot.
|
||||||
|
# If empty, private chats with the bot won't create a portal.
|
||||||
|
invite: []
|
||||||
|
# Whether or not to bridge state change messages in relaybot private chats.
|
||||||
|
state_changes: true
|
||||||
|
# When private_chat_invite is empty, this message is sent to users /starting the
|
||||||
|
# relaybot. Telegram's "markdown" is supported.
|
||||||
|
message: This is a Matrix bridge relaybot and does not support direct chats
|
||||||
|
# List of users to invite to all group chat portals created by the bridge.
|
||||||
|
group_chat_invite: []
|
||||||
|
# Whether or not the relaybot should not bridge events in unbridged group chats.
|
||||||
|
# If false, portals will be created when the relaybot receives messages, just like normal
|
||||||
|
# users. This behavior is usually not desirable, as it interferes with manually bridging
|
||||||
|
# the chat to another room.
|
||||||
|
ignore_unbridged_group_chat: true
|
||||||
|
# Whether or not to allow creating portals from Telegram.
|
||||||
|
authless_portals: true
|
||||||
|
# Whether or not to allow Telegram group admins to use the bot commands.
|
||||||
|
whitelist_group_admins: true
|
||||||
|
# Whether or not to ignore incoming events sent by the relay bot.
|
||||||
|
ignore_own_incoming_events: true
|
||||||
|
# List of usernames/user IDs who are also allowed to use the bot commands.
|
||||||
|
whitelist:
|
||||||
|
- myusername
|
||||||
|
- 12345678
|
||||||
|
# Telegram config
|
||||||
|
telegram:
|
||||||
|
# Get your own API keys at https://my.telegram.org/apps
|
||||||
|
api_id: TG_API_ID
|
||||||
|
api_hash: TG_API_HASH
|
||||||
|
# (Optional) Create your own bot at https://t.me/BotFather
|
||||||
|
bot_token: disabled
|
||||||
|
# Should the bridge request missed updates from Telegram when restarting?
|
||||||
|
catch_up: true
|
||||||
|
# Should incoming updates be handled sequentially to make sure order is preserved on Matrix?
|
||||||
|
sequential_updates: true
|
||||||
|
exit_on_update_error: false
|
||||||
|
# Interval to force refresh the connection (full reconnect). 0 disables it.
|
||||||
|
force_refresh_interval_seconds: 0
|
||||||
|
# Telethon connection options.
|
||||||
|
connection:
|
||||||
|
# The timeout in seconds to be used when connecting.
|
||||||
|
timeout: 120
|
||||||
|
# How many times the reconnection should retry, either on the initial connection or when
|
||||||
|
# Telegram disconnects us. May be set to a negative or null value for infinite retries, but
|
||||||
|
# this is not recommended, since the program can get stuck in an infinite loop.
|
||||||
|
retries: 5
|
||||||
|
# The delay in seconds to sleep between automatic reconnections.
|
||||||
|
retry_delay: 1
|
||||||
|
# The threshold below which the library should automatically sleep on flood wait errors
|
||||||
|
# (inclusive). For instance, if a FloodWaitError for 17s occurs and flood_sleep_threshold
|
||||||
|
# is 20s, the library will sleep automatically. If the error was for 21s, it would raise
|
||||||
|
# the error instead. Values larger than a day (86400) will be changed to a day.
|
||||||
|
flood_sleep_threshold: 60
|
||||||
|
# How many times a request should be retried. Request are retried when Telegram is having
|
||||||
|
# internal issues, when there is a FloodWaitError less than flood_sleep_threshold, or when
|
||||||
|
# there's a migrate error. May take a negative or null value for infinite retries, but this
|
||||||
|
# is not recommended, since some requests can always trigger a call fail (such as searching
|
||||||
|
# for messages).
|
||||||
|
request_retries: 5
|
||||||
|
# Use IPv6 for Telethon connection
|
||||||
|
use_ipv6: false
|
||||||
|
# Device info sent to Telegram.
|
||||||
|
device_info:
|
||||||
|
# "auto" = OS name+version.
|
||||||
|
device_model: mautrix-telegram
|
||||||
|
# "auto" = Telethon version.
|
||||||
|
system_version: auto
|
||||||
|
# "auto" = mautrix-telegram version.
|
||||||
|
app_version: auto
|
||||||
|
lang_code: en
|
||||||
|
system_lang_code: en
|
||||||
|
# Custom server to connect to.
|
||||||
|
server:
|
||||||
|
# Set to true to use these server settings. If false, will automatically
|
||||||
|
# use production server assigned by Telegram. Set to false in production.
|
||||||
|
enabled: false
|
||||||
|
# The DC ID to connect to.
|
||||||
|
dc: 2
|
||||||
|
# The IP to connect to.
|
||||||
|
ip: 149.154.167.40
|
||||||
|
# The port to connect to. 443 may not work, 80 is better and both are equally secure.
|
||||||
|
port: 80
|
||||||
|
# Telethon proxy configuration.
|
||||||
|
# You must install PySocks from pip for proxies to work.
|
||||||
|
proxy:
|
||||||
|
# Allowed types: disabled, socks4, socks5, http, mtproxy
|
||||||
|
type: disabled
|
||||||
|
# Proxy IP address and port.
|
||||||
|
address: 127.0.0.1
|
||||||
|
port: 1080
|
||||||
|
# Whether or not to perform DNS resolving remotely. Only for socks/http proxies.
|
||||||
|
rdns: true
|
||||||
|
# Proxy authentication (optional). Put MTProxy secret in password field.
|
||||||
|
username: ""
|
||||||
|
password: ""
|
||||||
|
# Python logging configuration.
|
||||||
|
#
|
||||||
|
# See section 16.7.2 of the Python documentation for more info:
|
||||||
|
# https://docs.python.org/3.6/library/logging.config.html#configuration-dictionary-schema
|
||||||
|
logging:
|
||||||
|
version: 1
|
||||||
|
formatters:
|
||||||
|
colored:
|
||||||
|
(): mautrix_telegram.util.ColorFormatter
|
||||||
|
format: "[%(asctime)s] [%(levelname)s@%(name)s] %(message)s"
|
||||||
|
normal:
|
||||||
|
format: "[%(asctime)s] [%(levelname)s@%(name)s] %(message)s"
|
||||||
|
handlers:
|
||||||
|
console:
|
||||||
|
class: logging.StreamHandler
|
||||||
|
formatter: colored
|
||||||
|
loggers:
|
||||||
|
mau:
|
||||||
|
level: INFO
|
||||||
|
telethon:
|
||||||
|
level: INFO
|
||||||
|
aiohttp:
|
||||||
|
level: INFO
|
||||||
|
root:
|
||||||
|
level: DEBUG
|
||||||
|
handlers: [console]
|
|
@ -24,9 +24,6 @@ form_secret: "SECRET_form_secret"
|
||||||
signing_key_path: "/secrets/janky.solutions.signing.key"
|
signing_key_path: "/secrets/janky.solutions.signing.key"
|
||||||
trusted_key_servers:
|
trusted_key_servers:
|
||||||
- server_name: "matrix.org"
|
- server_name: "matrix.org"
|
||||||
auto_join_rooms:
|
|
||||||
- "#jankysolutions:janky.solutions"
|
|
||||||
- "#general:janky.solutions"
|
|
||||||
public_baseurl: https://matrix.janky.solutions
|
public_baseurl: https://matrix.janky.solutions
|
||||||
ip_range_whitelist: [10.5.1.245,10.5.1.1]
|
ip_range_whitelist: [10.5.1.245,10.5.1.1]
|
||||||
# oidc_providers:
|
# oidc_providers:
|
||||||
|
@ -45,6 +42,7 @@ password_config:
|
||||||
enabled: false
|
enabled: false
|
||||||
app_service_config_files:
|
app_service_config_files:
|
||||||
- /data/facebook.yaml
|
- /data/facebook.yaml
|
||||||
|
- /data/telegram.yaml
|
||||||
- /data/signal.yaml
|
- /data/signal.yaml
|
||||||
media_storage_providers:
|
media_storage_providers:
|
||||||
- module: s3_storage_provider.S3StorageProviderBackend
|
- module: s3_storage_provider.S3StorageProviderBackend
|
||||||
|
|
|
@ -37,7 +37,7 @@ spec:
|
||||||
- secretRef:
|
- secretRef:
|
||||||
name: synapse-janky-bot
|
name: synapse-janky-bot
|
||||||
containers:
|
containers:
|
||||||
- image: matrixdotorg/synapse:v1.119.0
|
- image: matrixdotorg/synapse:v1.117.0
|
||||||
name: synapse
|
name: synapse
|
||||||
resources: {}
|
resources: {}
|
||||||
volumeMounts:
|
volumeMounts:
|
||||||
|
|
|
@ -38,7 +38,7 @@ spec:
|
||||||
name: synapse-janky-solutions
|
name: synapse-janky-solutions
|
||||||
- name: initialize-bridge-secrets
|
- name: initialize-bridge-secrets
|
||||||
image: docker.io/library/python:3
|
image: docker.io/library/python:3
|
||||||
command: ["python", "/init/initialize-secrets.py", "facebook.yaml", "signal.yaml"]
|
command: ["python", "/init/initialize-secrets.py", "facebook.yaml", "telegram.yaml", "signal.yaml"]
|
||||||
volumeMounts:
|
volumeMounts:
|
||||||
- name: init
|
- name: init
|
||||||
mountPath: /init
|
mountPath: /init
|
||||||
|
@ -64,6 +64,8 @@ spec:
|
||||||
mountPath: /config
|
mountPath: /config
|
||||||
- name: secrets
|
- name: secrets
|
||||||
mountPath: /secrets
|
mountPath: /secrets
|
||||||
|
- name: appservices
|
||||||
|
mountPath: /appservices
|
||||||
env:
|
env:
|
||||||
- name: SYNAPSE_SERVER_NAME
|
- name: SYNAPSE_SERVER_NAME
|
||||||
value: matrix.janky.solutions
|
value: matrix.janky.solutions
|
||||||
|
|
|
@ -3,8 +3,9 @@ kind: Kustomization
|
||||||
namespace: matrix
|
namespace: matrix
|
||||||
resources:
|
resources:
|
||||||
- namespace.yaml
|
- namespace.yaml
|
||||||
# - bridge-facebook.yaml
|
- bridge-facebook.yaml
|
||||||
- bridge-signal.yaml
|
- bridge-signal.yaml
|
||||||
|
- bridge-telegram.yaml
|
||||||
- janky.bot-homeserver.yaml
|
- janky.bot-homeserver.yaml
|
||||||
- janky.solutions-homeserver.yaml
|
- janky.solutions-homeserver.yaml
|
||||||
- secrets.yaml
|
- secrets.yaml
|
||||||
|
@ -21,6 +22,7 @@ configMapGenerator:
|
||||||
- name: appservices-janky-solutions
|
- name: appservices-janky-solutions
|
||||||
files:
|
files:
|
||||||
- appservices-janky.solutions/facebook.yaml
|
- appservices-janky.solutions/facebook.yaml
|
||||||
|
- appservices-janky.solutions/telegram.yaml
|
||||||
- appservices-janky.solutions/signal.yaml
|
- appservices-janky.solutions/signal.yaml
|
||||||
- name: mas-janky-solutions
|
- name: mas-janky-solutions
|
||||||
files:
|
files:
|
||||||
|
|
|
@ -82,8 +82,8 @@ upstream_oauth2:
|
||||||
action: suggest
|
action: suggest
|
||||||
template: "{{ user.name }}"
|
template: "{{ user.name }}"
|
||||||
email:
|
email:
|
||||||
action: require
|
action: suggest
|
||||||
template: "{{ user.name }}@janky.solutions"
|
template: "{{ user.email }}"
|
||||||
set_email_verification: always
|
set_email_verification: always
|
||||||
account:
|
account:
|
||||||
email_change_allowed: true
|
email_change_allowed: false
|
||||||
|
|
|
@ -3,9 +3,8 @@ kind: Kustomization
|
||||||
namespace: meshtastic
|
namespace: meshtastic
|
||||||
resources:
|
resources:
|
||||||
- namespace.yaml
|
- namespace.yaml
|
||||||
- matrix-bridge-meshtastic.yaml
|
- matrix-meshtastic-bridge.yaml
|
||||||
- secrets.yaml
|
secretGenerator:
|
||||||
configMapGenerator:
|
- name: matrix-meshtastic-bridge
|
||||||
- name: matrix-bridge-meshtastic
|
|
||||||
files:
|
files:
|
||||||
- matrix-bridge-meshtastic/config.json
|
- matrix-meshtastic-bridge/config.json
|
||||||
|
|
|
@ -1,44 +0,0 @@
|
||||||
apiVersion: apps/v1
|
|
||||||
kind: StatefulSet
|
|
||||||
metadata:
|
|
||||||
name: matrix-bridge-meshtastic
|
|
||||||
spec:
|
|
||||||
selector:
|
|
||||||
matchLabels:
|
|
||||||
app: matrix-bridge-meshtastic
|
|
||||||
serviceName: matrix-bridge-meshtastic
|
|
||||||
replicas: 1
|
|
||||||
template:
|
|
||||||
metadata:
|
|
||||||
labels:
|
|
||||||
app: matrix-bridge-meshtastic
|
|
||||||
spec:
|
|
||||||
containers:
|
|
||||||
- name: matrix-bridge-meshtastic
|
|
||||||
image: git.janky.solutions/finn/matrix-bridge-meshtastic:sha-fae2a30
|
|
||||||
securityContext:
|
|
||||||
privileged: true
|
|
||||||
env:
|
|
||||||
- name: MATRIX_BRIDGE_MESHTASTIC_CONFIG
|
|
||||||
value: /config/config.json
|
|
||||||
envFrom:
|
|
||||||
- secretRef:
|
|
||||||
name: matrix-bridge-meshtastic
|
|
||||||
volumeMounts:
|
|
||||||
- name: config
|
|
||||||
mountPath: /config
|
|
||||||
- name: data
|
|
||||||
mountPath: /data
|
|
||||||
volumes:
|
|
||||||
- name: config
|
|
||||||
configMap:
|
|
||||||
name: matrix-bridge-meshtastic
|
|
||||||
volumeClaimTemplates:
|
|
||||||
- metadata:
|
|
||||||
name: data
|
|
||||||
spec:
|
|
||||||
storageClassName: longhorn
|
|
||||||
accessModes: ["ReadWriteOnce"]
|
|
||||||
resources:
|
|
||||||
requests:
|
|
||||||
storage: 1Gi
|
|
|
@ -1,11 +0,0 @@
|
||||||
{
|
|
||||||
"matrix": {
|
|
||||||
"user": "@meshtastic-test:janky.bot",
|
|
||||||
"room": "!VRoqFXTXJCHdTDdilP:janky.solutions",
|
|
||||||
"state": "/data/matrix.db"
|
|
||||||
},
|
|
||||||
"meshtastic": {
|
|
||||||
"address": "10.5.0.214"
|
|
||||||
},
|
|
||||||
"db": "/data/bridge.db"
|
|
||||||
}
|
|
56
k8s/meshtastic/matrix-meshtastic-bridge.yaml
Normal file
56
k8s/meshtastic/matrix-meshtastic-bridge.yaml
Normal file
|
@ -0,0 +1,56 @@
|
||||||
|
apiVersion: apps/v1
|
||||||
|
kind: StatefulSet
|
||||||
|
metadata:
|
||||||
|
name: matrix-meshtastic-bridge
|
||||||
|
spec:
|
||||||
|
selector:
|
||||||
|
matchLabels:
|
||||||
|
app: matrix-meshtastic-bridge
|
||||||
|
serviceName: matrix-meshtastic-bridge
|
||||||
|
replicas: 1
|
||||||
|
template:
|
||||||
|
metadata:
|
||||||
|
labels:
|
||||||
|
app: matrix-meshtastic-bridge
|
||||||
|
spec:
|
||||||
|
tolerations:
|
||||||
|
- key: rtlsdr
|
||||||
|
value: "true"
|
||||||
|
effect: NoSchedule
|
||||||
|
containers:
|
||||||
|
- name: matrix-meshtastic-bridge
|
||||||
|
image: git.janky.solutions/finn/matrix-meshtastic-bridge:sha-008f7cd
|
||||||
|
securityContext:
|
||||||
|
privileged: true
|
||||||
|
env:
|
||||||
|
- name: RUST_LOG
|
||||||
|
value: debug
|
||||||
|
- name: CONFIG_PATH
|
||||||
|
value: /config/config.json
|
||||||
|
resources:
|
||||||
|
limits:
|
||||||
|
janky.solutions/meshtastic: "1"
|
||||||
|
volumeMounts:
|
||||||
|
- name: config
|
||||||
|
mountPath: /config
|
||||||
|
- name: ttyacm0
|
||||||
|
mountPath: /dev/ttyACM0
|
||||||
|
- name: data
|
||||||
|
mountPath: /data
|
||||||
|
volumes:
|
||||||
|
- name: config
|
||||||
|
secret:
|
||||||
|
secretName: matrix-meshtastic-bridge
|
||||||
|
- name: ttyacm0
|
||||||
|
hostPath:
|
||||||
|
type: CharDevice
|
||||||
|
path: /dev/ttyACM0
|
||||||
|
volumeClaimTemplates:
|
||||||
|
- metadata:
|
||||||
|
name: data
|
||||||
|
spec:
|
||||||
|
storageClassName: longhorn
|
||||||
|
accessModes: ["ReadWriteOnce"]
|
||||||
|
resources:
|
||||||
|
requests:
|
||||||
|
storage: 1Gi
|
12
k8s/meshtastic/matrix-meshtastic-bridge/config.json
Normal file
12
k8s/meshtastic/matrix-meshtastic-bridge/config.json
Normal file
|
@ -0,0 +1,12 @@
|
||||||
|
{
|
||||||
|
"matrix": {
|
||||||
|
"username": "@meshtastic-test:janky.bot",
|
||||||
|
"room": "!VRoqFXTXJCHdTDdilP:janky.solutions",
|
||||||
|
"user": "@finn:janky.solutions",
|
||||||
|
"state": "/data/matrix.db"
|
||||||
|
},
|
||||||
|
"meshtastic": {
|
||||||
|
"device": "/dev/ttyACM0"
|
||||||
|
},
|
||||||
|
"db": "sqlite:///data/bridge.db"
|
||||||
|
}
|
|
@ -16,7 +16,7 @@ spec:
|
||||||
spec:
|
spec:
|
||||||
containers:
|
containers:
|
||||||
- name: miniflux
|
- name: miniflux
|
||||||
image: docker.io/miniflux/miniflux:2.2.3
|
image: docker.io/miniflux/miniflux:2.2.1
|
||||||
imagePullPolicy: Always
|
imagePullPolicy: Always
|
||||||
resources: {}
|
resources: {}
|
||||||
envFrom:
|
envFrom:
|
||||||
|
|
|
@ -64,7 +64,7 @@ spec:
|
||||||
- metadata:
|
- metadata:
|
||||||
name: storage
|
name: storage
|
||||||
spec:
|
spec:
|
||||||
# storageClassName: longhorn
|
storageClassName: longhorn
|
||||||
accessModes: ["ReadWriteOnce"]
|
accessModes: ["ReadWriteOnce"]
|
||||||
resources:
|
resources:
|
||||||
requests:
|
requests:
|
||||||
|
|
|
@ -27,7 +27,7 @@ spec:
|
||||||
app: mysql
|
app: mysql
|
||||||
spec:
|
spec:
|
||||||
containers:
|
containers:
|
||||||
- image: docker.io/library/mysql:5.7
|
- image: docker.io/library/mysql:9
|
||||||
name: mysql
|
name: mysql
|
||||||
resources: {}
|
resources: {}
|
||||||
ports:
|
ports:
|
||||||
|
@ -50,7 +50,7 @@ spec:
|
||||||
- metadata:
|
- metadata:
|
||||||
name: storage
|
name: storage
|
||||||
spec:
|
spec:
|
||||||
# storageClassName: longhorn
|
storageClassName: longhorn
|
||||||
accessModes: ["ReadWriteOnce"]
|
accessModes: ["ReadWriteOnce"]
|
||||||
resources:
|
resources:
|
||||||
requests:
|
requests:
|
||||||
|
|
|
@ -9,6 +9,9 @@ resources:
|
||||||
- thanos.yaml
|
- thanos.yaml
|
||||||
- alerts-longhorn.yaml
|
- alerts-longhorn.yaml
|
||||||
- matrix-alertmanager-receiver.yaml
|
- matrix-alertmanager-receiver.yaml
|
||||||
|
images:
|
||||||
|
- name: quay.io/thanos/thanos
|
||||||
|
newTag: v0.36.1
|
||||||
secretGenerator:
|
secretGenerator:
|
||||||
- name: additional-scrape-configs
|
- name: additional-scrape-configs
|
||||||
options:
|
options:
|
||||||
|
|
|
@ -31,7 +31,7 @@ spec:
|
||||||
name: matrix-alertmanager-receiver
|
name: matrix-alertmanager-receiver
|
||||||
containers:
|
containers:
|
||||||
- name: matrix-alertmanager-receiver
|
- name: matrix-alertmanager-receiver
|
||||||
image: docker.io/metio/matrix-alertmanager-receiver:2024.11.20
|
image: docker.io/metio/matrix-alertmanager-receiver:2024.10.16
|
||||||
args: ["--config-path", "/config/config.yaml"]
|
args: ["--config-path", "/config/config.yaml"]
|
||||||
resources:
|
resources:
|
||||||
limits:
|
limits:
|
||||||
|
|
|
@ -30,9 +30,12 @@
|
||||||
static_configs:
|
static_configs:
|
||||||
- targets:
|
- targets:
|
||||||
- ubnt:9001 # mongod-exporter
|
- ubnt:9001 # mongod-exporter
|
||||||
|
- ubnt:9130 # unifi-exporter
|
||||||
- rpi4-build:8080
|
- rpi4-build:8080
|
||||||
- ci-runner-0:8080
|
- ci-runner-0:8080
|
||||||
- ci-runner-1:8080
|
- ci-runner-1:8080
|
||||||
|
- ci-runner-2:8080
|
||||||
|
- ci-runner-3:8080
|
||||||
- docker:9170 # docker hub prometheus exporter
|
- docker:9170 # docker hub prometheus exporter
|
||||||
- jellyfin:8096 # jellyfin
|
- jellyfin:8096 # jellyfin
|
||||||
- signald:9595 # signald on signald
|
- signald:9595 # signald on signald
|
||||||
|
@ -64,8 +67,8 @@
|
||||||
- forgejo-runner-1:9080
|
- forgejo-runner-1:9080
|
||||||
- forgejo-runner-2:9080
|
- forgejo-runner-2:9080
|
||||||
- forgejo-runner-3:9080
|
- forgejo-runner-3:9080
|
||||||
|
- forgejo-runner-4:9080
|
||||||
- monitoring-0:9080
|
- monitoring-0:9080
|
||||||
- freepbx:9080
|
|
||||||
- job_name: node-exporter
|
- job_name: node-exporter
|
||||||
static_configs:
|
static_configs:
|
||||||
- targets:
|
- targets:
|
||||||
|
@ -79,6 +82,8 @@
|
||||||
- signald:9100
|
- signald:9100
|
||||||
- ci-runner-0:9100
|
- ci-runner-0:9100
|
||||||
- ci-runner-1:9100
|
- ci-runner-1:9100
|
||||||
|
- ci-runner-2:9100
|
||||||
|
- ci-runner-3:9100
|
||||||
- media-ingest:9100
|
- media-ingest:9100
|
||||||
- mc:9100
|
- mc:9100
|
||||||
- http:9100
|
- http:9100
|
||||||
|
@ -86,26 +91,27 @@
|
||||||
- mx1.janky.email:9100
|
- mx1.janky.email:9100
|
||||||
- dns:9100
|
- dns:9100
|
||||||
- hypervisor-d:9100
|
- hypervisor-d:9100
|
||||||
|
- livingroom-tv:9100
|
||||||
- mobile-proxy:9100
|
- mobile-proxy:9100
|
||||||
- monitoring-0:9100
|
- monitoring-0:9100
|
||||||
- forgejo-runner-0:9100
|
- forgejo-runner-0:9100
|
||||||
- forgejo-runner-1:9100
|
- forgejo-runner-1:9100
|
||||||
- forgejo-runner-2:9100
|
- forgejo-runner-2:9100
|
||||||
- forgejo-runner-3:9100
|
- forgejo-runner-3:9100
|
||||||
- freepbx:9100
|
- forgejo-runner-4:9100
|
||||||
- job_name: minio
|
- job_name: minio
|
||||||
authorization:
|
authorization:
|
||||||
credentials_file: /etc/prometheus/secrets/scrape-secrets/minio.token
|
credentials_file: /etc/prometheus/secrets/scrape-secrets/minio.token
|
||||||
metrics_path: /minio/v2/metrics/cluster
|
metrics_path: /minio/v2/metrics/cluster
|
||||||
static_configs:
|
static_configs:
|
||||||
- targets: ['minio:9000']
|
- targets: ['minio:9000']
|
||||||
# - job_name: 'home-assistant'
|
- job_name: 'home-assistant'
|
||||||
# metrics_path: /api/prometheus
|
metrics_path: /api/prometheus
|
||||||
# authorization:
|
authorization:
|
||||||
# credentials_file: /etc/prometheus/secrets/scrape-secrets/home-assistant.token
|
credentials_file: /etc/prometheus/secrets/scrape-secrets/home-assistant.token
|
||||||
# static_configs:
|
static_configs:
|
||||||
# - targets:
|
- targets:
|
||||||
# - home-assistant:8123
|
- home-assistant:8123
|
||||||
- job_name: forgejo
|
- job_name: forgejo
|
||||||
authorization:
|
authorization:
|
||||||
credentials_file: /etc/prometheus/secrets/scrape-secrets/forgejo.token
|
credentials_file: /etc/prometheus/secrets/scrape-secrets/forgejo.token
|
||||||
|
|
|
@ -18,7 +18,7 @@ spec:
|
||||||
spec:
|
spec:
|
||||||
containers:
|
containers:
|
||||||
- name: thanos-querier
|
- name: thanos-querier
|
||||||
image: quay.io/thanos/thanos:v0.36.1
|
image: quay.io/thanos/thanos:latest
|
||||||
args:
|
args:
|
||||||
- query
|
- query
|
||||||
- --http-address
|
- --http-address
|
||||||
|
@ -120,7 +120,7 @@ spec:
|
||||||
valueFrom:
|
valueFrom:
|
||||||
fieldRef:
|
fieldRef:
|
||||||
fieldPath: status.hostIP
|
fieldPath: status.hostIP
|
||||||
image: quay.io/thanos/thanos:v0.36.1
|
image: quay.io/thanos/thanos:latest
|
||||||
livenessProbe:
|
livenessProbe:
|
||||||
failureThreshold: 8
|
failureThreshold: 8
|
||||||
httpGet:
|
httpGet:
|
||||||
|
|
|
@ -17,7 +17,7 @@ spec:
|
||||||
spec:
|
spec:
|
||||||
containers:
|
containers:
|
||||||
- name: bitwarden-cli
|
- name: bitwarden-cli
|
||||||
image: ghcr.io/charlesthomas/bitwarden-cli:2024.11.1
|
image: ghcr.io/charlesthomas/bitwarden-cli:2024.9.0
|
||||||
imagePullPolicy: IfNotPresent
|
imagePullPolicy: IfNotPresent
|
||||||
envFrom:
|
envFrom:
|
||||||
- secretRef:
|
- secretRef:
|
||||||
|
|
|
@ -4147,9 +4147,6 @@ spec:
|
||||||
type: string
|
type: string
|
||||||
projectSlug:
|
projectSlug:
|
||||||
type: string
|
type: string
|
||||||
recursive:
|
|
||||||
default: false
|
|
||||||
type: boolean
|
|
||||||
secretsPath:
|
secretsPath:
|
||||||
default: /
|
default: /
|
||||||
type: string
|
type: string
|
||||||
|
@ -7508,27 +7505,7 @@ spec:
|
||||||
type: array
|
type: array
|
||||||
selector:
|
selector:
|
||||||
description: The Secret Selector (k8s source) for the Push Secret
|
description: The Secret Selector (k8s source) for the Push Secret
|
||||||
maxProperties: 1
|
|
||||||
minProperties: 1
|
|
||||||
properties:
|
properties:
|
||||||
generatorRef:
|
|
||||||
description: Point to a generator to create a Secret.
|
|
||||||
properties:
|
|
||||||
apiVersion:
|
|
||||||
default: generators.external-secrets.io/v1alpha1
|
|
||||||
description: Specify the apiVersion of the generator resource
|
|
||||||
type: string
|
|
||||||
kind:
|
|
||||||
description: Specify the Kind of the resource, e.g. Password,
|
|
||||||
ACRAccessToken etc.
|
|
||||||
type: string
|
|
||||||
name:
|
|
||||||
description: Specify the name of the generator resource
|
|
||||||
type: string
|
|
||||||
required:
|
|
||||||
- kind
|
|
||||||
- name
|
|
||||||
type: object
|
|
||||||
secret:
|
secret:
|
||||||
description: Select a Secret to Push.
|
description: Select a Secret to Push.
|
||||||
properties:
|
properties:
|
||||||
|
@ -7539,6 +7516,8 @@ spec:
|
||||||
required:
|
required:
|
||||||
- name
|
- name
|
||||||
type: object
|
type: object
|
||||||
|
required:
|
||||||
|
- secret
|
||||||
type: object
|
type: object
|
||||||
template:
|
template:
|
||||||
description: Template defines a blueprint for the created Secret resource.
|
description: Template defines a blueprint for the created Secret resource.
|
||||||
|
@ -10989,9 +10968,6 @@ spec:
|
||||||
type: string
|
type: string
|
||||||
projectSlug:
|
projectSlug:
|
||||||
type: string
|
type: string
|
||||||
recursive:
|
|
||||||
default: false
|
|
||||||
type: boolean
|
|
||||||
secretsPath:
|
secretsPath:
|
||||||
default: /
|
default: /
|
||||||
type: string
|
type: string
|
||||||
|
@ -13695,8 +13671,8 @@ metadata:
|
||||||
app.kubernetes.io/instance: external-secrets
|
app.kubernetes.io/instance: external-secrets
|
||||||
app.kubernetes.io/managed-by: Helm
|
app.kubernetes.io/managed-by: Helm
|
||||||
app.kubernetes.io/name: external-secrets
|
app.kubernetes.io/name: external-secrets
|
||||||
app.kubernetes.io/version: v0.10.5
|
app.kubernetes.io/version: v0.10.4
|
||||||
helm.sh/chart: external-secrets-0.10.5
|
helm.sh/chart: external-secrets-0.10.4
|
||||||
name: external-secrets
|
name: external-secrets
|
||||||
namespace: external-secrets
|
namespace: external-secrets
|
||||||
---
|
---
|
||||||
|
@ -13707,8 +13683,8 @@ metadata:
|
||||||
app.kubernetes.io/instance: external-secrets
|
app.kubernetes.io/instance: external-secrets
|
||||||
app.kubernetes.io/managed-by: Helm
|
app.kubernetes.io/managed-by: Helm
|
||||||
app.kubernetes.io/name: external-secrets-cert-controller
|
app.kubernetes.io/name: external-secrets-cert-controller
|
||||||
app.kubernetes.io/version: v0.10.5
|
app.kubernetes.io/version: v0.10.4
|
||||||
helm.sh/chart: external-secrets-0.10.5
|
helm.sh/chart: external-secrets-0.10.4
|
||||||
name: external-secrets-cert-controller
|
name: external-secrets-cert-controller
|
||||||
namespace: external-secrets
|
namespace: external-secrets
|
||||||
---
|
---
|
||||||
|
@ -13719,8 +13695,8 @@ metadata:
|
||||||
app.kubernetes.io/instance: external-secrets
|
app.kubernetes.io/instance: external-secrets
|
||||||
app.kubernetes.io/managed-by: Helm
|
app.kubernetes.io/managed-by: Helm
|
||||||
app.kubernetes.io/name: external-secrets-webhook
|
app.kubernetes.io/name: external-secrets-webhook
|
||||||
app.kubernetes.io/version: v0.10.5
|
app.kubernetes.io/version: v0.10.4
|
||||||
helm.sh/chart: external-secrets-0.10.5
|
helm.sh/chart: external-secrets-0.10.4
|
||||||
name: external-secrets-webhook
|
name: external-secrets-webhook
|
||||||
namespace: external-secrets
|
namespace: external-secrets
|
||||||
---
|
---
|
||||||
|
@ -13731,8 +13707,8 @@ metadata:
|
||||||
app.kubernetes.io/instance: external-secrets
|
app.kubernetes.io/instance: external-secrets
|
||||||
app.kubernetes.io/managed-by: Helm
|
app.kubernetes.io/managed-by: Helm
|
||||||
app.kubernetes.io/name: external-secrets
|
app.kubernetes.io/name: external-secrets
|
||||||
app.kubernetes.io/version: v0.10.5
|
app.kubernetes.io/version: v0.10.4
|
||||||
helm.sh/chart: external-secrets-0.10.5
|
helm.sh/chart: external-secrets-0.10.4
|
||||||
name: external-secrets-leaderelection
|
name: external-secrets-leaderelection
|
||||||
namespace: external-secrets
|
namespace: external-secrets
|
||||||
rules:
|
rules:
|
||||||
|
@ -13769,8 +13745,8 @@ metadata:
|
||||||
app.kubernetes.io/instance: external-secrets
|
app.kubernetes.io/instance: external-secrets
|
||||||
app.kubernetes.io/managed-by: Helm
|
app.kubernetes.io/managed-by: Helm
|
||||||
app.kubernetes.io/name: external-secrets-cert-controller
|
app.kubernetes.io/name: external-secrets-cert-controller
|
||||||
app.kubernetes.io/version: v0.10.5
|
app.kubernetes.io/version: v0.10.4
|
||||||
helm.sh/chart: external-secrets-0.10.5
|
helm.sh/chart: external-secrets-0.10.4
|
||||||
name: external-secrets-cert-controller
|
name: external-secrets-cert-controller
|
||||||
rules:
|
rules:
|
||||||
- apiGroups:
|
- apiGroups:
|
||||||
|
@ -13843,8 +13819,8 @@ metadata:
|
||||||
app.kubernetes.io/instance: external-secrets
|
app.kubernetes.io/instance: external-secrets
|
||||||
app.kubernetes.io/managed-by: Helm
|
app.kubernetes.io/managed-by: Helm
|
||||||
app.kubernetes.io/name: external-secrets
|
app.kubernetes.io/name: external-secrets
|
||||||
app.kubernetes.io/version: v0.10.5
|
app.kubernetes.io/version: v0.10.4
|
||||||
helm.sh/chart: external-secrets-0.10.5
|
helm.sh/chart: external-secrets-0.10.4
|
||||||
name: external-secrets-controller
|
name: external-secrets-controller
|
||||||
rules:
|
rules:
|
||||||
- apiGroups:
|
- apiGroups:
|
||||||
|
@ -13954,8 +13930,8 @@ metadata:
|
||||||
app.kubernetes.io/instance: external-secrets
|
app.kubernetes.io/instance: external-secrets
|
||||||
app.kubernetes.io/managed-by: Helm
|
app.kubernetes.io/managed-by: Helm
|
||||||
app.kubernetes.io/name: external-secrets
|
app.kubernetes.io/name: external-secrets
|
||||||
app.kubernetes.io/version: v0.10.5
|
app.kubernetes.io/version: v0.10.4
|
||||||
helm.sh/chart: external-secrets-0.10.5
|
helm.sh/chart: external-secrets-0.10.4
|
||||||
rbac.authorization.k8s.io/aggregate-to-admin: "true"
|
rbac.authorization.k8s.io/aggregate-to-admin: "true"
|
||||||
rbac.authorization.k8s.io/aggregate-to-edit: "true"
|
rbac.authorization.k8s.io/aggregate-to-edit: "true"
|
||||||
name: external-secrets-edit
|
name: external-secrets-edit
|
||||||
|
@ -13998,8 +13974,8 @@ metadata:
|
||||||
app.kubernetes.io/instance: external-secrets
|
app.kubernetes.io/instance: external-secrets
|
||||||
app.kubernetes.io/managed-by: Helm
|
app.kubernetes.io/managed-by: Helm
|
||||||
app.kubernetes.io/name: external-secrets
|
app.kubernetes.io/name: external-secrets
|
||||||
app.kubernetes.io/version: v0.10.5
|
app.kubernetes.io/version: v0.10.4
|
||||||
helm.sh/chart: external-secrets-0.10.5
|
helm.sh/chart: external-secrets-0.10.4
|
||||||
servicebinding.io/controller: "true"
|
servicebinding.io/controller: "true"
|
||||||
name: external-secrets-servicebindings
|
name: external-secrets-servicebindings
|
||||||
rules:
|
rules:
|
||||||
|
@ -14019,8 +13995,8 @@ metadata:
|
||||||
app.kubernetes.io/instance: external-secrets
|
app.kubernetes.io/instance: external-secrets
|
||||||
app.kubernetes.io/managed-by: Helm
|
app.kubernetes.io/managed-by: Helm
|
||||||
app.kubernetes.io/name: external-secrets
|
app.kubernetes.io/name: external-secrets
|
||||||
app.kubernetes.io/version: v0.10.5
|
app.kubernetes.io/version: v0.10.4
|
||||||
helm.sh/chart: external-secrets-0.10.5
|
helm.sh/chart: external-secrets-0.10.4
|
||||||
rbac.authorization.k8s.io/aggregate-to-admin: "true"
|
rbac.authorization.k8s.io/aggregate-to-admin: "true"
|
||||||
rbac.authorization.k8s.io/aggregate-to-edit: "true"
|
rbac.authorization.k8s.io/aggregate-to-edit: "true"
|
||||||
rbac.authorization.k8s.io/aggregate-to-view: "true"
|
rbac.authorization.k8s.io/aggregate-to-view: "true"
|
||||||
|
@ -14060,8 +14036,8 @@ metadata:
|
||||||
app.kubernetes.io/instance: external-secrets
|
app.kubernetes.io/instance: external-secrets
|
||||||
app.kubernetes.io/managed-by: Helm
|
app.kubernetes.io/managed-by: Helm
|
||||||
app.kubernetes.io/name: external-secrets
|
app.kubernetes.io/name: external-secrets
|
||||||
app.kubernetes.io/version: v0.10.5
|
app.kubernetes.io/version: v0.10.4
|
||||||
helm.sh/chart: external-secrets-0.10.5
|
helm.sh/chart: external-secrets-0.10.4
|
||||||
name: external-secrets-leaderelection
|
name: external-secrets-leaderelection
|
||||||
namespace: external-secrets
|
namespace: external-secrets
|
||||||
roleRef:
|
roleRef:
|
||||||
|
@ -14080,8 +14056,8 @@ metadata:
|
||||||
app.kubernetes.io/instance: external-secrets
|
app.kubernetes.io/instance: external-secrets
|
||||||
app.kubernetes.io/managed-by: Helm
|
app.kubernetes.io/managed-by: Helm
|
||||||
app.kubernetes.io/name: external-secrets-cert-controller
|
app.kubernetes.io/name: external-secrets-cert-controller
|
||||||
app.kubernetes.io/version: v0.10.5
|
app.kubernetes.io/version: v0.10.4
|
||||||
helm.sh/chart: external-secrets-0.10.5
|
helm.sh/chart: external-secrets-0.10.4
|
||||||
name: external-secrets-cert-controller
|
name: external-secrets-cert-controller
|
||||||
roleRef:
|
roleRef:
|
||||||
apiGroup: rbac.authorization.k8s.io
|
apiGroup: rbac.authorization.k8s.io
|
||||||
|
@ -14099,8 +14075,8 @@ metadata:
|
||||||
app.kubernetes.io/instance: external-secrets
|
app.kubernetes.io/instance: external-secrets
|
||||||
app.kubernetes.io/managed-by: Helm
|
app.kubernetes.io/managed-by: Helm
|
||||||
app.kubernetes.io/name: external-secrets
|
app.kubernetes.io/name: external-secrets
|
||||||
app.kubernetes.io/version: v0.10.5
|
app.kubernetes.io/version: v0.10.4
|
||||||
helm.sh/chart: external-secrets-0.10.5
|
helm.sh/chart: external-secrets-0.10.4
|
||||||
name: external-secrets-controller
|
name: external-secrets-controller
|
||||||
roleRef:
|
roleRef:
|
||||||
apiGroup: rbac.authorization.k8s.io
|
apiGroup: rbac.authorization.k8s.io
|
||||||
|
@ -14118,9 +14094,9 @@ metadata:
|
||||||
app.kubernetes.io/instance: external-secrets
|
app.kubernetes.io/instance: external-secrets
|
||||||
app.kubernetes.io/managed-by: Helm
|
app.kubernetes.io/managed-by: Helm
|
||||||
app.kubernetes.io/name: external-secrets-webhook
|
app.kubernetes.io/name: external-secrets-webhook
|
||||||
app.kubernetes.io/version: v0.10.5
|
app.kubernetes.io/version: v0.10.4
|
||||||
external-secrets.io/component: webhook
|
external-secrets.io/component: webhook
|
||||||
helm.sh/chart: external-secrets-0.10.5
|
helm.sh/chart: external-secrets-0.10.4
|
||||||
name: external-secrets-webhook
|
name: external-secrets-webhook
|
||||||
namespace: external-secrets
|
namespace: external-secrets
|
||||||
---
|
---
|
||||||
|
@ -14131,9 +14107,9 @@ metadata:
|
||||||
app.kubernetes.io/instance: external-secrets
|
app.kubernetes.io/instance: external-secrets
|
||||||
app.kubernetes.io/managed-by: Helm
|
app.kubernetes.io/managed-by: Helm
|
||||||
app.kubernetes.io/name: external-secrets-webhook
|
app.kubernetes.io/name: external-secrets-webhook
|
||||||
app.kubernetes.io/version: v0.10.5
|
app.kubernetes.io/version: v0.10.4
|
||||||
external-secrets.io/component: webhook
|
external-secrets.io/component: webhook
|
||||||
helm.sh/chart: external-secrets-0.10.5
|
helm.sh/chart: external-secrets-0.10.4
|
||||||
name: external-secrets-webhook
|
name: external-secrets-webhook
|
||||||
namespace: external-secrets
|
namespace: external-secrets
|
||||||
spec:
|
spec:
|
||||||
|
@ -14154,8 +14130,8 @@ metadata:
|
||||||
app.kubernetes.io/instance: external-secrets
|
app.kubernetes.io/instance: external-secrets
|
||||||
app.kubernetes.io/managed-by: Helm
|
app.kubernetes.io/managed-by: Helm
|
||||||
app.kubernetes.io/name: external-secrets
|
app.kubernetes.io/name: external-secrets
|
||||||
app.kubernetes.io/version: v0.10.5
|
app.kubernetes.io/version: v0.10.4
|
||||||
helm.sh/chart: external-secrets-0.10.5
|
helm.sh/chart: external-secrets-0.10.4
|
||||||
name: external-secrets
|
name: external-secrets
|
||||||
namespace: external-secrets
|
namespace: external-secrets
|
||||||
spec:
|
spec:
|
||||||
|
@ -14171,8 +14147,8 @@ spec:
|
||||||
app.kubernetes.io/instance: external-secrets
|
app.kubernetes.io/instance: external-secrets
|
||||||
app.kubernetes.io/managed-by: Helm
|
app.kubernetes.io/managed-by: Helm
|
||||||
app.kubernetes.io/name: external-secrets
|
app.kubernetes.io/name: external-secrets
|
||||||
app.kubernetes.io/version: v0.10.5
|
app.kubernetes.io/version: v0.10.4
|
||||||
helm.sh/chart: external-secrets-0.10.5
|
helm.sh/chart: external-secrets-0.10.4
|
||||||
spec:
|
spec:
|
||||||
automountServiceAccountToken: true
|
automountServiceAccountToken: true
|
||||||
containers:
|
containers:
|
||||||
|
@ -14181,7 +14157,7 @@ spec:
|
||||||
- --metrics-addr=:8080
|
- --metrics-addr=:8080
|
||||||
- --loglevel=info
|
- --loglevel=info
|
||||||
- --zap-time-encoding=epoch
|
- --zap-time-encoding=epoch
|
||||||
image: oci.external-secrets.io/external-secrets/external-secrets:v0.10.5
|
image: oci.external-secrets.io/external-secrets/external-secrets:v0.10.4
|
||||||
imagePullPolicy: IfNotPresent
|
imagePullPolicy: IfNotPresent
|
||||||
name: external-secrets
|
name: external-secrets
|
||||||
ports:
|
ports:
|
||||||
|
@ -14209,8 +14185,8 @@ metadata:
|
||||||
app.kubernetes.io/instance: external-secrets
|
app.kubernetes.io/instance: external-secrets
|
||||||
app.kubernetes.io/managed-by: Helm
|
app.kubernetes.io/managed-by: Helm
|
||||||
app.kubernetes.io/name: external-secrets-cert-controller
|
app.kubernetes.io/name: external-secrets-cert-controller
|
||||||
app.kubernetes.io/version: v0.10.5
|
app.kubernetes.io/version: v0.10.4
|
||||||
helm.sh/chart: external-secrets-0.10.5
|
helm.sh/chart: external-secrets-0.10.4
|
||||||
name: external-secrets-cert-controller
|
name: external-secrets-cert-controller
|
||||||
namespace: external-secrets
|
namespace: external-secrets
|
||||||
spec:
|
spec:
|
||||||
|
@ -14226,8 +14202,8 @@ spec:
|
||||||
app.kubernetes.io/instance: external-secrets
|
app.kubernetes.io/instance: external-secrets
|
||||||
app.kubernetes.io/managed-by: Helm
|
app.kubernetes.io/managed-by: Helm
|
||||||
app.kubernetes.io/name: external-secrets-cert-controller
|
app.kubernetes.io/name: external-secrets-cert-controller
|
||||||
app.kubernetes.io/version: v0.10.5
|
app.kubernetes.io/version: v0.10.4
|
||||||
helm.sh/chart: external-secrets-0.10.5
|
helm.sh/chart: external-secrets-0.10.4
|
||||||
spec:
|
spec:
|
||||||
automountServiceAccountToken: true
|
automountServiceAccountToken: true
|
||||||
containers:
|
containers:
|
||||||
|
@ -14243,7 +14219,7 @@ spec:
|
||||||
- --loglevel=info
|
- --loglevel=info
|
||||||
- --zap-time-encoding=epoch
|
- --zap-time-encoding=epoch
|
||||||
- --enable-partial-cache=true
|
- --enable-partial-cache=true
|
||||||
image: oci.external-secrets.io/external-secrets/external-secrets:v0.10.5
|
image: oci.external-secrets.io/external-secrets/external-secrets:v0.10.4
|
||||||
imagePullPolicy: IfNotPresent
|
imagePullPolicy: IfNotPresent
|
||||||
name: cert-controller
|
name: cert-controller
|
||||||
ports:
|
ports:
|
||||||
|
@ -14276,8 +14252,8 @@ metadata:
|
||||||
app.kubernetes.io/instance: external-secrets
|
app.kubernetes.io/instance: external-secrets
|
||||||
app.kubernetes.io/managed-by: Helm
|
app.kubernetes.io/managed-by: Helm
|
||||||
app.kubernetes.io/name: external-secrets-webhook
|
app.kubernetes.io/name: external-secrets-webhook
|
||||||
app.kubernetes.io/version: v0.10.5
|
app.kubernetes.io/version: v0.10.4
|
||||||
helm.sh/chart: external-secrets-0.10.5
|
helm.sh/chart: external-secrets-0.10.4
|
||||||
name: external-secrets-webhook
|
name: external-secrets-webhook
|
||||||
namespace: external-secrets
|
namespace: external-secrets
|
||||||
spec:
|
spec:
|
||||||
|
@ -14293,8 +14269,8 @@ spec:
|
||||||
app.kubernetes.io/instance: external-secrets
|
app.kubernetes.io/instance: external-secrets
|
||||||
app.kubernetes.io/managed-by: Helm
|
app.kubernetes.io/managed-by: Helm
|
||||||
app.kubernetes.io/name: external-secrets-webhook
|
app.kubernetes.io/name: external-secrets-webhook
|
||||||
app.kubernetes.io/version: v0.10.5
|
app.kubernetes.io/version: v0.10.4
|
||||||
helm.sh/chart: external-secrets-0.10.5
|
helm.sh/chart: external-secrets-0.10.4
|
||||||
spec:
|
spec:
|
||||||
automountServiceAccountToken: true
|
automountServiceAccountToken: true
|
||||||
containers:
|
containers:
|
||||||
|
@ -14308,7 +14284,7 @@ spec:
|
||||||
- --healthz-addr=:8081
|
- --healthz-addr=:8081
|
||||||
- --loglevel=info
|
- --loglevel=info
|
||||||
- --zap-time-encoding=epoch
|
- --zap-time-encoding=epoch
|
||||||
image: oci.external-secrets.io/external-secrets/external-secrets:v0.10.5
|
image: oci.external-secrets.io/external-secrets/external-secrets:v0.10.4
|
||||||
imagePullPolicy: IfNotPresent
|
imagePullPolicy: IfNotPresent
|
||||||
name: webhook
|
name: webhook
|
||||||
ports:
|
ports:
|
||||||
|
|
|
@ -1,30 +1,27 @@
|
||||||
- op: add
|
|
||||||
path: /spec/storage
|
|
||||||
value:
|
|
||||||
volumeClaimTemplate:
|
|
||||||
spec:
|
|
||||||
resources:
|
|
||||||
requests:
|
|
||||||
storage: 50Gi
|
|
||||||
- op: add
|
- op: add
|
||||||
path: /spec/additionalScrapeConfigs
|
path: /spec/additionalScrapeConfigs
|
||||||
value:
|
value:
|
||||||
key: scrape-configs.yaml
|
key: scrape-configs.yaml
|
||||||
name: additional-scrape-configs
|
name: additional-scrape-configs
|
||||||
optional: true
|
optional: true
|
||||||
- op: add
|
|
||||||
path: /spec/externalUrl
|
|
||||||
value: https://prometheus.k8s.home.finn.io
|
|
||||||
- op: add
|
|
||||||
path: /spec/retention
|
|
||||||
value: 72h
|
|
||||||
- op: add
|
- op: add
|
||||||
path: /spec/secrets
|
path: /spec/secrets
|
||||||
value: [scrape-secrets]
|
value: [scrape-secrets]
|
||||||
|
- op: add
|
||||||
|
path: /spec/externalUrl
|
||||||
|
value: https://prometheus.k8s.home.finn.io
|
||||||
- op: add
|
- op: add
|
||||||
path: /spec/thanos
|
path: /spec/thanos
|
||||||
value:
|
value:
|
||||||
image: quay.io/thanos/thanos:v0.36.1
|
image: quay.io/thanos/thanos:v0.36.0
|
||||||
objectStorageConfig:
|
objectStorageConfig:
|
||||||
key: thanos.yaml
|
key: thanos.yaml
|
||||||
name: thanos-objstore
|
name: thanos-objstore
|
||||||
|
- op: add
|
||||||
|
path: /spec/storage
|
||||||
|
value:
|
||||||
|
volumeClaimTemplate:
|
||||||
|
spec:
|
||||||
|
resources:
|
||||||
|
requests:
|
||||||
|
storage: 20Gi
|
||||||
|
|
|
@ -2,9 +2,8 @@ apiVersion: kustomize.config.k8s.io/v1beta1
|
||||||
kind: Kustomization
|
kind: Kustomization
|
||||||
resources:
|
resources:
|
||||||
- longhorn
|
- longhorn
|
||||||
- cert-manager
|
|
||||||
- external-secrets
|
|
||||||
- kube-prometheus
|
|
||||||
- openbao
|
|
||||||
- postgres-operator
|
- postgres-operator
|
||||||
|
- cert-manager
|
||||||
|
- openbao
|
||||||
|
- kube-prometheus
|
||||||
- secrets-store-csi-driver
|
- secrets-store-csi-driver
|
||||||
|
|
|
@ -3,7 +3,7 @@ kind: Kustomization
|
||||||
namespace: longhorn-system
|
namespace: longhorn-system
|
||||||
resources:
|
resources:
|
||||||
- https://github.com/longhorn/longhorn/releases/download/v1.6.2/longhorn.yaml
|
- https://github.com/longhorn/longhorn/releases/download/v1.6.2/longhorn.yaml
|
||||||
# - secrets.yaml
|
- secrets.yaml
|
||||||
- backup.yaml
|
- backup.yaml
|
||||||
- ingress.yaml
|
- ingress.yaml
|
||||||
- servicemonitor.yaml
|
- servicemonitor.yaml
|
||||||
|
|
|
@ -4,7 +4,7 @@ namespace: postgres-operator
|
||||||
resources:
|
resources:
|
||||||
- namespace.yaml
|
- namespace.yaml
|
||||||
- github.com/zalando/postgres-operator/manifests?ref=v1.13.0
|
- github.com/zalando/postgres-operator/manifests?ref=v1.13.0
|
||||||
# - secrets.yaml
|
- secrets.yaml
|
||||||
configMapGenerator:
|
configMapGenerator:
|
||||||
- name: postgres-operator
|
- name: postgres-operator
|
||||||
behavior: merge
|
behavior: merge
|
||||||
|
|
|
@ -18,7 +18,7 @@ spec:
|
||||||
fsGroup: 1001
|
fsGroup: 1001
|
||||||
fsGroupChangePolicy: "OnRootMismatch"
|
fsGroupChangePolicy: "OnRootMismatch"
|
||||||
containers:
|
containers:
|
||||||
- image: ghcr.io/shlinkio/shlink:4.3.0
|
- image: ghcr.io/shlinkio/shlink:4.2.2
|
||||||
name: shlink
|
name: shlink
|
||||||
resources: {}
|
resources: {}
|
||||||
ports:
|
ports:
|
||||||
|
|
|
@ -18,7 +18,7 @@ spec:
|
||||||
fsGroupChangePolicy: "OnRootMismatch"
|
fsGroupChangePolicy: "OnRootMismatch"
|
||||||
containers:
|
containers:
|
||||||
- name: snipe
|
- name: snipe
|
||||||
image: snipe/snipe-it:v7.1.14
|
image: snipe/snipe-it:v7.0.13
|
||||||
ports:
|
ports:
|
||||||
- containerPort: 80
|
- containerPort: 80
|
||||||
name: web
|
name: web
|
||||||
|
|
|
@ -10,30 +10,3 @@
|
||||||
apt:
|
apt:
|
||||||
name: [ufw]
|
name: [ufw]
|
||||||
state: absent
|
state: absent
|
||||||
|
|
||||||
- name: check which users exist
|
|
||||||
ansible.builtin.user:
|
|
||||||
name: "{{ item }}"
|
|
||||||
loop: ["root", "finn", "debian"]
|
|
||||||
check_mode: true
|
|
||||||
register: users
|
|
||||||
|
|
||||||
- name: Ensure SSH key is set
|
|
||||||
ansible.posix.authorized_key:
|
|
||||||
user: "{{ item.item }}"
|
|
||||||
state: present
|
|
||||||
key: ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIJs5PJ6jQF7Sx3T1b1+NBXt4JRsnjGnWv8+bCf4RpwGM finn@taint
|
|
||||||
loop: "{{ users.results }}"
|
|
||||||
loop_control:
|
|
||||||
label: "{{ item.item }}"
|
|
||||||
when: item.state | d('') == 'present'
|
|
||||||
|
|
||||||
- name: Invalidate old SSH key
|
|
||||||
ansible.posix.authorized_key:
|
|
||||||
user: "{{ item.item }}"
|
|
||||||
state: absent
|
|
||||||
key: ssh-ed25519 AAAAC3NzaC1lZDI1NTE5AAAAIDMaJrZWSrAYTaCoGhW+o8HivmBj5oZi7Dei73FtCl0d finn@taint
|
|
||||||
loop: "{{ users.results }}"
|
|
||||||
loop_control:
|
|
||||||
label: "{{ item.item }}"
|
|
||||||
when: item.state | d('') == 'present'
|
|
||||||
|
|
|
@ -2,13 +2,12 @@ SECRET_KEY={{ lookup('ansible.builtin.ini', 'pdns_admin_secret section=pdns file
|
||||||
OIDC_OAUTH_ENABLED=true
|
OIDC_OAUTH_ENABLED=true
|
||||||
OIDC_OAUTH_KEY=powerdnsadmin
|
OIDC_OAUTH_KEY=powerdnsadmin
|
||||||
OIDC_OAUTH_SECRET={{ lookup('ansible.builtin.ini', 'oidc_secret section=pdns file=secrets/' + inventory_hostname + '.ini') }}
|
OIDC_OAUTH_SECRET={{ lookup('ansible.builtin.ini', 'oidc_secret section=pdns file=secrets/' + inventory_hostname + '.ini') }}
|
||||||
OIDC_OAUTH_API_URL=https://auth.janky.solutions/realms/janky.solutions/protocol/openid-connect/
|
OIDC_OAUTH_API_URL=https://auth.janky.solutions/auth/realms/janky.solutions/protocol/openid-connect/
|
||||||
OIDC_OAUTH_METADATA_URL=https://auth.janky.solutions/realms/janky.solutions/.well-known/openid-configuration
|
OIDC_OAUTH_METADATA_URL=https://auth.janky.solutions/auth/realms/janky.solutions/.well-known/openid-configuration
|
||||||
OIDC_OAUTH_LOGOUT_URL=https://auth.janky.solutions/realms/janky.solutions/protocol/openid-connect/logout
|
OIDC_OAUTH_LOGOUT_URL=https://auth.janky.solutions/auth/realms/janky.solutions/protocol/openid-connect/logout
|
||||||
OIDC_OAUTH_USERNAME=preferred_username
|
OIDC_OAUTH_USERNAME=preferred_username
|
||||||
OIDC_OAUTH_FIRSTNAME=given_name
|
OIDC_OAUTH_FIRSTNAME=given_name
|
||||||
OIDC_OAUTH_LAST_NAME=family_name
|
OIDC_OAUTH_LAST_NAME=family_name
|
||||||
OIDC_OAUTH_EMAIL=email
|
OIDC_OAUTH_EMAIL=email
|
||||||
OIDC_OAUTH_SCOPE=openid email
|
|
||||||
SIGNUP_ENABLED=false
|
SIGNUP_ENABLED=false
|
||||||
LOCAL_DB_ENABLED=false
|
LOCAL_DB_ENABLED=false
|
||||||
|
|
|
@ -5,8 +5,6 @@ Wants=network.target
|
||||||
[Service]
|
[Service]
|
||||||
Type=simple
|
Type=simple
|
||||||
ExecStartPre=/usr/bin/podman pull docker.io/powerdnsadmin/pda-legacy:latest
|
ExecStartPre=/usr/bin/podman pull docker.io/powerdnsadmin/pda-legacy:latest
|
||||||
ExecStartPre=-/usr/bin/podman stop powerdns-admin
|
|
||||||
ExecStartPre=-/usr/bin/podman rm powerdns-admin
|
|
||||||
ExecStart=/usr/bin/podman run --rm -v pda-data:/data -p 9191:80 --env-file /etc/powerdns-admin.env --name powerdns-admin docker.io/powerdnsadmin/pda-legacy:latest
|
ExecStart=/usr/bin/podman run --rm -v pda-data:/data -p 9191:80 --env-file /etc/powerdns-admin.env --name powerdns-admin docker.io/powerdnsadmin/pda-legacy:latest
|
||||||
Restart=always
|
Restart=always
|
||||||
|
|
||||||
|
|
Loading…
Reference in a new issue