146 Commits

Author SHA1 Message Date
01279dd023 Merge pull request 'Update octodns/octodns Docker tag to v2024.05' (#91) from renovate/octodns-octodns-2024.x into main
Reviewed-on: #91
2024-05-08 13:29:51 +00:00
d6ce07a8a0 Merge pull request 'Update ghcr.io/mealie-recipes/mealie Docker tag to v1.6.0' (#92) from renovate/ghcr.io-mealie-recipes-mealie-1.x into main
Reviewed-on: #92
2024-05-08 13:28:59 +00:00
6eb617086a Update ghcr.io/mealie-recipes/mealie Docker tag to v1.6.0 2024-05-07 12:00:58 +00:00
8137bf8f1b Update apps/immich/kustomization.yaml 2024-05-06 17:59:00 +00:00
5f1dcaabba Update octodns/octodns Docker tag to v2024.05 2024-05-06 15:30:45 +00:00
ca15a6497c Add apps/media/ingress.yaml 2024-05-04 12:10:12 +00:00
095d2d6392 remove limits 2024-05-04 12:47:10 +02:00
b2993c9395 Merge pull request 'Update homeassistant/home-assistant Docker tag to v2024.5' (#86) from renovate/homeassistant-home-assistant-2024.x into main
Reviewed-on: #86
2024-05-04 09:06:57 +00:00
d7b0f658de Merge pull request 'Update actualbudget/actual-server Docker tag to v24.5.0' (#89) from renovate/actualbudget-actual-server-24.x into main
Reviewed-on: #89
2024-05-04 09:06:26 +00:00
391c71729b Update actualbudget/actual-server Docker tag to v24.5.0 2024-05-03 17:00:53 +00:00
bee5dd0c0b Update owncloud/ocis Docker tag to v5.0.3 2024-05-02 16:30:46 +00:00
25ab46e69a change base image for k8s conformity 2024-05-02 17:26:36 +02:00
123412e073 small naming mistake 2024-05-02 17:15:30 +02:00
39818887fa add gitea actions 2024-05-02 17:12:43 +02:00
0700609568 Update homeassistant/home-assistant Docker tag to v2024.5 2024-05-01 19:30:44 +00:00
198b24132e Update Helm release metallb to v0.14.5 2024-04-27 09:13:11 +00:00
f6e45d089b Update docker.io/bitnami/sealed-secrets-controller Docker tag to v0.26.2 2024-04-27 09:12:35 +00:00
23eab57208 Update Helm release cloudnative-pg to v0.21.0 2024-04-25 12:01:00 +00:00
a94521f197 update ocis 2024-04-21 12:59:19 +02:00
38f58d86c9 new versiob 2024-04-21 12:26:44 +02:00
76d1c51157 improve thanos/prometheus retention 2024-04-20 19:04:44 +02:00
7aaeeded89 update and improve grafana 2024-04-20 18:37:19 +02:00
9b93016f93 bump immich version 2024-04-20 18:34:52 +02:00
aaf624bb42 bump immich version 2024-04-20 18:27:01 +02:00
8536d91288 Update Helm release immich to v0.6.0 2024-04-20 16:26:29 +00:00
3f62bee199 reduce gitea load by ditching redis 2024-04-20 18:22:16 +02:00
f9f39818a1 Merge pull request 'Update ghcr.io/mealie-recipes/mealie Docker tag to v1.5.1' (#77) from renovate/ghcr.io-mealie-recipes-mealie-1.x into main
Reviewed-on: #77
2024-04-17 16:29:25 +00:00
a73e6dc4db Update ghcr.io/mealie-recipes/mealie Docker tag to v1.5.1 2024-04-17 14:01:07 +00:00
1df7abf987 Merge pull request 'Update ghcr.io/gethomepage/homepage Docker tag to v0.8.12' (#76) from renovate/ghcr.io-gethomepage-homepage-0.x into main
Reviewed-on: #76
2024-04-17 12:54:36 +00:00
0e1bb58c24 Update ghcr.io/gethomepage/homepage Docker tag to v0.8.12 2024-04-17 12:54:36 +00:00
fcd2d2eaa2 Merge pull request 'Update ghcr.io/mealie-recipes/mealie Docker tag to v1.5.0' (#75) from renovate/ghcr.io-mealie-recipes-mealie-1.x into main
Reviewed-on: #75
2024-04-17 12:54:14 +00:00
455790d3c6 Update ghcr.io/mealie-recipes/mealie Docker tag to v1.5.0 2024-04-16 16:31:00 +00:00
cdbcdba25d Update Helm release gitea to v10.1.4 2024-04-16 10:58:27 +00:00
9dcb06678b remove old filesync deployments (nextcloud) 2024-04-16 12:56:54 +02:00
a4fe0a7fe4 add homepage as a deployment 2024-04-16 12:43:33 +02:00
ece9faa60c Merge pull request 'Update octodns/octodns Docker tag to v2024.04' (#72) from renovate/octodns-octodns-2024.x into main
Reviewed-on: #72
2024-04-16 07:54:27 +00:00
d4bea2994c Merge pull request 'Update Helm release traefik to v27' (#66) from renovate/traefik-27.x into main
Reviewed-on: #66
2024-04-16 07:53:47 +00:00
0ec3bf9ea8 Update Helm release traefik to v27 2024-04-12 08:01:16 +00:00
0c5760b22b Update octodns/octodns Docker tag to v2024.04 2024-04-10 16:30:48 +00:00
e144722d59 fix cnpg syncing issues 2024-04-10 14:01:57 +02:00
bf6e7aa10c mabye like that? 2024-04-06 14:33:57 +02:00
ae53c44428 fix servicemonitors 2024-04-06 14:24:06 +02:00
05d5b02347 Update actualbudget/actual-server Docker tag to v24.4.0 2024-04-06 12:22:05 +00:00
337237a0f8 Update ghcr.io/mealie-recipes/mealie Docker tag to v1.4.0 2024-04-06 12:21:39 +00:00
ccc4b13c35 Update adguard/adguardhome Docker tag to v0.107.48 2024-04-06 12:13:15 +00:00
a6a9c7c217 update home assistant and fix prometheus 2024-04-06 14:12:07 +02:00
bc0f29f028 update immich 2024-04-03 14:11:08 +02:00
e2c9d73728 update to dashboards 2024-04-01 13:23:46 +02:00
442c07f031 bad configmap 2024-04-01 13:11:11 +02:00
8fd9fa6f11 better dashboards 2024-04-01 12:21:50 +02:00
516d7e8e09 like that? 2024-04-01 11:57:06 +02:00
acf9d34b10 Merge branch 'main' of ssh://git.kluster.moll.re:2222/remoll/k3s-infra 2024-04-01 11:47:11 +02:00
3ffead0a14 try fixing grafana 2024-04-01 11:47:01 +02:00
b6bdc09efc Update docker.io/bitnami/sealed-secrets-controller Docker tag to v0.26.1 2024-04-01 09:33:23 +00:00
49b21cde52 proper backup config 2024-03-31 19:37:18 +02:00
deed24aa01 try fixing homeassistant again 2024-03-31 19:28:19 +02:00
9cfb98248d update immich 2024-03-31 19:08:14 +02:00
7bc4beefce Update Helm release cloudnative-pg to v0.20.2 2024-03-31 15:19:09 +00:00
ce9ff68c26 Update binwiederhier/ntfy Docker tag to v2.10.0 2024-03-31 15:18:06 +00:00
8249e7ef01 Update adguard/adguardhome Docker tag to v0.107.46 2024-03-31 15:15:00 +00:00
14e65df483 Update Helm release metallb to v0.14.4 2024-03-31 15:14:18 +00:00
f6fef4278b enable wal for grafana? 2024-03-29 00:55:34 +01:00
ef50df8386 slight mistake 2024-03-28 19:45:27 +01:00
b6df7604ed add missing references 2024-03-28 19:22:59 +01:00
a03d869d0c added dashboards 2024-03-28 19:20:28 +01:00
1063349fbe use sealedsecret 2024-03-28 19:17:19 +01:00
b88c212b57 now with correct secret 2024-03-28 19:10:01 +01:00
38a522a8d6 cleaner monitoring 2024-03-28 19:07:42 +01:00
046936f8f6 fix 2024-03-28 14:04:07 +01:00
309cbc08f5 so? 2024-03-28 13:55:57 +01:00
08b4c7eb5e switch ocis to nfs-provisioner 2024-03-28 13:52:44 +01:00
58e632e0b8 migrate mealie pvc 2024-03-28 13:21:50 +01:00
30d02edebc update rss 2024-03-28 13:19:53 +01:00
e30bfe64ae dum dum 2024-03-28 12:59:51 +01:00
764a3eafb7 switch some apps over to nfs-client 2024-03-28 12:40:48 +01:00
eff07665de add nfs-provisioner with sensible path template 2024-03-28 12:29:16 +01:00
571aebe78d now? 2024-03-27 14:15:13 +01:00
91a2ae5fe8 annoying 2024-03-27 14:13:22 +01:00
f12c21ef18 update vikunja 2024-03-27 14:03:55 +01:00
2a96b288bf or like that? 2024-03-27 09:39:58 +01:00
6f3a5aeab2 okey 2024-03-27 09:37:51 +01:00
b001bd3efc maybe like that? 2024-03-27 09:36:22 +01:00
b54794df35 dum-dum 2024-03-27 09:19:00 +01:00
51c8f7c092 fix the db location 2024-03-27 09:15:25 +01:00
cfb1a87a5b now with correct api path 2024-03-27 09:07:01 +01:00
10483431c6 trying todos like that 2024-03-27 09:04:40 +01:00
3a9450da9d now? 2024-03-27 08:34:48 +01:00
374e23ba1e trying to fix immich 2024-03-27 08:32:46 +01:00
66f703f5e1 update to correct location 2024-03-27 08:25:53 +01:00
4b05b53d72 small fixes 2024-03-27 00:38:34 +01:00
cfbc7fcd0d disable typesense 2024-03-27 00:31:41 +01:00
ffed2aea50 add media back 2024-03-27 00:27:57 +01:00
e674bf5b94 slim down the file sync 2024-03-27 00:12:50 +01:00
133af74ae0 missing namespace resource 2024-03-27 00:05:55 +01:00
f648064304 remove nfs-client 2024-03-26 23:50:27 +01:00
c7180f793a trying like that 2024-03-26 22:58:17 +01:00
4fcdaad297 move prometheus to its own config 2024-03-26 22:13:02 +01:00
f4b99ca037 now perhaps? 2024-03-26 11:16:33 +01:00
588bf774f9 or like that? 2024-03-26 10:58:44 +01:00
e18c661dbd typo 2024-03-26 10:57:18 +01:00
7d65ffea6a remove ocis:// 2024-03-26 10:56:34 +01:00
e460b5324a try differently configured todos 2024-03-26 10:55:25 +01:00
6fe166e60c manage todos 2024-03-24 15:31:59 +01:00
6ceb3816fb cleanup with regards to upcoming migration 2024-03-23 11:45:11 +01:00
19b63263e6 whoopsie 2024-03-22 14:57:17 +01:00
20d46d89d2 also manage ocis 2024-03-22 14:54:30 +01:00
7aee6c7cf0 basic auth maybe? 2024-03-22 14:53:29 +01:00
443da20ff9 steps towards a completely managed cluster 2024-03-20 23:45:08 +01:00
84a47b15b6 increase renovate frequency 2024-03-12 21:28:35 +01:00
40259ee57e Update apps/immich/kustomization.yaml 2024-03-12 14:01:08 +00:00
619368a2fd Merge pull request 'Update homeassistant/home-assistant Docker tag to v2024.3' (#54) from renovate/homeassistant-home-assistant-2024.x into main
Reviewed-on: #54
2024-03-12 09:04:37 +00:00
3288966b95 Merge pull request 'Update octodns/octodns Docker tag to v2024.03' (#55) from renovate/octodns-octodns-2024.x into main
Reviewed-on: #55
2024-03-12 09:04:16 +00:00
d12d50b906 Update apps/immich/kustomization.yaml 2024-03-12 09:03:55 +00:00
c7f0221062 Update octodns/octodns Docker tag to v2024.03 2024-03-12 09:02:04 +00:00
7819867091 Update homeassistant/home-assistant Docker tag to v2024.3 2024-03-12 09:01:41 +00:00
dd4c3d7a36 Update apps/immich/kustomization.yaml 2024-03-12 08:37:11 +00:00
e66905402e Merge pull request 'Update Helm release immich to v0.4.0' (#47) from renovate/immich-0.x into main
Reviewed-on: #47
2024-03-12 08:35:56 +00:00
1bdb4522c3 Merge pull request 'Update ghcr.io/mealie-recipes/mealie Docker tag to v1.3.2' (#53) from renovate/ghcr.io-mealie-recipes-mealie-1.x into main
Reviewed-on: #53
2024-03-12 08:32:10 +00:00
b5845479c2 Update ghcr.io/mealie-recipes/mealie Docker tag to v1.3.2 2024-03-10 19:01:42 +00:00
f2f31c4f4e Merge pull request 'Update binwiederhier/ntfy Docker tag to v2.9.0' (#52) from renovate/binwiederhier-ntfy-2.x into main
Reviewed-on: #52
2024-03-10 09:57:10 +00:00
ded829500c Update binwiederhier/ntfy Docker tag to v2.9.0 2024-03-09 11:04:03 +00:00
f762f5451b Merge pull request 'Update adguard/adguardhome Docker tag to v0.107.45' (#51) from renovate/adguard-adguardhome-0.x into main
Reviewed-on: #51
2024-03-08 07:42:50 +00:00
709f21998e Update adguard/adguardhome Docker tag to v0.107.45 2024-03-07 18:01:21 +00:00
47f091be83 Merge pull request 'Update actualbudget/actual-server Docker tag to v24.3.0' (#48) from renovate/actualbudget-actual-server-24.x into main
Reviewed-on: #48
2024-03-07 17:31:17 +00:00
da8be916bf fix bad naming 2024-03-07 13:21:05 +01:00
ad67acb9e7 again 2024-03-07 13:17:50 +01:00
5a7b5a82d7 maybe the service was misconfigured 2024-03-07 13:16:14 +01:00
2c32db61ec why? 2024-03-07 13:13:54 +01:00
141b80d15c man... 2024-03-07 13:11:08 +01:00
bf1d4badbe or directly use the dns name 2024-03-07 13:08:29 +01:00
be48049e22 fix bad syntax 2024-03-07 13:01:21 +01:00
3a629284f3 perhaps now 2024-03-07 12:59:04 +01:00
28c92e727f last chance 2024-03-06 14:48:14 +01:00
9a65c531f1 now? 2024-03-06 14:37:23 +01:00
52a086df73 come on 2024-03-06 14:34:19 +01:00
b728e21a15 expose grpc of store 2024-03-06 14:31:04 +01:00
da32c9c2ce neew 2024-03-06 14:25:47 +01:00
846390600e let's try with query as well 2024-03-06 14:24:07 +01:00
18d7a6b4cb or maybe like that? 2024-03-06 11:34:15 +01:00
31c8e91502 actually don't specify data 2024-03-06 11:31:15 +01:00
f0adf6b5db change user of prometheus to make thanos happy 2024-03-06 08:14:55 +01:00
b24ae9c698 with correct image 2024-03-05 16:44:42 +01:00
f3c108e362 fix 2024-03-05 16:41:54 +01:00
d2a8d92864 also use thanos object store 2024-03-05 16:39:15 +01:00
10816c4bd9 Update actualbudget/actual-server Docker tag to v24.3.0 2024-03-03 20:01:34 +00:00
aca0d4ba21 Update Helm release immich to v0.4.0 2024-03-03 20:01:27 +00:00
142 changed files with 1118 additions and 1857 deletions

6
.gitignore vendored
View File

@@ -1,2 +1,6 @@
# Kubernetes secrets
*.secret.yaml
charts/
main.key
# Helm Chart files
charts/

3
.gitmodules vendored
View File

@@ -1,3 +1,6 @@
[submodule "infrastructure/external-dns/octodns"]
path = infrastructure/external-dns/octodns
url = ssh://git@git.kluster.moll.re:2222/remoll/dns.git
[submodule "apps/monitoring/dashboards"]
path = apps/monitoring/dashboards
url = ssh://git@git.kluster.moll.re:2222/remoll/grafana-dashboards.git

View File

@@ -1,7 +1,6 @@
# Kluster setup and IaaC using argoCD
### Initial setup
#### Requirements:
- A running k3s instance
@@ -28,5 +27,21 @@ The app-of-apps will bootstrap a fully featured cluster with the following compo
- immich
- ...
#### Recap
- install sealedsecrets see [README](./infrastructure/sealedsecrets/README.md)
```bash
kubectl apply -k infrastructure/sealedsecrets
kubectl apply -f infrastructure/sealedsecrets/main.key
kubectl delete pod -n kube-system -l name=sealed-secrets-controller
```
- install argocd
```bash
kubectl apply -k infrastructure/argocd
```
- wait...
### Adding an application
todo

View File

@@ -1,4 +1,4 @@
apiVersion: traefik.containo.us/v1alpha1
apiVersion: traefik.io/v1alpha1
kind: IngressRouteTCP
metadata:
name: adguard-tls-ingress

View File

@@ -10,7 +10,7 @@ resources:
images:
- name: adguard/adguardhome
newName: adguard/adguardhome
newTag: v0.107.44
newTag: v0.107.48
namespace: adguard

View File

@@ -0,0 +1,48 @@
apiVersion: apps/v1
kind: StatefulSet
metadata:
name: ocis-statefulset
spec:
selector:
matchLabels:
app: ocis
serviceName: ocis-web
replicas: 1
template:
metadata:
labels:
app: ocis
spec:
containers:
- name: ocis
image: ocis
resources:
limits:
memory: "1Gi"
cpu: "1000m"
env:
- name: OCIS_INSECURE
value: "true"
- name: OCIS_URL
value: "https://ocis.kluster.moll.re"
- name: OCIS_LOG_LEVEL
value: "debug"
ports:
- containerPort: 9200
volumeMounts:
- name: config
mountPath: /etc/ocis
# - name: ocis-config-file
# mountPath: /etc/ocis/config.yaml
- name: data
mountPath: /var/lib/ocis
volumes:
# - name: ocis-config
# persistentVolumeClaim:
# claimName: ocis-config
- name: config
secret:
secretName: ocis-config
- name: data
persistentVolumeClaim:
claimName: ocis

18
apps/files/ingress.yaml Normal file
View File

@@ -0,0 +1,18 @@
apiVersion: traefik.io/v1alpha1
kind: IngressRoute
metadata:
name: ocis-ingressroute
spec:
entryPoints:
- websecure
routes:
- match: Host(`ocis.kluster.moll.re`)
kind: Rule
services:
- name: ocis-web
port: 9200
scheme: https
tls:
certResolver: default-tls

View File

@@ -0,0 +1,16 @@
apiVersion: kustomize.config.k8s.io/v1beta1
kind: Kustomization
resources:
- namespace.yaml
- ingress.yaml
- service.yaml
- pvc.yaml
- deployment.yaml
- ocis-config.sealedsecret.yaml
namespace: files
images:
- name: ocis
newName: owncloud/ocis
newTag: "5.0.3"

File diff suppressed because one or more lines are too long

View File

@@ -1,11 +1,11 @@
kind: PersistentVolumeClaim
apiVersion: v1
metadata:
name: syncthing-claim
name: ocis
spec:
storageClassName: nfs-client
storageClassName: "nfs-client"
accessModes:
- ReadWriteOnce
resources:
requests:
storage: 10Gi
storage: 150Gi

10
apps/files/service.yaml Normal file
View File

@@ -0,0 +1,10 @@
apiVersion: v1
kind: Service
metadata:
name: ocis-web
spec:
selector:
app: ocis
ports:
- port: 9200
targetPort: 9200

View File

@@ -22,13 +22,13 @@ spec:
- name: TZ
value: Europe/Berlin
volumeMounts:
- name: actualbudget-data-nfs
- name: data
mountPath: /data
ports:
- containerPort: 5006
name: http
protocol: TCP
volumes:
- name: actualbudget-data-nfs
- name: data
persistentVolumeClaim:
claimName: actualbudget-data-nfs
claimName: data

View File

@@ -1,4 +1,4 @@
apiVersion: traefik.containo.us/v1alpha1
apiVersion: traefik.io/v1alpha1
kind: IngressRoute
metadata:
name: actualbudget

View File

@@ -1,27 +1,11 @@
apiVersion: v1
kind: PersistentVolume
metadata:
name: "actualbudget-data-nfs"
spec:
capacity:
storage: "5Gi"
accessModes:
- ReadWriteOnce
nfs:
path: /export/kluster/actualbudget
server: 192.168.1.157
---
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
name: "actualbudget-data-nfs"
name: "data"
spec:
storageClassName: ""
storageClassName: "nfs-client"
accessModes:
- ReadWriteOnce
resources:
requests:
storage: "5Gi"
volumeName: actualbudget-data-nfs

View File

@@ -13,4 +13,4 @@ resources:
images:
- name: actualbudget
newName: actualbudget/actual-server
newTag: 24.2.0
newTag: 24.5.0

View File

@@ -1,4 +1,3 @@
apiVersion: apps/v1
kind: Deployment
metadata:
@@ -22,7 +21,7 @@ spec:
- name: TZ
value: Europe/Berlin
volumeMounts:
- name: config
- name: config-dir
mountPath: /config
resources:
requests:
@@ -32,6 +31,7 @@ spec:
cpu: "2"
memory: "1Gi"
volumes:
- name: config
- name: config-dir
persistentVolumeClaim:
claimName: homeassistant-nfs
claimName: config

View File

@@ -1,4 +1,4 @@
apiVersion: traefik.containo.us/v1alpha1
apiVersion: traefik.io/v1alpha1
kind: IngressRoute
metadata:
name: homeassistant-ingress
@@ -6,7 +6,7 @@ spec:
entryPoints:
- websecure
routes:
- match: Host(`home.kluster.moll.re`)
- match: Host(`home.kluster.moll.re`) && !Path(`/api/prometheus`)
middlewares:
- name: homeassistant-websocket
kind: Rule
@@ -15,9 +15,8 @@ spec:
port: 8123
tls:
certResolver: default-tls
---
apiVersion: traefik.containo.us/v1alpha1
apiVersion: traefik.io/v1alpha1
kind: Middleware
metadata:
name: homeassistant-websocket
@@ -27,6 +26,3 @@ spec:
X-Forwarded-Proto: "https"
# enable websockets
Upgrade: "websocket"

View File

@@ -9,8 +9,10 @@ resources:
- pvc.yaml
- service.yaml
- deployment.yaml
- servicemonitor.yaml
images:
- name: homeassistant/home-assistant
newName: homeassistant/home-assistant
newTag: "2024.2"
newTag: "2024.5"

View File

@@ -1,28 +1,11 @@
---
apiVersion: v1
kind: PersistentVolume
metadata:
name: homeassistant-nfs
spec:
# storageClassName: slow
capacity:
storage: "1Gi"
# volumeMode: Filesystem
accessModes:
- ReadWriteOnce
nfs:
path: /kluster/homeassistant
server: 192.168.1.157
---
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
name: homeassistant-nfs
name: config
spec:
storageClassName: ""
storageClassName: "nfs-client"
accessModes:
- ReadWriteOnce
resources:
requests:
storage: "1Gi"
volumeName: homeassistant-nfs

View File

@@ -7,4 +7,5 @@ spec:
app: homeassistant
ports:
- port: 8123
targetPort: 8123
targetPort: 8123
name: http

View File

@@ -0,0 +1,13 @@
apiVersion: monitoring.coreos.com/v1
kind: ServiceMonitor
metadata:
name: homeassistant-servicemonitor
labels:
app: homeassistant
spec:
selector:
matchLabels:
app: homeassistant
endpoints:
- port: http
path: /api/prometheus

View File

@@ -0,0 +1,98 @@
apiVersion: v1
kind: ConfigMap
metadata:
name: config
labels:
app.kubernetes.io/name: homepage
data:
kubernetes.yaml: "" #|
# mode: cluster
settings.yaml: |
title: "Homepage"
background: https://images.unsplash.com/photo-1547327132-5d20850c62b5?q=80&w=3870&auto=format&fit=crop
cardBlur: sm
#settings.yaml: |
# providers:
# longhorn:
# url: https://longhorn.my.network
custom.css: ""
custom.js: ""
bookmarks.yaml: |
- Developer:
- Github:
- abbr: GH
href: https://github.com/moll-re
services.yaml: |
- Media:
- Jellyfin backend:
href: https://media-backend.kluster.moll.re
ping: media-backend.kluster.moll.re
- Jellyfin vue:
href: https://media.kluster.moll.re
ping: media.kluster.moll.re
- Immich:
href: https://immich.kluster.moll.re
ping: immich.kluster.moll.re
- Productivity:
- OwnCloud:
href: https://ocis.kluster.moll.re
ping: ocis.kluster.moll.re
- ToDo:
href: https://todos.kluster.moll.re
ping: todos.kluster.moll.re
- Finance:
href: https://finance.kluster.moll.re
ping: finance.kluster.moll.re
- Home:
- Home Assistant:
href: https://home.kluster.moll.re
ping: home.kluster.moll.re
- Grafana:
href: https://grafana.kluster.moll.re
ping: grafana.kluster.moll.re
- Recipes:
href: https://recipes.kluster.moll.re
ping: recipes.kluster.moll.re
- Infra:
- Gitea:
href: https://git.kluster.moll.re
ping: git.kluster.moll.re
- ArgoCD:
href: https://argocd.kluster.moll.re
ping: argocd.kluster.moll.re
widgets.yaml: |
# - kubernetes:
# cluster:
# show: true
# cpu: true
# memory: true
# showLabel: true
# label: "cluster"
# nodes:
# show: true
# cpu: true
# memory: true
# showLabel: true
- search:
provider: duckduckgo
- openmeteo:
label: Zürich # optional
latitude: 47.24236
longitude: 8.30439
units: metric # or imperial
cache: 30 # Time in minutes to cache API responses, to stay within limits
format: # optional, Intl.NumberFormat options
maximumFractionDigits: 1
- datetime:
locale: de
format:
dateStyle: long
timeStyle: short
- adguard:
url: http://adguard-home-web.adguard-home:3000
docker.yaml: ""

View File

@@ -0,0 +1,64 @@
apiVersion: apps/v1
kind: Deployment
metadata:
name: homepage
labels:
app.kubernetes.io/name: homepage
spec:
revisionHistoryLimit: 3
replicas: 1
strategy:
type: RollingUpdate
selector:
matchLabels:
app.kubernetes.io/name: homepage
template:
metadata:
labels:
app.kubernetes.io/name: homepage
spec:
# serviceAccountName: homepage
# automountServiceAccountToken: true
dnsPolicy: ClusterFirst
# enableServiceLinks: true
containers:
- name: homepage
image: homepage
imagePullPolicy: Always
ports:
- name: http
containerPort: 3000
protocol: TCP
volumeMounts:
- mountPath: /app/config/custom.js
name: config
subPath: custom.js
- mountPath: /app/config/custom.css
name: config
subPath: custom.css
- mountPath: /app/config/bookmarks.yaml
name: config
subPath: bookmarks.yaml
- mountPath: /app/config/docker.yaml
name: config
subPath: docker.yaml
- mountPath: /app/config/kubernetes.yaml
name: config
subPath: kubernetes.yaml
- mountPath: /app/config/services.yaml
name: config
subPath: services.yaml
- mountPath: /app/config/settings.yaml
name: config
subPath: settings.yaml
- mountPath: /app/config/widgets.yaml
name: config
subPath: widgets.yaml
- mountPath: /app/config/logs
name: logs
volumes:
- name: config
configMap:
name: config
- name: logs
emptyDir: {}

View File

@@ -0,0 +1,16 @@
apiVersion: traefik.io/v1alpha1
kind: IngressRoute
metadata:
name: homepage-ingressroute
spec:
entryPoints:
- websecure
routes:
- match: Host(`start.kluster.moll.re`)
kind: Rule
services:
- name: homepage-web
port: 3000
tls:
certResolver: default-tls

View File

@@ -0,0 +1,17 @@
apiVersion: kustomize.config.k8s.io/v1beta1
kind: Kustomization
namespace: homepage
resources:
- namespace.yaml
- deployment.yaml
- service.yaml
- configmap.yaml
- ingress.yaml
images:
- name: homepage
newName: ghcr.io/gethomepage/homepage
newTag: v0.8.12

View File

@@ -0,0 +1,15 @@
apiVersion: v1
kind: Service
metadata:
name: homepage-web
labels:
app.kubernetes.io/name: homepage
spec:
type: ClusterIP
ports:
- port: 3000
targetPort: http
protocol: TCP
name: http
selector:
app.kubernetes.io/name: homepage

View File

@@ -1,4 +1,4 @@
apiVersion: traefik.containo.us/v1alpha1
apiVersion: traefik.io/v1alpha1
kind: Middleware
metadata:
name: stripprefix
@@ -7,7 +7,7 @@ spec:
prefixes:
- /api
---
apiVersion: traefik.containo.us/v1alpha1
apiVersion: traefik.io/v1alpha1
kind: Middleware
metadata:
name: websocket
@@ -18,7 +18,7 @@ spec:
# enable websockets
Upgrade: "websocket"
---
apiVersion: traefik.containo.us/v1alpha1
apiVersion: traefik.io/v1alpha1
kind: IngressRoute
metadata:
name: immich-ingressroute

View File

@@ -12,14 +12,13 @@ namespace: immich
helmCharts:
- name: immich
releaseName: immich
version: 0.3.1
version: 0.6.0
valuesFile: values.yaml
repo: https://immich-app.github.io/immich-charts
images:
- name: ghcr.io/immich-app/immich-machine-learning
newTag: v1.95.1
newTag: v1.103.1
- name: ghcr.io/immich-app/immich-server
newTag: v1.95.1
newTag: v1.103.1

View File

@@ -10,7 +10,7 @@ spec:
initdb:
owner: immich
database: immich
secret:
secret:
name: postgres-password
postgresql:
@@ -19,7 +19,12 @@ spec:
storage:
size: 1Gi
storageClass: nfs-client
pvcTemplate:
storageClassName: ""
resources:
requests:
storage: "1Gi"
volumeName: immich-postgres
monitoring:
enablePodMonitor: true

View File

@@ -24,3 +24,17 @@ spec:
requests:
storage: "50Gi"
volumeName: immich-nfs
---
apiVersion: v1
kind: PersistentVolume
metadata:
name: immich-postgres
spec:
capacity:
storage: "1Gi"
accessModes:
- ReadWriteOnce
nfs:
path: /kluster/immich-postgres
server: 192.168.1.157
# later used by cnpg

View File

@@ -1,4 +1,4 @@
apiVersion: traefik.containo.us/v1alpha1
apiVersion: traefik.io/v1alpha1
kind: IngressRoute
metadata:
name: jellyfin-vue-ingress
@@ -17,7 +17,7 @@ spec:
tls:
certResolver: default-tls
---
apiVersion: traefik.containo.us/v1alpha1
apiVersion: traefik.io/v1alpha1
kind: IngressRoute
metadata:
name: jellyfin-backend-ingress
@@ -37,7 +37,7 @@ spec:
tls:
certResolver: default-tls
---
apiVersion: traefik.containo.us/v1alpha1
apiVersion: traefik.io/v1alpha1
kind: Middleware
metadata:
name: jellyfin-websocket
@@ -48,7 +48,7 @@ spec:
Connection: keep-alive, Upgrade
Upgrade: WebSocket
---
apiVersion: traefik.containo.us/v1alpha1
apiVersion: traefik.io/v1alpha1
kind: Middleware
metadata:
name: jellyfin-server-headers
@@ -60,4 +60,4 @@ spec:
accessControlAllowMethods: [ "GET","HEAD","OPTIONS" ] # "POST","PUT"
accessControlAllowOriginList:
- "*"
accessControlMaxAge: 100
accessControlMaxAge: 100

View File

@@ -10,7 +10,6 @@ resources:
- web.deployment.yaml
- web.service.yaml
- ingress.yaml
- jellyfin.servicemonitor.yaml
images:
- name: jellyfin/jellyfin

View File

@@ -1,39 +1,21 @@
apiVersion: v1
kind: PersistentVolume
metadata:
namespace: media
name: jellyfin-config-nfs
spec:
capacity:
storage: "1Gi"
accessModes:
- ReadWriteOnce
nfs:
path: /export/kluster/jellyfin-config
server: 192.168.1.157
---
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
namespace: media
name: jellyfin-config-nfs
name: config
spec:
storageClassName: ""
storageClassName: "nfs-client"
accessModes:
- ReadWriteOnce
resources:
requests:
storage: "1Gi"
volumeName: jellyfin-config-nfs
---
apiVersion: v1
kind: PersistentVolume
metadata:
namespace: media
name: jellyfin-data-nfs
name: media
spec:
capacity:
storage: "1Ti"
@@ -46,8 +28,7 @@ spec:
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
namespace: media
name: jellyfin-data-nfs
name: media
spec:
storageClassName: ""
accessModes:
@@ -55,4 +36,4 @@ spec:
resources:
requests:
storage: "1Ti"
volumeName: jellyfin-data-nfs
volumeName: media

View File

@@ -25,9 +25,9 @@ spec:
- name: TZ
value: Europe/Berlin
volumeMounts:
- name: jellyfin-config
- name: config
mountPath: /config
- name: jellyfin-data
- name: media
mountPath: /media
livenessProbe:
httpGet:
@@ -36,10 +36,10 @@ spec:
initialDelaySeconds: 100
periodSeconds: 15
volumes:
- name: jellyfin-config
- name: config
persistentVolumeClaim:
claimName: jellyfin-config-nfs
- name: jellyfin-data
claimName: config
- name: media
persistentVolumeClaim:
claimName: jellyfin-data-nfs
claimName: media

View File

@@ -0,0 +1,17 @@
---
apiVersion: bitnami.com/v1alpha1
kind: SealedSecret
metadata:
creationTimestamp: null
name: grafana-admin-secret
namespace: monitoring
spec:
encryptedData:
password: AgBe8isrCWd5MuaQq5CpA+P3fDizCCDo23BVauaBJLuMRIYbVwpfahaJW7Ocj3LTXwdeVVPBrOk2D6vESUXu6I0EWc3y/NFN4ZezScxMcjmeaAb+z1zWwdH0FynTPJYOxv1fis1FDTkXDmGy3FXo5NDK9ET899TtulKFkh7UqSxdrRWbD3pegJgqKGPIqDCTAxZN/ssiccfWGS4lHqQBJkXn8DeampcKwjOCvgaBdilF03GoSfpgsqa2Iw2SfTDEobWBWVMMK/RB3/Oi/YJkGwMW3ECUxvTDam8gb0RFA1xjWXoYTLVVP5fK7q7x63ns51HebloxAP1GBrt138N/iDrfbGfjNP8Lx0NFl5y5bTgYN/z8DVTOFf90xxWe+YYERdwllg0Ci1JLNbA+NszXTD4L/HC7a8XuBfjRzxMTeymNjR76jzfPkH6v1EvesOduTfSrahPgS0qS+eGOier1rHxj3EBRhOScY1ut5Bq4oJMNId9nMVbVa6xyq2HyxuJHXV+j6h5FGHmEXn9gIR7wGp8RhtPhKgVGLrHcbHZ5Th2E7eomz1T2NK/ezNP8ZhcwOj/lyGywlW0vhU798zpWhMf57k2OPeuMlfs8Y8y74epBdyBjsrMR4EDctF8RZR3vraxENiMJ6kk1gqKj04ir6HwL7blqwiybIFFnJrp2j7MzgjS4SQ687qMX5Zf5XT03aEE+9W9Epy73tT7zVQKdENCQlcm5
user: AgAdiOivMn0d+nYjYycMZz9QSiS/9QqwHPJQMHkE7/IOou+CJtBknlETNtdv84KZgBQTucufYqu3LR3djOBpdnQsYbIXDxPFgRZQ11pwu/sO2EGifDk218yyzzfZMvx1FL7JL4LI1rKoiHycZowCwsAjEtlICVOOYv1/Plki+6MHXiAGG4r/yUhugGx3VLLX+Poq8oaTeHndgSsFXJege8SfgYR4TsC7pQgsM1UQEFncGIhJYTD2ashmUxFJ+7CJjHqPR0lFRrZXmFvPwTYTCMT+tnSHnCFWtTht8cEi1NxA4kD/eKEX0rOol15EUZnFUws2WqWI634TbyGwZ7km/Yw4XoDxiQR4ar6ulkqb/djcc3cWDYE7PF1m1c+r3iog85S5CSfZ5EvdCHHrbPN9uO2gmoRQWiR5qI70YMxBSnkeLZWN05O1vUuopdXFDTafY7YskxLEdIGHGqFUpUrJZOvBB0zNBdHGgYxFzb5pNmMCC5LPlOuoKjV4yskh9Tgovz06aAvsPxn2WWx6NOJambeziKB5OmSKvPsFofViyGBekVAWSWtt9yJe6lu5OKpBEiA6xhGhQ4ZryTXu9wvVALuPSIwBFITv85sIxjJb80qhJ51wb12QgzLLcPby0HSanyBI1M4jfsXWpK8gIAbDNO+eD7z3PhD9Y/5hPqYKXZ37Geyq23xiyxG8XDj6cL+Ie6k8XipayI4=
template:
metadata:
creationTimestamp: null
name: grafana-admin-secret
namespace: monitoring
type: Opaque

View File

@@ -1,5 +1,5 @@
kind: IngressRoute
apiVersion: traefik.containo.us/v1alpha1
apiVersion: traefik.io/v1alpha1
metadata:
name: grafana-ingress
spec:

View File

@@ -1,4 +1,3 @@
replicas: 1
## Create a headless service for the deployment
@@ -10,13 +9,6 @@ headlessService: false
##
service:
enabled: true
type: ClusterIP
port: 80
targetPort: 3000
# targetPort: 4181 To be used with a proxy extraContainer
annotations: {}
labels: {}
portName: service
serviceMonitor:
## If true, a ServiceMonitor CRD is created for a prometheus operator
@@ -24,42 +16,54 @@ serviceMonitor:
##
enabled: false
ingress:
enabled: false
persistence:
type: pvc
enabled: true
# storageClassName: default
accessModes:
- ReadWriteOnce
size: 10Gi
# annotations: {}
finalizers:
- kubernetes.io/pvc-protection
# selectorLabels: {}
## Sub-directory of the PV to mount. Can be templated.
# subPath: ""
## Name of an existing PVC. Can be templated.
existingClaim: grafana-nfs
## If persistence is not enabled, this allows to mount the
## local storage in-memory to improve performance
##
inMemory:
# credentials
admin:
existingSecret: grafana-admin-secret
userKey: user
passwordKey: password
datasources:
datasources.yaml:
apiVersion: 1
datasources:
- name: Thanos
type: prometheus
url: http://thanos-querier.prometheus.svc:9090
isDefault: true
- name: Prometheus
type: prometheus
url: http://prometheus.prometheus.svc:9090
isDefault: false
dashboardProviders:
dashboardproviders.yaml:
apiVersion: 1
providers:
- name: 'default'
orgId: 1
folder: ''
type: file
disableDeletion: false
editable: true
options:
path: /var/lib/grafana/dashboards/default
## Reference to external ConfigMap per provider. Use provider name as key and ConfigMap name as value.
## A provider dashboards must be defined either by external ConfigMaps or in values.yaml, not in both.
## ConfigMap data example:
##
## data:
## example-dashboard.json: |
## RAW_JSON
##
dashboardsConfigMaps:
default: grafana-dashboards
grafana.ini:
wal: true
default_theme: dark
unified_alerting:
enabled: false
## The maximum usage on memory medium EmptyDir would be
## the minimum value between the SizeLimit specified
## here and the sum of memory limits of all containers in a pod
##
# sizeLimit: 300Mi
initChownData:
## If false, data ownership will not be reset at startup
## This allows the prometheus-server to be run with an arbitrary user
##
enabled: true
# Administrator credentials when not using an existing secret (see below)
adminUser: admin
# adminPassword: strongpassword

View File

@@ -6,29 +6,15 @@ namespace: monitoring
resources:
- namespace.yaml
- grafana.pvc.yaml
- influxdb.pvc.yaml
# - influxdb.pvc.yaml
- grafana.ingress.yaml
# prometheus-operator crds
- https://github.com/prometheus-operator/prometheus-operator/releases/download/v0.70.0/bundle.yaml
- prometheus.yaml
- thanos-objstore-config.sealedsecret.yaml
- grafana-admin.sealedsecret.yaml
- dashboards/
helmCharts:
- releaseName: grafana
name: grafana
repo: https://grafana.github.io/helm-charts
version: 7.3.0
version: 7.3.9
valuesFile: grafana.values.yaml
- releaseName: influxdb
name: influxdb2
repo: https://helm.influxdata.com/
version: 2.1.2
valuesFile: influxdb.values.yaml
- releaseName: telegraf-speedtest
name: telegraf
repo: https://helm.influxdata.com/
version: 1.8.39
valuesFile: telegraf-speedtest.values.yaml

View File

@@ -1,16 +0,0 @@
---
apiVersion: bitnami.com/v1alpha1
kind: SealedSecret
metadata:
creationTimestamp: null
name: thanos-objstore-config
namespace: monitoring
spec:
encryptedData:
thanos.yaml: AgCXlr7NO2DoH1R0ngtDFi8rgJaDnW5WSmOMjvXF4GMcEjnn1kwQMLkF0Xz1BUB5GlQkTAg+ZjCWGMlfycBmUnZb+koZK3X1YLsk1BxBxtuSqhj35iQYxKQ7rAlsz7FxUQjK2oiJkFeQmo/rwcw6l6vZJ73+THYSebR9mLQ/H0pnmJM3ldLX4iWL2H8BZ7ftOYdXO7Xv0lk2k2L4O4LgnB1Uedpyk0HLVxAv3VdVU/RFpHm5Q7kudrCMm9ENcJG7qIWuii8GkysvEefbo2phgKn1Zr5XR6SyekuW2e6FyHe9us5Pv5HnJ6Z2+ZyewygaGgHiRqtxRMaLbahICewfSHwyGzeAD2kdgwVyJYXxVPV9qKQvZmj0ZDCDZ5K548mSUq7nNXSI9M9AJBTKUoqb2FXK3pqn4yh9M1l+7Pmno5Fs22blAyGsRqO32GxrYvEXPpdSeqHRjOMYTnbPuteGRKcvmSEUSuHzkeoTzU1Jh4Sg0ygtQUNIKtbwhJm1XpbJ0oaR5ukWMxPfpDv+B5FmrDsU/I+o62+NtCLQLkK6MoRBFiJ1kymtKkM3vQ1CVg4Vtc5Gc2D6mMu5K8kEuUODweBb8qPnYH7ULfTYORldj3d+Fb2mGF5mAU6xHMzbocsdgZpbAzUP/FfJmMMDWf4aW3LJ1mBjUD06KAwPsQvbTm6VInrdXh2QVb4UIp41kbyK8sanHrvh3bprHloxt8OnTZ2HQl+XN+kxYirkVkL34lIlk7KdYCWqO7QqH0ncd9WF0f9mpPGbxo3J
template:
metadata:
creationTimestamp: null
name: thanos-objstore-config
namespace: monitoring
type: Opaque

View File

@@ -1,5 +0,0 @@
### Runninf `occ` commands:
```
su -s /bin/bash www-data -c "php occ user:list"
```

View File

@@ -1,16 +0,0 @@
apiVersion: traefik.containo.us/v1alpha1
kind: IngressRoute
metadata:
name: nextcloud-ingressroute
spec:
entryPoints:
- websecure
routes:
- match: Host(`nextcloud.kluster.moll.re`)
kind: Rule
services:
- name: nextcloud
port: 8080
tls:
certResolver: default-tls

View File

@@ -1,16 +0,0 @@
apiVersion: kustomize.config.k8s.io/v1beta1
kind: Kustomization
resources:
- namespace.yaml
- ingress.yaml
- pvc.yaml
- postgres.sealedsecret.yaml
namespace: nextcloud
helmCharts:
- name: nextcloud
releaseName: nextcloud
version: 4.5.5
valuesFile: values.yaml
repo: https://nextcloud.github.io/helm/

View File

@@ -1,22 +0,0 @@
{
"kind": "SealedSecret",
"apiVersion": "bitnami.com/v1alpha1",
"metadata": {
"name": "postgres-password",
"namespace": "nextcloud",
"creationTimestamp": null
},
"spec": {
"template": {
"metadata": {
"name": "postgres-password",
"namespace": "nextcloud",
"creationTimestamp": null
}
},
"encryptedData": {
"password": "AgCTmvBe9YFnyWOdz02rxr0hTXnWuVLeUt5dpieWMzl4cVMBj7WcyyODWtNd+eQOLARRssGNZAP4C9gH90iVRFAW1aU+NeA76oceXE5Kiiqoc8T30wE5FC6/UbTjQYRH520NF4wcCQKm//iH8o5uI2+NxZW4goeuShibXK9sijFVNXxUuTeXTmaSJjEPyB+pnmPwjzw+qjhkJJADefh9oryy5+t9ecCwXDiI/2ce2n1Vawm/Nq6/0rZMUSsF8XSiTFczKMunuGMhxGEyyx/I8NZd4XMXGSnBo0YZF7jR9+eRHIjuenPHq1kfEid2Ps4fhFSE8mEecnK7w5xE3r0XeTNHQcTId1yYneK/LQfcRkzInuRddytTwTAmsoSjROcjKjAvtyZSM81pFWJsMQ7bSVXOC0K2wvEz9khDT0RIoR/8tMh2G737F15raTe9Ggbgy3DHst4mYIpoWV/slHrOF0vR9j7X+MRN9R1cVtI1coof/tVSWQsLvv0AJfB4/6dUl+i/yNO/j+4c3WolGwqyXd+oxsZK1VrSwSCBZwBO17BmePJL2QsPVRdutq06TrlvGqP4wXySH9LRuHr3sWgr2VuDV00w+UvuU7ExI+16dWh7jrn/rvIBQSJlHDhl5+VpyM0WTMy5kSfO6nits73ZzT7BAoSU7AeQOMj3t+cUiEq9f9dk7em7QxWMuWg6QIJ+ZZ2+CCBms4rSE4x2glOxanNX/HktQg==",
"username": "AgCxJKzhsF7yNJesK5oLJP62kjFnX4UUNQ2NrHl02Hv6MAzi/AUEV3uJSXXIi3H/uMJSMxRpJQjIDsrznYVI0YHOoz1M8/y1dx8xotFv/i0XByI9sMuGtesop7ncmQbEPMaJ3pqTJyaGkEwcsEMGmwwYiRfJHmEhhCYtzEc5IAnx+nmk//HYsrSWKpJGSWl0LvdMJsnsTxrWoJjaYTW3J0Of3VOOmgkuwIFKyXW9S2cUbAco8xVYchbyiHc8LXbS3izyAidRzg1OWyqvTGMIKJDQZ3ibIiXheon5ZeYjj0fkEkv3TrB7WoKdo0090OY1eHabqAPHT8aP+WG1g6TAzbJEtg+zFfYDKIw5Tp1WkRlsD2me4HycGuZbsaXgP5vWlxF5+rULUzUgxfmTRmYTl0H8kIlmUrusZwxR5ZXnSuBJ3n3AMEjmpmTTALakxEFEPDJJoVbgcViLtANwk72yu15FlOxczT22uyW8FMkj9kYzcq/+2a/EjaTo62SnUYJ3UTQXvgMKML1yJD+zym2+xscPNmwZFBPN5BQ/64ru/Z51nWB20fWFgW3Rw67jEQMajmVclmUcASWOjHzO87feEprHeilTH+224IHzpmC4aLz/JtIP9EEvqfDUr3fRrxcgtT1DgxV37vPj6Pqn47MHr39AA850CxjFmb1VcwfH6ygXABFlxnVByZDn7xCyBNswtKJqtw=="
}
}
}

View File

@@ -1,25 +0,0 @@
apiVersion: v1
kind: PersistentVolume
metadata:
name: nextcloud-nfs
spec:
capacity:
storage: "150Gi"
accessModes:
- ReadWriteOnce
nfs:
path: /kluster/nextcloud
server: 192.168.1.157
---
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
name: nextcloud-nfs
spec:
storageClassName: ""
accessModes:
- ReadWriteOnce
resources:
requests:
storage: "150Gi"
volumeName: nextcloud-nfs

View File

@@ -1,17 +0,0 @@
## Running occ commands
Sometimes you need to run a command on the Nextcloud container directly. You can do that by running commands as the user www-data via the kubectl exec command.
```
# $NEXTCLOUD_POD should be the name of *your* nextcloud pod :)
kubectl exec $NEXTCLOUD_POD -- su -s /bin/sh www-data -c "php occ myocccomand"
```
Here are some examples below.
Putting Nextcloud into maintanence mode
Some admin actions require you to put your Nextcloud instance into
(e.g. backups):
```
# $NEXTCLOUD_POD should be the name of *your* nextcloud pod :)
kubectl exec $NEXTCLOUD_POD -- su -s /bin/sh www-data -c "php occ maintenance:mode --on"
```

View File

@@ -1,171 +0,0 @@
## Official nextcloud image version
## ref: https://hub.docker.com/r/library/nextcloud/tags/
image:
tag: "28"
ingress:
enabled: false
nextcloud:
host: nextcloud.kluster.moll.re
username: admin
password: changeme
## Use an existing secret
existingSecret:
enabled: false
update: 0
# If web server is not binding default port, you can define it
# containerPort: 8080
datadir: /var/www/html/data
persistence:
subPath:
mail:
enabled: false
# PHP Configuration files
# Will be injected in /usr/local/etc/php/conf.d for apache image and in /usr/local/etc/php-fpm.d when nginx.enabled: true
phpConfigs: {}
# Default config files
# IMPORTANT: Will be used only if you put extra configs, otherwise default will come from nextcloud itself
# Default confgurations can be found here: https://github.com/nextcloud/docker/tree/master/16.0/apache/config
defaultConfigs:
# To protect /var/www/html/config
.htaccess: true
# Redis default configuration
redis.config.php: true
# Apache configuration for rewrite urls
apache-pretty-urls.config.php: true
# Define APCu as local cache
apcu.config.php: true
# Apps directory configs
apps.config.php: true
# Used for auto configure database
autoconfig.php: true
# SMTP default configuration
smtp.config.php: true
# Extra config files created in /var/www/html/config/
# ref: https://docs.nextcloud.com/server/15/admin_manual/configuration_server/config_sample_php_parameters.html#multiple-config-php-file
configs: {}
# For example, to use S3 as primary storage
# ref: https://docs.nextcloud.com/server/13/admin_manual/configuration_files/primary_storage.html#simple-storage-service-s3
#
# configs:
# s3.config.php: |-
# <?php
# $CONFIG = array (
# 'objectstore' => array(
# 'class' => '\\OC\\Files\\ObjectStore\\S3',
# 'arguments' => array(
# 'bucket' => 'my-bucket',
# 'autocreate' => true,
# 'key' => 'xxx',
# 'secret' => 'xxx',
# 'region' => 'us-east-1',
# 'use_ssl' => true
# )
# )
# );
nginx:
## You need to set an fpm version of the image for nextcloud if you want to use nginx!
enabled: false
internalDatabase:
enabled: true
name: nextcloud
##
## External database configuration
##
externalDatabase:
enabled: true
## Supported database engines: mysql or postgresql
type: postgresql
## Database host
host: postgres-postgresql.postgres
## Database user
# user: nextcloud
# ## Database password
# password: test
## Database name
database: nextcloud
## Use a existing secret
existingSecret:
enabled: true
secretName: postgres-password
usernameKey: username
passwordKey: password
##
## MariaDB chart configuration
##
mariadb:
enabled: false
postgresql:
enabled: false
redis:
enabled: false
## Cronjob to execute Nextcloud background tasks
## ref: https://docs.nextcloud.com/server/latest/admin_manual/configuration_server/background_jobs_configuration.html#webcron
##
cronjob:
enabled: false
## Enable persistence using Persistent Volume Claims
## ref: http://kubernetes.io/docs/user-guide/persistent-volumes/
##
persistence:
# Nextcloud Data (/var/www/html)
enabled: true
annotations: {}
## If defined, PVC must be created manually before volume will be bound
existingClaim: nextcloud-nfs
## Use an additional pvc for the data directory rather than a subpath of the default PVC
## Useful to store data on a different storageClass (e.g. on slower disks)
nextcloudData:
enabled: false
resources:
# We usually recommend not to specify default resources and to leave this as a conscious
# choice for the user. This also increases chances charts run on environments with little
# resources, such as Minikube. If you do want to specify resources, uncomment the following
# lines, adjust them as necessary, and remove the curly braces after 'resources:'.
limits:
cpu: 2000m
memory: 2Gi
requests:
cpu: 100m
memory: 128Mi
livenessProbe:
enabled: true
# disable when upgrading from a previous chart version
## Enable pod autoscaling using HorizontalPodAutoscaler
## ref: https://kubernetes.io/docs/tasks/run-application/horizontal-pod-autoscale/
##
hpa:
enabled: false
## Prometheus Exporter / Metrics
##
metrics:
enabled: false
rbac:
enabled: false

View File

@@ -1,4 +1,4 @@
apiVersion: traefik.containo.us/v1alpha1
apiVersion: traefik.io/v1alpha1
kind: Middleware
metadata:
name: websocket
@@ -9,7 +9,7 @@ spec:
# enable websockets
Upgrade: "websocket"
---
apiVersion: traefik.containo.us/v1alpha1
apiVersion: traefik.io/v1alpha1
kind: IngressRoute
metadata:
name: ntfy-ingressroute

View File

@@ -13,4 +13,4 @@ resources:
images:
- name: binwiederhier/ntfy
newName: binwiederhier/ntfy
newTag: v2.8.0
newTag: v2.10.0

View File

@@ -34,4 +34,4 @@ spec:
volumes:
- name: mealie-data
persistentVolumeClaim:
claimName: mealie-data
claimName: mealie

View File

@@ -1,4 +1,4 @@
apiVersion: traefik.containo.us/v1alpha1
apiVersion: traefik.io/v1alpha1
kind: IngressRoute
metadata:
name: mealie-ingressroute

View File

@@ -12,5 +12,5 @@ resources:
images:
- name: mealie
newTag: v1.2.0
newTag: v1.6.0
newName: ghcr.io/mealie-recipes/mealie

View File

@@ -1,12 +1,12 @@
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
name: mealie-data
name: mealie
spec:
resources:
requests:
storage: 1Gi
storage: 5Gi
volumeMode: Filesystem
storageClassName: nfs-client
storageClassName: "nfs-client"
accessModes:
- ReadWriteOnce

View File

@@ -18,9 +18,9 @@ spec:
ports:
- containerPort: 7070
volumeMounts:
- name: rss-data
- name: data
mountPath: /data
volumes:
- name: rss-data
- name: data
persistentVolumeClaim:
claimName: rss-claim
claimName: data

View File

@@ -1,4 +1,4 @@
apiVersion: traefik.containo.us/v1alpha1
apiVersion: traefik.io/v1alpha1
kind: IngressRoute
metadata:
name: rss-ingressroute

View File

@@ -1,4 +1,4 @@
apiVersion: v1
kind: Namespace
metadata:
name: placeholder
name: placeholder

View File

@@ -1,9 +1,9 @@
kind: PersistentVolumeClaim
apiVersion: v1
metadata:
name: rss-claim
name: data
spec:
storageClassName: nfs-client
storageClassName: "nfs-client"
accessModes:
- ReadWriteOnce
resources:

View File

@@ -1,35 +0,0 @@
apiVersion: apps/v1
kind: Deployment
metadata:
name: syncthing
spec:
selector:
matchLabels:
app: syncthing
template:
metadata:
labels:
app: syncthing
spec:
containers:
- name: syncthing
image: syncthing
resources:
limits:
memory: "256Mi"
cpu: "500m"
ports:
- containerPort: 8384
protocol: TCP
name: syncthing-web
- containerPort: 22000
protocol: TCP
- containerPort: 22000
protocol: UDP
volumeMounts:
- name: persistence
mountPath: /var/syncthing
volumes:
- name: persistence
persistentVolumeClaim:
claimName: syncthing-claim

View File

@@ -1,16 +0,0 @@
apiVersion: traefik.containo.us/v1alpha1
kind: IngressRoute
metadata:
name: rss-ingressroute
spec:
entryPoints:
- websecure
routes:
- match: Host(`syncthing.kluster.moll.re`)
kind: Rule
services:
- name: syncthing-web
port: 8384
tls:
certResolver: default-tls

View File

@@ -1,18 +0,0 @@
apiVersion: kustomize.config.k8s.io/v1beta1
kind: Kustomization
namespace: syncthing
resources:
- namespace.yaml
- pvc.yaml
- deployment.yaml
- service.yaml
- ingress.yaml
- servicemonitor.yaml
- syncthing-api.sealedsecret.yaml
images:
- name: syncthing
newName: syncthing/syncthing
newTag: "1.27"

View File

@@ -1,46 +0,0 @@
apiVersion: v1
kind: Service
metadata:
name: syncthing-web
labels:
app: syncthing
spec:
selector:
app: syncthing
type: ClusterIP
ports:
- port: 8384
targetPort: 8384
name: syncthing-web
---
apiVersion: v1
kind: Service
metadata:
name: syncthing-listen
annotations:
metallb.universe.tf/allow-shared-ip: syncthing-service
spec:
selector:
app: syncthing
type: LoadBalancer
LoadBalancerIP: 192.168.3.4
ports:
- port: 22000
targetPort: 22000
protocol: TCP
---
apiVersion: v1
kind: Service
metadata:
name: syncthing-discover
annotations:
metallb.universe.tf/allow-shared-ip: syncthing-service
spec:
selector:
app: syncthing
type: LoadBalancer
LoadBalancerIP: 192.168.3.4
ports:
- port: 22000
targetPort: 22000
protocol: UDP

View File

@@ -1,17 +0,0 @@
apiVersion: monitoring.coreos.com/v1
kind: ServiceMonitor
metadata:
name: syncthing-servicemonitor
labels:
app: syncthing
spec:
selector:
matchLabels:
app: syncthing
endpoints:
- port: syncthing-web
path: /metrics
bearerTokenSecret:
name: syncthing-api
key: token
namespace: syncthing

View File

@@ -1,16 +0,0 @@
---
apiVersion: bitnami.com/v1alpha1
kind: SealedSecret
metadata:
creationTimestamp: null
name: syncthing-api
namespace: syncthing
spec:
encryptedData:
token: AgC1hG1aguLIWBgA1R19MGrXDq7BONAldMEXtCeGXLO9Xar08f7qFqprtRJAMOID4trUEBMAkF96m7rH7QHTpO0WzRLrJctLi7U6NgESUJBDxusqjij3RAANS69Xt27mu2oa+rhm605CfFJT6Gpx/2CxrFtUD3yCijilDnEVvw4WvTLHvVQMCd8cM8ZDlpBsSYbxvtCUN1+B02DCucLpMphspxV2SGPAdc04xQD7d0vUhNLekFi0xSgu0jiRGVDHOG5Egd9d/BGeNOBgiUVxJxqqdXc6EmkslcSUtMQJ5luSxjogf+p3jdOqt4aPpUeR8sSPb6OSEIZD/Cfs9X4akHdpUAqkycu+V24lDxeHWAtIviCMBPttrwNAEytgwqaiT0U4UmL5GqR97jpmy3Tx+jYKuXkt4Igb6VByreuL9aZacRrqRhCCgbg95Y/UrYlLAbZYOI/+KsFzB5akGpZXUDcW9h2IkTUmcT+QxWXqEoNpoTI5qAnKiu/9T5elDKghjMHYX+CnPj+rXlQIJzX7NkZ0Q6HpKQ4B2Vd1Ewkvadf963jBodUe7WiMt8UeYgzCa33F4U23JjExIrL8t3r8MQ/IIdtfUvyz6Da1vp5hjpBUnUCk8rca/6VC3GO1GP3DLdIXiZQY1OOTHJlyLG7+bIL35zVfkmLMzmlIdaFsfeYiL4P+hYRbLABPAJk8lY7MEdiczpvI9HlmFVatJaPrFJwx9jyhzqIOq5eGt0OIkFt+fw==
template:
metadata:
creationTimestamp: null
name: syncthing-api
namespace: syncthing
type: Opaque

6
apps/todos/README.md Normal file
View File

@@ -0,0 +1,6 @@
### Adding a user
```bash
kubectl exec -it -n todos deployments/todos-vikunja -- /app/vikunja/vikunja user create -u <username> -e "<user-email>"
```
Password will be prompted.

21
apps/todos/ingress.yaml Normal file
View File

@@ -0,0 +1,21 @@
apiVersion: traefik.io/v1alpha1
kind: IngressRoute
metadata:
name: todos-ingressroute
spec:
entryPoints:
- websecure
routes:
- match: Host(`todos.kluster.moll.re`) && PathPrefix(`/api/v1`)
kind: Rule
services:
- name: todos-api
port: 3456
- match: Host(`todos.kluster.moll.re`) && PathPrefix(`/`)
kind: Rule
services:
- name: todos-frontend
port: 80
tls:
certResolver: default-tls

View File

@@ -0,0 +1,18 @@
apiVersion: kustomize.config.k8s.io/v1beta1
kind: Kustomization
namespace: todos
resources:
- namespace.yaml
- pvc.yaml
- ingress.yaml
# helmCharts:
# - name: vikunja
# version: 0.1.5
# repo: https://charts.oecis.io
# valuesFile: values.yaml
# releaseName: todos
# managed by argocd directly

View File

@@ -0,0 +1,4 @@
apiVersion: v1
kind: Namespace
metadata:
name: placeholder

12
apps/todos/pvc.yaml Normal file
View File

@@ -0,0 +1,12 @@
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
name: data
spec:
resources:
requests:
storage: 5Gi
volumeMode: Filesystem
storageClassName: "nfs-client"
accessModes:
- ReadWriteOnce

51
apps/todos/values.yaml Normal file
View File

@@ -0,0 +1,51 @@
######################
# VIKUNJA COMPONENTS #
######################
# You can find the default values that this `values.yaml` overrides, in the comment at the top of this file.
api:
enabled: true
image:
tag: 0.22.1
persistence:
# This is your Vikunja data will live, you can either let
# the chart create a new PVC for you or provide an existing one.
data:
enabled: true
existingClaim: data
accessMode: ReadWriteOnce
size: 10Gi
mountPath: /app/vikunja/files
ingress:
main:
enabled: false
configMaps:
# The configuration for Vikunja's api.
# https://vikunja.io/docs/config-options/
config:
enabled: true
data:
config.yml: |
service:
frontendUrl: https://todos.kluster.moll.re
database:
type: sqlite
path: /app/vikunja/files/vikunja.db
registration: false
env:
frontend:
enabled: true
image:
tag: 0.22.1
ingress:
main:
enabled: false
postgresql:
enabled: false
redis:
enabled: false
typesense:
enabled: false

View File

@@ -1,4 +1,4 @@
apiVersion: traefik.containo.us/v1alpha1
apiVersion: traefik.io/v1alpha1
kind: IngressRoute
metadata:
namespace: whoami

View File

@@ -12,7 +12,9 @@ spec:
destination:
server: https://kubernetes.default.svc
namespace: argocd
syncPolicy:
automated:
prune: true
# selfHeal: true
# syncPolicy:
# automated:
# prune: true
# selfHeal: false
# DO NOT AUTO SYNC THE APP OF APPS.
# all other apps are auto-synced, but adding new apps should be done manually.

View File

@@ -1,5 +1,5 @@
---
apiVersion: traefik.containo.us/v1alpha1
apiVersion: traefik.io/v1alpha1
kind: IngressRoute
metadata:
name: argocd-ingressroute

View File

@@ -0,0 +1,15 @@
# How to restore
1. Port forward the rest api for gcloud
```bash
kubectl port-forward -n backup service/rclone-gcloud 8000
```
2. Load the snapshots locally
```bash
restic -r rest:http://127.0.0.1:8000/kluster mount /mnt/restic
```
(The password is in a secret)
3. Copy relevant files to the correct location on the NAS

View File

@@ -4,7 +4,7 @@ kind: Kustomization
namespace: backup
# nameSuffix: -backup
resources:
- ../../base
- ../../cronjobs-base
# - ./restic-commands.yaml

View File

@@ -3,7 +3,7 @@ kind: Kustomization
namespace: backup
resources:
- ../../base
- ../../cronjobs-base
# patch the cronjob args field:

View File

@@ -1,6 +1,14 @@
apiVersion: kustomize.config.k8s.io/v1beta1
kind: Kustomization
namespace: backup
resources:
- rclone-config.sealedsecret.yaml
- namespace.yaml
- pvc.yaml
- restic-password.sealedsecret.yaml
- pvc.yaml
- rclone-config.sealedsecret.yaml
- rclone-gcloud.deployment.yaml
- cronjobs-overlays/prune/
- cronjobs-overlays/backup/

View File

@@ -0,0 +1,4 @@
apiVersion: v1
kind: Namespace
metadata:
name: placeholder

View File

@@ -1,3 +1,16 @@
apiVersion: v1
kind: PersistentVolume
metadata:
name: pg-backup-data
spec:
capacity:
storage: "50Gi"
accessModes:
- ReadWriteOnce
nfs:
path: /kluster/pg-backup
server: 192.168.1.157
---
kind: PersistentVolumeClaim
apiVersion: v1
@@ -5,9 +18,10 @@ metadata:
name: postgres-backup-claim
spec:
storageClassName: nfs-client
storageClassName: ""
accessModes:
- ReadWriteOnce
resources:
requests:
storage: 10Gi
volumeName: pg-backup-data

View File

@@ -11,7 +11,7 @@ resources:
images:
- name: octodns
newName: octodns/octodns # has all plugins
newTag: "2024.02"
newTag: "2024.05"
- name: git
newName: alpine/git

View File

@@ -0,0 +1,11 @@
apiVersion: kustomize.config.k8s.io/v1beta1
kind: Kustomization
namespace: external
resources:
- namespace.yaml
- omv-s3.ingress.yaml
- openmediavault.ingress.yaml
- proxmox.ingress.yaml

View File

@@ -0,0 +1,4 @@
apiVersion: v1
kind: Namespace
metadata:
name: placeholder

View File

@@ -1,8 +1,7 @@
apiVersion: traefik.containo.us/v1alpha1
apiVersion: traefik.io/v1alpha1
kind: IngressRoute
metadata:
name: omv-s3-ingressroute
namespace: external
spec:
entryPoints:
- websecure
@@ -20,7 +19,6 @@ apiVersion: v1
kind: Endpoints
metadata:
name: omv-s3
namespace: external
subsets:
- addresses:
- ip: 192.168.1.157
@@ -31,7 +29,6 @@ apiVersion: v1
kind: Service
metadata:
name: omv-s3
namespace: external
spec:
ports:
- port: 9000

View File

@@ -1,4 +1,4 @@
apiVersion: traefik.containo.us/v1alpha1
apiVersion: traefik.io/v1alpha1
kind: IngressRoute
metadata:
name: omv-ingressroute

View File

@@ -1,4 +1,4 @@
apiVersion: traefik.containo.us/v1alpha1
apiVersion: traefik.io/v1alpha1
kind: IngressRoute
metadata:
name: proxmox-ingressroute
@@ -19,7 +19,7 @@ spec:
tls:
certResolver: default-tls
---
apiVersion: traefik.containo.us/v1alpha1
apiVersion: traefik.io/v1alpha1
kind: Middleware
metadata:
name: proxmox-websocket

View File

@@ -0,0 +1,31 @@
apiVersion: apps/v1
kind: Deployment
metadata:
name: actions-runner
spec:
selector:
matchLabels:
app: actions-runner
template:
metadata:
labels:
app: actions-runner
spec:
hostname: kube-runner
containers:
- name: actions-runner
image: actions-runner
resources:
requests:
memory: "128Mi"
cpu: "500m"
env:
- name: GITEA_INSTANCE_URL
value: "https://git.kluster.moll.re"
- name: GITEA_RUNNER_REGISTRATION_TOKEN
valueFrom:
secretKeyRef:
name: actions-runner-secret
key: runner-token
- name: GITEA_RUNNER_LABELS
value: k8s

View File

@@ -0,0 +1,16 @@
---
apiVersion: bitnami.com/v1alpha1
kind: SealedSecret
metadata:
creationTimestamp: null
name: actions-runner-secret
namespace: gitea
spec:
encryptedData:
runner-token: AgBHwek/Aj/0oOnI/bnZ4FgtRoeJw4tIKvcDzBhaPdQ7bMVHyHUKYUNP7lkPgZrIN+7rhMY7C/j13iGWx4iTdhTgipLiJvyZ70pXKLSix4IpcypJTElggWkW0JW79x1HyJfBtn9iJiHnEZXPi7sEnyKhA0asAOR0ae8NS6mxxei0TIImaPaC2RHL6MOi40xsXpHz2ZaVhDQaTSRWjv0U6+WkCGcueqM2HLYfF1gqqkzGCjjhdOTK1CKvIvApZ5n8x6x94IiywCXJraDCwLz+acF2c2vA/Jb/3p7TwyyRZ5uIF5LZufhTJ6+5sFJSReHYxO4CpPA8KvM880vtiEjN7LxVo/Jruj2459OvjviKZS03ZwLHHrjanom1+HA9Sx2ffRLiR5ayGkfj/6kvpIRt5x1F7BbPp+a0LXuxJX+1nGDyEa1D1WzVKvZASav6/v7cXcom/nKGO91Zb8qHlOv7ZTs5guGQ9G9VCOHOG8szwpW3ZmQwWfFoWsShzqbDqszBYOGeIjIiDllLzTZ8A9dv9J2ELngZ1IPGIkfpQNEW8hsbNXTYhdVIrkh7BIFkRWfYDNWxqZd4iE6XllQcT1rqndusgiNEJX2r+P4nT8dPewATXQ79wzvZU3kB+VHzM8cLymlVGADi7v/qTY9RcrhuE0oMLzHRShr6JU05VfLGbMsttrYKmW7smvBp3lRJitO5A8+r8cRniS1+Xr8mIx87vCvnoWSH6BKkl9pCdDeCGylAWfkJN9UpkaKg
template:
metadata:
creationTimestamp: null
name: actions-runner-secret
namespace: gitea
type: Opaque

View File

@@ -71,7 +71,7 @@ spec:
selector:
app: drone-server
---
apiVersion: traefik.containo.us/v1alpha1
apiVersion: traefik.io/v1alpha1
kind: IngressRoute
metadata:
name: drone-server-ingress

View File

@@ -1,4 +1,4 @@
apiVersion: traefik.containo.us/v1alpha1
apiVersion: traefik.io/v1alpha1
kind: IngressRoute
metadata:
name: gitea-ingress

View File

@@ -111,6 +111,15 @@ gitea:
SSH_PORT: 2222
actions:
ENABLED: true
session:
PROVIDER: db
cache:
ADAPTER: memory
queue:
TYPE: level
indexer:
ISSUE_INDEXER_TYPE: bleve
REPO_INDEXER_ENABLED: true
@@ -154,12 +163,7 @@ postgresql:
redis-cluster:
enabled: true
usePassword: false
cluster:
nodes: 3 # default: 6
replicas: 0 # default: 1
enabled: false
postgresql-ha:
enabled: false

View File

@@ -8,14 +8,21 @@ resources:
- drone-kube-runner.deployment.yaml
- drone-server.deployment.yaml
- drone-server.sealedsecret.yaml
- actions.deployment.yaml
- actions.sealedsecret.yaml
namespace: gitea
images:
- name: actions-runner
newName: ghcr.io/christopherhx/gitea-actions-runner
newTag: v0.0.11
helmCharts:
- name: gitea
namespace: gitea # needs to be set explicitly for svc to be referenced correctly
releaseName: gitea
version: 10.1.3
version: 10.1.4
valuesFile: gitea.values.yaml
repo: https://dl.gitea.io/charts/

Some files were not shown because too many files have changed in this diff Show More