disable nvidia-gpu support from cluster
All checks were successful
renovate / renovate (push) Successful in 6m50s

This commit is contained in:
2025-10-12 03:06:30 +05:30
parent a7002f6f50
commit 083ad61920
10 changed files with 3 additions and 4 deletions

View File

@@ -0,0 +1,14 @@
---
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
name: open-webui-longhorn
namespace: default
spec:
accessModes:
- ReadWriteOnce
volumeMode: Filesystem
resources:
requests:
storage: 2Gi
storageClassName: longhorn

View File

@@ -0,0 +1,13 @@
---
apiVersion: v1
kind: Service
metadata:
name: open-webui-service
namespace: default
spec:
type: LoadBalancer
selector:
app: open-webui
ports:
- port: 4321
targetPort: 8080

View File

@@ -0,0 +1,32 @@
---
apiVersion: apps/v1
kind: Deployment
metadata:
name: open-webui
namespace: default
spec:
strategy:
type: Recreate
selector:
matchLabels:
app: open-webui
template:
metadata:
labels:
app: open-webui
spec:
containers:
- name: open-webui
image: ghcr.io/open-webui/open-webui:0.6.33
ports:
- containerPort: 8080
env:
- name: OLLAMA_BASE_URL
value: "http://10.0.0.14:3210"
volumeMounts:
- name: config
mountPath: /app/backend/data
volumes:
- name: config
persistentVolumeClaim:
claimName: open-webui-longhorn

View File

@@ -0,0 +1,18 @@
apiVersion: v1
kind: ConfigMap
metadata:
name: time-slicing-config
namespace: gpu-operator
data:
any: |-
version: v1
flags:
migStrategy: none
sharing:
timeSlicing:
resources:
- name: nvidia.com/gpu
replicas: 4
# remember to patch the cluster policy to use this configmap
# kubectl patch clusterpolicy/cluster-policy -n gpu-operator --type merge -p '{"spec": {"devicePlugin": {"config": {"name": "time-slicing-config", "default": "any"}}}}'

View File

@@ -0,0 +1,31 @@
---
apiVersion: helm.toolkit.fluxcd.io/v2
kind: HelmRelease
metadata:
name: gpu-operator
namespace: gpu-operator
spec:
interval: 24h
chart:
spec:
chart: gpu-operator
version: "v25.3.4"
sourceRef:
kind: HelmRepository
name: nvidia
namespace: flux-system
interval: 24h
install:
createNamespace: true
upgrade:
remediation:
remediateLastFailure: true
values:
driver:
enabled: false
toolkit:
env:
- name: CONTAINERD_SOCKET
value: /run/k3s/containerd/containerd.sock
- name: CONTAINERD_CONFIG
value: /var/lib/rancher/k3s/agent/etc/containerd/config.toml

View File

@@ -0,0 +1,9 @@
---
apiVersion: source.toolkit.fluxcd.io/v1
kind: HelmRepository
metadata:
name: nvidia
namespace: flux-system
spec:
interval: 24h
url: https://helm.ngc.nvidia.com/nvidia

View File

@@ -0,0 +1,14 @@
---
apiVersion: v1
kind: PersistentVolumeClaim
metadata:
name: ollama-longhorn
namespace: default
spec:
accessModes:
- ReadWriteOnce
volumeMode: Filesystem
resources:
requests:
storage: 6Gi
storageClassName: longhorn

View File

@@ -0,0 +1,40 @@
---
apiVersion: helm.toolkit.fluxcd.io/v2
kind: HelmRelease
metadata:
name: ollama
namespace: default
spec:
interval: 24h
chart:
spec:
chart: ollama
version: "1.31.0"
sourceRef:
kind: HelmRepository
name: ollama
namespace: flux-system
interval: 24h
install:
remediation:
retries: 3
upgrade:
remediation:
retries: 3
values:
ollama:
gpu:
enabled: true
type: nvidia
models:
pull:
- gemma3:4b
run:
- gemma3:4b
service:
type: LoadBalancer
port: 3210
runtimeClassName: nvidia
persistentVolume:
enabled: true
existingClaim: ollama-longhorn

View File

@@ -0,0 +1,9 @@
---
apiVersion: source.toolkit.fluxcd.io/v1
kind: HelmRepository
metadata:
name: ollama
namespace: flux-system
spec:
interval: 24h
url: https://otwld.github.io/ollama-helm/