4 Commits

Author SHA1 Message Date
b02d39fe47 disable gpu 2025-02-28 11:53:53 -06:00
a4395834e4 disable gpu 2025-02-28 11:52:12 -06:00
24a4e58054 disable extra server 2025-02-28 11:50:52 -06:00
renovate[bot]
6b67c2d037 Update Helm release headlamp to v0.29.1 (#1335)
Co-authored-by: renovate[bot] <29139614+renovate[bot]@users.noreply.github.com>
2025-02-28 11:45:17 -06:00
3 changed files with 111 additions and 114 deletions

View File

@@ -42,11 +42,8 @@ photoview:
value: qsv
resources:
requests:
gpu.intel.com/i915: 1
cpu: 10m
memory: 512Mi
limits:
gpu.intel.com/i915: 1
serviceAccount:
create: true
service:

View File

@@ -14,7 +14,7 @@ maintainers:
- name: alexlebens
dependencies:
- name: headlamp
version: 0.28.1
version: 0.29.1
repository: https://headlamp-k8s.github.io/headlamp/
icon: https://raw.githubusercontent.com/walkxcode/dashboard-icons/main/png/kubernetes-dashboard.png
appVersion: 0.27.0

View File

@@ -33,84 +33,84 @@ ollama:
requests:
cpu: 100m
memory: 1Gi
gpu.intel.com/i915: 1
# gpu.intel.com/i915: 1
limits:
cpu: 4
gpu.intel.com/i915: 1
server-2:
type: deployment
replicas: 1
strategy: Recreate
revisionHistoryLimit: 3
pod:
labels:
ollama-type: server
affinity:
podAntiAffinity:
requiredDuringSchedulingIgnoredDuringExecution:
- labelSelector:
matchExpressions:
- key: ollama-type
operator: In
values:
- server
topologyKey: kubernetes.io/hostname
containers:
main:
image:
repository: ollama/ollama
tag: 0.5.12
pullPolicy: IfNotPresent
env:
- name: OLLAMA_KEEP_ALIVE
value: 24h
- name: OLLAMA_HOST
value: 0.0.0.0
resources:
requests:
cpu: 100m
memory: 1Gi
gpu.intel.com/i915: 1
limits:
cpu: 4
gpu.intel.com/i915: 1
server-3:
type: deployment
replicas: 1
strategy: Recreate
revisionHistoryLimit: 3
pod:
labels:
ollama-type: server
affinity:
podAntiAffinity:
requiredDuringSchedulingIgnoredDuringExecution:
- labelSelector:
matchExpressions:
- key: ollama-type
operator: In
values:
- server
topologyKey: kubernetes.io/hostname
containers:
main:
image:
repository: ollama/ollama
tag: 0.5.12
pullPolicy: IfNotPresent
env:
- name: OLLAMA_KEEP_ALIVE
value: 24h
- name: OLLAMA_HOST
value: 0.0.0.0
resources:
requests:
cpu: 100m
memory: 1Gi
gpu.intel.com/i915: 1
limits:
cpu: 4
gpu.intel.com/i915: 1
# gpu.intel.com/i915: 1
# server-2:
# type: deployment
# replicas: 1
# strategy: Recreate
# revisionHistoryLimit: 3
# pod:
# labels:
# ollama-type: server
# affinity:
# podAntiAffinity:
# requiredDuringSchedulingIgnoredDuringExecution:
# - labelSelector:
# matchExpressions:
# - key: ollama-type
# operator: In
# values:
# - server
# topologyKey: kubernetes.io/hostname
# containers:
# main:
# image:
# repository: ollama/ollama
# tag: 0.5.12
# pullPolicy: IfNotPresent
# env:
# - name: OLLAMA_KEEP_ALIVE
# value: 24h
# - name: OLLAMA_HOST
# value: 0.0.0.0
# resources:
# requests:
# cpu: 100m
# memory: 1Gi
# gpu.intel.com/i915: 1
# limits:
# cpu: 4
# gpu.intel.com/i915: 1
# server-3:
# type: deployment
# replicas: 1
# strategy: Recreate
# revisionHistoryLimit: 3
# pod:
# labels:
# ollama-type: server
# affinity:
# podAntiAffinity:
# requiredDuringSchedulingIgnoredDuringExecution:
# - labelSelector:
# matchExpressions:
# - key: ollama-type
# operator: In
# values:
# - server
# topologyKey: kubernetes.io/hostname
# containers:
# main:
# image:
# repository: ollama/ollama
# tag: 0.5.12
# pullPolicy: IfNotPresent
# env:
# - name: OLLAMA_KEEP_ALIVE
# value: 24h
# - name: OLLAMA_HOST
# value: 0.0.0.0
# resources:
# requests:
# cpu: 100m
# memory: 1Gi
# gpu.intel.com/i915: 1
# limits:
# cpu: 4
# gpu.intel.com/i915: 1
web:
type: deployment
replicas: 1
@@ -179,20 +179,20 @@ ollama:
port: 11434
targetPort: 11434
protocol: HTTP
server-2:
controller: server-2
ports:
http:
port: 11434
targetPort: 11434
protocol: HTTP
server-3:
controller: server-3
ports:
http:
port: 11434
targetPort: 11434
protocol: HTTP
# server-2:
# controller: server-2
# ports:
# http:
# port: 11434
# targetPort: 11434
# protocol: HTTP
# server-3:
# controller: server-3
# ports:
# http:
# port: 11434
# targetPort: 11434
# protocol: HTTP
web:
controller: web
ports:
@@ -226,26 +226,26 @@ ollama:
main:
- path: /root/.ollama
readOnly: false
server-2:
storageClass: ceph-block
accessMode: ReadWriteOnce
size: 40Gi
retain: true
advancedMounts:
server-2:
main:
- path: /root/.ollama
readOnly: false
server-3:
storageClass: ceph-block
accessMode: ReadWriteOnce
size: 40Gi
retain: true
advancedMounts:
server-3:
main:
- path: /root/.ollama
readOnly: false
# server-2:
# storageClass: ceph-block
# accessMode: ReadWriteOnce
# size: 40Gi
# retain: true
# advancedMounts:
# server-2:
# main:
# - path: /root/.ollama
# readOnly: false
# server-3:
# storageClass: ceph-block
# accessMode: ReadWriteOnce
# size: 40Gi
# retain: true
# advancedMounts:
# server-3:
# main:
# - path: /root/.ollama
# readOnly: false
web-data:
storageClass: ceph-block
accessMode: ReadWriteOnce