lmstudio und webui im pod
This commit is contained in:
@@ -1,6 +1,7 @@
|
||||
apiVersion: v1
|
||||
kind: Pod
|
||||
metadata:
|
||||
hostname: ollama-webui
|
||||
name: ollama-open-webui
|
||||
labels:
|
||||
app: ollama-open-webui
|
||||
@@ -14,6 +15,7 @@ spec:
|
||||
# Ollama container
|
||||
- name: ollama
|
||||
image: docker.io/ollama/ollama:0.12.3-rocm
|
||||
hostname: ollama
|
||||
args:
|
||||
- serve
|
||||
env:
|
||||
@@ -36,6 +38,7 @@ spec:
|
||||
# Open WebUI container
|
||||
- name: open-webui
|
||||
image: ghcr.io/open-webui/open-webui:main
|
||||
hostname: webui
|
||||
ports:
|
||||
- containerPort: 8080
|
||||
volumeMounts:
|
||||
|
||||
Reference in New Issue
Block a user