Skip to content

Commit 1c43ca9

Browse files
authored
add image repo for lws and all backend runtime, like llama.cpp (#328)
1 parent fd77f80 commit 1c43ca9

File tree

7 files changed

+34
-11
lines changed

7 files changed

+34
-11
lines changed

Diff for: chart/templates/backends/llamacpp.yaml

+2-2
Original file line numberDiff line numberDiff line change
@@ -10,8 +10,8 @@ metadata:
1010
spec:
1111
command:
1212
- ./llama-server
13-
image: ghcr.io/ggerganov/llama.cpp
14-
version: server
13+
image: {{ .Values.backendRuntime.llamacpp.image.repository }}
14+
version: {{ .Values.backendRuntime.llamacpp.image.tag }}
1515
# Do not edit the preset argument name unless you know what you're doing.
1616
# Free to add more arguments with your requirements.
1717
recommendedConfigs:

Diff for: chart/templates/backends/ollama.yaml

+2-2
Original file line numberDiff line numberDiff line change
@@ -11,8 +11,8 @@ spec:
1111
command:
1212
- sh
1313
- -c
14-
image: ollama/ollama
15-
version: latest
14+
image: {{ .Values.backendRuntime.ollama.image.repository }}
15+
version: {{ .Values.backendRuntime.ollama.image.tag }}
1616
envs:
1717
- name: OLLAMA_HOST
1818
value: 0.0.0.0:8080

Diff for: chart/templates/backends/sglang.yaml

+2-2
Original file line numberDiff line numberDiff line change
@@ -12,8 +12,8 @@ spec:
1212
- python3
1313
- -m
1414
- sglang.launch_server
15-
image: lmsysorg/sglang
16-
version: v0.2.10-cu121
15+
image: {{ .Values.backendRuntime.sglang.image.repository }}
16+
version: {{ .Values.backendRuntime.sglang.image.tag }}
1717
# Do not edit the preset argument name unless you know what you're doing.
1818
# Free to add more arguments with your requirements.
1919
recommendedConfigs:

Diff for: chart/templates/backends/tgi.yaml

+2-2
Original file line numberDiff line numberDiff line change
@@ -8,8 +8,8 @@ metadata:
88
app.kubernetes.io/created-by: llmaz
99
name: tgi
1010
spec:
11-
image: ghcr.io/huggingface/text-generation-inference
12-
version: 2.3.1
11+
image: {{ .Values.backendRuntime.tgi.image.repository }}
12+
version: {{ .Values.backendRuntime.tgi.image.tag }}
1313
# Do not edit the preset argument name unless you know what you're doing.
1414
# Free to add more arguments with your requirements.
1515
recommendedConfigs:

Diff for: chart/templates/backends/vllm.yaml

+2-2
Original file line numberDiff line numberDiff line change
@@ -12,8 +12,8 @@ spec:
1212
- python3
1313
- -m
1414
- vllm.entrypoints.openai.api_server
15-
image: vllm/vllm-openai
16-
version: v0.7.3
15+
image: {{ .Values.backendRuntime.vllm.image.repository }}
16+
version: {{ .Values.backendRuntime.vllm.image.tag }}
1717
lifecycle:
1818
preStop:
1919
exec:

Diff for: chart/templates/lws/leaderworkerset.yaml

+1-1
Original file line numberDiff line numberDiff line change
@@ -16681,7 +16681,7 @@ spec:
1668116681
- --zap-log-level=2
1668216682
command:
1668316683
- /manager
16684-
image: registry.k8s.io/lws/lws:v0.5.0
16684+
image: {{ .Values.backendRuntime.image.repository }}:{{ .Values.backendRuntime.image.tag }}
1668516685
livenessProbe:
1668616686
httpGet:
1668716687
path: /healthz

Diff for: chart/values.global.yaml

+23
Original file line numberDiff line numberDiff line change
@@ -2,6 +2,29 @@ fullnameOverride: "llmaz"
22

33
backendRuntime:
44
install: true
5+
llamacpp:
6+
image:
7+
repository: ghcr.io/ggerganov/llama.cpp
8+
tag: server
9+
ollama:
10+
image:
11+
repository: ollama/ollama
12+
tag: latest
13+
sglang:
14+
image:
15+
repository: lmsysorg/sglang
16+
tag: v0.2.10-cu121
17+
tgi:
18+
image:
19+
repository: ghcr.io/huggingface/text-generation-inference
20+
tag: 2.3.1
21+
vllm:
22+
image:
23+
repository: vllm/vllm-openai
24+
tag: v0.7.3
525

626
leaderWorkerSet:
727
install: true
28+
image:
29+
repository: registry.k8s.io/lws/lws
30+
tag: v0.5.0

0 commit comments

Comments
 (0)