File tree 7 files changed +34
-11
lines changed
7 files changed +34
-11
lines changed Original file line number Diff line number Diff line change @@ -10,8 +10,8 @@ metadata:
10
10
spec :
11
11
command :
12
12
- ./llama-server
13
- image : ghcr.io/ggerganov/llama.cpp
14
- version : server
13
+ image : {{ .Values.backendRuntime.llamacpp.image.repository }}
14
+ version : {{ .Values.backendRuntime.llamacpp.image.tag }}
15
15
# Do not edit the preset argument name unless you know what you're doing.
16
16
# Free to add more arguments with your requirements.
17
17
recommendedConfigs :
Original file line number Diff line number Diff line change 11
11
command :
12
12
- sh
13
13
- -c
14
- image : ollama/ollama
15
- version : latest
14
+ image : {{ .Values.backendRuntime. ollama.image.repository }}
15
+ version : {{ .Values.backendRuntime.ollama.image.tag }}
16
16
envs :
17
17
- name : OLLAMA_HOST
18
18
value : 0.0.0.0:8080
Original file line number Diff line number Diff line change 12
12
- python3
13
13
- -m
14
14
- sglang.launch_server
15
- image : lmsysorg/ sglang
16
- version : v0.2.10-cu121
15
+ image : {{ .Values.backendRuntime. sglang.image.repository }}
16
+ version : {{ .Values.backendRuntime.sglang.image.tag }}
17
17
# Do not edit the preset argument name unless you know what you're doing.
18
18
# Free to add more arguments with your requirements.
19
19
recommendedConfigs :
Original file line number Diff line number Diff line change @@ -8,8 +8,8 @@ metadata:
8
8
app.kubernetes.io/created-by : llmaz
9
9
name : tgi
10
10
spec :
11
- image : ghcr.io/huggingface/text-generation-inference
12
- version : 2.3.1
11
+ image : {{ .Values.backendRuntime.tgi.image.repository }}
12
+ version : {{ .Values.backendRuntime.tgi.image.tag }}
13
13
# Do not edit the preset argument name unless you know what you're doing.
14
14
# Free to add more arguments with your requirements.
15
15
recommendedConfigs :
Original file line number Diff line number Diff line change 12
12
- python3
13
13
- -m
14
14
- vllm.entrypoints.openai.api_server
15
- image : vllm/vllm-openai
16
- version : v0.7.3
15
+ image : {{ .Values.backendRuntime. vllm.image.repository }}
16
+ version : {{ .Values.backendRuntime.vllm.image.tag }}
17
17
lifecycle :
18
18
preStop :
19
19
exec :
Original file line number Diff line number Diff line change @@ -16681,7 +16681,7 @@ spec:
16681
16681
- --zap-log-level=2
16682
16682
command:
16683
16683
- /manager
16684
- image: registry.k8s.io/lws/lws:v0.5.0
16684
+ image: {{ .Values.backendRuntime.image.repository }}:{{ .Values.backendRuntime.image.tag }}
16685
16685
livenessProbe:
16686
16686
httpGet:
16687
16687
path: /healthz
Original file line number Diff line number Diff line change @@ -2,6 +2,29 @@ fullnameOverride: "llmaz"
2
2
3
3
backendRuntime :
4
4
install : true
5
+ llamacpp :
6
+ image :
7
+ repository : ghcr.io/ggerganov/llama.cpp
8
+ tag : server
9
+ ollama :
10
+ image :
11
+ repository : ollama/ollama
12
+ tag : latest
13
+ sglang :
14
+ image :
15
+ repository : lmsysorg/sglang
16
+ tag : v0.2.10-cu121
17
+ tgi :
18
+ image :
19
+ repository : ghcr.io/huggingface/text-generation-inference
20
+ tag : 2.3.1
21
+ vllm :
22
+ image :
23
+ repository : vllm/vllm-openai
24
+ tag : v0.7.3
5
25
6
26
leaderWorkerSet :
7
27
install : true
28
+ image :
29
+ repository : registry.k8s.io/lws/lws
30
+ tag : v0.5.0
You can’t perform that action at this time.
0 commit comments