Replies: 3 comments 1 reply
-
|
The I don't think it makes much sense to have it as you either use what the model and/or chat template defaults to, or set it in the UI, alternatively pass it through the API. |
Beta Was this translation helpful? Give feedback.
-
|
I understand the point about using the web UI or curl for ad-hoc prompts, but that doesn't fully cover certain specialized use cases. Template defaults is indeed a way to achieve my goal but not generic enough - or a bit complex to modify since models carry their own. For instance, with large-context LLMs, one could configure an easy-to-use for everybody web-llama-server to share local knowledge and turn it into a dedicated chatbot for a local community. In such scenarios, preloading a system prompt—depending on its size—could be highly efficient and practical for this tailored, always-on setup. Would it be worth reconsidering the --system-prompt-file option for these kinds of deployments? |
Beta Was this translation helpful? Give feedback.
-
|
I successfully preloaded a huge system prompt (260kB txt -> ~100kT) using Qwen3 (ctx:256kT). Here's the changes, I have some questions for polishing it:
https://github.com/ggml-org/llama.cpp/compare/master...d-a-v:llama.cpp:serversystemprompt?expand=1 |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
The option
--system-prompt-fileis available withllama-clibut not withllama-server.What would be the way to get such a feature with the current available options in
llama-server?I tried to get and update the chat-template from model with no luck.
Is it worth asking for a feature request ?
Beta Was this translation helpful? Give feedback.
All reactions