Replies: 2 comments 1 reply
-
|
Great idea! Hoping to contribute where I can. |
Beta Was this translation helpful? Give feedback.
1 reply
-
|
The GPT4All desktop app provides a local OpenAI-compatible server with LocalDocs support, and there is also a Python binding for our backend although it hasn't been updated in a while. We are unfortunately stuck on an old version of llama.cpp because we rely on many patches for the Kompute backend, which we have not had time to rebase on the new backend API. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
Proposing a living doc about all the frameworks that work with (or should work with)
llama.cpp, at any level.The list is long so let's keep it roughly sorted by decreasing community contributions or stars or something ✌️ (direct edits from contributors / suggestions of edits in comments highly welcome, I've probably made a gazillion mistakes and omissions already!)
Part of the goal is to identify which projects would benefit from a documentation update or small patches for direct support. For instance, a few Python projects only document the (amazing)
llama-cpp-pythonbindings and could use instructions on how to also use llama-server (our canonical OpenAI-compatible server)Projects with some integration (non exhaustive list!)
llama.cppcontributors: @ggerganov, @slaren, @JohannesGaessler, @ngxson & too many to count / large overlap 🤗🤗🤗🤗🤗🤗🤗libllama?llama.cppcontributors: @cebtenzzre (PRs) 🤗llama-server(see Hugging Face Inference Endpoints now supports GGUF out of the box! #9669, revshare goes to ggml.ai)llama.cppcontributors: @ngxson (PRs) 🤗llama.cpp/ modified serverllama.cppcontributors: @jart (PRs) 🤗libllama+ includes llama.cpp's JSON schema conversionllama.cppcontributors: @abetlen (PRs) 🤗libllama?llama.cppcontributors: @mudler (PRs) 🤗llama-servercan only partially use Ollama models (custom incompatible chat template format)ggmllibllama(?)llama-server(OpenAI+base_urloverride: example)llama-server(OpenAI like integration w/api_base): needs docsLlamaCPP)llama-server(using OpenAI +base_urioverride): needs docsllama-server(OpenAI-compatible endpoint doc): needs docsllama-server: not yet / needs help ❌llama-server: not yet / needs help ❌base_urloverride?llama-server: not yet / needs help ❌llama-cpp-python(labelledllama.cpp)llama-server(OpenAI Chat Model Node + overrideBase URL/ API key): needs docsllama-server: not yet / needs help ❌base_urlparamllama-serverorOllama(use OpenAI + overridebase_url&api_key): needs docsllama-server: not yet / needs help ❌base_urioverridellama-server(use OpenAI + overridebase_uri): needs docsllama-server: not yet / needs help ❌base_urloverridellama-server✅llama-cpp-pythonllama-server: not yet / needs help ❌llama-serverpowered)llama-cpp(using LLMEndpointConfig +llm_base_url)Projects w/o integration
Beta Was this translation helpful? Give feedback.
All reactions