Replies: 2 comments
-
It looks like you might be running the wrong executable: warning: not compiled with GPU offload support, --gpu-layers option will be ignored When building with CMake like you showed above I believe the executable files will be placed in $ build/bin/llama-server -np 4 -c 8192 -n 512 -ngl 100 -m ../text-generation-webui/models/gemma-2-27b-it-SimPO-37K-Q4_K_M.gguf It could be that you previously built a |
Beta Was this translation helpful? Give feedback.
0 replies
-
@TeaCult, do clean build in new folder with inital enable -DGGML_GPU=ON |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I have built as instructed, my cuda libs are working in text-generaton-webui and torch but its not working with llama-cli or llama-server :
Could you help me out please
I am on archlinux Linux LLAMA 6.4.6-arch1-1 #1 SMP PREEMPT_DYNAMIC Mon, 24 Jul 2023 20:19:38 +0000 x86_64 GNU/Linux
| NVIDIA-SMI 535.86.05 Driver Version: 535.86.05 CUDA Version: 12.2
Beta Was this translation helpful? Give feedback.
All reactions