-
Notifications
You must be signed in to change notification settings - Fork 97
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
HIP out of memory #208
Comments
remove --gpu-only from ur command |
This indeed works but will slow down generation significantly and is not the point of my question. I also tried loading CLIP encoder into RAM to give the flux model all space available in the VRAM, without success I still get OOM errors. |
how did u compare the speed ! , what ever if u insist to use --gpu-only u must have card can hold the size of all models u load !
|
The Q_8 (needs 16gb) or at least the Q_6 (needs 12gb) model should fit into 16gb VRAM, so why do I get OOM Errors, then? |
Trying to run the
flux1-dev-Q8_0.gguf
on my 6900XT with 16GB of ram.This should work in theory, right? But somehow I run out of memory.
Do I have a false assumption here or is something wrong with my workflow?
oom_workflow.json
System Information
Devices
The text was updated successfully, but these errors were encountered: