From c6ad4f236379ce2d4aa533e5d283d1e455e84e8d Mon Sep 17 00:00:00 2001 From: City <125218114+city96@users.noreply.github.com> Date: Sun, 22 Dec 2024 19:00:03 +0100 Subject: [PATCH] Account for not having quantized weights #188 --- loader.py | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-) diff --git a/loader.py b/loader.py index 7c8120e..0f983f9 100644 --- a/loader.py +++ b/loader.py @@ -81,8 +81,9 @@ def gguf_sd_loader(path, handle_prefix="model.diffusion_model.", return_arch=Fal # mark largest tensor for vram estimation qsd = {k:v for k,v in state_dict.items() if is_quantized(v)} - max_key = max(qsd.keys(), key=lambda k: qsd[k].numel()) - state_dict[max_key].is_largest_weight = True + if len(qsd) > 0: + max_key = max(qsd.keys(), key=lambda k: qsd[k].numel()) + state_dict[max_key].is_largest_weight = True # sanity check debug print print("\nggml_sd_loader:")