Replies: 1 comment
-
@xdd130 Does your model have any weight of shape 10304? Since 10304=64*161, you should set -gs=64 in the command. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I try used the AutoGPTQ tool to quantize the Qwen2.5-3B-Instruct model to 3 bits. I successfully obtained the model in GPTQ format, but when I compiled the script using T-MAC:
I got the following error:
Does this indicate that there is a problem with my quantization step or that t-mac does not support the direct use of 3-bit models? What additional operations do I need to do?
Beta Was this translation helpful? Give feedback.
All reactions