Skip to content

Commit 7b64b33

Browse files
authored
Added compressed quantization status (#341)
Signed-off-by: Rishin Raj <[email protected]>
1 parent fc89e8b commit 7b64b33

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

QEfficient/transformers/quantizers/quantizer_compressed_tensors.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -232,13 +232,13 @@ def __init__(
232232
"Currently only single quantization group is supported, please raise an issue with model details for support!"
233233
)
234234

235-
if quantization_status != "frozen":
235+
if quantization_status not in {"compressed", "frozen"}:
236236
raise NotImplementedError(f"expected quantization_status=`frozen`, got {quantization_status}")
237237

238238
if kv_cache_scheme:
239239
raise NotImplementedError(f"Expected kv_cache_scheme=None, got {kv_cache_scheme}")
240240

241-
if format not in ["naive-quantized", "float-quantized"]:
241+
if format not in {"naive-quantized", "float-quantized"}:
242242
raise NotImplementedError(
243243
f"Expected quantization format in ['naive_quantized', 'float-quantized'] got {format}"
244244
)

0 commit comments

Comments
 (0)