16
16
"tests/llmcompressor/transformers/obcq/obcq_configs/consec_runs/gpu"
17
17
)
18
18
19
- quantization_config = CompressedTensorsConfig (run_compressed = False )
20
-
21
19
22
20
class TestConsecutiveRuns (unittest .TestCase ):
21
+ quantization_config = CompressedTensorsConfig (run_compressed = False )
22
+
23
23
def _test_consecutive_runs (
24
24
self , tolerance : float , num_calibration_samples : int = 16
25
25
):
@@ -45,7 +45,7 @@ def _test_consecutive_runs(
45
45
first_model = AutoModelForCausalLM .from_pretrained (
46
46
self .output_first ,
47
47
device_map = "auto" ,
48
- quantization_config = quantization_config ,
48
+ quantization_config = self . quantization_config ,
49
49
)
50
50
51
51
layer_0_sparse = tensor_sparsity (
@@ -78,7 +78,7 @@ def _test_consecutive_runs(
78
78
second_model = AutoModelForCausalLM .from_pretrained (
79
79
self .output_second ,
80
80
device_map = "auto" ,
81
- quantization_config = quantization_config ,
81
+ quantization_config = self . quantization_config ,
82
82
)
83
83
84
84
layer_0_sparse = tensor_sparsity (
@@ -141,7 +141,7 @@ def setUp(self):
141
141
self .model = AutoModelForCausalLM .from_pretrained (
142
142
self .model ,
143
143
device_map = self .device ,
144
- quantization_config = quantization_config ,
144
+ quantization_config = self . quantization_config ,
145
145
)
146
146
147
147
self .output = "./oneshot_output"
0 commit comments