We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent b8429a5 commit 1104f44Copy full SHA for 1104f44
test/3x/torch/quantization/weight_only/test_mixed_algos.py
@@ -18,8 +18,8 @@ def run_fn(model):
18
class TestMixedTwoAlgo:
19
def test_mixed_gptq_and_rtn(self):
20
with patch.object(logger, "info") as mock_info:
21
- rtn_config = RTNConfig(quant_lm_head=True)
22
- gptq_config = GPTQConfig(double_quant_bits=4, white_list=["transformer.*"])
+ rtn_config = RTNConfig(white_list=[".*mlp.*"])
+ gptq_config = GPTQConfig(double_quant_bits=4, white_list=[".*attn.*"])
23
combined_config = rtn_config + gptq_config
24
logger.info(combined_config)
25
0 commit comments