Skip to content

Commit 27f3e26

Browse files
authored
Adapt autoround format (#2038)
Signed-off-by: Kaihui-intel <[email protected]>
1 parent 7775768 commit 27f3e26

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

neural_compressor/transformers/models/modeling_auto.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -325,7 +325,7 @@ def load_low_bit(cls, pretrained_model_name_or_path, *model_args, **kwargs):
325325
quantization_config = TeqConfig.from_dict(quantization_config)
326326
elif quantization_config["quant_method"] == "gptq":
327327
quantization_config = GPTQConfig.from_dict(quantization_config)
328-
elif quantization_config["quant_method"] == "autoround":
328+
elif quantization_config["quant_method"] in ["autoround", "intel/auto-round"]:
329329
quantization_config = AutoRoundConfig.from_dict(quantization_config)
330330

331331
assert quantization_config is not None, "Detect this model is not a low-bit model."

0 commit comments

Comments
 (0)