Skip to content

Commit 4707914

Browse files
committed
Fixed issue with Transformers
1 parent 277d39a commit 4707914

File tree

2 files changed

+3
-3
lines changed

2 files changed

+3
-3
lines changed

optimum/intel/openvino/configuration.py

+2-2
Original file line numberDiff line numberDiff line change
@@ -84,7 +84,7 @@ def __init__(
8484
compression: Union[List[Dict], Dict, None] = None,
8585
input_info: Optional[List] = None,
8686
save_onnx_model: bool = False,
87-
quantization_config: Optional[QuantizationConfigMixin] = None,
87+
weight_quantization_config: Optional[QuantizationConfigMixin] = None,
8888
**kwargs,
8989
):
9090
super().__init__()
@@ -93,7 +93,7 @@ def __init__(
9393
self.save_onnx_model = save_onnx_model
9494
self._enable_standard_onnx_export_option()
9595
self.optimum_version = kwargs.pop("optimum_version", None)
96-
self.quantization_config = quantization_config
96+
self.weight_quantization_config = weight_quantization_config
9797

9898
def add_input_info(self, model_inputs: Dict, force_batch_one: bool = False):
9999
self.input_info = [

optimum/intel/openvino/quantization.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -318,7 +318,7 @@ def _quantize_ovcausallm(
318318
save_directory.mkdir(parents=True, exist_ok=True)
319319

320320
if weights_only:
321-
quantization_config = None if ov_config is None else ov_config.quantization_config
321+
quantization_config = None if ov_config is None else ov_config.weight_quantization_config
322322
if quantization_config is None:
323323
# Use default 8-bit compression
324324
self.model.model = nncf.compress_weights(self.model.model)

0 commit comments

Comments
 (0)