Skip to content

Commit f1c9d6f

Browse files
committed
fix style
1 parent 70468a6 commit f1c9d6f

File tree

3 files changed

+8
-7
lines changed

3 files changed

+8
-7
lines changed

optimum/intel/openvino/modeling_decoder.py

-1
Original file line numberDiff line numberDiff line change
@@ -242,7 +242,6 @@ def _from_transformers(
242242
quantization_config: Optional[Union[OVWeightQuantizationConfig, Dict]] = None,
243243
**kwargs,
244244
):
245-
246245
if config.model_type.replace("_", "-") not in _SUPPORTED_ARCHITECTURES:
247246
logger.warning(
248247
f"This architecture : {config.model_type} was not validated, only :{', '.join(_SUPPORTED_ARCHITECTURES)} architectures were "

optimum/intel/openvino/quantization.py

+2-4
Original file line numberDiff line numberDiff line change
@@ -588,9 +588,8 @@ def _int4_weight_only_quantization(
588588
if model.export_feature != "text-generation":
589589
raise ValueError("Only `OVModelForCausalLM` are supported for now")
590590

591-
592591
quantization_config = quantization_config or _check_default_4bit_configs(model.config)
593-
592+
594593
# Data-free weight-only quantization to asymmetric INT4
595594
if quantization_config is None:
596595
quantization_config = OVWeightQuantizationConfig(bits=4, sym=False)
@@ -632,7 +631,7 @@ def _weight_only_quantization(model: OVBaseModel, quantization_config: Union[OVW
632631
mode = CompressWeightsMode.INT8_SYM if config.sym else CompressWeightsMode.INT8_ASYM
633632
else:
634633
mode = CompressWeightsMode.INT4_SYM if config.sym else CompressWeightsMode.INT4_ASYM
635-
634+
636635
model.model = nncf.compress_weights(
637636
ov_model,
638637
mode=mode,
@@ -644,4 +643,3 @@ def _weight_only_quantization(model: OVBaseModel, quantization_config: Union[OVW
644643
ignored_scope=ignored_scope,
645644
dataset=dataset,
646645
)
647-

tests/openvino/test_quantization.py

+6-2
Original file line numberDiff line numberDiff line change
@@ -349,7 +349,9 @@ def test_ovmodel_load_with_compressed_weights(self, model_cls, model_type):
349349
def test_ovmodel_4bit_auto_compression(self, model_cls, model_type, expected_ov_int8, expected_ov_int4):
350350
with tempfile.TemporaryDirectory() as tmp_dir:
351351
model_id = MODEL_NAMES[model_type]
352-
model = model_cls.from_pretrained(model_id, export=True, quantization_config=OVWeightQuantizationConfig(bits=4))
352+
model = model_cls.from_pretrained(
353+
model_id, export=True, quantization_config=OVWeightQuantizationConfig(bits=4)
354+
)
353355
tokenizer = AutoTokenizer.from_pretrained(model_id)
354356
if tokenizer.pad_token is None:
355357
tokenizer.pad_token = tokenizer.eos_token
@@ -404,7 +406,9 @@ def transform_fn(data, tokenizer):
404406
model = model_cls.from_pretrained(
405407
model_id,
406408
export=True,
407-
quantization_config=OVWeightQuantizationConfig(bits=4, sym=True, group_size=-1, ratio=0.8, dataset=quantization_dataset),
409+
quantization_config=OVWeightQuantizationConfig(
410+
bits=4, sym=True, group_size=-1, ratio=0.8, dataset=quantization_dataset
411+
),
408412
)
409413

410414
_, num_int8, num_int4 = get_num_quantized_nodes(model)

0 commit comments

Comments
 (0)