Skip to content

Commit d89f59e

Browse files
Align optimum quantization calls with the latest API
1 parent 8221cb3 commit d89f59e

File tree

3 files changed

+10
-7
lines changed

3 files changed

+10
-7
lines changed

notebooks/code-language-id/code-language-id.ipynb

+4-3
Original file line numberDiff line numberDiff line change
@@ -165,7 +165,7 @@
165165
"import evaluate\n",
166166
"from transformers import pipeline, AutoTokenizer, AutoModelForSequenceClassification\n",
167167
"from optimum.intel import OVModelForSequenceClassification\n",
168-
"from optimum.intel.openvino import OVConfig, OVQuantizer\n",
168+
"from optimum.intel.openvino import OVConfig, OVQuantizer, OVWeightQuantizationConfig\n",
169169
"from huggingface_hub.utils import RepositoryNotFoundError"
170170
]
171171
},
@@ -559,7 +559,8 @@
559559
"base_model = AutoModelForSequenceClassification.from_pretrained(MODEL_ID)\n",
560560
"\n",
561561
"quantizer = OVQuantizer.from_pretrained(base_model)\n",
562-
"quantization_config = OVConfig()"
562+
"quantization_config = OVWeightQuantizationConfig()\n",
563+
"ov_config = OVConfig(quantization_config=quantization_config)"
563564
]
564565
},
565566
{
@@ -793,7 +794,7 @@
793794
],
794795
"source": [
795796
"quantizer.quantize(\n",
796-
" quantization_config=quantization_config,\n",
797+
" ov_config=ov_config,\n",
797798
" calibration_dataset=calibration_sample,\n",
798799
" save_directory=QUANTIZED_MODEL_LOCAL_PATH,\n",
799800
")"

notebooks/dolly-2-instruction-following/dolly-2-instruction-following.ipynb

+3-2
Original file line numberDiff line numberDiff line change
@@ -288,7 +288,7 @@
288288
],
289289
"source": [
290290
"import gc\n",
291-
"from optimum.intel import OVQuantizer\n",
291+
"from optimum.intel import OVQuantizer, OVConfig, OVWeightQuantizationConfig\n",
292292
"\n",
293293
"compressed_model_path = Path(f\"{model_path}_compressed\")\n",
294294
"\n",
@@ -304,7 +304,8 @@
304304
"if to_compress.value:\n",
305305
" if not compressed_model_path.exists():\n",
306306
" quantizer = OVQuantizer.from_pretrained(ov_model)\n",
307-
" quantizer.quantize(save_directory=compressed_model_path, weights_only=True)\n",
307+
" ov_config = OVConfig(quantization_config=OVWeightQuantizationConfig(bits=8))\n",
308+
" quantizer.quantize(save_directory=compressed_model_path, ov_config=ov_config)\n",
308309
" del quantizer\n",
309310
" gc.collect()\n",
310311
"\n",

notebooks/named-entity-recognition/named-entity-recognition.ipynb

+3-2
Original file line numberDiff line numberDiff line change
@@ -315,7 +315,7 @@
315315
],
316316
"source": [
317317
"from functools import partial\n",
318-
"from optimum.intel import OVQuantizer\n",
318+
"from optimum.intel import OVQuantizer, OVConfig, OVQuantizationConfig\n",
319319
"\n",
320320
"from optimum.intel import OVModelForTokenClassification\n",
321321
"\n",
@@ -341,10 +341,11 @@
341341
"quantized_ner_model_dir = \"quantized_ner_model\"\n",
342342
"\n",
343343
"# Apply static quantization and save the resulting model in the OpenVINO IR format\n",
344+
"ov_config = OVConfig(quantization_config=OVQuantizationConfig(num_samples=len(calibration_dataset)))\n",
344345
"quantizer.quantize(\n",
345346
" calibration_dataset=calibration_dataset,\n",
346347
" save_directory=quantized_ner_model_dir,\n",
347-
" subset_size=len(calibration_dataset),\n",
348+
" ov_config=ov_config,\n",
348349
")"
349350
]
350351
},

0 commit comments

Comments
 (0)