We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent a76be08 commit 8e4a81eCopy full SHA for 8e4a81e
optimum/intel/openvino/quantization.py
@@ -44,7 +44,6 @@
44
from transformers.pytorch_utils import Conv1D
45
from transformers.utils import is_accelerate_available
46
47
-from optimum.exporters.onnx.convert import check_dummy_inputs_are_allowed
48
from optimum.exporters.tasks import TasksManager
49
from optimum.quantization_base import OptimumQuantizer
50
@@ -524,6 +523,8 @@ def _quantize_torchmodel(
524
523
525
quantization_config = ov_config.quantization_config
526
if isinstance(quantization_config, OVWeightQuantizationConfig):
+ from optimum.exporters.onnx.convert import check_dummy_inputs_are_allowed
527
+
528
if stateful:
529
# patch model before weight compression
530
model = patch_model_with_bettertransformer(model)
0 commit comments