Skip to content

Commit 9689dca

Browse files
[TorchFX] Constant folding is removed from nncf.quantize and OpenVINOQuantizer
1 parent d400543 commit 9689dca

File tree

12 files changed

+36849
-28757
lines changed

12 files changed

+36849
-28757
lines changed

nncf/experimental/torch/fx/quantization/quantize_model.py

-1
Original file line numberDiff line numberDiff line change
@@ -86,7 +86,6 @@ def quantize_impl(
8686
advanced_parameters=advanced_parameters,
8787
)
8888

89-
# To make it easier for bias correction algorithms.
9089
apply_quantization_transformations(copied_model)
9190

9291
nncf_graph = NNCFGraphFactory.create(copied_model)

nncf/experimental/torch/fx/quantization/quantizer/openvino_quantizer.py

-2
Original file line numberDiff line numberDiff line change
@@ -38,7 +38,6 @@
3838
from nncf.common.utils.api_marker import api
3939
from nncf.experimental.torch.fx.nncf_graph_builder import GraphConverter
4040
from nncf.experimental.torch.fx.node_utils import get_graph_node_by_name
41-
from nncf.experimental.torch.fx.transformations import fold_constant_except_qdq
4241
from nncf.quantization.advanced_parameters import FP8QuantizationParameters
4342
from nncf.quantization.advanced_parameters import QuantizationParameters
4443
from nncf.quantization.algorithms.min_max.algorithm import MinMaxQuantization
@@ -362,5 +361,4 @@ def transform_for_annotation(self, model: torch.fx.GraphModule) -> torch.fx.Grap
362361
:param model: Given torch.fx.GraphModule to transform before the annotation.
363362
:return: The transformed torch.fx.GraphModule ready for the annotation.
364363
"""
365-
fold_constant_except_qdq(model)
366364
return model

nncf/experimental/torch/fx/transformations.py

-1
Original file line numberDiff line numberDiff line change
@@ -733,7 +733,6 @@ def apply_quantization_transformations(model: torch.fx.GraphModule) -> None:
733733
# to make it easier for algorithms to work
734734
# with the target graph BatchNorm operations
735735
# are being fused
736-
fold_constant_except_qdq(model)
737736
_fuse_conv_bn_(model)
738737

739738

nncf/version.py

+1-1
Original file line numberDiff line numberDiff line change
@@ -9,7 +9,7 @@
99
# See the License for the specific language governing permissions and
1010
# limitations under the License.
1111

12-
__version__ = "2.16.0"
12+
__version__ = "2.16.0.dev0+d400543fbdirty"
1313

1414

1515
BKC_TORCH_SPEC = "==2.6.*"

tests/torch/data/reference_graphs/fx/experimental/OpenVINOQuantizer/swin_v2_s.dot

+7,308-5,262
Large diffs are not rendered by default.

tests/torch/data/reference_graphs/fx/experimental/OpenVINOQuantizer/vit_b_16.dot

+2,011-2,009
Large diffs are not rendered by default.

tests/torch/data/reference_graphs/fx/experimental/ao_export_quantization_OpenVINOQuantizer/swin_v2_s.dot

+6,800-4,850
Large diffs are not rendered by default.

tests/torch/data/reference_graphs/fx/experimental/ao_export_quantization_OpenVINOQuantizer/vit_b_16.dot

+1,789-1,787
Large diffs are not rendered by default.

tests/torch/data/reference_graphs/fx/post_quantization_compressed/swin_v2_s.dot

+7,308-5,262
Large diffs are not rendered by default.

tests/torch/data/reference_graphs/fx/post_quantization_compressed/vit_b_16.dot

+2,011-2,009
Large diffs are not rendered by default.

tests/torch/data/reference_graphs/fx/quantized/swin_v2_s.dot

+7,510-5,464
Large diffs are not rendered by default.

tests/torch/data/reference_graphs/fx/quantized/vit_b_16.dot

+2,111-2,109
Large diffs are not rendered by default.

0 commit comments

Comments
 (0)