Skip to content

Commit e54dcd2

Browse files
committedMar 13, 2024
fix typo
1 parent afc23d0 commit e54dcd2

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed
 

‎docs/source/inference.mdx

+1-1
Original file line numberDiff line numberDiff line change
@@ -108,7 +108,7 @@ optimum-cli export openvino --model gpt2 --weight-format int8 ov_model
108108
This type of optimization allows to reduce the memory footprint and inference latency.
109109

110110

111-
By default the quantization scheme will be [assymmetric](https://github.com/openvinotoolkit/nncf/blob/develop/docs/compression_algorithms/Quantization.md#asymmetric-quantization), to make it [symmetric](https://github.com/openvinotoolkit/nncf/blob/develop/docs/compression_algorithms/Quantization.md#symmetric-quantization) you can add `--sym`.
111+
By default the quantization scheme will be [asymmetric](https://github.com/openvinotoolkit/nncf/blob/develop/docs/compression_algorithms/Quantization.md#asymmetric-quantization), to make it [symmetric](https://github.com/openvinotoolkit/nncf/blob/develop/docs/compression_algorithms/Quantization.md#symmetric-quantization) you can add `--sym`.
112112

113113
For INT4 quantization you can also specify the following arguments :
114114
* The `--group-size` parameter will define the group size to use for quantization, `-1` it will results in per-column quantization.

0 commit comments

Comments
 (0)