Skip to content

Commit 0f45751

Browse files
authored
Update README (#549)
1 parent 5120f75 commit 0f45751

File tree

1 file changed

+2
-2
lines changed

1 file changed

+2
-2
lines changed

README.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -78,10 +78,10 @@ It is possible to export your model to the [OpenVINO](https://docs.openvino.ai/2
7878
optimum-cli export openvino --model gpt2 ov_model
7979
```
8080

81-
If you add `--int8`, the model linear and embedding weights will be quantized to INT8, the activations will be kept in floating point precision.
81+
You can also apply 8-bit weight-only quantization when exporting your model : the model linear and embedding weights will be quantized to INT8, the activations will be kept in floating point precision.
8282

8383
```plain
84-
optimum-cli export openvino --model gpt2 --int8 ov_model
84+
optimum-cli export openvino --model gpt2 --weight-format int8 ov_model
8585
```
8686

8787
To apply quantization on both weights and activations, you can find more information in the [documentation](https://huggingface.co/docs/optimum/main/en/intel/optimization_ov).

0 commit comments

Comments
 (0)