Skip to content

Commit 9cefecf

Browse files
Update docs/source/inference.mdx
Co-authored-by: Alexander Kozlov <alexander.kozlov@intel.com>
1 parent c0bc7ba commit 9cefecf

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

docs/source/inference.mdx

+1-1
Original file line numberDiff line numberDiff line change
@@ -114,7 +114,7 @@ For INT4 quantization you can also specify the following arguments :
114114

115115
Smaller `group_size` and `ratio` of usually improve accuracy at the sacrifice of the model size and inference latency.
116116

117-
You can also apply apply 8-bit quantization on your model's weight when loading your model by setting the `load_in_8bit=True` argument when calling the `from_pretrained()` method.
117+
You can also apply 8-bit quantization on your model's weight when loading your model by setting the `load_in_8bit=True` argument when calling the `from_pretrained()` method.
118118

119119
```python
120120
from optimum.intel import OVModelForCausalLM

0 commit comments

Comments
 (0)