We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
1 parent c0bc7ba commit 9cefecfCopy full SHA for 9cefecf
docs/source/inference.mdx
@@ -114,7 +114,7 @@ For INT4 quantization you can also specify the following arguments :
114
115
Smaller `group_size` and `ratio` of usually improve accuracy at the sacrifice of the model size and inference latency.
116
117
-You can also apply apply 8-bit quantization on your model's weight when loading your model by setting the `load_in_8bit=True` argument when calling the `from_pretrained()` method.
+You can also apply 8-bit quantization on your model's weight when loading your model by setting the `load_in_8bit=True` argument when calling the `from_pretrained()` method.
118
119
```python
120
from optimum.intel import OVModelForCausalLM
0 commit comments