Skip to content

Commit 11857b6

Browse files
committed
Update documentation
1 parent f4ec215 commit 11857b6

File tree

1 file changed

+2
-4
lines changed

1 file changed

+2
-4
lines changed

docs/source/optimization_ov.mdx

+2-4
Original file line numberDiff line numberDiff line change
@@ -74,17 +74,15 @@ model = OVModelForCausalLM.from_pretrained(model_id, load_in_8bit=True)
7474

7575
> **NOTE:** `load_in_8bit` is enabled by default for models larger than 1 billion parameters.
7676

77-
For the 4-bit weight quantization you can use `load_in_4bit` option. The `quantization_config` can be used to controll the optimization parameters, for example:
77+
For the 4-bit weight quantization you can use yhe `quantization_config` to specify the optimization parameters, for example:
7878

7979
```python
8080
from optimum.intel import OVModelForCausalLM, OVWeightQuantizationConfig
81-
import nncf
8281

8382
model = OVModelForCausalLM.from_pretrained(
8483
model_id,
8584
export=True,
86-
load_in_4bit=True,
87-
quantization_config=OVWeightQuantizationConfig(mode=nncf.CompressWeightsMode.INT4_ASYM, ratio=0.8, dataset="ptb"),
85+
quantization_config=OVWeightQuantizationConfig(bits=4, sym=False, ratio=0.8, dataset="ptb"),
8886
)
8987
```
9088

0 commit comments

Comments
 (0)