Skip to content

Commit 78b3d8c

Browse files
authored
Bump transformers version (huggingface#622)
* Bump transformers version * Add generation config * update setup * fix * add _convert_tokens_to_ids
1 parent 3879e4e commit 78b3d8c

File tree

3 files changed

+12
-7
lines changed

3 files changed

+12
-7
lines changed

optimum/intel/neural_compressor/modeling_base.py

+3
Original file line numberDiff line numberDiff line change
@@ -32,6 +32,7 @@
3232
AutoModelForSequenceClassification,
3333
AutoModelForTokenClassification,
3434
AutoModelForVision2Seq,
35+
GenerationConfig,
3536
GenerationMixin,
3637
PretrainedConfig,
3738
)
@@ -83,6 +84,8 @@ def __init__(
8384
self._device = getattr(self.model, "device", None) or torch.device(
8485
"cuda:0" if torch.cuda.is_available() else "cpu"
8586
)
87+
self.generation_config = GenerationConfig.from_model_config(config)
88+
8689
# Registers the INCModelForXXX classes into the transformers AutoModel classes to avoid warnings when creating
8790
# a pipeline https://github.com/huggingface/transformers/blob/cad61b68396a1a387287a8e2e2fef78a25b79383/src/transformers/pipelines/base.py#L863
8891
AutoConfig.register(self.base_model_prefix, AutoConfig)

setup.py

+2-2
Original file line numberDiff line numberDiff line change
@@ -28,7 +28,7 @@
2828

2929
INSTALL_REQUIRE = [
3030
"torch>=1.11",
31-
"transformers>=4.36.0,<4.39.0",
31+
"transformers>=4.36.0,<4.40.0",
3232
"optimum @ git+https://github.com/huggingface/optimum.git#egg=optimum",
3333
"datasets>=1.4.0",
3434
"sentencepiece",
@@ -61,7 +61,7 @@
6161
"openvino": ["openvino>=2023.3", "nncf>=2.8.1"],
6262
"openvino-tokenizers": ["openvino-tokenizers[transformers]"],
6363
"nncf": ["nncf>=2.8.1"],
64-
"ipex": ["intel-extension-for-pytorch"],
64+
"ipex": ["intel-extension-for-pytorch", "transformers>=4.36.0,<4.39.0"],
6565
"diffusers": ["diffusers"],
6666
"quality": QUALITY_REQUIRE,
6767
"tests": TESTS_REQUIRE,

tests/openvino/test_modeling.py

+7-5
Original file line numberDiff line numberDiff line change
@@ -501,7 +501,7 @@ class OVModelForCausalLMIntegrationTest(unittest.TestCase):
501501
"qwen",
502502
"qwen2",
503503
"stablelm",
504-
# "starcoder2", # TODO: enable with next transformers release
504+
"starcoder2",
505505
"phi",
506506
)
507507
GENERATION_LENGTH = 100
@@ -525,10 +525,8 @@ def test_compare_to_transformers(self, model_arch):
525525

526526
model_kwargs = {}
527527
if model_arch in self.REMOTE_CODE_MODELS:
528-
model_kwargs = {
529-
"config": AutoConfig.from_pretrained(model_id, trust_remote_code=True),
530-
"trust_remote_code": True,
531-
}
528+
model_kwargs = {"trust_remote_code": True}
529+
532530
ov_model = OVModelForCausalLM.from_pretrained(model_id, export=True, ov_config=F32_CONFIG, **model_kwargs)
533531
self.assertIsInstance(ov_model.config, PretrainedConfig)
534532
self.assertTrue(ov_model.use_cache)
@@ -572,6 +570,10 @@ def test_pipeline(self, model_arch):
572570
"trust_remote_code": True,
573571
}
574572
tokenizer = AutoTokenizer.from_pretrained(model_id, trust_remote_code=model_arch in self.REMOTE_CODE_MODELS)
573+
574+
if model_arch == "qwen":
575+
tokenizer._convert_tokens_to_ids = lambda x: 0
576+
575577
model = OVModelForCausalLM.from_pretrained(
576578
model_id, export=True, use_cache=False, compile=False, **model_kwargs
577579
)

0 commit comments

Comments
 (0)