Skip to content

Commit 92a03d5

Browse files
committed
use speicfic version for minicpmv test
1 parent 8a37b61 commit 92a03d5

File tree

1 file changed

+16
-11
lines changed

1 file changed

+16
-11
lines changed

tests/openvino/test_modeling.py

+16-11
Original file line numberDiff line numberDiff line change
@@ -44,13 +44,13 @@
4444
AutoModelForCTC,
4545
AutoModelForImageClassification,
4646
AutoModelForMaskedLM,
47-
AutoProcessor,
4847
AutoModelForQuestionAnswering,
4948
AutoModelForSeq2SeqLM,
5049
AutoModelForSequenceClassification,
5150
AutoModelForSpeechSeq2Seq,
5251
AutoModelForTokenClassification,
5352
AutoModelForVision2Seq,
53+
AutoProcessor,
5454
AutoTokenizer,
5555
GenerationConfig,
5656
Pix2StructForConditionalGeneration,
@@ -1868,15 +1868,14 @@ def test_compare_with_and_without_past_key_values(self):
18681868

18691869

18701870
class OVModelForVisualCausalLMIntegrationTest(unittest.TestCase):
1871-
SUPPORTED_ARCHITECTURES = [
1872-
"llava",
1873-
"minicpmv"
1874-
]
1871+
SUPPORTED_ARCHITECTURES = ["llava"]
18751872

18761873
REMOTE_CODE_MODELS = ["minicpmv"]
18771874

18781875
if is_transformers_version(">=", "4.40.0"):
18791876
SUPPORTED_ARCHITECTURES += ["llava_next"]
1877+
if is_transformers_version(">=", "4.45.0"):
1878+
SUPPORTED_ARCHITECTURES += ["minicpmv"]
18801879
TASK = "image-text-to-text"
18811880

18821881
IMAGE = Image.open(
@@ -1904,20 +1903,24 @@ def test_compare_to_transformers(self, model_arch):
19041903
if "llava" in model_arch:
19051904
prompt = "<image>\n What is shown in this image?"
19061905
elif "minicpmv" in model_arch:
1907-
prompt = "<|im_start|>user\n(<image>./</image>)\n What is shown in this image?<|im_end|>\n<|im_start|>assistant\n"
1906+
prompt = "<|im_start|>user\n(<image>./</image>)\n What is shown in this image?<|im_end|>\n<|im_start|>assistant\n"
19081907
model_id = MODEL_NAMES[model_arch]
19091908
processor = AutoProcessor.from_pretrained(model_id, trust_remote_code=model_arch in self.REMOTE_CODE_MODELS)
1910-
transformers_model = self.get_transformer_model_class(model_arch).from_pretrained(model_id, trust_remote_code=model_arch in self.REMOTE_CODE_MODELS)
1909+
transformers_model = self.get_transformer_model_class(model_arch).from_pretrained(
1910+
model_id, trust_remote_code=model_arch in self.REMOTE_CODE_MODELS
1911+
)
19111912
inputs = processor(images=[self.IMAGE.resize((600, 600))], text=[prompt], return_tensors="pt")
1912-
ov_model = OVModelForVisualCausalLM.from_pretrained(model_id, export=True, trust_remote_code=model_arch in self.REMOTE_CODE_MODELS)
1913+
ov_model = OVModelForVisualCausalLM.from_pretrained(
1914+
model_id, export=True, trust_remote_code=model_arch in self.REMOTE_CODE_MODELS
1915+
)
19131916
self.assertIsInstance(ov_model, MODEL_TYPE_TO_CLS_MAPPING[ov_model.config.model_type])
19141917
self.assertIsInstance(ov_model.vision_embeddings, OVVisionEmbedding)
19151918
self.assertIsInstance(ov_model.language_model, OVModelWithEmbedForCausalLM)
19161919
for additional_part in ov_model.additional_parts:
19171920
self.assertTrue(hasattr(ov_model, additional_part))
19181921
self.assertIsInstance(getattr(ov_model, additional_part), MODEL_PARTS_CLS_MAPPING[additional_part])
19191922
self.assertIsInstance(ov_model.config, PretrainedConfig)
1920-
if not "minicpmv" in model_arch:
1923+
if "minicpmv" not in model_arch:
19211924
set_seed(SEED)
19221925
with torch.no_grad():
19231926
transformers_outputs = transformers_model(**inputs)
@@ -1939,7 +1942,7 @@ def test_compare_to_transformers(self, model_arch):
19391942
set_seed(SEED)
19401943
transformers_outputs = transformers_model.generate(**inputs, generation_config=gen_config)
19411944
if model_arch == "minicpmv":
1942-
ov_outputs = ov_outputs[:, inputs["input_ids"].shape[1]:]
1945+
ov_outputs = ov_outputs[:, inputs["input_ids"].shape[1] :]
19431946
self.assertTrue(
19441947
torch.equal(ov_outputs, transformers_outputs),
19451948
f"generation config : {gen_config}, transformers output {transformers_outputs}, ov_model output {ov_outputs}",
@@ -1953,7 +1956,9 @@ def test_compare_to_transformers(self, model_arch):
19531956
@parameterized.expand(SUPPORTED_ARCHITECTURES)
19541957
def test_generate_utils(self, model_arch):
19551958
model_id = MODEL_NAMES[model_arch]
1956-
model = OVModelForVisualCausalLM.from_pretrained(model_id, export=True, trust_remote_code=model_arch in self.REMOTE_CODE_MODELS)
1959+
model = OVModelForVisualCausalLM.from_pretrained(
1960+
model_id, export=True, trust_remote_code=model_arch in self.REMOTE_CODE_MODELS
1961+
)
19571962
preprocessor = AutoProcessor.from_pretrained(model_id, trust_remote_code=model_arch in self.REMOTE_CODE_MODELS)
19581963
if "llava" in model_arch:
19591964
question = "<image>\nDescribe image"

0 commit comments

Comments
 (0)