@@ -2110,9 +2110,9 @@ class OVModelForVisualCausalLMIntegrationTest(unittest.TestCase):
2110
2110
if is_transformers_version (">=" , "4.40.0" ):
2111
2111
SUPPORTED_ARCHITECTURES += ["llava_next" , "nanollava" ]
2112
2112
if is_transformers_version (">=" , "4.45.0" ):
2113
- SUPPORTED_ARCHITECTURES += ["minicpmv" , "internvl2" , "phi3_v" , "qwen2_vl" ]
2113
+ SUPPORTED_ARCHITECTURES += ["minicpmv" , "internvl2" , "phi3_v" , "qwen2_vl" , "maira2" ]
2114
2114
TASK = "image-text-to-text"
2115
- REMOTE_CODE_MODELS = ["internvl2" , "minicpmv" , "nanollava" , "phi3_v" ]
2115
+ REMOTE_CODE_MODELS = ["internvl2" , "minicpmv" , "nanollava" , "phi3_v" , "maira2" ]
2116
2116
2117
2117
IMAGE = Image .open (
2118
2118
requests .get (
@@ -2192,7 +2192,7 @@ def test_compare_to_transformers(self, model_arch):
2192
2192
with torch .no_grad ():
2193
2193
transformers_outputs = transformers_model (** transformers_inputs )
2194
2194
self .assertTrue (
2195
- torch .allclose (ov_outputs .logits , transformers_outputs .logits , atol = 1e-4 ),
2195
+ torch .allclose (ov_outputs .logits , transformers_outputs .logits , atol = 4e-3 ),
2196
2196
f"Max abs diff { (torch .abs (ov_outputs .logits - transformers_outputs .logits ).max ())} " ,
2197
2197
)
2198
2198
0 commit comments