@@ -980,25 +980,28 @@ class OVModelForCausalLMIntegrationTest(unittest.TestCase):
980
980
if is_transformers_version (">=" , "4.40.0" ):
981
981
SUPPORTED_ARCHITECTURES += (
982
982
"gemma" ,
983
- "gemma2" ,
984
983
"olmo" ,
985
984
"stablelm" ,
986
985
"starcoder2" ,
987
986
"dbrx" ,
988
- "phi3" ,
989
987
"cohere" ,
990
988
"qwen2" ,
991
989
"qwen2-moe" ,
992
990
"arctic" ,
993
- "exaone" ,
994
- "mistral-nemo" ,
995
- "minicpm3" ,
996
- "glm" ,
997
- "granite" ,
998
- "granite-moe" ,
999
991
)
1000
992
1001
- # openvino 2025.0 required for disabling check_trace
993
+ if is_transformers_version (">=" , "4.41.0" ):
994
+ SUPPORTED_ARCHITECTURES += ("phi3" ,)
995
+
996
+ if is_transformers_version (">=" , "4.43.0" ):
997
+ SUPPORTED_ARCHITECTURES += ("gemma2" , "exaone" )
998
+
999
+ if is_transformers_version (">=" , "4.44.0" ):
1000
+ SUPPORTED_ARCHITECTURES += ("granite" , "granite-moe" )
1001
+
1002
+ if is_transformers_version (">=" , "4.46.0" ):
1003
+ SUPPORTED_ARCHITECTURES += ("glm" , "mistral-nemo" , "minicpm3" )
1004
+ # openvino 2025.0 required for disabling check_trace
1002
1005
if is_openvino_version (">=" , "2025.0" ):
1003
1006
SUPPORTED_ARCHITECTURES += ("deepseek" ,)
1004
1007
@@ -2115,9 +2118,9 @@ class OVModelForVisualCausalLMIntegrationTest(unittest.TestCase):
2115
2118
if is_transformers_version (">=" , "4.40.0" ):
2116
2119
SUPPORTED_ARCHITECTURES += ["llava_next" , "nanollava" ]
2117
2120
if is_transformers_version (">=" , "4.45.0" ):
2118
- SUPPORTED_ARCHITECTURES += ["minicpmv" , "internvl2" , "phi3_v" , "qwen2_vl" ]
2121
+ SUPPORTED_ARCHITECTURES += ["minicpmv" , "internvl2" , "phi3_v" , "qwen2_vl" , "maira2" ]
2119
2122
TASK = "image-text-to-text"
2120
- REMOTE_CODE_MODELS = ["internvl2" , "minicpmv" , "nanollava" , "phi3_v" ]
2123
+ REMOTE_CODE_MODELS = ["internvl2" , "minicpmv" , "nanollava" , "phi3_v" , "maira2" ]
2121
2124
2122
2125
IMAGE = Image .open (
2123
2126
requests .get (
@@ -2197,7 +2200,7 @@ def test_compare_to_transformers(self, model_arch):
2197
2200
with torch .no_grad ():
2198
2201
transformers_outputs = transformers_model (** transformers_inputs )
2199
2202
self .assertTrue (
2200
- torch .allclose (ov_outputs .logits , transformers_outputs .logits , atol = 1e-4 ),
2203
+ torch .allclose (ov_outputs .logits , transformers_outputs .logits , atol = 4e-3 ),
2201
2204
f"Max abs diff { (torch .abs (ov_outputs .logits - transformers_outputs .logits ).max ())} " ,
2202
2205
)
2203
2206
0 commit comments