44
44
OVModelOpenCLIPForZeroShotImageClassification ,
45
45
OVModelOpenCLIPText ,
46
46
OVModelOpenCLIPVisual ,
47
+ OVSanaPipeline ,
47
48
OVSentenceTransformer ,
48
49
OVStableDiffusion3Pipeline ,
49
50
OVStableDiffusionPipeline ,
@@ -107,6 +108,7 @@ class OVCLIExportTestCase(unittest.TestCase):
107
108
"flux" : 4 if is_tokenizers_version ("<" , "0.20" ) or is_openvino_version (">=" , "2024.5" ) else 0 ,
108
109
"flux-fill" : 2 if is_tokenizers_version ("<" , "0.20" ) or is_openvino_version (">=" , "2024.5" ) else 0 ,
109
110
"llava" : 2 if is_tokenizers_version ("<" , "0.20" ) or is_openvino_version (">=" , "2024.5" ) else 0 ,
111
+ "sana" : 2 if is_tokenizers_version ("<" , "0.20.0" ) or is_openvino_version (">=" , "2024.5" ) else 0 ,
110
112
}
111
113
112
114
SUPPORTED_SD_HYBRID_ARCHITECTURES = [
@@ -118,7 +120,7 @@ class OVCLIExportTestCase(unittest.TestCase):
118
120
if is_transformers_version (">=" , "4.45" ):
119
121
SUPPORTED_SD_HYBRID_ARCHITECTURES .append (("stable-diffusion-3" , 9 , 65 ))
120
122
SUPPORTED_SD_HYBRID_ARCHITECTURES .append (("flux" , 7 , 56 ))
121
- SUPPORTED_SD_HYBRID_ARCHITECTURES .append (("sana" , 7 , 56 ))
123
+ SUPPORTED_SD_HYBRID_ARCHITECTURES .append (("sana" , 19 , 53 ))
122
124
123
125
SUPPORTED_QUANTIZATION_ARCHITECTURES = [
124
126
(
@@ -348,9 +350,15 @@ def test_exporters_cli_int8(self, task: str, model_type: str):
348
350
models = [model .encoder , model .decoder ]
349
351
if task .endswith ("with-past" ):
350
352
models .append (model .decoder_with_past )
351
- elif model_type .startswith ("stable-diffusion" ) or model_type .startswith ("flux" ):
353
+ elif (
354
+ model_type .startswith ("stable-diffusion" )
355
+ or model_type .startswith ("flux" )
356
+ or model_type .startswith ("sana" )
357
+ ):
352
358
models = [model .unet or model .transformer , model .vae_encoder , model .vae_decoder ]
353
- models .append (model .text_encoder if model_type == "stable-diffusion" else model .text_encoder_2 )
359
+ models .append (
360
+ model .text_encoder if model_type in ["stable-diffusion" , "sana" ] else model .text_encoder_2
361
+ )
354
362
elif task .startswith ("image-text-to-text" ):
355
363
models = [model .language_model , model .vision_embeddings ]
356
364
else :
0 commit comments