4444 OVModelOpenCLIPForZeroShotImageClassification ,
4545 OVModelOpenCLIPText ,
4646 OVModelOpenCLIPVisual ,
47+ OVSanaPipeline ,
4748 OVSentenceTransformer ,
4849 OVStableDiffusion3Pipeline ,
4950 OVStableDiffusionPipeline ,
@@ -107,6 +108,7 @@ class OVCLIExportTestCase(unittest.TestCase):
107108 "flux" : 4 if is_tokenizers_version ("<" , "0.20" ) or is_openvino_version (">=" , "2024.5" ) else 0 ,
108109 "flux-fill" : 2 if is_tokenizers_version ("<" , "0.20" ) or is_openvino_version (">=" , "2024.5" ) else 0 ,
109110 "llava" : 2 if is_tokenizers_version ("<" , "0.20" ) or is_openvino_version (">=" , "2024.5" ) else 0 ,
111+ "sana" : 2 if is_tokenizers_version ("<" , "0.20.0" ) or is_openvino_version (">=" , "2024.5" ) else 0 ,
110112 }
111113
112114 SUPPORTED_SD_HYBRID_ARCHITECTURES = [
@@ -118,7 +120,7 @@ class OVCLIExportTestCase(unittest.TestCase):
118120 if is_transformers_version (">=" , "4.45" ):
119121 SUPPORTED_SD_HYBRID_ARCHITECTURES .append (("stable-diffusion-3" , 9 , 65 ))
120122 SUPPORTED_SD_HYBRID_ARCHITECTURES .append (("flux" , 7 , 56 ))
121- SUPPORTED_SD_HYBRID_ARCHITECTURES .append (("sana" , 7 , 56 ))
123+ SUPPORTED_SD_HYBRID_ARCHITECTURES .append (("sana" , 19 , 53 ))
122124
123125 SUPPORTED_QUANTIZATION_ARCHITECTURES = [
124126 (
@@ -348,9 +350,15 @@ def test_exporters_cli_int8(self, task: str, model_type: str):
348350 models = [model .encoder , model .decoder ]
349351 if task .endswith ("with-past" ):
350352 models .append (model .decoder_with_past )
351- elif model_type .startswith ("stable-diffusion" ) or model_type .startswith ("flux" ):
353+ elif (
354+ model_type .startswith ("stable-diffusion" )
355+ or model_type .startswith ("flux" )
356+ or model_type .startswith ("sana" )
357+ ):
352358 models = [model .unet or model .transformer , model .vae_encoder , model .vae_decoder ]
353- models .append (model .text_encoder if model_type == "stable-diffusion" else model .text_encoder_2 )
359+ models .append (
360+ model .text_encoder if model_type in ["stable-diffusion" , "sana" ] else model .text_encoder_2
361+ )
354362 elif task .startswith ("image-text-to-text" ):
355363 models = [model .language_model , model .vision_embeddings ]
356364 else :
0 commit comments