|
4 | 4 | # LICENSE file in the root directory of this source tree. |
5 | 5 |
|
6 | 6 |
|
7 | | -import unittest |
| 7 | +from typing import Tuple |
8 | 8 |
|
9 | 9 | import torch |
10 | 10 | from diffusers.models.transformers import SD3Transformer2DModel |
|
13 | 13 | from executorch.backends.arm.test.models.stable_diffusion.stable_diffusion_module_test_configs import ( |
14 | 14 | SD3Transformer2DModel_init_dict, |
15 | 15 | ) |
16 | | -from executorch.backends.arm.test.tester.arm_tester import ArmTester |
| 16 | +from executorch.backends.arm.test.tester.test_pipeline import ( |
| 17 | + TosaPipelineFP, |
| 18 | + TosaPipelineINT, |
| 19 | + VgfPipeline, |
| 20 | +) |
| 21 | + |
| 22 | +input_t4 = Tuple[torch.Tensor, torch.Tensor, torch.Tensor, torch.Tensor] |
17 | 23 |
|
18 | 24 |
|
19 | | -class TestSD3Transformer2DModel(unittest.TestCase): |
| 25 | +class TestSD3Transformer2DModel: |
20 | 26 | """ |
21 | 27 | Test class of AutoenSD3Transformer2DModelcoderKL. |
22 | 28 | SD3Transformer2DModel is the transformer model used by Stable Diffusion 3.5 Medium |
@@ -93,48 +99,88 @@ def forward(self, *args, **kwargs): |
93 | 99 |
|
94 | 100 | return sd35_transformer2D_model, sd35_transformer2D_model_inputs |
95 | 101 |
|
96 | | - def test_SD3Transformer2DModel_tosa_FP(self): |
97 | | - sd35_transformer2D_model, sd35_transformer2D_model_inputs = ( |
98 | | - self.prepare_model_and_inputs() |
99 | | - ) |
100 | | - with torch.no_grad(): |
101 | | - ( |
102 | | - ArmTester( |
103 | | - sd35_transformer2D_model, |
104 | | - example_inputs=sd35_transformer2D_model_inputs, |
105 | | - compile_spec=common.get_tosa_compile_spec(tosa_spec="TOSA-1.0+FP"), |
106 | | - ) |
107 | | - .export() |
108 | | - .to_edge_transform_and_lower() |
109 | | - .check_count(self.ops_after_partitioner_FP) |
110 | | - .to_executorch() |
111 | | - .run_method_and_compare_outputs( |
112 | | - inputs=sd35_transformer2D_model_inputs, |
113 | | - rtol=1.0, # TODO: MLETORCH-875: Reduce tolerance of SD3Transformer2DModel with FP and INT |
114 | | - atol=4.0, |
115 | | - ) |
116 | | - ) |
117 | 102 |
|
118 | | - def test_SD3Transformer2DModel_tosa_INT(self): |
119 | | - sd35_transformer2D_model, sd35_transformer2D_model_inputs = ( |
120 | | - self.prepare_model_and_inputs() |
| 103 | +def test_SD3Transformer2DModel_tosa_FP(): |
| 104 | + sd35_transformer2D_model, sd35_transformer2D_model_inputs = ( |
| 105 | + TestSD3Transformer2DModel().prepare_model_and_inputs() |
| 106 | + ) |
| 107 | + with torch.no_grad(): |
| 108 | + pipeline = TosaPipelineFP[input_t4]( |
| 109 | + sd35_transformer2D_model, |
| 110 | + sd35_transformer2D_model_inputs, |
| 111 | + aten_op=[], |
| 112 | + exir_op=[], |
| 113 | + use_to_edge_transform_and_lower=True, |
| 114 | + rtol=1.0, # TODO: MLETORCH-875: Reduce tolerance of SD3Transformer2DModel with FP and INT |
| 115 | + atol=4.0, |
121 | 116 | ) |
122 | | - with torch.no_grad(): |
123 | | - ( |
124 | | - ArmTester( |
125 | | - sd35_transformer2D_model, |
126 | | - example_inputs=sd35_transformer2D_model_inputs, |
127 | | - compile_spec=common.get_tosa_compile_spec(tosa_spec="TOSA-1.0+INT"), |
128 | | - ) |
129 | | - .quantize() |
130 | | - .export() |
131 | | - .to_edge_transform_and_lower() |
132 | | - .check_count(self.ops_after_partitioner_INT) |
133 | | - .to_executorch() |
134 | | - .run_method_and_compare_outputs( |
135 | | - inputs=sd35_transformer2D_model_inputs, |
136 | | - qtol=1.0, # TODO: MLETORCH-875: Reduce tolerance of SD3Transformer2DModel with FP and INT |
137 | | - rtol=1.0, |
138 | | - atol=4.0, |
139 | | - ) |
140 | | - ) |
| 117 | + pipeline.change_args( |
| 118 | + "check_count.exir", TestSD3Transformer2DModel.ops_after_partitioner_FP |
| 119 | + ) |
| 120 | + pipeline.run() |
| 121 | + |
| 122 | + |
| 123 | +def test_SD3Transformer2DModel_tosa_INT(): |
| 124 | + sd35_transformer2D_model, sd35_transformer2D_model_inputs = ( |
| 125 | + TestSD3Transformer2DModel().prepare_model_and_inputs() |
| 126 | + ) |
| 127 | + with torch.no_grad(): |
| 128 | + pipeline = TosaPipelineINT[input_t4]( |
| 129 | + sd35_transformer2D_model, |
| 130 | + sd35_transformer2D_model_inputs, |
| 131 | + aten_op=[], |
| 132 | + exir_op=[], |
| 133 | + use_to_edge_transform_and_lower=True, |
| 134 | + qtol=1.0, # TODO: MLETORCH-875: Reduce tolerance of SD3Transformer2DModel with FP and INT |
| 135 | + rtol=1.0, |
| 136 | + atol=4.0, |
| 137 | + ) |
| 138 | + pipeline.change_args( |
| 139 | + "check_count.exir", TestSD3Transformer2DModel.ops_after_partitioner_INT |
| 140 | + ) |
| 141 | + pipeline.run() |
| 142 | + |
| 143 | + |
| 144 | +@common.SkipIfNoModelConverter |
| 145 | +def test_SD3Transformer2DModel_vgf_FP(): |
| 146 | + sd35_transformer2D_model, sd35_transformer2D_model_inputs = ( |
| 147 | + TestSD3Transformer2DModel().prepare_model_and_inputs() |
| 148 | + ) |
| 149 | + with torch.no_grad(): |
| 150 | + pipeline = VgfPipeline[input_t4]( |
| 151 | + sd35_transformer2D_model, |
| 152 | + sd35_transformer2D_model_inputs, |
| 153 | + aten_op=[], |
| 154 | + exir_op=[], |
| 155 | + tosa_version="TOSA-1.0+FP", |
| 156 | + use_to_edge_transform_and_lower=True, |
| 157 | + rtol=1.0, # TODO: MLETORCH-875: Reduce tolerance of SD3Transformer2DModel with FP and INT |
| 158 | + atol=4.0, |
| 159 | + ) |
| 160 | + pipeline.change_args( |
| 161 | + "check_count.exir", TestSD3Transformer2DModel.ops_after_partitioner_FP |
| 162 | + ) |
| 163 | + pipeline.run() |
| 164 | + |
| 165 | + |
| 166 | +@common.SkipIfNoModelConverter |
| 167 | +def test_SD3Transformer2DModel_vgf_INT(): |
| 168 | + sd35_transformer2D_model, sd35_transformer2D_model_inputs = ( |
| 169 | + TestSD3Transformer2DModel().prepare_model_and_inputs() |
| 170 | + ) |
| 171 | + with torch.no_grad(): |
| 172 | + pipeline = VgfPipeline[input_t4]( |
| 173 | + sd35_transformer2D_model, |
| 174 | + sd35_transformer2D_model_inputs, |
| 175 | + aten_op=[], |
| 176 | + exir_op=[], |
| 177 | + tosa_version="TOSA-1.0+INT", |
| 178 | + use_to_edge_transform_and_lower=True, |
| 179 | + qtol=1.0, # TODO: MLETORCH-875: Reduce tolerance of SD3Transformer2DModel with FP and INT |
| 180 | + rtol=1.0, |
| 181 | + atol=4.0, |
| 182 | + ) |
| 183 | + pipeline.change_args( |
| 184 | + "check_count.exir", TestSD3Transformer2DModel.ops_after_partitioner_INT |
| 185 | + ) |
| 186 | + pipeline.run() |
0 commit comments