|
4 | 4 | # LICENSE file in the root directory of this source tree.
|
5 | 5 |
|
6 | 6 |
|
7 |
| -import unittest |
| 7 | +from typing import Tuple |
8 | 8 |
|
9 | 9 | import torch
|
10 | 10 | from diffusers.models.transformers import SD3Transformer2DModel
|
|
13 | 13 | from executorch.backends.arm.test.models.stable_diffusion.stable_diffusion_module_test_configs import (
|
14 | 14 | SD3Transformer2DModel_init_dict,
|
15 | 15 | )
|
16 |
| -from executorch.backends.arm.test.tester.arm_tester import ArmTester |
| 16 | +from executorch.backends.arm.test.tester.test_pipeline import ( |
| 17 | + TosaPipelineFP, |
| 18 | + TosaPipelineINT, |
| 19 | + VgfPipeline, |
| 20 | +) |
| 21 | + |
| 22 | +input_t4 = Tuple[torch.Tensor, torch.Tensor, torch.Tensor, torch.Tensor] |
17 | 23 |
|
18 | 24 |
|
19 |
| -class TestSD3Transformer2DModel(unittest.TestCase): |
| 25 | +class TestSD3Transformer2DModel: |
20 | 26 | """
|
21 | 27 | Test class of AutoenSD3Transformer2DModelcoderKL.
|
22 | 28 | SD3Transformer2DModel is the transformer model used by Stable Diffusion 3.5 Medium
|
@@ -93,48 +99,88 @@ def forward(self, *args, **kwargs):
|
93 | 99 |
|
94 | 100 | return sd35_transformer2D_model, sd35_transformer2D_model_inputs
|
95 | 101 |
|
96 |
| - def test_SD3Transformer2DModel_tosa_FP(self): |
97 |
| - sd35_transformer2D_model, sd35_transformer2D_model_inputs = ( |
98 |
| - self.prepare_model_and_inputs() |
99 |
| - ) |
100 |
| - with torch.no_grad(): |
101 |
| - ( |
102 |
| - ArmTester( |
103 |
| - sd35_transformer2D_model, |
104 |
| - example_inputs=sd35_transformer2D_model_inputs, |
105 |
| - compile_spec=common.get_tosa_compile_spec(tosa_spec="TOSA-1.0+FP"), |
106 |
| - ) |
107 |
| - .export() |
108 |
| - .to_edge_transform_and_lower() |
109 |
| - .check_count(self.ops_after_partitioner_FP) |
110 |
| - .to_executorch() |
111 |
| - .run_method_and_compare_outputs( |
112 |
| - inputs=sd35_transformer2D_model_inputs, |
113 |
| - rtol=1.0, # TODO: MLETORCH-875: Reduce tolerance of SD3Transformer2DModel with FP and INT |
114 |
| - atol=4.0, |
115 |
| - ) |
116 |
| - ) |
117 | 102 |
|
118 |
| - def test_SD3Transformer2DModel_tosa_INT(self): |
119 |
| - sd35_transformer2D_model, sd35_transformer2D_model_inputs = ( |
120 |
| - self.prepare_model_and_inputs() |
| 103 | +def test_SD3Transformer2DModel_tosa_FP(): |
| 104 | + sd35_transformer2D_model, sd35_transformer2D_model_inputs = ( |
| 105 | + TestSD3Transformer2DModel().prepare_model_and_inputs() |
| 106 | + ) |
| 107 | + with torch.no_grad(): |
| 108 | + pipeline = TosaPipelineFP[input_t4]( |
| 109 | + sd35_transformer2D_model, |
| 110 | + sd35_transformer2D_model_inputs, |
| 111 | + aten_op=[], |
| 112 | + exir_op=[], |
| 113 | + use_to_edge_transform_and_lower=True, |
| 114 | + rtol=1.0, # TODO: MLETORCH-875: Reduce tolerance of SD3Transformer2DModel with FP and INT |
| 115 | + atol=4.0, |
121 | 116 | )
|
122 |
| - with torch.no_grad(): |
123 |
| - ( |
124 |
| - ArmTester( |
125 |
| - sd35_transformer2D_model, |
126 |
| - example_inputs=sd35_transformer2D_model_inputs, |
127 |
| - compile_spec=common.get_tosa_compile_spec(tosa_spec="TOSA-1.0+INT"), |
128 |
| - ) |
129 |
| - .quantize() |
130 |
| - .export() |
131 |
| - .to_edge_transform_and_lower() |
132 |
| - .check_count(self.ops_after_partitioner_INT) |
133 |
| - .to_executorch() |
134 |
| - .run_method_and_compare_outputs( |
135 |
| - inputs=sd35_transformer2D_model_inputs, |
136 |
| - qtol=1.0, # TODO: MLETORCH-875: Reduce tolerance of SD3Transformer2DModel with FP and INT |
137 |
| - rtol=1.0, |
138 |
| - atol=4.0, |
139 |
| - ) |
140 |
| - ) |
| 117 | + pipeline.change_args( |
| 118 | + "check_count.exir", TestSD3Transformer2DModel.ops_after_partitioner_FP |
| 119 | + ) |
| 120 | + pipeline.run() |
| 121 | + |
| 122 | + |
| 123 | +def test_SD3Transformer2DModel_tosa_INT(): |
| 124 | + sd35_transformer2D_model, sd35_transformer2D_model_inputs = ( |
| 125 | + TestSD3Transformer2DModel().prepare_model_and_inputs() |
| 126 | + ) |
| 127 | + with torch.no_grad(): |
| 128 | + pipeline = TosaPipelineINT[input_t4]( |
| 129 | + sd35_transformer2D_model, |
| 130 | + sd35_transformer2D_model_inputs, |
| 131 | + aten_op=[], |
| 132 | + exir_op=[], |
| 133 | + use_to_edge_transform_and_lower=True, |
| 134 | + qtol=1.0, # TODO: MLETORCH-875: Reduce tolerance of SD3Transformer2DModel with FP and INT |
| 135 | + rtol=1.0, |
| 136 | + atol=4.0, |
| 137 | + ) |
| 138 | + pipeline.change_args( |
| 139 | + "check_count.exir", TestSD3Transformer2DModel.ops_after_partitioner_INT |
| 140 | + ) |
| 141 | + pipeline.run() |
| 142 | + |
| 143 | + |
| 144 | +@common.SkipIfNoModelConverter |
| 145 | +def test_SD3Transformer2DModel_vgf_FP(): |
| 146 | + sd35_transformer2D_model, sd35_transformer2D_model_inputs = ( |
| 147 | + TestSD3Transformer2DModel().prepare_model_and_inputs() |
| 148 | + ) |
| 149 | + with torch.no_grad(): |
| 150 | + pipeline = VgfPipeline[input_t4]( |
| 151 | + sd35_transformer2D_model, |
| 152 | + sd35_transformer2D_model_inputs, |
| 153 | + aten_op=[], |
| 154 | + exir_op=[], |
| 155 | + tosa_version="TOSA-1.0+FP", |
| 156 | + use_to_edge_transform_and_lower=True, |
| 157 | + rtol=1.0, # TODO: MLETORCH-875: Reduce tolerance of SD3Transformer2DModel with FP and INT |
| 158 | + atol=4.0, |
| 159 | + ) |
| 160 | + pipeline.change_args( |
| 161 | + "check_count.exir", TestSD3Transformer2DModel.ops_after_partitioner_FP |
| 162 | + ) |
| 163 | + pipeline.run() |
| 164 | + |
| 165 | + |
| 166 | +@common.SkipIfNoModelConverter |
| 167 | +def test_SD3Transformer2DModel_vgf_INT(): |
| 168 | + sd35_transformer2D_model, sd35_transformer2D_model_inputs = ( |
| 169 | + TestSD3Transformer2DModel().prepare_model_and_inputs() |
| 170 | + ) |
| 171 | + with torch.no_grad(): |
| 172 | + pipeline = VgfPipeline[input_t4]( |
| 173 | + sd35_transformer2D_model, |
| 174 | + sd35_transformer2D_model_inputs, |
| 175 | + aten_op=[], |
| 176 | + exir_op=[], |
| 177 | + tosa_version="TOSA-1.0+INT", |
| 178 | + use_to_edge_transform_and_lower=True, |
| 179 | + qtol=1.0, # TODO: MLETORCH-875: Reduce tolerance of SD3Transformer2DModel with FP and INT |
| 180 | + rtol=1.0, |
| 181 | + atol=4.0, |
| 182 | + ) |
| 183 | + pipeline.change_args( |
| 184 | + "check_count.exir", TestSD3Transformer2DModel.ops_after_partitioner_INT |
| 185 | + ) |
| 186 | + pipeline.run() |
0 commit comments