From cd954156d883e1a9ad7eae6aafbf1a1d72684dd8 Mon Sep 17 00:00:00 2001 From: Bilgin Cagatay Date: Fri, 15 Nov 2024 14:25:52 -0800 Subject: [PATCH 1/3] [EE/BE] T206498038 Adding the legacy flow tests to add op --- backends/xnnpack/test/ops/add.py | 250 +++++++++++++++++-------------- 1 file changed, 139 insertions(+), 111 deletions(-) diff --git a/backends/xnnpack/test/ops/add.py b/backends/xnnpack/test/ops/add.py index 784a9d3bbf4..6c26acff36f 100644 --- a/backends/xnnpack/test/ops/add.py +++ b/backends/xnnpack/test/ops/add.py @@ -43,17 +43,20 @@ def forward(self, x): return out1, out2 def _test_add(self, inputs): - ( - Tester(self.Add(), inputs) - .export() - .check_count({"torch.ops.aten.add.Tensor": 4}) - .to_edge_transform_and_lower() - .check_count({"torch.ops.higher_order.executorch_call_delegate": 1}) - .check_not(["executorch_exir_dialects_edge__ops_aten_add_Tensor"]) - .to_executorch() - .serialize() - .run_method_and_compare_outputs() - ) + for legacy in (True, False): + tester = Tester(self.Add(), inputs) + tester.export() + tester.check_count({"torch.ops.aten.add.Tensor": 4}) + if legacy: + tester.to_edge() + tester.partition() + else: + tester.to_edge_transform_and_lower() + tester.check_count({"torch.ops.higher_order.executorch_call_delegate": 1}) + tester.check_not(["executorch_exir_dialects_edge__ops_aten_add_Tensor"]) + tester.to_executorch() + tester.serialize() + tester.run_method_and_compare_outputs() def test_fp16_add(self): inputs = (torch.randn(1).to(torch.float16), torch.randn(1).to(torch.float16)) @@ -65,95 +68,110 @@ def test_fp32_add(self): def test_fp32_add_constant(self): inputs = (torch.randn(4, 4, 4),) - ( - Tester(self.AddConstant(torch.randn(4, 4, 4)), inputs) - .export() - .check_count({"torch.ops.aten.add.Tensor": 4}) - .to_edge_transform_and_lower() - .check_count({"torch.ops.higher_order.executorch_call_delegate": 1}) - .check_not(["executorch_exir_dialects_edge__ops_aten_add_Tensor"]) - .to_executorch() - .serialize() - .run_method_and_compare_outputs() - ) + for legacy in (True, False): + tester = Tester(self.AddConstant(torch.randn(4, 4, 4)), inputs) + tester.export() + tester.check_count({"torch.ops.aten.add.Tensor": 4}) + if legacy: + tester.to_edge() + tester.partition() + else: + tester.to_edge_transform_and_lower() + tester.check_count({"torch.ops.higher_order.executorch_call_delegate": 1}) + tester.check_not(["executorch_exir_dialects_edge__ops_aten_add_Tensor"]) + tester.to_executorch() + tester.serialize() + tester.run_method_and_compare_outputs() def test_qs8_add_constant(self): inputs = (torch.randn(4, 4, 4),) - ( - Tester(self.AddConstant(torch.randn(4, 4, 4)), inputs) - .quantize() - .export() - .check_count({"torch.ops.aten.add.Tensor": 4}) - .to_edge_transform_and_lower() - .check_count({"torch.ops.higher_order.executorch_call_delegate": 1}) - .check_not(["executorch_exir_dialects_edge__ops_aten_add_Tensor"]) - .to_executorch() - .serialize() - .run_method_and_compare_outputs() - ) + for legacy in (True, False): + tester = Tester(self.AddConstant(torch.randn(4, 4, 4)), inputs) + tester.quantize() + tester.export() + tester.check_count({"torch.ops.aten.add.Tensor": 4}) + if legacy: + tester.to_edge() + tester.partition() + else: + tester.to_edge_transform_and_lower() + tester.check_count({"torch.ops.higher_order.executorch_call_delegate": 1}) + tester.check_not(["executorch_exir_dialects_edge__ops_aten_add_Tensor"]) + tester.to_executorch() + tester.serialize() + tester.run_method_and_compare_outputs() def test_qs8_add(self): inputs = (torch.randn(1, 1, 4, 4), torch.randn(1, 1, 4, 4)) - ( - Tester(self.Add(), inputs) - .quantize() - .export() - .check_count({"torch.ops.aten.add.Tensor": 4}) - .check(["torch.ops.quantized_decomposed"]) - .to_edge_transform_and_lower() - .check_count({"torch.ops.higher_order.executorch_call_delegate": 1}) - .check_not( + for legacy in (True, False): + tester = Tester(self.Add(), inputs) + tester.quantize() + tester.export() + tester.check_count({"torch.ops.aten.add.Tensor": 4}) + tester.check(["torch.ops.quantized_decomposed"]) + if legacy: + tester.to_edge() + tester.partition() + else: + tester.to_edge_transform_and_lower() + tester.check_count({"torch.ops.higher_order.executorch_call_delegate": 1}) + tester.check_not( [ "executorch_exir_dialects_edge__ops_aten_add_Tensor", "torch.ops.quantized_decomposed", ] ) - .to_executorch() - .serialize() - .run_method_and_compare_outputs() - ) + tester.to_executorch() + tester.serialize() + tester.run_method_and_compare_outputs() def test_qs8_add2(self): inputs = (torch.randn(1, 1, 4, 4),) - ( - Tester(self.Add2(), inputs) - .quantize() - .export() - .check_count({"torch.ops.aten.add.Tensor": 1}) - .check(["torch.ops.quantized_decomposed"]) - .to_edge_transform_and_lower() - .check_count({"torch.ops.higher_order.executorch_call_delegate": 1}) - .check_not( + for legacy in (True, False): + tester = Tester(self.Add2(), inputs) + tester.quantize() + tester.export() + tester.check_count({"torch.ops.aten.add.Tensor": 1}) + tester.check(["torch.ops.quantized_decomposed"]) + if legacy: + tester.to_edge() + tester.partition() + else: + tester.to_edge_transform_and_lower() + tester.check_count({"torch.ops.higher_order.executorch_call_delegate": 1}) + tester.check_not( [ "executorch_exir_dialects_edge__ops_aten_add_Tensor", "torch.ops.quantized_decomposed", ] ) - .to_executorch() - .serialize() - .run_method_and_compare_outputs() - ) + tester.to_executorch() + tester.serialize() + tester.run_method_and_compare_outputs() def test_qs8_add3(self): inputs = (torch.randn(1, 1, 4, 4), torch.randn(1, 1, 4, 1)) - ( - Tester(self.Add(), inputs) - .quantize() - .export() - .check_count({"torch.ops.aten.add.Tensor": 4}) - .check(["torch.ops.quantized_decomposed"]) - .to_edge_transform_and_lower() - .check_count({"torch.ops.higher_order.executorch_call_delegate": 1}) - .check_not( + for legacy in (True, False): + tester = Tester(self.Add(), inputs) + tester.quantize() + tester.export() + tester.check_count({"torch.ops.aten.add.Tensor": 4}) + tester.check(["torch.ops.quantized_decomposed"]) + if legacy: + tester.to_edge() + tester.partition() + else: + tester.to_edge_transform_and_lower() + tester.check_count({"torch.ops.higher_order.executorch_call_delegate": 1}) + tester.check_not( [ "executorch_exir_dialects_edge__ops_aten_add_Tensor", "torch.ops.quantized_decomposed", ] ) - .to_executorch() - .serialize() - .run_method_and_compare_outputs() - ) + tester.to_executorch() + tester.serialize() + tester.run_method_and_compare_outputs() class AddRelu(torch.nn.Module): def forward(self, x, y): @@ -162,35 +180,41 @@ def forward(self, x, y): def test_fp32_add_relu(self): inputs = (torch.randn(1, 1, 4, 4), torch.randn(1, 1, 4, 4)) - ( - Tester(self.AddRelu(), inputs) - .export() - .check_count({"torch.ops.aten.add.Tensor": 1}) - .check_count({"torch.ops.aten.relu.default": 1}) - .to_edge_transform_and_lower() - .check_not(["executorch_exir_dialects_edge__ops_aten_add_Tensor"]) - .check_not(["executorch_exir_dialects_edge__ops_aten_relu_default"]) - .check_count({"torch.ops.higher_order.executorch_call_delegate": 1}) - .to_executorch() - .serialize() - .run_method_and_compare_outputs() - ) + for legacy in (True, False): + tester = Tester(self.AddRelu(), inputs) + tester.export() + tester.check_count({"torch.ops.aten.add.Tensor": 1}) + tester.check_count({"torch.ops.aten.relu.default": 1}) + if legacy: + tester.to_edge() + tester.partition() + else: + tester.to_edge_transform_and_lower() + tester.check_not(["executorch_exir_dialects_edge__ops_aten_add_Tensor"]) + tester.check_not(["executorch_exir_dialects_edge__ops_aten_relu_default"]) + tester.check_count({"torch.ops.higher_order.executorch_call_delegate": 1}) + tester.to_executorch() + tester.serialize() + tester.run_method_and_compare_outputs() def test_qs8_add_relu(self): inputs = (torch.randn(1, 1, 4, 4), torch.randn(1, 1, 4, 4)) - ( - Tester(self.AddRelu(), inputs) - .quantize() - .export() - .check_count({"torch.ops.aten.add.Tensor": 1}) - .check_count({"torch.ops.aten.relu.default": 1}) - .check(["torch.ops.quantized_decomposed"]) - .to_edge_transform_and_lower() - .check_count({"torch.ops.higher_order.executorch_call_delegate": 1}) - .to_executorch() - .serialize() - .run_method_and_compare_outputs() - ) + for legacy in (True, False): + tester = Tester(self.AddRelu(), inputs) + tester.quantize() + tester.export() + tester.check_count({"torch.ops.aten.add.Tensor": 1}) + tester.check_count({"torch.ops.aten.relu.default": 1}) + tester.check(["torch.ops.quantized_decomposed"]) + if legacy: + tester.to_edge() + tester.partition() + else: + tester.to_edge_transform_and_lower() + tester.check_count({"torch.ops.higher_order.executorch_call_delegate": 1}) + tester.to_executorch() + tester.serialize() + tester.run_method_and_compare_outputs() def test_qs8_add_relu_seq(self): class AddReLU(torch.nn.Module): @@ -220,17 +244,21 @@ def forward(self, x, z): ), ) - ( - Tester(self.AddRelu(), inputs) - .quantize() - .export() - .check_count( + for legacy in (True, False): + tester = Tester(self.AddRelu(), inputs) + tester.quantize() + tester.export() + tester.check_count( {"torch.ops.aten.add.Tensor": 1, "torch.ops.aten.relu.default": 1} ) - .check(["torch.ops.quantized_decomposed"]) - .to_edge_transform_and_lower() - .check_count({"torch.ops.higher_order.executorch_call_delegate": 1}) - .to_executorch() - .serialize() - .run_method_and_compare_outputs() - ) + tesster.check(["torch.ops.quantized_decomposed"]) + if legacy: + tester.to_edge() + tester.partition() + else: + tester.to_edge_transform_and_lower() + tester.check_count({"torch.ops.higher_order.executorch_call_delegate": 1}) + tester.to_executorch() + tester.serialize() + tester.run_method_and_compare_outputs() + From 27516458b105c931fcf3efbcad9e45776888796f Mon Sep 17 00:00:00 2001 From: Bilgin Cagatay Date: Fri, 15 Nov 2024 14:51:58 -0800 Subject: [PATCH 2/3] fixing typo --- backends/xnnpack/test/ops/add.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/backends/xnnpack/test/ops/add.py b/backends/xnnpack/test/ops/add.py index 6c26acff36f..f25842f97f6 100644 --- a/backends/xnnpack/test/ops/add.py +++ b/backends/xnnpack/test/ops/add.py @@ -251,7 +251,7 @@ def forward(self, x, z): tester.check_count( {"torch.ops.aten.add.Tensor": 1, "torch.ops.aten.relu.default": 1} ) - tesster.check(["torch.ops.quantized_decomposed"]) + tester.check(["torch.ops.quantized_decomposed"]) if legacy: tester.to_edge() tester.partition() From fda0b1f0029691c8e203bddfcd7c1f6071788a19 Mon Sep 17 00:00:00 2001 From: Bilgin Cagatay Date: Sat, 16 Nov 2024 16:09:11 -0800 Subject: [PATCH 3/3] fix linter --- backends/xnnpack/test/ops/add.py | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-) diff --git a/backends/xnnpack/test/ops/add.py b/backends/xnnpack/test/ops/add.py index f25842f97f6..c67d21d2ccf 100644 --- a/backends/xnnpack/test/ops/add.py +++ b/backends/xnnpack/test/ops/add.py @@ -211,7 +211,9 @@ def test_qs8_add_relu(self): tester.partition() else: tester.to_edge_transform_and_lower() - tester.check_count({"torch.ops.higher_order.executorch_call_delegate": 1}) + tester.check_count( + {"torch.ops.higher_order.executorch_call_delegate": 1} + ) tester.to_executorch() tester.serialize() tester.run_method_and_compare_outputs() @@ -261,4 +263,3 @@ def forward(self, x, z): tester.to_executorch() tester.serialize() tester.run_method_and_compare_outputs() -