|
8 | 8 |
|
9 | 9 | from typing import Tuple |
10 | 10 |
|
11 | | -import pytest |
12 | | - |
13 | 11 | import torch |
14 | 12 | from executorch.backends.arm.quantizer.arm_quantizer import ( |
15 | 13 | get_symmetric_a16w8_quantization_config, |
@@ -313,12 +311,8 @@ def test_linear_16a8w_tosa_INT(test_data: torch.Tensor): |
313 | 311 | pipeline.run() |
314 | 312 |
|
315 | 313 |
|
316 | | -@common.parametrize("test_data", test_data_rank1_INT | test_data_rank4_INT) |
| 314 | +@common.parametrize("test_data", test_data_all_16a8w) |
317 | 315 | @common.XfailIfNoCorstone300 |
318 | | -@pytest.mark.xfail( |
319 | | - reason="Ethos-U55 A16W8 linear: int16 matmul not yet supported; pending backend support or linear->conv1x1 lowering. See: https://github.com/pytorch/executorch/issues/13947", |
320 | | - strict=False, |
321 | | -) |
322 | 316 | def test_linear_16a8w_u55_INT16(test_data: torch.Tensor): |
323 | 317 | """Test linear operation with 16A8W quantization on U55 (16-bit activations, 8-bit weights)""" |
324 | 318 | test_data, out_features, has_bias, per_channel_quantization = test_data() |
@@ -347,12 +341,8 @@ def test_linear_16a8w_u55_INT16(test_data: torch.Tensor): |
347 | 341 | pipeline.run() |
348 | 342 |
|
349 | 343 |
|
350 | | -@common.parametrize("test_data", test_data_rank1_INT | test_data_rank4_INT) |
| 344 | +@common.parametrize("test_data", test_data_all_16a8w) |
351 | 345 | @common.XfailIfNoCorstone320 |
352 | | -@pytest.mark.xfail( |
353 | | - reason="Ethos-U55 A16W8 linear: int16 matmul not yet supported; pending backend support or linear->conv1x1 lowering. See: https://github.com/pytorch/executorch/issues/13947", |
354 | | - strict=False, |
355 | | -) |
356 | 346 | def test_linear_16a8w_u85_INT16(test_data: torch.Tensor): |
357 | 347 | """Test linear operation with 16A8W quantization on U85 (16-bit activations, 8-bit weights)""" |
358 | 348 | test_data, out_features, has_bias, per_channel_quantization = test_data() |
|
0 commit comments