|
8 | 8 |
|
9 | 9 | from typing import Tuple
|
10 | 10 |
|
11 |
| -import pytest |
12 |
| - |
13 | 11 | import torch
|
14 | 12 | from executorch.backends.arm.quantizer.arm_quantizer import (
|
15 | 13 | get_symmetric_a16w8_quantization_config,
|
@@ -313,12 +311,8 @@ def test_linear_16a8w_tosa_INT(test_data: torch.Tensor):
|
313 | 311 | pipeline.run()
|
314 | 312 |
|
315 | 313 |
|
316 |
| -@common.parametrize("test_data", test_data_rank1_INT | test_data_rank4_INT) |
| 314 | +@common.parametrize("test_data", test_data_all_16a8w) |
317 | 315 | @common.XfailIfNoCorstone300
|
318 |
| -@pytest.mark.xfail( |
319 |
| - reason="Ethos-U55 A16W8 linear: int16 matmul not yet supported; pending backend support or linear->conv1x1 lowering. See: https://github.com/pytorch/executorch/issues/13947", |
320 |
| - strict=False, |
321 |
| -) |
322 | 316 | def test_linear_16a8w_u55_INT16(test_data: torch.Tensor):
|
323 | 317 | """Test linear operation with 16A8W quantization on U55 (16-bit activations, 8-bit weights)"""
|
324 | 318 | test_data, out_features, has_bias, per_channel_quantization = test_data()
|
@@ -347,12 +341,8 @@ def test_linear_16a8w_u55_INT16(test_data: torch.Tensor):
|
347 | 341 | pipeline.run()
|
348 | 342 |
|
349 | 343 |
|
350 |
| -@common.parametrize("test_data", test_data_rank1_INT | test_data_rank4_INT) |
| 344 | +@common.parametrize("test_data", test_data_all_16a8w) |
351 | 345 | @common.XfailIfNoCorstone320
|
352 |
| -@pytest.mark.xfail( |
353 |
| - reason="Ethos-U55 A16W8 linear: int16 matmul not yet supported; pending backend support or linear->conv1x1 lowering. See: https://github.com/pytorch/executorch/issues/13947", |
354 |
| - strict=False, |
355 |
| -) |
356 | 346 | def test_linear_16a8w_u85_INT16(test_data: torch.Tensor):
|
357 | 347 | """Test linear operation with 16A8W quantization on U85 (16-bit activations, 8-bit weights)"""
|
358 | 348 | test_data, out_features, has_bias, per_channel_quantization = test_data()
|
|
0 commit comments