@@ -50,7 +50,6 @@ namespace test
5050{
5151namespace validation
5252{
53- using framework::dataset::make;
5453namespace
5554{
5655
@@ -63,11 +62,6 @@ const auto NeonActivationFunctionsDataset = concat(datasets::ActivationFunctions
6362
6463/* * Input data sets. */
6564const auto ActivationDataset = combine(combine(framework::dataset::make(" InPlace" , { false , true }), NeonActivationFunctionsDataset), framework::dataset::make(" AlphaBeta" , { 0 .5f , 1 .f }));
66- const auto ActivationDatasetForPaddingAfterConfigure = combine(
67- make (" InPlace" , { false , true }),
68- NeonActivationFunctionsDataset,
69- make (" AlphaBeta" , { 0 .5f })
70- );
7165
7266template <typename T, ARM_COMPUTE_REQUIRES_TA(arm_compute::utils::traits::is_floating_point<T>::value)>
7367void test_float_sqrt_boundary_value ()
@@ -187,8 +181,6 @@ DATA_TEST_CASE(Validate, framework::DatasetMode::ALL, zip(zip(zip(
187181
188182template <typename T>
189183using NEActivationLayerFixture = ActivationValidationFixture<Tensor, Accessor, NEActivationLayer, T>;
190- template <typename T>
191- using NEActivationLayerWithPaddingFixture = ActivationWithPaddingValidationFixture<Tensor, Accessor, NEActivationLayer, T>;
192184
193185TEST_SUITE (Float)
194186#ifdef ARM_COMPUTE_ENABLE_FP16
@@ -212,25 +204,6 @@ FIXTURE_DATA_TEST_CASE(RunSmall, NEActivationLayerFixture<half>, framework::Data
212204 framework::ARM_COMPUTE_PRINT_INFO ();
213205 }
214206}
215-
216- FIXTURE_DATA_TEST_CASE (PaddingAfterConfigure, NEActivationLayerWithPaddingFixture<half>, framework::DatasetMode::ALL,
217- combine (
218- make (" Shape" , TensorShape{ 7U , 7U , 17U , 2U }),
219- ActivationDatasetForPaddingAfterConfigure,
220- make(" DataType" , DataType::F16))
221- )
222- {
223- if (CPUInfo::get ().has_fp16 ())
224- {
225- // Validate output
226- validate (Accessor (_target), _reference, helper::relative_tolerance (_data_type, _function), 0 .f , helper::absolute_tolerance (_data_type, _function));
227- }
228- else
229- {
230- ARM_COMPUTE_TEST_INFO (" Device does not support fp16 vector operations. Test SKIPPED." );
231- framework::ARM_COMPUTE_PRINT_INFO ();
232- }
233- }
234207TEST_SUITE_END () // FP16
235208#endif /* ARM_COMPUTE_ENABLE_FP16 */
236209
@@ -239,45 +212,28 @@ TEST_CASE(SqrtBoundaryValue, framework::DatasetMode::ALL)
239212{
240213 test_float_sqrt_boundary_value<float >();
241214}
242- FIXTURE_DATA_TEST_CASE (RunSmall, NEActivationLayerFixture<float >, framework::DatasetMode::ALL, combine(combine(datasets::SmallShapes(), ActivationDataset), make(" DataType" ,
215+ FIXTURE_DATA_TEST_CASE (RunSmall, NEActivationLayerFixture<float >, framework::DatasetMode::ALL, combine(combine(datasets::SmallShapes(), ActivationDataset), framework::dataset:: make(" DataType" ,
243216 DataType::F32)))
244217
245218{
246219 // Validate output
247220 validate (Accessor (_target), _reference, helper::relative_tolerance (_data_type, _function), 0 .f , helper::absolute_tolerance (_data_type, _function));
248221}
249-
250- FIXTURE_DATA_TEST_CASE (PaddingAfterConfigure, NEActivationLayerWithPaddingFixture<float >, framework::DatasetMode::ALL,
251- combine (
252- make (" Shape" , TensorShape{ 7U , 7U , 17U , 2U }),
253- ActivationDatasetForPaddingAfterConfigure,
254- make(" DataType" , DataType::F32))
255- )
256- {
257- validate (Accessor (_target), _reference, helper::relative_tolerance (_data_type, _function), 0 .f , helper::absolute_tolerance (_data_type, _function));
258- }
259222// Run only on SME Devices to stress Logistic SME kernel
260223#ifdef ARM_COMPUTE_ENABLE_SME2
261224TEST_SUITE (SME)
262- const auto LogisticDataset = combine(make(" InPlace" , { false }), make(" Function" , ActivationLayerInfo::ActivationFunction::LOGISTIC), make(" AlphaBeta" , { 1 .f }));
263- FIXTURE_DATA_TEST_CASE (RunLogistic5D, NEActivationLayerFixture<float >, framework::DatasetMode::ALL, combine(datasets::Tiny5dShapes(), LogisticDataset, make(" DataType" ,
225+ const auto LogsisticDataset = combine(framework::dataset:: make(" InPlace" , { false }), framework::dataset:: make(" Function" , ActivationLayerInfo::ActivationFunction::LOGISTIC), framework::dataset:: make(" AlphaBeta" , { 1 .f }));
226+ FIXTURE_DATA_TEST_CASE (RunLogistic5D, NEActivationLayerFixture<float >, framework::DatasetMode::ALL, combine(datasets::Tiny5dShapes(), LogsisticDataset, framework::dataset:: make(" DataType" ,
264227 DataType::F32)))
265228
266229{
267230 // Validate output
268231 validate (Accessor (_target), _reference, helper::relative_tolerance (_data_type, _function), 0 .f , helper::absolute_tolerance (_data_type, _function));
269232}
270233
271- FIXTURE_DATA_TEST_CASE (RunLogisticSME, NEActivationLayerFixture<float >, framework::DatasetMode::ALL, combine(datasets::LogisticSMEStressShapesFp32(), LogisticDataset, make(" DataType" ,
234+ FIXTURE_DATA_TEST_CASE (RunLogisticSME, NEActivationLayerFixture<float >, framework::DatasetMode::ALL, combine(datasets::LogisticSMEStressShapesFp32(), LogsisticDataset, framework::dataset:: make(" DataType" ,
272235 DataType::F32)))
273236
274- {
275- // Validate output
276- validate (Accessor (_target), _reference, helper::relative_tolerance (_data_type, _function), 0 .f , helper::absolute_tolerance (_data_type, _function));
277- }
278- FIXTURE_DATA_TEST_CASE (PaddingAfterConfigure, NEActivationLayerWithPaddingFixture<float >, framework::DatasetMode::ALL,
279- combine (datasets::LogisticSMEStressShapesFp32(), LogisticDataset, make(" DataType" , DataType::F32)))
280-
281237{
282238 // Validate output
283239 validate (Accessor (_target), _reference, helper::relative_tolerance (_data_type, _function), 0 .f , helper::absolute_tolerance (_data_type, _function));
@@ -289,8 +245,6 @@ TEST_SUITE_END() // Float
289245
290246template <typename T>
291247using NEActivationLayerQuantizedFixture = ActivationValidationQuantizedFixture<Tensor, Accessor, NEActivationLayer, T>;
292- template <typename T>
293- using NEActivationLayerWithPaddingQuantizedFixture = ActivationWithPaddingValidationQuantizedFixture<Tensor, Accessor, NEActivationLayer, T>;
294248
295249/* * Input data sets. */
296250const auto QuantizedActivationFunctionsDataset = framework::dataset::make(" ActivationFunction" ,
@@ -309,13 +263,6 @@ const auto QuantizedActivationFunctionsDataset = framework::dataset::make("Activ
309263const auto QuantizedActivationDataset = combine(combine(framework::dataset::make(" InPlace" , { false }),
310264 concat (QuantizedActivationFunctionsDataset, framework::dataset::make(" ActivationFunction" , ActivationLayerInfo::ActivationFunction::HARD_SWISH))),
311265 framework::dataset::make (" AlphaBeta" , { 0 .5f , 1 .f }));
312- const auto QuantizedActivationDatasetForPaddingAfterConfigure = combine(
313- make (" InPlace" , { false }),
314- concat(QuantizedActivationFunctionsDataset,
315- make (" ActivationFunction" , ActivationLayerInfo::ActivationFunction::HARD_SWISH)
316- ),
317- make(" AlphaBeta" , { 0 .5f })
318- );
319266
320267TEST_SUITE (Quantized)
321268TEST_SUITE (QASYMM8)
@@ -327,17 +274,6 @@ FIXTURE_DATA_TEST_CASE(RunSmall, NEActivationLayerQuantizedFixture<uint8_t>, fra
327274 // Validate output
328275 validate (Accessor (_target), _reference, helper::tolerance_qasymm8 (_function));
329276}
330- FIXTURE_DATA_TEST_CASE (PaddingAfterConfigure, NEActivationLayerWithPaddingQuantizedFixture<uint8_t >, framework::DatasetMode::ALL,
331- combine (
332- make (" Shape" , TensorShape{ 7U , 7U , 17U , 2U }),
333- QuantizedActivationDatasetForPaddingAfterConfigure,
334- make(" DataType" , DataType::QASYMM8),
335- make(" QuantizationInfo" , { QuantizationInfo (0 .1f , 128 .0f ) }
336- )))
337- {
338- // Validate output
339- validate (Accessor (_target), _reference, helper::tolerance_qasymm8 (_function));
340- }
341277TEST_SUITE_END () // QASYMM8
342278
343279TEST_SUITE (QASYMM8_SIGNED)
@@ -349,17 +285,6 @@ FIXTURE_DATA_TEST_CASE(RunSmall, NEActivationLayerQuantizedFixture<int8_t>, fram
349285 // Validate output
350286 validate (Accessor (_target), _reference, helper::tolerance_qasymm8 (_function));
351287}
352- FIXTURE_DATA_TEST_CASE (PaddingAfterConfigure, NEActivationLayerWithPaddingQuantizedFixture<int8_t >, framework::DatasetMode::ALL,
353- combine (
354- make (" Shape" , TensorShape{ 7U , 7U , 17U , 2U }),
355- QuantizedActivationDatasetForPaddingAfterConfigure,
356- make(" DataType" , DataType::QASYMM8_SIGNED),
357- make(" QuantizationInfo" , { QuantizationInfo (0 .5f , 10 .0f ) }
358- )))
359- {
360- // Validate output
361- validate (Accessor (_target), _reference, helper::tolerance_qasymm8 (_function));
362- }
363288TEST_SUITE_END () // QASYMM8_SIGNED
364289
365290/* * Input data sets. */
@@ -372,12 +297,6 @@ const auto Int16QuantizedActivationFunctionsDataset = framework::dataset::make("
372297const auto Int16QuantizedActivationDataset = combine(combine(framework::dataset::make(" InPlace" , { false }), Int16QuantizedActivationFunctionsDataset),
373298 framework::dataset::make (" AlphaBeta" , { 0 .5f , 1 .f }));
374299
375- const auto Int16QuantizedActivationDatasetForPaddingAfterConfigure = combine(
376- make (" InPlace" , { false }),
377- Int16QuantizedActivationFunctionsDataset,
378- make(" AlphaBeta" , { 0 .5f })
379- );
380-
381300TEST_SUITE (QSYMM16)
382301FIXTURE_DATA_TEST_CASE (RunSmall, NEActivationLayerQuantizedFixture<int16_t >, framework::DatasetMode::ALL, combine(combine(combine(datasets::SmallShapes(), Int16QuantizedActivationDataset),
383302 framework::dataset::make (" DataType" ,
@@ -387,17 +306,6 @@ FIXTURE_DATA_TEST_CASE(RunSmall, NEActivationLayerQuantizedFixture<int16_t>, fra
387306 // Validate output
388307 validate (Accessor (_target), _reference, tolerance_qsymm16);
389308}
390- FIXTURE_DATA_TEST_CASE (PaddingAfterConfigure, NEActivationLayerWithPaddingQuantizedFixture<int16_t >, framework::DatasetMode::ALL,
391- combine (
392- make (" Shape" , TensorShape{ 7U , 7U , 17U , 2U }),
393- Int16QuantizedActivationDatasetForPaddingAfterConfigure,
394- make(" DataType" , DataType::QSYMM16),
395- make(" QuantizationInfo" , { QuantizationInfo (1 .f / 32768 .f , 0 .f ) }))
396- )
397- {
398- // Validate output
399- validate (Accessor (_target), _reference, tolerance_qsymm16);
400- }
401309TEST_SUITE_END () // QSYMM16
402310TEST_SUITE_END () // Quantized
403311
0 commit comments