diff options
Diffstat (limited to 'tests/validation')
-rw-r--r-- | tests/validation/NEON/ActivationLayer.cpp | 19 | ||||
-rw-r--r-- | tests/validation/NEON/PReluLayer.cpp | 69 | ||||
-rw-r--r-- | tests/validation/fixtures/ActivationLayerFixture.h | 13 | ||||
-rw-r--r-- | tests/validation/reference/ElementwiseOperations.cpp | 30 |
4 files changed, 92 insertions, 39 deletions
diff --git a/tests/validation/NEON/ActivationLayer.cpp b/tests/validation/NEON/ActivationLayer.cpp index 8c18d47da9..1b9278988a 100644 --- a/tests/validation/NEON/ActivationLayer.cpp +++ b/tests/validation/NEON/ActivationLayer.cpp @@ -263,6 +263,25 @@ FIXTURE_DATA_TEST_CASE(RunLarge, NEActivationLayerQuantizedFixture<uint8_t>, fra } TEST_SUITE_END() // QASYMM8 +TEST_SUITE(QASYMM8_SIGNED) +FIXTURE_DATA_TEST_CASE(RunSmall, NEActivationLayerQuantizedFixture<int8_t>, framework::DatasetMode::PRECOMMIT, combine(combine(combine(datasets::SmallShapes(), QuantizedActivationDataset), + framework::dataset::make("DataType", + DataType::QASYMM8_SIGNED)), + framework::dataset::make("QuantizationInfo", { QuantizationInfo(0.5f, 10.0f) }))) +{ + // Validate output + validate(Accessor(_target), _reference, tolerance_qasymm8); +} +FIXTURE_DATA_TEST_CASE(RunLarge, NEActivationLayerQuantizedFixture<int8_t>, framework::DatasetMode::NIGHTLY, combine(combine(combine(datasets::LargeShapes(), QuantizedActivationDataset), + framework::dataset::make("DataType", + DataType::QASYMM8_SIGNED)), + framework::dataset::make("QuantizationInfo", { QuantizationInfo(0.5f, 10.0f) }))) +{ + // Validate output + validate(Accessor(_target), _reference, tolerance_qasymm8); +} +TEST_SUITE_END() // QASYMM8_SIGNED + /** Input data sets. */ const auto Int16QuantizedActivationFunctionsDataset = framework::dataset::make("ActivationFunction", { ActivationLayerInfo::ActivationFunction::LOGISTIC, ActivationLayerInfo::ActivationFunction::TANH diff --git a/tests/validation/NEON/PReluLayer.cpp b/tests/validation/NEON/PReluLayer.cpp index d9604f94aa..0630a057d6 100644 --- a/tests/validation/NEON/PReluLayer.cpp +++ b/tests/validation/NEON/PReluLayer.cpp @@ -42,12 +42,15 @@ namespace validation { namespace { -RelativeTolerance<float> tolerance_fp32(0.000001f); +RelativeTolerance<float> tolerance_fp32(0.000001f); +AbsoluteTolerance<int8_t> tolerance_s8(1); /** Input data sets **/ const auto PReluLayerQASYMM8Dataset = combine(combine(framework::dataset::make("DataType", DataType::QASYMM8), framework::dataset::make("DataType", DataType::QASYMM8)), framework::dataset::make("DataType", DataType::QASYMM8)); +const auto PReluLayerQASYMM8SignedDataset = combine(combine(framework::dataset::make("DataType", DataType::QASYMM8_SIGNED), framework::dataset::make("DataType", DataType::QASYMM8_SIGNED)), + framework::dataset::make("DataType", DataType::QASYMM8_SIGNED)); const auto PReluLayerFP32Dataset = combine(combine(framework::dataset::make("DataType", DataType::F32), framework::dataset::make("DataType", DataType::F32)), framework::dataset::make("DataType", DataType::F32)); @@ -101,23 +104,6 @@ using NEPReluLayerQuantizedFixture = PReluLayerValidationQuantizedFixture<Tensor TEST_SUITE(Quantized) TEST_SUITE(QASYMM8) -DATA_TEST_CASE(Configuration, framework::DatasetMode::ALL, datasets::SmallShapes(), - shape) -{ - // Create tensors - Tensor ref_src1 = create_tensor<Tensor>(shape, DataType::QASYMM8); - Tensor ref_src2 = create_tensor<Tensor>(shape, DataType::QASYMM8); - Tensor dst = create_tensor<Tensor>(shape, DataType::QASYMM8); - - // Create and Configure function - NEPReluLayer prelu; - prelu.configure(&ref_src1, &ref_src2, &dst); - - // Validate valid region - const ValidRegion valid_region = shape_to_valid_region(shape); - validate(dst.info()->valid_region(), valid_region); -} - FIXTURE_DATA_TEST_CASE(RunSmall, NEPReluLayerQuantizedFixture<uint8_t>, framework::DatasetMode::ALL, combine(combine(combine(combine(datasets::SmallShapes(), PReluLayerQASYMM8Dataset), framework::dataset::make("QuantizationInfo", { QuantizationInfo(5.f / 255.f, 20) })), @@ -141,8 +127,34 @@ FIXTURE_DATA_TEST_CASE(RunLarge, NEPReluLayerQuantizedFixture<uint8_t>, framewor // Validate output validate(Accessor(_target), _reference, tolerance_fp32, 0.01); } -TEST_SUITE_END() -TEST_SUITE_END() +TEST_SUITE_END() // QASYMM8 + +TEST_SUITE(QASYMM8_SIGNED) +FIXTURE_DATA_TEST_CASE(RunSmall, NEPReluLayerQuantizedFixture<int8_t>, framework::DatasetMode::ALL, combine(combine(combine(combine(datasets::SmallShapes(), + PReluLayerQASYMM8SignedDataset), + framework::dataset::make("QuantizationInfo", { QuantizationInfo(0.5f, 20) })), + framework::dataset::make("QuantizationInfo", { QuantizationInfo(0.5f, 10) })), + framework::dataset::make("QuantizationInfo", { QuantizationInfo(0.5f, 5) })) + + ) +{ + // Validate output + validate(Accessor(_target), _reference, tolerance_s8, 0.01); +} + +FIXTURE_DATA_TEST_CASE(RunLarge, NEPReluLayerQuantizedFixture<int8_t>, framework::DatasetMode::NIGHTLY, combine(combine(combine(combine(datasets::LargeShapes(), + PReluLayerQASYMM8SignedDataset), + framework::dataset::make("QuantizationInfo", { QuantizationInfo(0.5f, 20) })), + framework::dataset::make("QuantizationInfo", { QuantizationInfo(0.5f, 10) })), + framework::dataset::make("QuantizationInfo", { QuantizationInfo(0.5f, 5) })) + + ) +{ + // Validate output + validate(Accessor(_target), _reference, tolerance_s8, 0.01); +} +TEST_SUITE_END() // QASYMM8_SIGNED +TEST_SUITE_END() // Quantized TEST_SUITE(Float) #ifdef __ARM_FEATURE_FP16_VECTOR_ARITHMETIC @@ -162,23 +174,6 @@ TEST_SUITE_END() // FP16 #endif /* __ARM_FEATURE_FP16_VECTOR_ARITHMETIC */ TEST_SUITE(FP32) -DATA_TEST_CASE(Configuration, framework::DatasetMode::ALL, datasets::SmallShapes(), - shape) -{ - // Create tensors - Tensor ref_src1 = create_tensor<Tensor>(shape, DataType::F32); - Tensor ref_src2 = create_tensor<Tensor>(shape, DataType::F32); - Tensor dst = create_tensor<Tensor>(shape, DataType::F32); - - // Create and Configure function - NEPReluLayer prelu; - prelu.configure(&ref_src1, &ref_src2, &dst); - - // Validate valid region - const ValidRegion valid_region = shape_to_valid_region(shape); - validate(dst.info()->valid_region(), valid_region); -} - FIXTURE_DATA_TEST_CASE(RunSmall, NEPReluLayerFixture<float>, framework::DatasetMode::ALL, combine(datasets::SmallShapes(), PReluLayerFP32Dataset)) { // Validate output diff --git a/tests/validation/fixtures/ActivationLayerFixture.h b/tests/validation/fixtures/ActivationLayerFixture.h index f6d43ddd89..3294986519 100644 --- a/tests/validation/fixtures/ActivationLayerFixture.h +++ b/tests/validation/fixtures/ActivationLayerFixture.h @@ -150,8 +150,9 @@ protected: private: QuantizationInfo calculate_output_quantization_info(DataType dt, const ActivationLayerInfo &act_info, const QuantizationInfo &default_qinfo) { - auto qasymm8_max = float(std::numeric_limits<uint8_t>::max()) + 1.f; - auto qsymm16_max = float(std::numeric_limits<int16_t>::max()) + 1.f; + auto qasymm8_max = float(std::numeric_limits<uint8_t>::max()) + 1.f; + auto qasymm8_signed_max = float(std::numeric_limits<int8_t>::max()) + 1.f; + auto qsymm16_max = float(std::numeric_limits<int16_t>::max()) + 1.f; switch(act_info.activation()) { @@ -164,6 +165,10 @@ private: { return QuantizationInfo(1.f / (0.5 * qasymm8_max), int(0.5 * qasymm8_max)); } + else if(dt == DataType::QASYMM8_SIGNED) + { + return QuantizationInfo(1.f / qasymm8_signed_max, 0); + } else { return default_qinfo; @@ -177,6 +182,10 @@ private: { return QuantizationInfo(1.f / qasymm8_max, 0); } + else if(dt == DataType::QASYMM8_SIGNED) + { + return QuantizationInfo(1.f / (2.f * qasymm8_signed_max), -int(qasymm8_signed_max)); + } else { return default_qinfo; diff --git a/tests/validation/reference/ElementwiseOperations.cpp b/tests/validation/reference/ElementwiseOperations.cpp index 7b39e18bd9..bd6eec3688 100644 --- a/tests/validation/reference/ElementwiseOperations.cpp +++ b/tests/validation/reference/ElementwiseOperations.cpp @@ -183,6 +183,36 @@ SimpleTensor<uint8_t> arithmetic_operation(ArithmeticOperation op, const SimpleT return dst; } } +template <> +SimpleTensor<int8_t> arithmetic_operation(ArithmeticOperation op, const SimpleTensor<int8_t> &src1, const SimpleTensor<int8_t> &src2, SimpleTensor<int8_t> &dst, ConvertPolicy convert_policy) +{ + if(dst.data_type() == DataType::QASYMM8_SIGNED) + { + SimpleTensor<float> src1_tmp = convert_from_asymmetric(src1); + SimpleTensor<float> src2_tmp = convert_from_asymmetric(src2); + SimpleTensor<float> dst_tmp(TensorShape::broadcast_shape(src1.shape(), src2.shape()), dst.data_type()); + + Coordinates id_src1{}; + Coordinates id_src2{}; + Coordinates id_dst{}; + + BroadcastUnroll<Coordinates::num_max_dimensions>::unroll(op, src1_tmp, src2_tmp, dst_tmp, convert_policy, id_src1, id_src2, id_dst); + + dst = convert_to_asymmetric<int8_t>(dst_tmp, dst.quantization_info()); + return dst; + } + else + { + // DataType::S8 + Coordinates id_src1{}; + Coordinates id_src2{}; + Coordinates id_dst{}; + + BroadcastUnroll<Coordinates::num_max_dimensions>::unroll(op, src1, src2, dst, convert_policy, id_src1, id_src2, id_dst); + + return dst; + } +} template <> SimpleTensor<int16_t> arithmetic_operation(ArithmeticOperation op, const SimpleTensor<int16_t> &src1, const SimpleTensor<int16_t> &src2, SimpleTensor<int16_t> &dst, ConvertPolicy convert_policy) |