From 3689fcd5915cd902cb4ea5f618f2a6e42f6dc4a1 Mon Sep 17 00:00:00 2001 From: Manuel Bottini Date: Fri, 14 Jun 2019 17:18:12 +0100 Subject: COMPMID-2408: Add QSYMM16 support for ElementwiseAddition for NEON Change-Id: I22991e9369ffba9b51a94522ff4977933e887b94 Signed-off-by: Manuel Bottini Reviewed-on: https://review.mlplatform.org/c/1352 Comments-Addressed: Arm Jenkins Reviewed-by: Michele Di Giorgio Tested-by: Arm Jenkins Reviewed-by: Giuseppe Rossini --- tests/validation/Helpers.cpp | 26 ++++++++++++ tests/validation/Helpers.h | 19 +++++++++ tests/validation/NEON/ArithmeticAddition.cpp | 46 ++++++++++++++++++++-- .../validation/reference/ArithmeticOperations.cpp | 30 ++++++++++++-- 4 files changed, 114 insertions(+), 7 deletions(-) (limited to 'tests/validation') diff --git a/tests/validation/Helpers.cpp b/tests/validation/Helpers.cpp index 31d6bfae07..360859e487 100644 --- a/tests/validation/Helpers.cpp +++ b/tests/validation/Helpers.cpp @@ -132,6 +132,32 @@ SimpleTensor convert_to_asymmetric(const SimpleTensor &src, cons return dst; } +template <> +SimpleTensor convert_to_symmetric(const SimpleTensor &src, const QuantizationInfo &quantization_info) +{ + SimpleTensor dst{ src.shape(), DataType::QSYMM16, 1, quantization_info }; + const UniformQuantizationInfo &qinfo = quantization_info.uniform(); + + for(int i = 0; i < src.num_elements(); ++i) + { + dst[i] = quantize_qsymm16(src[i], qinfo); + } + return dst; +} + +template <> +SimpleTensor convert_from_symmetric(const SimpleTensor &src) +{ + const UniformQuantizationInfo &quantization_info = src.quantization_info().uniform(); + SimpleTensor dst{ src.shape(), DataType::F32, 1, QuantizationInfo(), src.data_layout() }; + + for(int i = 0; i < src.num_elements(); ++i) + { + dst[i] = dequantize_qsymm16(src[i], quantization_info); + } + return dst; +} + template void matrix_multiply(const SimpleTensor &a, const SimpleTensor &b, SimpleTensor &out) { diff --git a/tests/validation/Helpers.h b/tests/validation/Helpers.h index 2e8c667a41..44dd7a9b81 100644 --- a/tests/validation/Helpers.h +++ b/tests/validation/Helpers.h @@ -194,6 +194,25 @@ SimpleTensor convert_from_asymmetric(const SimpleTensor &src); */ SimpleTensor convert_to_asymmetric(const SimpleTensor &src, const QuantizationInfo &quantization_info); +/** Convert quantized simple tensor into float using tensor quantization information. + * + * @param[in] src Quantized tensor. + * + * @return Float tensor. + */ +template +SimpleTensor convert_from_symmetric(const SimpleTensor &src); + +/** Convert float simple tensor into quantized using specified quantization information. + * + * @param[in] src Float tensor. + * @param[in] quantization_info Quantification information. + * + * @return Quantized tensor. + */ +template +SimpleTensor convert_to_symmetric(const SimpleTensor &src, const QuantizationInfo &quantization_info); + /** Matrix multiply between 2 float simple tensors * * @param[in] a Input tensor A diff --git a/tests/validation/NEON/ArithmeticAddition.cpp b/tests/validation/NEON/ArithmeticAddition.cpp index 4a72dfc923..8d8a327f69 100644 --- a/tests/validation/NEON/ArithmeticAddition.cpp +++ b/tests/validation/NEON/ArithmeticAddition.cpp @@ -45,7 +45,7 @@ namespace { #ifndef __aarch64__ constexpr AbsoluteTolerance tolerance_qasymm8(1); /**< Tolerance value for comparing reference's output against implementation's output for quantized data types */ -#endif //__aarch64__ +#endif //__aarch64__ /** Input data sets **/ const auto ArithmeticAdditionU8Dataset = combine(combine(framework::dataset::make("DataType", DataType::U8), framework::dataset::make("DataType", DataType::U8)), framework::dataset::make("DataType", @@ -60,6 +60,8 @@ const auto ArithmeticAdditionFP32Dataset = combine(combine(framework::dataset::m framework::dataset::make("DataType", DataType::F32)); const auto ArithmeticAdditionQASYMM8Dataset = combine(combine(framework::dataset::make("DataType", DataType::QASYMM8), framework::dataset::make("DataType", DataType::QASYMM8)), framework::dataset::make("DataType", DataType::QASYMM8)); +const auto ArithmeticAdditionQSYMM16Dataset = combine(combine(framework::dataset::make("DataType", DataType::QSYMM16), framework::dataset::make("DataType", DataType::QSYMM16)), + framework::dataset::make("DataType", DataType::QSYMM16)); } // namespace TEST_SUITE(NEON) @@ -275,9 +277,9 @@ FIXTURE_DATA_TEST_CASE(RunSmall, framework::DatasetMode::PRECOMMIT, combine(combine(combine(combine(combine(datasets::SmallShapes(), ArithmeticAdditionQASYMM8Dataset), framework::dataset::make("ConvertPolicy", { ConvertPolicy::SATURATE })), - framework::dataset::make("QuantizationInfo", { QuantizationInfo(5.f / 255.f, 20) })), - framework::dataset::make("QuantizationInfo", { QuantizationInfo(2.f / 255.f, 10) })), - framework::dataset::make("QuantizationInfo", { QuantizationInfo(1.f / 255.f, 5) }))) + framework::dataset::make("Src0QInfo", { QuantizationInfo(5.f / 255.f, 20) })), + framework::dataset::make("Src1QInfo", { QuantizationInfo(2.f / 255.f, 10) })), + framework::dataset::make("OutQInfo", { QuantizationInfo(1.f / 255.f, 5) }))) { // Validate output #ifdef __aarch64__ @@ -287,6 +289,42 @@ FIXTURE_DATA_TEST_CASE(RunSmall, #endif //__aarch64__ } TEST_SUITE_END() // QASYMM8 +TEST_SUITE(QSYMM16) +DATA_TEST_CASE(Configuration, framework::DatasetMode::ALL, combine(datasets::SmallShapes(), framework::dataset::make("ConvertPolicy", { ConvertPolicy::SATURATE })), + shape, policy) +{ + // Create tensors + Tensor ref_src1 = create_tensor(shape, DataType::QSYMM16); + Tensor ref_src2 = create_tensor(shape, DataType::QSYMM16); + Tensor dst = create_tensor(shape, DataType::QSYMM16); + + // Create and Configure function + NEArithmeticAddition add; + add.configure(&ref_src1, &ref_src2, &dst, policy); + + // Validate valid region + const ValidRegion valid_region = shape_to_valid_region(shape); + validate(dst.info()->valid_region(), valid_region); + + // Validate padding + validate(ref_src1.info()->padding(), PaddingSize()); + validate(ref_src2.info()->padding(), PaddingSize()); + validate(dst.info()->padding(), PaddingSize()); +} + +FIXTURE_DATA_TEST_CASE(RunSmall, + NEArithmeticAdditionQuantizedFixture, + framework::DatasetMode::PRECOMMIT, + combine(combine(combine(combine(combine(datasets::SmallShapes(), ArithmeticAdditionQSYMM16Dataset), + framework::dataset::make("ConvertPolicy", { ConvertPolicy::SATURATE })), + framework::dataset::make("Src0QInfo", { QuantizationInfo(1.f / 32768.f, 0), QuantizationInfo(5.f / 32768.f, 0) })), + framework::dataset::make("Src1QInfo", { QuantizationInfo(2.f / 32768.f, 0), QuantizationInfo(5.f / 32768.f, 0) })), + framework::dataset::make("OutQInfo", { QuantizationInfo(5.f / 32768.f, 0) }))) +{ + // Validate output + validate(Accessor(_target), _reference); +} +TEST_SUITE_END() // QSYMM16 TEST_SUITE_END() // Quantized TEST_SUITE_END() // ArithmeticAddition diff --git a/tests/validation/reference/ArithmeticOperations.cpp b/tests/validation/reference/ArithmeticOperations.cpp index a6205af2c6..abd4f31d72 100644 --- a/tests/validation/reference/ArithmeticOperations.cpp +++ b/tests/validation/reference/ArithmeticOperations.cpp @@ -124,8 +124,32 @@ SimpleTensor arithmetic_operation(ArithmeticOperation op, const SimpleT } } -template SimpleTensor arithmetic_operation(ArithmeticOperation op, const SimpleTensor &src1, const SimpleTensor &src2, SimpleTensor &dst, - ConvertPolicy convert_policy); +template <> +SimpleTensor arithmetic_operation(ArithmeticOperation op, const SimpleTensor &src1, const SimpleTensor &src2, SimpleTensor &dst, ConvertPolicy convert_policy) +{ + Coordinates id_src1{}; + Coordinates id_src2{}; + Coordinates id_dst{}; + + if(dst.data_type() == DataType::QSYMM16) + { + SimpleTensor src1_tmp = convert_from_symmetric(src1); + SimpleTensor src2_tmp = convert_from_symmetric(src2); + SimpleTensor dst_tmp(TensorShape::broadcast_shape(src1.shape(), src2.shape()), dst.data_type()); + + BroadcastUnroll::unroll(op, src1_tmp, src2_tmp, dst_tmp, convert_policy, id_src1, id_src2, id_dst); + + dst = convert_to_symmetric(dst_tmp, dst.quantization_info()); + return dst; + } + else + { + // DataType::S16 + BroadcastUnroll::unroll(op, src1, src2, dst, convert_policy, id_src1, id_src2, id_dst); + return dst; + } +} + template SimpleTensor arithmetic_operation(ArithmeticOperation op, const SimpleTensor &src1, const SimpleTensor &src2, SimpleTensor &dst, ConvertPolicy convert_policy); template SimpleTensor arithmetic_operation(ArithmeticOperation op, const SimpleTensor &src1, const SimpleTensor &src2, SimpleTensor &dst, ConvertPolicy convert_policy); template SimpleTensor arithmetic_operation(ArithmeticOperation op, const SimpleTensor &src1, const SimpleTensor &src2, SimpleTensor &dst, ConvertPolicy convert_policy); @@ -133,7 +157,7 @@ template SimpleTensor arithmetic_operation(ArithmeticOperation op, const template SimpleTensor arithmetic_operation(ArithmeticOperation op, const SimpleTensor &src1, const SimpleTensor &src2, DataType dst_data_type, ConvertPolicy convert_policy) { - ARM_COMPUTE_ERROR_ON_MSG(dst_data_type == DataType::QASYMM8, "For QASYMM8, the quantized output tensor should be passed directly."); + ARM_COMPUTE_ERROR_ON_MSG(is_data_type_quantized(dst_data_type), "For quantized input data types, the quantized output tensor should be passed directly."); SimpleTensor dst(TensorShape::broadcast_shape(src1.shape(), src2.shape()), dst_data_type); arithmetic_operation(op, src1, src2, dst, convert_policy); -- cgit v1.2.1