From 66cc12f27dcc2171e1c6f5b357e95e54689a48f0 Mon Sep 17 00:00:00 2001 From: Michel Iwaniec Date: Thu, 7 Dec 2017 17:26:40 +0000 Subject: IVGCVSW-839: Add QASYMM8 validation tests for CL ActivationLayer Change-Id: Ia564ccc6c5901aad2037fa84f69ef913b4118b24 Reviewed-on: https://eu-gerrit-1.euhpc.arm.com/112421 Reviewed-by: Georgios Pinitas Tested-by: BSG Visual Compute Jenkins server to access repositories on http://mpd-gerrit.cambridge.arm.com Reviewed-by: Anthony Barbier --- tests/validation/CL/ActivationLayer.cpp | 30 +++++++++++- tests/validation/fixtures/ActivationLayerFixture.h | 57 ++++++++++++++++------ tests/validation/reference/ActivationLayer.cpp | 9 ++++ 3 files changed, 81 insertions(+), 15 deletions(-) (limited to 'tests/validation') diff --git a/tests/validation/CL/ActivationLayer.cpp b/tests/validation/CL/ActivationLayer.cpp index d3f55d9959..5ceaecaaa9 100644 --- a/tests/validation/CL/ActivationLayer.cpp +++ b/tests/validation/CL/ActivationLayer.cpp @@ -237,7 +237,7 @@ TEST_SUITE_END() template using CLActivationLayerFixedPointFixture = ActivationValidationFixedPointFixture; -TEST_SUITE(Quantized) +TEST_SUITE(FixedPoint) TEST_SUITE(QS8) // We test for fixed point precision [3,5] because [1,2] and [6,7] ranges cause // overflowing issues in most of the transcendentals functions. @@ -280,6 +280,34 @@ FIXTURE_DATA_TEST_CASE(RunLarge, CLActivationLayerFixedPointFixture, fr TEST_SUITE_END() TEST_SUITE_END() +template +using CLActivationLayerQuantizedFixture = ActivationValidationQuantizedFixture; + +/** Input data sets. */ +const auto QuantizedActivationDataset = combine(combine(framework::dataset::make("InPlace", { false, true }), framework::dataset::make("ActivationFunction", { ActivationLayerInfo::ActivationFunction::LU_BOUNDED_RELU })), + framework::dataset::make("AlphaBeta", { 0.5f, 1.f })); + +TEST_SUITE(Quantized) +TEST_SUITE(QASYMM8) +FIXTURE_DATA_TEST_CASE(RunSmall, CLActivationLayerQuantizedFixture, framework::DatasetMode::PRECOMMIT, combine(combine(combine(datasets::SmallShapes(), QuantizedActivationDataset), + framework::dataset::make("DataType", + DataType::QASYMM8)), + framework::dataset::make("QuantizationInfo", { QuantizationInfo(0.1f, 128.0f) }))) +{ + // Validate output + validate(CLAccessor(_target), _reference, tolerance(_function, _data_type)); +} +FIXTURE_DATA_TEST_CASE(RunLarge, CLActivationLayerQuantizedFixture, framework::DatasetMode::NIGHTLY, combine(combine(combine(datasets::LargeShapes(), QuantizedActivationDataset), + framework::dataset::make("DataType", + DataType::QASYMM8)), + framework::dataset::make("QuantizationInfo", { QuantizationInfo(0.1f, 128.0f) }))) +{ + // Validate output + validate(CLAccessor(_target), _reference, tolerance(_function, _data_type)); +} +TEST_SUITE_END() +TEST_SUITE_END() + TEST_SUITE_END() TEST_SUITE_END() } // namespace validation diff --git a/tests/validation/fixtures/ActivationLayerFixture.h b/tests/validation/fixtures/ActivationLayerFixture.h index 75e625f715..e212c7bd9b 100644 --- a/tests/validation/fixtures/ActivationLayerFixture.h +++ b/tests/validation/fixtures/ActivationLayerFixture.h @@ -43,20 +43,21 @@ namespace test namespace validation { template -class ActivationValidationFixedPointFixture : public framework::Fixture +class ActivationValidationGenericFixture : public framework::Fixture { public: template - void setup(TensorShape shape, bool in_place, ActivationLayerInfo::ActivationFunction function, float alpha_beta, DataType data_type, int fractional_bits) + void setup(TensorShape shape, bool in_place, ActivationLayerInfo::ActivationFunction function, float alpha_beta, DataType data_type, int fractional_bits, QuantizationInfo quantization_info) { - _fractional_bits = fractional_bits; - _data_type = data_type; - _function = function; + _fractional_bits = fractional_bits; + _quantization_info = quantization_info; + _data_type = data_type; + _function = function; ActivationLayerInfo info(function, alpha_beta, alpha_beta); - _target = compute_target(shape, in_place, info, data_type, fractional_bits); - _reference = compute_reference(shape, info, data_type, fractional_bits); + _target = compute_target(shape, in_place, info, data_type, fractional_bits, quantization_info); + _reference = compute_reference(shape, info, data_type, fractional_bits, quantization_info); } protected: @@ -71,6 +72,10 @@ protected: std::uniform_real_distribution<> distribution(min_bound, max_bound); library->fill(tensor, distribution, 0); } + else if(is_data_type_quantized_asymmetric(tensor.data_type())) + { + library->fill_tensor_uniform(tensor, 0); + } else { int min_bound = 0; @@ -81,11 +86,11 @@ protected: } } - TensorType compute_target(const TensorShape &shape, bool in_place, ActivationLayerInfo info, DataType data_type, int fixed_point_position = 0) + TensorType compute_target(const TensorShape &shape, bool in_place, ActivationLayerInfo info, DataType data_type, int fixed_point_position, QuantizationInfo quantization_info) { // Create tensors - TensorType src = create_tensor(shape, data_type, 1, fixed_point_position); - TensorType dst = create_tensor(shape, data_type, 1, fixed_point_position); + TensorType src = create_tensor(shape, data_type, 1, fixed_point_position, quantization_info); + TensorType dst = create_tensor(shape, data_type, 1, fixed_point_position, quantization_info); // Create and configure function FunctionType act_layer; @@ -123,10 +128,10 @@ protected: } } - SimpleTensor compute_reference(const TensorShape &shape, ActivationLayerInfo info, DataType data_type, int fixed_point_position = 0) + SimpleTensor compute_reference(const TensorShape &shape, ActivationLayerInfo info, DataType data_type, int fixed_point_position, QuantizationInfo quantization_info) { // Create reference - SimpleTensor src{ shape, data_type, 1, fixed_point_position }; + SimpleTensor src{ shape, data_type, 1, fixed_point_position, quantization_info }; // Fill reference fill(src); @@ -137,20 +142,44 @@ protected: TensorType _target{}; SimpleTensor _reference{}; int _fractional_bits{}; + QuantizationInfo _quantization_info{}; DataType _data_type{}; ActivationLayerInfo::ActivationFunction _function{}; }; template -class ActivationValidationFixture : public ActivationValidationFixedPointFixture +class ActivationValidationFixture : public ActivationValidationGenericFixture { public: template void setup(TensorShape shape, bool in_place, ActivationLayerInfo::ActivationFunction function, float alpha_beta, DataType data_type) { - ActivationValidationFixedPointFixture::setup(shape, in_place, function, alpha_beta, data_type, 0); + ActivationValidationGenericFixture::setup(shape, in_place, function, alpha_beta, data_type, 0, QuantizationInfo()); } }; + +template +class ActivationValidationFixedPointFixture : public ActivationValidationGenericFixture +{ +public: + template + void setup(TensorShape shape, bool in_place, ActivationLayerInfo::ActivationFunction function, float alpha_beta, DataType data_type, int fractional_bits) + { + ActivationValidationGenericFixture::setup(shape, in_place, function, alpha_beta, data_type, fractional_bits, QuantizationInfo()); + } +}; + +template +class ActivationValidationQuantizedFixture : public ActivationValidationGenericFixture +{ +public: + template + void setup(TensorShape shape, bool in_place, ActivationLayerInfo::ActivationFunction function, float alpha_beta, DataType data_type, QuantizationInfo quantization_info) + { + ActivationValidationGenericFixture::setup(shape, in_place, function, alpha_beta, data_type, 0, quantization_info); + } +}; + } // namespace validation } // namespace test } // namespace arm_compute diff --git a/tests/validation/reference/ActivationLayer.cpp b/tests/validation/reference/ActivationLayer.cpp index 2243e6ff59..df7f6534bc 100644 --- a/tests/validation/reference/ActivationLayer.cpp +++ b/tests/validation/reference/ActivationLayer.cpp @@ -154,6 +154,15 @@ SimpleTensor activation_layer(const SimpleTensor &src, ActivationLayerInfo return dst; } +template <> +SimpleTensor activation_layer(const SimpleTensor &src, ActivationLayerInfo info) +{ + SimpleTensor src_tmp = convert_from_asymmetric(src); + SimpleTensor dst_tmp = activation_layer(src_tmp, info); + SimpleTensor dst = convert_to_asymmetric(dst_tmp, src.quantization_info()); + return dst; +} + template SimpleTensor activation_layer(const SimpleTensor &src, ActivationLayerInfo info); template SimpleTensor activation_layer(const SimpleTensor &src, ActivationLayerInfo info); template SimpleTensor activation_layer(const SimpleTensor &src, ActivationLayerInfo info); -- cgit v1.2.1