From 4e09b3839206254d0df56095ad0762718a764c9c Mon Sep 17 00:00:00 2001 From: Michele Di Giorgio Date: Wed, 5 Jul 2017 18:20:02 +0100 Subject: COMPMID-439: Implement NEON Quantization Layer. Change-Id: Iefbb421915e56d880d6a3e20c113913560f6ca10 Reviewed-on: http://mpd-gerrit.cambridge.arm.com/79934 Tested-by: Kaizen Reviewed-by: Georgios Pinitas --- .../core/NEON/kernels/NEQuantizationLayerKernel.h | 68 ++++++++++++ .../runtime/NEON/functions/NEQuantizationLayer.h | 67 ++++++++++++ .../NEON/kernels/NEQuantizationLayerKernel.cpp | 106 ++++++++++++++++++ src/runtime/NEON/functions/NEQuantizationLayer.cpp | 56 ++++++++++ tests/validation_new/CPP/QuantizationLayer.cpp | 85 +++++++++++++++ tests/validation_new/CPP/QuantizationLayer.h | 44 ++++++++ tests/validation_new/NEON/QuantizationLayer.cpp | 98 +++++++++++++++++ .../fixtures/QuantizationLayerFixture.h | 120 +++++++++++++++++++++ 8 files changed, 644 insertions(+) create mode 100644 arm_compute/core/NEON/kernels/NEQuantizationLayerKernel.h create mode 100644 arm_compute/runtime/NEON/functions/NEQuantizationLayer.h create mode 100644 src/core/NEON/kernels/NEQuantizationLayerKernel.cpp create mode 100644 src/runtime/NEON/functions/NEQuantizationLayer.cpp create mode 100644 tests/validation_new/CPP/QuantizationLayer.cpp create mode 100644 tests/validation_new/CPP/QuantizationLayer.h create mode 100644 tests/validation_new/NEON/QuantizationLayer.cpp create mode 100644 tests/validation_new/fixtures/QuantizationLayerFixture.h diff --git a/arm_compute/core/NEON/kernels/NEQuantizationLayerKernel.h b/arm_compute/core/NEON/kernels/NEQuantizationLayerKernel.h new file mode 100644 index 0000000000..9867e1db03 --- /dev/null +++ b/arm_compute/core/NEON/kernels/NEQuantizationLayerKernel.h @@ -0,0 +1,68 @@ +/* + * Copyright (c) 2017 ARM Limited. + * + * SPDX-License-Identifier: MIT + * + * Permission is hereby granted, free of charge, to any person obtaining a copy + * of this software and associated documentation files (the "Software"), to + * deal in the Software without restriction, including without limitation the + * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or + * sell copies of the Software, and to permit persons to whom the Software is + * furnished to do so, subject to the following conditions: + * + * The above copyright notice and this permission notice shall be included in all + * copies or substantial portions of the Software. + * + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE + * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE + * SOFTWARE. + */ +#ifndef __ARM_COMPUTE_NEQUANTIZATIONLAYERKERNEL_H__ +#define __ARM_COMPUTE_NEQUANTIZATIONLAYERKERNEL_H__ + +#include "arm_compute/core/NEON/INEKernel.h" + +namespace arm_compute +{ +class ITensor; + +/** Interface for the quantization layer kernel. */ +class NEQuantizationLayerKernel : public INEKernel +{ +public: + /** Default constructor */ + NEQuantizationLayerKernel(); + /** Prevent instances of this class from being copied (As this class contains pointers) */ + NEQuantizationLayerKernel(const NEQuantizationLayerKernel &) = delete; + /** Prevent instances of this class from being copied (As this class contains pointers) */ + NEQuantizationLayerKernel &operator=(const NEQuantizationLayerKernel &) = delete; + /** Default Move Constructor. */ + NEQuantizationLayerKernel(NEQuantizationLayerKernel &&) = default; + /** Default move assignment operator. */ + NEQuantizationLayerKernel &operator=(NEQuantizationLayerKernel &&) = default; + /** Default destructor */ + ~NEQuantizationLayerKernel() = default; + /** Set the input, output, min and max. + * + * @param[in] input Source tensor. Data types supported: F32. + * @param[out] output Destination tensor. Data types supported: U8. + * @param[in] min Pointer to the minimum value of the input tensor. + * @param[in] max Pointer to the maximum value of the input tensor. + */ + void configure(const ITensor *input, ITensor *output, const float *min, const float *max); + + // Inherited methods overridden: + void run(const Window &window) override; + +private: + const ITensor *_input; + ITensor *_output; + const float *_min; + const float *_max; +}; +} +#endif /*__ARM_COMPUTE_NEQUANTIZATIONLAYERKERNEL_H__ */ diff --git a/arm_compute/runtime/NEON/functions/NEQuantizationLayer.h b/arm_compute/runtime/NEON/functions/NEQuantizationLayer.h new file mode 100644 index 0000000000..ad0f68c58c --- /dev/null +++ b/arm_compute/runtime/NEON/functions/NEQuantizationLayer.h @@ -0,0 +1,67 @@ +/* + * Copyright (c) 2017 ARM Limited. + * + * SPDX-License-Identifier: MIT + * + * Permission is hereby granted, free of charge, to any person obtaining a copy + * of this software and associated documentation files (the "Software"), to + * deal in the Software without restriction, including without limitation the + * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or + * sell copies of the Software, and to permit persons to whom the Software is + * furnished to do so, subject to the following conditions: + * + * The above copyright notice and this permission notice shall be included in all + * copies or substantial portions of the Software. + * + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE + * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE + * SOFTWARE. + */ +#ifndef __ARM_COMPUTE_NEQUANTIZATIONLAYER_H__ +#define __ARM_COMPUTE_NEQUANTIZATIONLAYER_H__ + +#include "arm_compute/runtime/IFunction.h" + +#include "arm_compute/core/NEON/kernels/NEMinMaxLocationKernel.h" +#include "arm_compute/core/NEON/kernels/NEQuantizationLayerKernel.h" +#include "arm_compute/runtime/Tensor.h" + +#include "arm_compute/core/Types.h" + +namespace arm_compute +{ +class ITensor; + +/** Basic function to simulate a quantization layer. This function calls the following NEON kernels: + * + * -# @ref NEMinMaxKernel + * -# @ref NEQuantizationLayerKernel + * + */ +class NEQuantizationLayer : public IFunction +{ +public: + /** Default constructor */ + NEQuantizationLayer(); + /** Set the input and output tensors. + * + * @param[in] input Source tensor. Data types supported: F32 + * @param[out] output Destination tensor. Data types supported: F32 + */ + void configure(const ITensor *input, ITensor *output); + + // Inherited methods overridden: + void run() override; + +private: + NEQuantizationLayerKernel _quantize_kernel; + NEMinMaxKernel _min_max_kernel; + float _min; + float _max; +}; +} +#endif /* __ARM_COMPUTE_NEQUANTIZATIONLAYER_H__ */ diff --git a/src/core/NEON/kernels/NEQuantizationLayerKernel.cpp b/src/core/NEON/kernels/NEQuantizationLayerKernel.cpp new file mode 100644 index 0000000000..fbf7e1ba87 --- /dev/null +++ b/src/core/NEON/kernels/NEQuantizationLayerKernel.cpp @@ -0,0 +1,106 @@ +/* + * Copyright (c) 2017 ARM Limited. + * + * SPDX-License-Identifier: MIT + * + * Permission is hereby granted, free of charge, to any person obtaining a copy + * of this software and associated documentation files (the "Software"), to + * deal in the Software without restriction, including without limitation the + * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or + * sell copies of the Software, and to permit persons to whom the Software is + * furnished to do so, subject to the following conditions: + * + * The above copyright notice and this permission notice shall be included in all + * copies or substantial portions of the Software. + * + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE + * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE + * SOFTWARE. + */ +#include "arm_compute/core/NEON/kernels/NEQuantizationLayerKernel.h" + +#include "arm_compute/core/Error.h" +#include "arm_compute/core/Helpers.h" +#include "arm_compute/core/TensorInfo.h" +#include "arm_compute/core/Utils.h" +#include "arm_compute/core/Validate.h" +#include "arm_compute/core/Window.h" + +#include + +using namespace arm_compute; + +NEQuantizationLayerKernel::NEQuantizationLayerKernel() + : _input(nullptr), _output(nullptr), _min(nullptr), _max(nullptr) +{ +} + +void NEQuantizationLayerKernel::configure(const ITensor *input, ITensor *output, const float *min, const float *max) +{ + ARM_COMPUTE_ERROR_ON_DATA_TYPE_CHANNEL_NOT_IN(input, 1, DataType::F32); + ARM_COMPUTE_ERROR_ON_NULLPTR(output); + + // Output tensor auto initialization if not yet initialized + auto_init_if_empty(*output->info(), input->info()->tensor_shape(), 1, DataType::U8, 0); + + ARM_COMPUTE_ERROR_ON_DATA_TYPE_CHANNEL_NOT_IN(output, 1, DataType::U8); + ARM_COMPUTE_ERROR_ON_MISMATCHING_SHAPES(input, output); + + _input = input; + _output = output; + _min = min; + _max = max; + + constexpr unsigned int num_elems_processed_per_iteration = 8; + + // Configure window + Window win = calculate_max_window(*input->info(), Steps(num_elems_processed_per_iteration)); + AccessWindowHorizontal output_access(output->info(), 0, num_elems_processed_per_iteration); + update_window_and_padding(win, AccessWindowHorizontal(input->info(), 0, num_elems_processed_per_iteration), output_access); + output_access.set_valid_region(win, input->info()->valid_region()); + + INEKernel::configure(win); +} + +void NEQuantizationLayerKernel::run(const Window &window) +{ + ARM_COMPUTE_ERROR_ON_UNCONFIGURED_KERNEL(this); + ARM_COMPUTE_ERROR_ON_INVALID_SUBWINDOW(INEKernel::window(), window); + + Iterator input(_input, window); + Iterator output(_output, window); + + const float32x4_t vmin = vdupq_n_f32(*_min); + const float32x4_t inv_range = vdupq_n_f32(1.0f / (*_max - *_min)); + const float32x4_t quantization_max = vdupq_n_f32(255.0f); + const float32x4_t quantization_mul = vdupq_n_f32(256.0f); + + // Uniformly map values to range 8bit integers, i.e. [min, max] -> [0, 255] + execute_window_loop(window, [&](const Coordinates & id) + { + float32x4x2_t val = vld2q_f32(reinterpret_cast(input.ptr())); + // Map float values to range [0.0, 1.0] + val.val[0] = vsubq_f32(val.val[0], vmin); + val.val[1] = vsubq_f32(val.val[1], vmin); + val.val[0] = vmulq_f32(val.val[0], inv_range); + val.val[1] = vmulq_f32(val.val[1], inv_range); + + // Quantize + val.val[0] = vmulq_f32(val.val[0], quantization_mul); + val.val[1] = vmulq_f32(val.val[1], quantization_mul); + val.val[0] = vminq_f32(val.val[0], quantization_max); + val.val[1] = vminq_f32(val.val[1], quantization_max); + + const uint32x4_t val_u32_low = vcvtq_u32_f32(val.val[0]); + const uint32x4_t val_u32_high = vcvtq_u32_f32(val.val[1]); + const uint16x4x2_t val_u16 = vzip_u16(vmovn_u32(val_u32_low), vmovn_u32(val_u32_high)); + + const uint8x8_t quantized = vmovn_u16(vcombine_u16(val_u16.val[0], val_u16.val[1])); + vst1_u8(reinterpret_cast(output.ptr()), quantized); + }, + input, output); +} diff --git a/src/runtime/NEON/functions/NEQuantizationLayer.cpp b/src/runtime/NEON/functions/NEQuantizationLayer.cpp new file mode 100644 index 0000000000..46b9d7d707 --- /dev/null +++ b/src/runtime/NEON/functions/NEQuantizationLayer.cpp @@ -0,0 +1,56 @@ +/* + * Copyright (c) 2017 ARM Limited. + * + * SPDX-License-Identifier: MIT + * + * Permission is hereby granted, free of charge, to any person obtaining a copy + * of this software and associated documentation files (the "Software"), to + * deal in the Software without restriction, including without limitation the + * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or + * sell copies of the Software, and to permit persons to whom the Software is + * furnished to do so, subject to the following conditions: + * + * The above copyright notice and this permission notice shall be included in all + * copies or substantial portions of the Software. + * + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE + * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE + * SOFTWARE. + */ + +#include "arm_compute/runtime/NEON/functions/NEQuantizationLayer.h" + +#include "arm_compute/core/Types.h" +#include "arm_compute/runtime/NEON/NEScheduler.h" + +using namespace arm_compute; + +NEQuantizationLayer::NEQuantizationLayer() + : _quantize_kernel(), _min_max_kernel(), _min(0.f), _max(0.f) +{ +} + +void NEQuantizationLayer::configure(const ITensor *input, ITensor *output) +{ + // Configure min-max kernel + _min_max_kernel.configure(input, &_min, &_max); + + // Configure quantize kernel + _quantize_kernel.configure(input, output, &_min, &_max); +} + +void NEQuantizationLayer::run() +{ + // Reset min and max + _min_max_kernel.reset(); + + // Run min and max kernel + NEScheduler::get().schedule(&_min_max_kernel, Window::DimY); + + // Run quantize kernel + NEScheduler::get().schedule(&_quantize_kernel, Window::DimY); +} diff --git a/tests/validation_new/CPP/QuantizationLayer.cpp b/tests/validation_new/CPP/QuantizationLayer.cpp new file mode 100644 index 0000000000..d61e75a3a9 --- /dev/null +++ b/tests/validation_new/CPP/QuantizationLayer.cpp @@ -0,0 +1,85 @@ +/* + * Copyright (c) 2017 ARM Limited. + * + * SPDX-License-Identifier: MIT + * + * Permission is hereby granted, free of charge, to any person obtaining a copy + * of this software and associated documentation files (the "Software"), to + * deal in the Software without restriction, including without limitation the + * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or + * sell copies of the Software, and to permit persons to whom the Software is + * furnished to do so, subject to the following conditions: + * + * The above copyright notice and this permission notice shall be included in all + * copies or substantial portions of the Software. + * + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE + * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE + * SOFTWARE. + */ +#include "QuantizationLayer.h" + +namespace arm_compute +{ +namespace test +{ +namespace validation +{ +namespace reference +{ +void compute_min_max(const SimpleTensor &src, float *min, float *max) +{ + // Set min and max to first pixel + float tmp_min = src[0]; + float tmp_max = src[0]; + + // Look for min and max values + for(int i = 1; i < src.num_elements(); ++i) + { + if(src[i] < tmp_min) + { + tmp_min = src[i]; + } + if(src[i] > tmp_max) + { + tmp_max = src[i]; + } + } + + *min = tmp_min; + *max = tmp_max; +} + +template ::value, int>::type> +SimpleTensor quantization_layer(const SimpleTensor &src) +{ + // Create reference + SimpleTensor dst{ src.shape(), DataType::U8 }; + + // Compute min and max of the tensor using Min-Max layer + float min = 0.f; + float max = 0.f; + + compute_min_max(src, &min, &max); + + const float range = max - min; + + for(int i = 0; i < src.num_elements(); ++i) + { + // map values to range [0.0, 1.0] + const float normalized = (src[i] - min) / range; + dst[i] = static_cast(std::min(255.0f, normalized * 256.0f)); + } + + return dst; +} + +template SimpleTensor quantization_layer(const SimpleTensor &src); +} // namespace reference +} // namespace validation +} // namespace test +} // namespace arm_compute diff --git a/tests/validation_new/CPP/QuantizationLayer.h b/tests/validation_new/CPP/QuantizationLayer.h new file mode 100644 index 0000000000..5fd2379259 --- /dev/null +++ b/tests/validation_new/CPP/QuantizationLayer.h @@ -0,0 +1,44 @@ +/* + * Copyright (c) 2017 ARM Limited. + * + * SPDX-License-Identifier: MIT + * + * Permission is hereby granted, free of charge, to any person obtaining a copy + * of this software and associated documentation files (the "Software"), to + * deal in the Software without restriction, including without limitation the + * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or + * sell copies of the Software, and to permit persons to whom the Software is + * furnished to do so, subject to the following conditions: + * + * The above copyright notice and this permission notice shall be included in all + * copies or substantial portions of the Software. + * + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE + * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE + * SOFTWARE. + */ +#ifndef __ARM_COMPUTE_TEST_QUANTIZATION_LAYER_H__ +#define __ARM_COMPUTE_TEST_QUANTIZATION_LAYER_H__ + +#include "tests/validation_new/Helpers.h" +#include "tests/validation_new/SimpleTensor.h" + +namespace arm_compute +{ +namespace test +{ +namespace validation +{ +namespace reference +{ +template ::value, int>::type = 0> +SimpleTensor quantization_layer(const SimpleTensor &src); +} // namespace reference +} // namespace validation +} // namespace test +} // namespace arm_compute +#endif /* __ARM_COMPUTE_TEST_QUANTIZATION_LAYER_H__ */ diff --git a/tests/validation_new/NEON/QuantizationLayer.cpp b/tests/validation_new/NEON/QuantizationLayer.cpp new file mode 100644 index 0000000000..8b2acd6331 --- /dev/null +++ b/tests/validation_new/NEON/QuantizationLayer.cpp @@ -0,0 +1,98 @@ +/* + * Copyright (c) 2017 ARM Limited. + * + * SPDX-License-Identifier: MIT + * + * Permission is hereby granted, free of charge, to any person obtaining a copy + * of this software and associated documentation files (the "Software"), to + * deal in the Software without restriction, including without limitation the + * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or + * sell copies of the Software, and to permit persons to whom the Software is + * furnished to do so, subject to the following conditions: + * + * The above copyright notice and this permission notice shall be included in all + * copies or substantial portions of the Software. + * + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE + * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE + * SOFTWARE. + */ +#include "arm_compute/core/Types.h" +#include "arm_compute/runtime/NEON/functions/NEQuantizationLayer.h" +#include "arm_compute/runtime/Tensor.h" +#include "arm_compute/runtime/TensorAllocator.h" +#include "framework/Asserts.h" +#include "framework/Macros.h" +#include "framework/datasets/Datasets.h" +#include "tests/NEON/Accessor.h" +#include "tests/PaddingCalculator.h" +#include "tests/datasets_new/ShapeDatasets.h" +#include "tests/validation_new/Validation.h" +#include "tests/validation_new/fixtures/QuantizationLayerFixture.h" + +namespace arm_compute +{ +namespace test +{ +namespace validation +{ +namespace +{ +/** Tolerance for quantization */ +constexpr AbsoluteTolerance tolerance_u8(1); +} // namespace + +TEST_SUITE(NEON) +TEST_SUITE(QuantizationLayer) + +DATA_TEST_CASE(Configuration, framework::DatasetMode::ALL, combine(concat(datasets::Small2DShapes(), datasets::Large2DShapes()), framework::dataset::make("DataType", DataType::F32)), shape, data_type) +{ + // Create tensors + Tensor src = create_tensor(shape, data_type); + Tensor dst = create_tensor(shape, DataType::U8); + + ARM_COMPUTE_EXPECT(src.info()->is_resizable(), framework::LogLevel::ERRORS); + ARM_COMPUTE_EXPECT(dst.info()->is_resizable(), framework::LogLevel::ERRORS); + + // Create and configure function + NEQuantizationLayer quant_layer; + quant_layer.configure(&src, &dst); + + // Validate valid region + const ValidRegion valid_region = shape_to_valid_region(shape); + validate(src.info()->valid_region(), valid_region); + validate(dst.info()->valid_region(), valid_region); + + // Validate padding + const PaddingSize padding = PaddingCalculator(shape.x(), 8).required_padding(); + validate(src.info()->padding(), padding); + validate(dst.info()->padding(), padding); +} + +template +using NEQuantizationLayerFixture = QuantizationValidationFixture; + +TEST_SUITE(Float) +TEST_SUITE(FP32) +FIXTURE_DATA_TEST_CASE(RunSmall, NEQuantizationLayerFixture, framework::DatasetMode::PRECOMMIT, combine(datasets::Small2DShapes(), framework::dataset::make("DataType", DataType::F32))) +{ + // Validate output + validate(Accessor(_target), _reference, tolerance_u8); +} +FIXTURE_DATA_TEST_CASE(RunLarge, NEQuantizationLayerFixture, framework::DatasetMode::NIGHTLY, combine(datasets::Large2DShapes(), framework::dataset::make("DataType", DataType::F32))) +{ + // Validate output + validate(Accessor(_target), _reference, tolerance_u8); +} +TEST_SUITE_END() +TEST_SUITE_END() + +TEST_SUITE_END() +TEST_SUITE_END() +} // namespace validation +} // namespace test +} // namespace arm_compute diff --git a/tests/validation_new/fixtures/QuantizationLayerFixture.h b/tests/validation_new/fixtures/QuantizationLayerFixture.h new file mode 100644 index 0000000000..469fd5f231 --- /dev/null +++ b/tests/validation_new/fixtures/QuantizationLayerFixture.h @@ -0,0 +1,120 @@ +/* + * Copyright (c) 2017 ARM Limited. + * + * SPDX-License-Identifier: MIT + * + * Permission is hereby granted, free of charge, to any person obtaining a copy + * of this software and associated documentation files (the "Software"), to + * deal in the Software without restriction, including without limitation the + * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or + * sell copies of the Software, and to permit persons to whom the Software is + * furnished to do so, subject to the following conditions: + * + * The above copyright notice and this permission notice shall be included in all + * copies or substantial portions of the Software. + * + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE + * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE + * SOFTWARE. + */ +#ifndef ARM_COMPUTE_TEST_QUANTIZATION_LAYER_FIXTURE +#define ARM_COMPUTE_TEST_QUANTIZATION_LAYER_FIXTURE + +#include "arm_compute/core/TensorShape.h" +#include "arm_compute/core/Types.h" +#include "arm_compute/runtime/Tensor.h" +#include "framework/Asserts.h" +#include "framework/Fixture.h" +#include "tests/AssetsLibrary.h" +#include "tests/Globals.h" +#include "tests/IAccessor.h" +#include "tests/validation_new/CPP/QuantizationLayer.h" + +#include + +namespace arm_compute +{ +namespace test +{ +namespace validation +{ +template +class QuantizationValidationFixedPointFixture : public framework::Fixture +{ +public: + template + void setup(TensorShape shape, DataType data_type) + { + _target = compute_target(shape, data_type); + _reference = compute_reference(shape, data_type); + } + +protected: + template + void fill(U &&tensor) + { + library->fill_tensor_uniform(tensor, 0); + } + + TensorType compute_target(const TensorShape &shape, DataType data_type) + { + // Create tensors + TensorType src = create_tensor(shape, data_type); + TensorType dst = create_tensor(shape, DataType::U8); + + // Create and configure function + FunctionType quantization_layer; + quantization_layer.configure(&src, &dst); + + ARM_COMPUTE_EXPECT(src.info()->is_resizable(), framework::LogLevel::ERRORS); + ARM_COMPUTE_EXPECT(dst.info()->is_resizable(), framework::LogLevel::ERRORS); + + // Allocate tensors + src.allocator()->allocate(); + dst.allocator()->allocate(); + + ARM_COMPUTE_EXPECT(!src.info()->is_resizable(), framework::LogLevel::ERRORS); + ARM_COMPUTE_EXPECT(!dst.info()->is_resizable(), framework::LogLevel::ERRORS); + + // Fill tensors + fill(AccessorType(src)); + + // Compute function + quantization_layer.run(); + + return dst; + } + + SimpleTensor compute_reference(const TensorShape &shape, DataType data_type) + { + // Create reference + SimpleTensor src{ shape, data_type }; + + // Fill reference + fill(src); + + return reference::quantization_layer(src); + } + + TensorType _target{}; + SimpleTensor _reference{}; +}; + +template +class QuantizationValidationFixture : public QuantizationValidationFixedPointFixture +{ +public: + template + void setup(TensorShape shape, DataType data_type) + { + QuantizationValidationFixedPointFixture::setup(shape, data_type); + } +}; +} // namespace validation +} // namespace test +} // namespace arm_compute +#endif /* ARM_COMPUTE_TEST_QUANTIZATION_LAYER_FIXTURE */ -- cgit v1.2.1