/* * Copyright (c) 2017-2019 ARM Limited. * * SPDX-License-Identifier: MIT * * Permission is hereby granted, free of charge, to any person obtaining a copy * of this software and associated documentation files (the "Software"), to * deal in the Software without restriction, including without limitation the * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or * sell copies of the Software, and to permit persons to whom the Software is * furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice shall be included in all * copies or substantial portions of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE * SOFTWARE. */ #include "ReductionOperation.h" #include "tests/validation/Helpers.h" #include #include namespace arm_compute { namespace test { namespace validation { namespace reference { namespace { template OT reduce_operation(const T *ptr, int reduce_elements, ReductionOperation op, int stride) { using type = typename std::remove_cv::type; auto res = (op == ReductionOperation::PROD) ? type(1) : type(0); if(std::is_integral::value) { auto int_res = static_cast(res); for(int i = 0; i < reduce_elements; ++i) { auto elem = *(ptr + stride * i); switch(op) { case ReductionOperation::ARG_IDX_MIN: if(*(ptr + stride * static_cast(int_res)) > elem) { int_res = static_cast(i); } break; case ReductionOperation::ARG_IDX_MAX: if(*(ptr + stride * static_cast(int_res)) < elem) { int_res = static_cast(i); } break; case ReductionOperation::SUM_SQUARE: int_res += elem * elem; break; case ReductionOperation::MEAN_SUM: case ReductionOperation::SUM: int_res += elem; break; case ReductionOperation::PROD: int_res *= elem; break; default: ARM_COMPUTE_ERROR("Operation not supported"); } } if(op == ReductionOperation::MEAN_SUM && reduce_elements > 0) { int_res /= reduce_elements; } res = saturate_cast(int_res); } else { for(int i = 0; i < reduce_elements; ++i) { auto elem = *(ptr + stride * i); switch(op) { case ReductionOperation::ARG_IDX_MIN: if(*(ptr + stride * static_cast(res)) > elem) { res = static_cast(i); } break; case ReductionOperation::ARG_IDX_MAX: if(*(ptr + stride * static_cast(res)) < elem) { res = static_cast(i); } break; case ReductionOperation::SUM_SQUARE: res += elem * elem; break; case ReductionOperation::MEAN_SUM: case ReductionOperation::SUM: res += elem; break; case ReductionOperation::PROD: res *= elem; break; default: ARM_COMPUTE_ERROR("Operation not supported"); } } if(op == ReductionOperation::MEAN_SUM && reduce_elements > 0) { res /= reduce_elements; } } return res; } } // namespace template SimpleTensor compute_reduction_operation(const SimpleTensor &src, const TensorShape &dst_shape, unsigned int axis, ReductionOperation op) { // Create reference const bool is_arg_min_max = (op == ReductionOperation::ARG_IDX_MIN || op == ReductionOperation::ARG_IDX_MAX); DataType output_data_type = is_arg_min_max ? DataType::U32 : src.data_type(); SimpleTensor dst{ dst_shape, output_data_type, 1, src.quantization_info() }; const unsigned int src_width = src.shape().x(); const unsigned int src_height = src.shape().y(); const unsigned int src_depth = src.shape().z(); const unsigned int src_batch = src.shape()[3]; const int reduce_elems = src.shape()[axis]; switch(axis) { case 0: { const unsigned int upper_dims = src.shape().total_size_upper(1); for(unsigned int du = 0; du < upper_dims; ++du) { const T *src_row_ptr = src.data() + du * reduce_elems; dst[du] = reduce_operation(src_row_ptr, reduce_elems, op, 1); } } break; case 1: { const unsigned int upper_dims = src.shape().total_size_upper(2); for(unsigned int du = 0; du < upper_dims; ++du) { for(unsigned int x = 0; x < src_width; ++x) { const int in_offset = du * src_height * src_width + x; const int out_offset = du * src_width + x; const T *src_row_ptr = src.data() + in_offset; dst[out_offset] = reduce_operation(src_row_ptr, reduce_elems, op, src_width); } } } break; case 2: { const unsigned int upper_dims = src.shape().total_size_upper(3); for(unsigned int du = 0; du < upper_dims; ++du) { for(unsigned int x = 0; x < src_width; ++x) { for(unsigned int y = 0; y < src_height; ++y) { const int in_offset = du * src_depth * src_height * src_width + y * src_width + x; const int out_offset = du * src_width * src_height + y * src_width + x; const T *src_row_ptr = src.data() + in_offset; dst[out_offset] = reduce_operation(src_row_ptr, reduce_elems, op, src_height * src_width); } } } } break; case 3: { const unsigned int upper_dims = src.shape().total_size_upper(4); for(unsigned int du = 0; du < upper_dims; ++du) { for(unsigned int z = 0; z < src_depth; ++z) { for(unsigned int y = 0; y < src_height; ++y) { for(unsigned int x = 0; x < src_width; ++x) { const int in_offset = du * src_batch * src_depth * src_height * src_width + z * src_width * src_height + y * src_width + x; const int out_offset = du * src_depth * src_height * src_width + z * src_width * src_height + y * src_width + x; const T *src_row_ptr = src.data() + in_offset; dst[out_offset] = reduce_operation(src_row_ptr, reduce_elems, op, src_width * src_height * src_depth); } } } } } break; default: ARM_COMPUTE_ERROR("Unsupported reduction axis"); } return dst; } template SimpleTensor reduction_operation(const SimpleTensor &src, const TensorShape &dst_shape, unsigned int axis, ReductionOperation op) { return compute_reduction_operation(src, dst_shape, axis, op); } template <> SimpleTensor reduction_operation(const SimpleTensor &src, const TensorShape &dst_shape, unsigned int axis, ReductionOperation op) { if(src.data_type() == DataType::QASYMM8 && op != ReductionOperation::MEAN_SUM) { SimpleTensor src_f = convert_from_asymmetric(src); SimpleTensor dst_f = reference::reduction_operation(src_f, dst_shape, axis, op); return convert_to_asymmetric(dst_f, src.quantization_info()); } else { return compute_reduction_operation(src, dst_shape, axis, op); } } template SimpleTensor reduction_operation(const SimpleTensor &src, const TensorShape &dst_shape, unsigned int axis, ReductionOperation op); template SimpleTensor reduction_operation(const SimpleTensor &src, const TensorShape &dst_shape, unsigned int axis, ReductionOperation op); template SimpleTensor reduction_operation(const SimpleTensor &src, const TensorShape &dst_shape, unsigned int axis, ReductionOperation op); template SimpleTensor reduction_operation(const SimpleTensor &src, const TensorShape &dst_shape, unsigned int axis, ReductionOperation op); template SimpleTensor reduction_operation(const SimpleTensor &src, const TensorShape &dst_shape, unsigned int axis, ReductionOperation op); } // namespace reference } // namespace validation } // namespace test } // namespace arm_compute