diff options
Diffstat (limited to 'arm_compute/runtime/NEON/functions/NEConcatenateLayer.h')
-rw-r--r-- | arm_compute/runtime/NEON/functions/NEConcatenateLayer.h | 60 |
1 files changed, 32 insertions, 28 deletions
diff --git a/arm_compute/runtime/NEON/functions/NEConcatenateLayer.h b/arm_compute/runtime/NEON/functions/NEConcatenateLayer.h index 8207589680..1600f85488 100644 --- a/arm_compute/runtime/NEON/functions/NEConcatenateLayer.h +++ b/arm_compute/runtime/NEON/functions/NEConcatenateLayer.h @@ -1,5 +1,5 @@ /* - * Copyright (c) 2018-2019 ARM Limited. + * Copyright (c) 2018-2021 Arm Limited. * * SPDX-License-Identifier: MIT * @@ -24,14 +24,10 @@ #ifndef ARM_COMPUTE_NECONCATENATELAYER_H #define ARM_COMPUTE_NECONCATENATELAYER_H -#include "arm_compute/runtime/IFunction.h" - -#include "arm_compute/core/NEON/INEKernel.h" #include "arm_compute/core/Types.h" -#include "arm_compute/core/utils/misc/Requires.h" +#include "arm_compute/runtime/IFunction.h" #include <memory> -#include <vector> namespace arm_compute { @@ -40,33 +36,49 @@ class ITensor; class ITensorInfo; class Status; -/** Basic function to execute concatenate tensors along a given axis. This function calls the following kernels: - * - * -# @ref NEWidthConcatenateLayerKernel (if underlying concatenation axis is 0). - * -# @ref NEHeightConcatenateLayerKernel (if underlying concatenation axis is 1). - * -# @ref NEDepthConcatenateLayerKernel (if underlying concatenation axis is 2). - * -# @ref NEBatchConcatenateLayerKernel (if underlying concatenation axis is 3). - */ +/** Basic function to execute concatenate tensors along a given axis */ class NEConcatenateLayer : public IFunction { public: /** Default constructor */ NEConcatenateLayer(); + /** Destructor */ + ~NEConcatenateLayer(); + /** Prevent instances of this class from being copied (As this class contains pointers) */ + NEConcatenateLayer(const NEConcatenateLayer &) = delete; + /** Default move constructor */ + NEConcatenateLayer(NEConcatenateLayer &&); + /** Prevent instances of this class from being copied (As this class contains pointers) */ + NEConcatenateLayer &operator=(const NEConcatenateLayer &) = delete; + /** Default move assignment operator */ + NEConcatenateLayer &operator=(NEConcatenateLayer &&); /** Initialise the kernel's inputs vector and output. * + * Valid data layouts: + * - All + * + * Valid data type configurations: + * |src |dst | + * |:--------------|:--------------| + * |QASYMM8 |QASYMM8 | + * |QASYMM8_SIGNED |QASYMM8_SIGNED | + * |F16 |F16 | + * |F32 |F32 | + * * @note Input and output tensor dimensions preconditions defer depending on the concatenation axis. - * @note Preconditions can be found respectively at @ref NEWidthConcatenateLayerKernel, @ref NEHeightConcatenateLayerKernel and @ref NEDepthConcatenateLayerKernel. + * @note Preconditions can be found respectively at @ref cpu::kernels::CpuConcatenateWidthKernel, @ref cpu::kernels::CpuConcatenateHeightKernel, + * @ref cpu::kernels::CpuConcatenateDepthKernel and @ref cpu::kernels::CpuConcatenateBatchKernel. * * @param[in,out] inputs_vector The vectors containing all the tensors to concatenate. Data types supported: QASYMM8/QASYMM8_SIGNED/F16/F32. * @param[out] output Output tensor. Data types supported: Same as @p input. * @param[in] axis Concatenation axis. Supported underlying concatenation axis are 0, 1, 2 and 3. */ - void configure(std::vector<ITensor *> inputs_vector, ITensor *output, size_t axis); void configure(std::vector<const ITensor *> inputs_vector, ITensor *output, size_t axis); /** Static function to check if given info will lead to a valid configuration of @ref NEConcatenateLayer * * @note Input and output tensor dimensions preconditions defer depending on the concatenation axis. - * @note Preconditions can be found respectively at @ref NEWidthConcatenateLayerKernel, @ref NEHeightConcatenateLayerKernel and @ref NEDepthConcatenateLayerKernel. + * @note Preconditions can be found respectively at @ref cpu::kernels::CpuConcatenateWidthKernel, @ref cpu::kernels::CpuConcatenateHeightKernel, + * @ref cpu::kernels::CpuConcatenateDepthKernel and @ref cpu::kernels::CpuConcatenateBatchKernel. * * @param[in] inputs_vector The vectors containing all the tensors info to concatenate. Data types supported: QASYMM8/QASYMM8_SIGNED/F16/F32. * @param[in] output Output tensor info. Data types supported: Same as @p input. @@ -74,23 +86,15 @@ public: * * @return a status */ - static Status validate(const std::vector<ITensorInfo *> &inputs_vector, const ITensorInfo *output, size_t axis); - static Status validate(const std::vector<const ITensorInfo *> &inputs_vector, const ITensorInfo *output, size_t axis); + static Status + validate(const std::vector<const ITensorInfo *> &inputs_vector, const ITensorInfo *output, size_t axis); // Inherited methods overridden: void run() override; private: - template <typename TensorType, REQUIRES_TA(std::is_same<typename std::remove_cv<TensorType>::type, ITensor>::value)> - void configure_internal(std::vector<TensorType *> &&inputs_vector, ITensor *output, size_t axis); - - template <typename TensorInfoType, REQUIRES_TA(std::is_same<typename std::remove_cv<TensorInfoType>::type, ITensorInfo>::value)> - static Status validate_internal(const std::vector<TensorInfoType *> &inputs_vector, const ITensorInfo *output, size_t axis); - -private: - std::vector<std::unique_ptr<INEKernel>> _concat_kernels; - unsigned int _num_inputs; - unsigned int _axis; + struct Impl; + std::unique_ptr<Impl> _impl; }; } // namespace arm_compute #endif /* ARM_COMPUTE_NECONCATENATELAYER_H */ |