diff options
author | Michele Di Giorgio <michele.digiorgio@arm.com> | 2020-10-27 10:56:31 +0000 |
---|---|---|
committer | Michele Di Giorgio <michele.digiorgio@arm.com> | 2021-01-20 16:39:53 +0000 |
commit | d556d7bafe6ad943f4aca0f5285ada7b8ce497f7 (patch) | |
tree | 11c7077daf97b46c47a4eac821830b37a7ce9e76 /src/runtime/NEON/functions/NEPoolingAssemblyDispatch.h | |
parent | 7d61ff041826782d14e67b7f5b7a2864905ff38b (diff) | |
download | ComputeLibrary-d556d7bafe6ad943f4aca0f5285ada7b8ce497f7.tar.gz |
Integrate improved pooling layer on NEON
Resolves COMPMID-4035
Change-Id: I559f8c4208fba9193dfe5012f03ddaf26c746215
Signed-off-by: Michele Di Giorgio <michele.digiorgio@arm.com>
Reviewed-on: https://review.mlplatform.org/c/ml/ComputeLibrary/+/4855
Tested-by: Arm Jenkins <bsgcomp@arm.com>
Reviewed-by: Georgios Pinitas <georgios.pinitas@arm.com>
Comments-Addressed: Arm Jenkins <bsgcomp@arm.com>
Diffstat (limited to 'src/runtime/NEON/functions/NEPoolingAssemblyDispatch.h')
-rw-r--r-- | src/runtime/NEON/functions/NEPoolingAssemblyDispatch.h | 149 |
1 files changed, 149 insertions, 0 deletions
diff --git a/src/runtime/NEON/functions/NEPoolingAssemblyDispatch.h b/src/runtime/NEON/functions/NEPoolingAssemblyDispatch.h new file mode 100644 index 0000000000..f6d232b931 --- /dev/null +++ b/src/runtime/NEON/functions/NEPoolingAssemblyDispatch.h @@ -0,0 +1,149 @@ +/* + * Copyright (c) 2021 Arm Limited. + * + * SPDX-License-Identifier: MIT + * + * Permission is hereby granted, free of charge, to any person obtaining a copy + * of this software and associated documentation files (the "Software"), to + * deal in the Software without restriction, including without limitation the + * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or + * sell copies of the Software, and to permit persons to whom the Software is + * furnished to do so, subject to the following conditions: + * + * The above copyright notice and this permission notice shall be included in all + * copies or substantial portions of the Software. + * + * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR + * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, + * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE + * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER + * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, + * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE + * SOFTWARE. + */ +#ifndef ARM_COMPUTE_NEPOOLINGASSEMBLYDISPATCH_H +#define ARM_COMPUTE_NEPOOLINGASSEMBLYDISPATCH_H + +#include "arm_compute/runtime/IFunction.h" +#include "arm_compute/runtime/IMemoryManager.h" +#include "arm_compute/runtime/MemoryGroup.h" +#include "arm_compute/runtime/NEON/INEOperator.h" +#include "arm_compute/runtime/Tensor.h" +#include "src/core/NEON/INEKernel.h" + +namespace arm_compute +{ +// Forward Declarations +class ITensor; +struct PoolingLayerInfo; + +/** Assembly kernel glue */ +class NEPoolingAssemblyDispatch : public IFunction +{ +public: + /** Constructor */ + NEPoolingAssemblyDispatch(std::shared_ptr<IMemoryManager> memory_manager = nullptr); + /** Prevent instances of this class from being copied (As this class contains pointers) */ + NEPoolingAssemblyDispatch(const NEPoolingAssemblyDispatch &) = delete; + /** Default move constructor */ + NEPoolingAssemblyDispatch(NEPoolingAssemblyDispatch &&); + /** Prevent instances of this class from being copied (As this class contains pointers) */ + NEPoolingAssemblyDispatch &operator=(const NEPoolingAssemblyDispatch &) = delete; + /** Default move assignment operator */ + NEPoolingAssemblyDispatch &operator=(NEPoolingAssemblyDispatch &&); + /** Destructor */ + ~NEPoolingAssemblyDispatch(); + + /** If supported create an assembly routine, else fallback to Compute Library function. + * + * @param[in] input Input tensor. Data types supported: QASYMM8/QASYMM8_SIGNED/F16/F32. + * @param[out] output Output tensor to store the result of pooling. Data types supported: same as @p input. + * @param[in] info Pooling meta-data + */ + void configure(const ITensor *input, ITensor *output, const PoolingLayerInfo &info); + + /** Indicates whether or not this function can be used to process the given parameters. + * + * @param[in] input Input tensor info. Data types supported: QASYMM8/QASYMM8_SIGNED/F16/F32. + * @param[in] output Output tensor to store the result of pooling. Data types supported: same as @p input. + * @param[in] info Pooling meta-data + * + * @return a status. + */ + static Status validate(const ITensorInfo *input, const ITensorInfo *output, const PoolingLayerInfo &info); + + /** Was the function successfully configured ? + * + * @return True if the function is configured and ready to run + */ + bool is_configured() const; + + // Inherited methods overridden: + void run() override; + +private: + /** Helper function to allocate memory for the workspace needed by the + * assembly kernels + * + * @param[in] workspace_size Total size of the workspace. + * @param[in] alignment Alignment requirement in bytes. + */ + void allocate_workspace(size_t workspace_size, size_t alignment); + + struct Impl; + std::unique_ptr<Impl> _impl; + + MemoryGroup _memory_group{}; + Tensor _workspace{}; +}; + +namespace experimental +{ +/** Basic function to run pooling assembly kernels */ +class NEPoolingAssemblyDispatch : public INEOperator +{ +public: + /** Constructor */ + NEPoolingAssemblyDispatch() = default; + /** Prevent instances of this class from being copied */ + NEPoolingAssemblyDispatch(const NEPoolingAssemblyDispatch &) = delete; + /** Default move constructor */ + NEPoolingAssemblyDispatch(NEPoolingAssemblyDispatch &&) = default; + /** Prevent instances of this class from being copied */ + NEPoolingAssemblyDispatch &operator=(const NEPoolingAssemblyDispatch &) = delete; + /** Default move assignment operator */ + NEPoolingAssemblyDispatch &operator=(NEPoolingAssemblyDispatch &&) = default; + /** Destructor */ + ~NEPoolingAssemblyDispatch(); + + /** If supported create an assembly routine, else fallback to Compute Library function. + * + * @param[in] input Input tensor. Data types supported: QASYMM8/QASYMM8_SIGNED/F16/F32. + * @param[out] output Output tensor to store the result of pooling. Data types supported: same as @p input. + * @param[in] info Pooling meta-data + */ + void configure(const ITensorInfo *input, ITensorInfo *output, const PoolingLayerInfo &info); + + /** Indicates whether or not this function can be used to process the given parameters. + * + * @param[in] input Input tensor info. Data types supported: QASYMM8/QASYMM8_SIGNED/F16/F32. + * @param[in] output Output tensor to store the result of pooling. Data types supported: same as @p input. + * @param[in] info Pooling meta-data + * + * @return a status. + */ + static Status validate(const ITensorInfo *input, const ITensorInfo *output, const PoolingLayerInfo &info); + /** Was the function successfully configured ? + * + * @return True if the function is configured and ready to run + */ + bool is_configured() const; + // Run method overriden + void run(ITensorPack &tensors) override; + +private: + bool _is_global_pooling_layer{ false }; +}; +} // namespace experimental +} // namespace arm_compute +#endif /* ARM_COMPUTE_NEPOOLINGASSEMBLYDISPATCH_H */ |