/* * Copyright (c) 2021 Arm Limited. * * SPDX-License-Identifier: MIT * * Permission is hereby granted, free of charge, to any person obtaining a copy * of this software and associated documentation files (the "Software"), to * deal in the Software without restriction, including without limitation the * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or * sell copies of the Software, and to permit persons to whom the Software is * furnished to do so, subject to the following conditions: * * The above copyright notice and this permission notice shall be included in all * copies or substantial portions of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE * SOFTWARE. */ #ifndef ARM_COMPUTE_NEPOOLINGASSEMBLYDISPATCH_H #define ARM_COMPUTE_NEPOOLINGASSEMBLYDISPATCH_H #include "arm_compute/runtime/IFunction.h" #include "arm_compute/runtime/IMemoryManager.h" #include "arm_compute/runtime/MemoryGroup.h" #include "arm_compute/runtime/NEON/INEOperator.h" #include "arm_compute/runtime/Tensor.h" #include "src/core/NEON/INEKernel.h" namespace arm_compute { // Forward Declarations class ITensor; struct PoolingLayerInfo; /** Assembly kernel glue */ class NEPoolingAssemblyDispatch : public IFunction { public: /** Constructor */ NEPoolingAssemblyDispatch(std::shared_ptr memory_manager = nullptr); /** Prevent instances of this class from being copied (As this class contains pointers) */ NEPoolingAssemblyDispatch(const NEPoolingAssemblyDispatch &) = delete; /** Default move constructor */ NEPoolingAssemblyDispatch(NEPoolingAssemblyDispatch &&); /** Prevent instances of this class from being copied (As this class contains pointers) */ NEPoolingAssemblyDispatch &operator=(const NEPoolingAssemblyDispatch &) = delete; /** Default move assignment operator */ NEPoolingAssemblyDispatch &operator=(NEPoolingAssemblyDispatch &&); /** Destructor */ ~NEPoolingAssemblyDispatch(); /** If supported create an assembly routine, else fallback to Compute Library function. * * @param[in] input Input tensor. Data types supported: QASYMM8/QASYMM8_SIGNED/F16/F32. * @param[out] output Output tensor to store the result of pooling. Data types supported: same as @p input. * @param[in] info Pooling meta-data */ void configure(const ITensor *input, ITensor *output, const PoolingLayerInfo &info); /** Indicates whether or not this function can be used to process the given parameters. * * @param[in] input Input tensor info. Data types supported: QASYMM8/QASYMM8_SIGNED/F16/F32. * @param[in] output Output tensor to store the result of pooling. Data types supported: same as @p input. * @param[in] info Pooling meta-data * * @return a status. */ static Status validate(const ITensorInfo *input, const ITensorInfo *output, const PoolingLayerInfo &info); /** Was the function successfully configured ? * * @return True if the function is configured and ready to run */ bool is_configured() const; // Inherited methods overridden: void run() override; private: /** Helper function to allocate memory for the workspace needed by the * assembly kernels * * @param[in] workspace_size Total size of the workspace. * @param[in] alignment Alignment requirement in bytes. */ void allocate_workspace(size_t workspace_size, size_t alignment); struct Impl; std::unique_ptr _impl; MemoryGroup _memory_group{}; Tensor _workspace{}; }; namespace experimental { /** Basic function to run pooling assembly kernels */ class NEPoolingAssemblyDispatch : public INEOperator { public: /** Constructor */ NEPoolingAssemblyDispatch() = default; /** Prevent instances of this class from being copied */ NEPoolingAssemblyDispatch(const NEPoolingAssemblyDispatch &) = delete; /** Default move constructor */ NEPoolingAssemblyDispatch(NEPoolingAssemblyDispatch &&) = default; /** Prevent instances of this class from being copied */ NEPoolingAssemblyDispatch &operator=(const NEPoolingAssemblyDispatch &) = delete; /** Default move assignment operator */ NEPoolingAssemblyDispatch &operator=(NEPoolingAssemblyDispatch &&) = default; /** Destructor */ ~NEPoolingAssemblyDispatch(); /** If supported create an assembly routine, else fallback to Compute Library function. * * @param[in] input Input tensor. Data types supported: QASYMM8/QASYMM8_SIGNED/F16/F32. * @param[out] output Output tensor to store the result of pooling. Data types supported: same as @p input. * @param[in] info Pooling meta-data */ void configure(const ITensorInfo *input, ITensorInfo *output, const PoolingLayerInfo &info); /** Indicates whether or not this function can be used to process the given parameters. * * @param[in] input Input tensor info. Data types supported: QASYMM8/QASYMM8_SIGNED/F16/F32. * @param[in] output Output tensor to store the result of pooling. Data types supported: same as @p input. * @param[in] info Pooling meta-data * * @return a status. */ static Status validate(const ITensorInfo *input, const ITensorInfo *output, const PoolingLayerInfo &info); /** Was the function successfully configured ? * * @return True if the function is configured and ready to run */ bool is_configured() const; // Run method overriden void run(ITensorPack &tensors) override; private: bool _is_global_pooling_layer{ false }; }; } // namespace experimental } // namespace arm_compute #endif /* ARM_COMPUTE_NEPOOLINGASSEMBLYDISPATCH_H */