diff options
Diffstat (limited to 'src/runtime/NEON/functions/NEPoolingAssemblyDispatch.h')
-rw-r--r-- | src/runtime/NEON/functions/NEPoolingAssemblyDispatch.h | 149 |
1 files changed, 0 insertions, 149 deletions
diff --git a/src/runtime/NEON/functions/NEPoolingAssemblyDispatch.h b/src/runtime/NEON/functions/NEPoolingAssemblyDispatch.h deleted file mode 100644 index f6d232b931..0000000000 --- a/src/runtime/NEON/functions/NEPoolingAssemblyDispatch.h +++ /dev/null @@ -1,149 +0,0 @@ -/* - * Copyright (c) 2021 Arm Limited. - * - * SPDX-License-Identifier: MIT - * - * Permission is hereby granted, free of charge, to any person obtaining a copy - * of this software and associated documentation files (the "Software"), to - * deal in the Software without restriction, including without limitation the - * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or - * sell copies of the Software, and to permit persons to whom the Software is - * furnished to do so, subject to the following conditions: - * - * The above copyright notice and this permission notice shall be included in all - * copies or substantial portions of the Software. - * - * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR - * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, - * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE - * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER - * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, - * OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE - * SOFTWARE. - */ -#ifndef ARM_COMPUTE_NEPOOLINGASSEMBLYDISPATCH_H -#define ARM_COMPUTE_NEPOOLINGASSEMBLYDISPATCH_H - -#include "arm_compute/runtime/IFunction.h" -#include "arm_compute/runtime/IMemoryManager.h" -#include "arm_compute/runtime/MemoryGroup.h" -#include "arm_compute/runtime/NEON/INEOperator.h" -#include "arm_compute/runtime/Tensor.h" -#include "src/core/NEON/INEKernel.h" - -namespace arm_compute -{ -// Forward Declarations -class ITensor; -struct PoolingLayerInfo; - -/** Assembly kernel glue */ -class NEPoolingAssemblyDispatch : public IFunction -{ -public: - /** Constructor */ - NEPoolingAssemblyDispatch(std::shared_ptr<IMemoryManager> memory_manager = nullptr); - /** Prevent instances of this class from being copied (As this class contains pointers) */ - NEPoolingAssemblyDispatch(const NEPoolingAssemblyDispatch &) = delete; - /** Default move constructor */ - NEPoolingAssemblyDispatch(NEPoolingAssemblyDispatch &&); - /** Prevent instances of this class from being copied (As this class contains pointers) */ - NEPoolingAssemblyDispatch &operator=(const NEPoolingAssemblyDispatch &) = delete; - /** Default move assignment operator */ - NEPoolingAssemblyDispatch &operator=(NEPoolingAssemblyDispatch &&); - /** Destructor */ - ~NEPoolingAssemblyDispatch(); - - /** If supported create an assembly routine, else fallback to Compute Library function. - * - * @param[in] input Input tensor. Data types supported: QASYMM8/QASYMM8_SIGNED/F16/F32. - * @param[out] output Output tensor to store the result of pooling. Data types supported: same as @p input. - * @param[in] info Pooling meta-data - */ - void configure(const ITensor *input, ITensor *output, const PoolingLayerInfo &info); - - /** Indicates whether or not this function can be used to process the given parameters. - * - * @param[in] input Input tensor info. Data types supported: QASYMM8/QASYMM8_SIGNED/F16/F32. - * @param[in] output Output tensor to store the result of pooling. Data types supported: same as @p input. - * @param[in] info Pooling meta-data - * - * @return a status. - */ - static Status validate(const ITensorInfo *input, const ITensorInfo *output, const PoolingLayerInfo &info); - - /** Was the function successfully configured ? - * - * @return True if the function is configured and ready to run - */ - bool is_configured() const; - - // Inherited methods overridden: - void run() override; - -private: - /** Helper function to allocate memory for the workspace needed by the - * assembly kernels - * - * @param[in] workspace_size Total size of the workspace. - * @param[in] alignment Alignment requirement in bytes. - */ - void allocate_workspace(size_t workspace_size, size_t alignment); - - struct Impl; - std::unique_ptr<Impl> _impl; - - MemoryGroup _memory_group{}; - Tensor _workspace{}; -}; - -namespace experimental -{ -/** Basic function to run pooling assembly kernels */ -class NEPoolingAssemblyDispatch : public INEOperator -{ -public: - /** Constructor */ - NEPoolingAssemblyDispatch() = default; - /** Prevent instances of this class from being copied */ - NEPoolingAssemblyDispatch(const NEPoolingAssemblyDispatch &) = delete; - /** Default move constructor */ - NEPoolingAssemblyDispatch(NEPoolingAssemblyDispatch &&) = default; - /** Prevent instances of this class from being copied */ - NEPoolingAssemblyDispatch &operator=(const NEPoolingAssemblyDispatch &) = delete; - /** Default move assignment operator */ - NEPoolingAssemblyDispatch &operator=(NEPoolingAssemblyDispatch &&) = default; - /** Destructor */ - ~NEPoolingAssemblyDispatch(); - - /** If supported create an assembly routine, else fallback to Compute Library function. - * - * @param[in] input Input tensor. Data types supported: QASYMM8/QASYMM8_SIGNED/F16/F32. - * @param[out] output Output tensor to store the result of pooling. Data types supported: same as @p input. - * @param[in] info Pooling meta-data - */ - void configure(const ITensorInfo *input, ITensorInfo *output, const PoolingLayerInfo &info); - - /** Indicates whether or not this function can be used to process the given parameters. - * - * @param[in] input Input tensor info. Data types supported: QASYMM8/QASYMM8_SIGNED/F16/F32. - * @param[in] output Output tensor to store the result of pooling. Data types supported: same as @p input. - * @param[in] info Pooling meta-data - * - * @return a status. - */ - static Status validate(const ITensorInfo *input, const ITensorInfo *output, const PoolingLayerInfo &info); - /** Was the function successfully configured ? - * - * @return True if the function is configured and ready to run - */ - bool is_configured() const; - // Run method overriden - void run(ITensorPack &tensors) override; - -private: - bool _is_global_pooling_layer{ false }; -}; -} // namespace experimental -} // namespace arm_compute -#endif /* ARM_COMPUTE_NEPOOLINGASSEMBLYDISPATCH_H */ |