10 #include <arm_compute/runtime/IFunction.h> 11 #include <arm_compute/runtime/Tensor.h> 12 #include <arm_compute/runtime/MemoryManagerOnDemand.h> 20 const TensorInfo& output,
21 const Convolution2dDescriptor& descriptor,
22 const TensorInfo& weights,
23 const Optional<TensorInfo>& biases,
24 bool isFastMathEnabled =
false,
25 const ActivationDescriptor* activationDescriptor =
nullptr);
34 std::shared_ptr<arm_compute::MemoryManagerOnDemand>& memoryManager,
35 const bool isFastMathENabled =
false);
42 std::unique_ptr<arm_compute::IFunction> m_ConvolutionLayer;
44 std::unique_ptr<arm_compute::Tensor> m_KernelTensor;
45 std::unique_ptr<arm_compute::Tensor> m_BiasTensor;
47 arm_compute::ConvolutionMethod m_ConvolutionMethod;
49 void FreeUnusedTensors();
void Execute() const override
NeonConvolution2dWorkload(const Convolution2dQueueDescriptor &descriptor, const WorkloadInfo &info, std::shared_ptr< arm_compute::MemoryManagerOnDemand > &memoryManager, const bool isFastMathENabled=false)
Copyright (c) 2021 ARM Limited and Contributors.
arm_compute::Status NeonConvolution2dWorkloadValidate(const TensorInfo &input, const TensorInfo &output, const Convolution2dDescriptor &descriptor, const TensorInfo &weights, const Optional< TensorInfo > &biases, bool isFastMathEnabled, const ActivationDescriptor *activationDescriptor)
arm_compute::ConvolutionMethod GetConvolutionMethod() const
Contains information about TensorInfos of a layer.