20 template <
typename Functor,
typename ParentDescriptor,
typename armnn::StringMapping::Id DebugString>
22 const ParentDescriptor& desc,
28 template <
typename Functor,
typename ParentDescriptor,
typename armnn::StringMapping::Id DebugString>
34 template <
typename Functor,
typename ParentDescriptor,
typename armnn::StringMapping::Id DebugString>
42 template <
typename Functor,
typename ParentDescriptor,
typename armnn::StringMapping::Id DebugString>
44 std::vector<ITensorHandle*> inputs, std::vector<ITensorHandle*> outputs)
const 55 std::unique_ptr<Decoder<InType>> input0 = MakeDecoder<InType>(inputInfo0, inputs[0]->Map());
56 std::unique_ptr<Decoder<InType>> input1 = MakeDecoder<InType>(inputInfo1, inputs[1]->Map());
57 std::unique_ptr<Encoder<OutType>> output= MakeEncoder<OutType>(outputInfo, outputs[0]->Map());
82 armnn::SubtractionQueueDescriptor,
83 armnn::StringMapping::RefSubtractionWorkload_Execute>;
90 armnn::MultiplicationQueueDescriptor,
91 armnn::StringMapping::RefMultiplicationWorkload_Execute>;
98 armnn::DivisionQueueDescriptor,
99 armnn::StringMapping::RefDivisionWorkload_Execute>;
106 armnn::MaximumQueueDescriptor,
107 armnn::StringMapping::RefMaximumWorkload_Execute>;
114 armnn::MinimumQueueDescriptor,
115 armnn::StringMapping::RefMinimumWorkload_Execute>;
const TensorShape & GetShape() const
CPU Execution: Reference C++ kernels.
static const StringMapping & Instance()
Copyright (c) 2021 ARM Limited and Contributors.
#define ARMNN_SCOPED_PROFILING_EVENT(backendId, name)
std::vector< ITensorHandle * > m_Inputs
void Execute() const override
RefElementwiseWorkload(const ParentDescriptor &descriptor, const WorkloadInfo &info)
void ExecuteAsync(ExecutionData &executionData) override
std::vector< ITensorHandle * > m_Outputs
Contains information about TensorInfos of a layer.
const TensorInfo & GetTensorInfo(const ITensorHandle *tensorHandle)
float32 helpers