23.05
|
Go to the documentation of this file.
20 template <
typename Functor,
typename ParentDescriptor,
typename armnn::StringMapping::Id DebugString>
22 const ParentDescriptor& desc,
28 template <
typename Functor,
typename ParentDescriptor,
typename armnn::StringMapping::Id DebugString>
31 Execute(m_Data.m_Inputs, m_Data.m_Outputs);
34 template <
typename Functor,
typename ParentDescriptor,
typename armnn::StringMapping::Id DebugString>
42 template <
typename Functor,
typename ParentDescriptor,
typename armnn::StringMapping::Id DebugString>
44 std::vector<ITensorHandle*> inputs, std::vector<ITensorHandle*> outputs)
const
55 std::unique_ptr<Decoder<InType>> input0 = MakeDecoder<InType>(inputInfo0, inputs[0]->
Map());
56 std::unique_ptr<Decoder<InType>> input1 = MakeDecoder<InType>(inputInfo1, inputs[1]->
Map());
57 std::unique_ptr<Encoder<OutType>> output= MakeEncoder<OutType>(outputInfo, outputs[0]->
Map());
@ RefMinimumWorkload_Execute
const TensorInfo & GetTensorInfo(const ITensorHandle *tensorHandle)
float32 helpers
@ RefMaximumWorkload_Execute
std::vector< ITensorHandle * > m_Inputs
@ RefAdditionWorkload_Execute
@ RefDivisionWorkload_Execute
Copyright (c) 2021 ARM Limited and Contributors.
@ RefMultiplicationWorkload_Execute
void Execute() const override
static const StringMapping & Instance()
RefElementwiseWorkload(const ParentDescriptor &descriptor, const WorkloadInfo &info)
#define ARMNN_SCOPED_PROFILING_EVENT(backendId, name)
@ RefSubtractionWorkload_Execute
void ExecuteAsync(ExecutionData &executionData) override
const TensorShape & GetShape() const
Contains information about TensorInfos of a layer.
std::vector< ITensorHandle * > m_Outputs
@ CpuRef
CPU Execution: Reference C++ kernels.