23.02
|
Go to the documentation of this file.
39 std::unique_ptr<Decoder<float>> decoderPtr = MakeDecoder<float>(inputTensorInfo, inputs[0]->
Map());
45 int32_t *output = GetOutputTensorData<int32_t>(outputs[0]);
49 int64_t *output = GetOutputTensorData<int64_t>(outputs[0]);
const TensorInfo & GetTensorInfo(const ITensorHandle *tensorHandle)
float32 helpers
void ArgMinMax(Decoder< float > &in, OUT *out, const TensorInfo &inputTensorInfo, const TensorInfo &outputTensorInfo, ArgMinMaxFunction function, int axis)
std::vector< ITensorHandle * > m_Inputs
int m_Axis
Axis to reduce across the input tensor.
void ExecuteAsync(ExecutionData &executionData) override
ArgMinMaxQueueDescriptor m_Data
Copyright (c) 2021 ARM Limited and Contributors.
ArgMinMaxFunction m_Function
Specify if the function is to find Min or Max.
#define ARMNN_SCOPED_PROFILING_EVENT(backendId, name)
RefArgMinMaxWorkload(const ArgMinMaxQueueDescriptor &descriptor, const WorkloadInfo &info)
void Execute() const override
Contains information about TensorInfos of a layer.
LayerDescriptor m_Parameters
std::vector< ITensorHandle * > m_Outputs
std::vector< ITensorHandle * > m_Outputs
@ CpuRef
CPU Execution: Reference C++ kernels.
DataType GetDataType() const
std::vector< ITensorHandle * > m_Inputs