32 std::vector<ITensorHandle*> outputs)
const 38 const TensorInfo& anchorsInfo = m_Anchors->GetTensorInfo();
45 auto boxEncodings = MakeDecoder<float>(boxEncodingsInfo, inputs[0]->Map());
46 auto scores = MakeDecoder<float>(scoresInfo, inputs[1]->Map());
47 auto anchors = MakeDecoder<float>(anchorsInfo, m_Anchors->Map(
false));
49 float* detectionBoxes =
reinterpret_cast<float*
>(outputs[0]->Map());
50 float* detectionClasses =
reinterpret_cast<float*
>(outputs[1]->Map());
51 float* detectionScores =
reinterpret_cast<float*
>(outputs[2]->Map());
52 float* numDetections =
reinterpret_cast<float*
>(outputs[3]->Map());
55 detectionBoxesInfo, detectionClassesInfo,
57 *boxEncodings, *scores, *anchors, detectionBoxes,
58 detectionClasses, detectionScores, numDetections);
CPU Execution: Reference C++ kernels.
void Execute() const override
void ExecuteAsync(WorkingMemDescriptor &workingMemDescriptor) override
Copyright (c) 2021 ARM Limited and Contributors.
RefDetectionPostProcessWorkload(const DetectionPostProcessQueueDescriptor &descriptor, const WorkloadInfo &info)
LayerDescriptor m_Parameters
#define ARMNN_SCOPED_PROFILING_EVENT(backendId, name)
std::vector< ITensorHandle * > m_Inputs
DetectionPostProcessQueueDescriptor m_Data
std::vector< ITensorHandle * > m_Outputs
std::vector< ITensorHandle * > m_Outputs
Contains information about TensorInfos of a layer.
std::vector< ITensorHandle * > m_Inputs
const TensorInfo & GetTensorInfo(const ITensorHandle *tensorHandle)
float32 helpers