33 std::vector<ITensorHandle*> outputs)
const 39 const TensorInfo& anchorsInfo = m_Anchors->GetTensorInfo();
46 auto boxEncodings = MakeDecoder<float>(boxEncodingsInfo, inputs[0]->Map());
47 auto scores = MakeDecoder<float>(scoresInfo, inputs[1]->Map());
48 auto anchors = MakeDecoder<float>(anchorsInfo, m_Anchors->Map(
false));
50 float* detectionBoxes =
reinterpret_cast<float*
>(outputs[0]->Map());
51 float* detectionClasses =
reinterpret_cast<float*
>(outputs[1]->Map());
52 float* detectionScores =
reinterpret_cast<float*
>(outputs[2]->Map());
53 float* numDetections =
reinterpret_cast<float*
>(outputs[3]->Map());
56 detectionBoxesInfo, detectionClassesInfo,
58 *boxEncodings, *scores, *anchors, detectionBoxes,
59 detectionClasses, detectionScores, numDetections);
CPU Execution: Reference C++ kernels.
void Execute() const override
Copyright (c) 2021 ARM Limited and Contributors.
RefDetectionPostProcessWorkload(const DetectionPostProcessQueueDescriptor &descriptor, const WorkloadInfo &info)
void ExecuteAsync(ExecutionData &executionData) override
LayerDescriptor m_Parameters
#define ARMNN_SCOPED_PROFILING_EVENT(backendId, name)
std::vector< ITensorHandle * > m_Inputs
DetectionPostProcessQueueDescriptor m_Data
std::vector< ITensorHandle * > m_Outputs
std::vector< ITensorHandle * > m_Outputs
Contains information about TensorInfos of a layer.
std::vector< ITensorHandle * > m_Inputs
const TensorInfo & GetTensorInfo(const ITensorHandle *tensorHandle)
float32 helpers