#define ARMNN_SCOPED_PROFILING_EVENT(backendId, name)
virtual void Execute() const override
float * GetOutputTensorDataFloat(unsigned int idx, const PayloadType &data)
static void ConvertFloat16To32(const void *srcFloat16Buffer, size_t numElements, float *dstFloat32Buffer)
const QueueDescriptor m_Data
CPU Execution: Reference C++ kernels.
const Half * GetInputTensorDataHalf(unsigned int idx, const PayloadType &data)
unsigned int GetNumElements() const
std::vector< ITensorHandle * > m_Inputs
const TensorInfo & GetTensorInfo(const ITensorHandle *tensorHandle)
float32 helpers