const BFloat16 * GetInputTensorDataBFloat16(unsigned int idx, const PayloadType &data)
CPU Execution: Reference C++ kernels.
static void ConvertBFloat16ToFloat32(const void *srcBFloat16Buffer, size_t numElements, float *dstFloat32Buffer)
const QueueDescriptor m_Data
Copyright (c) 2020 ARM Limited.
#define ARMNN_SCOPED_PROFILING_EVENT(backendId, name)
virtual void Execute() const override
std::vector< ITensorHandle * > m_Inputs
float * GetOutputTensorDataFloat(unsigned int idx, const PayloadType &data)
const TensorInfo & GetTensorInfo(const ITensorHandle *tensorHandle)
float32 helpers
unsigned int GetNumElements() const