const float * GetInputTensorDataFloat(unsigned int idx, const PayloadType &data)
CPU Execution: Reference C++ kernels.
const QueueDescriptor m_Data
Copyright (c) 2020 ARM Limited.
#define ARMNN_SCOPED_PROFILING_EVENT(backendId, name)
static void ConvertFloat32ToBFloat16(const float *srcFloat32Buffer, size_t numElements, void *dstBFloat16Buffer)
virtual void Execute() const override
std::vector< ITensorHandle * > m_Inputs
BFloat16 * GetOutputTensorDataBFloat16(unsigned int idx, const PayloadType &data)
const TensorInfo & GetTensorInfo(const ITensorHandle *tensorHandle)
float32 helpers
unsigned int GetNumElements() const