20 const TensorInfo& rWeightInfo = m_Weight->GetTensorInfo();
21 m_WeightShape = rWeightInfo.
GetShape();
22 m_WeightDecoder = MakeDecoder<float>(rWeightInfo, m_Weight->Map(
true));
26 m_Bias = std::make_unique<ScopedCpuTensorHandle>(*(descriptor.
m_Bias));
27 const TensorInfo& biasInfo = m_Bias->GetTensorInfo();
28 m_BiasDecoder = MakeDecoder<float>(biasInfo, m_Bias->Map(
true));
37 m_InputDecoder = MakeDecoder<float>(inputInfo);
40 m_OutputShape = outputInfo.GetShape();
41 m_OutputEncoder = MakeEncoder<float>(outputInfo);
46 m_NumActivations *= inputInfo.
GetShape()[i];
#define ARMNN_SCOPED_PROFILING_EVENT(backendId, name)
virtual void Execute() const override
unsigned int GetNumDimensions() const
void PostAllocationConfigure() override
LayerDescriptor m_Parameters
const ConstCpuTensorHandle * m_Bias
bool m_BiasEnabled
Enable/disable bias.
const FullyConnectedQueueDescriptor m_Data
RefFullyConnectedWorkload(const FullyConnectedQueueDescriptor &descriptor, const WorkloadInfo &info)
bool m_TransposeWeightMatrix
Enable/disable transpose weight matrix.
CPU Execution: Reference C++ kernels.
std::vector< ITensorHandle * > m_Outputs
std::vector< ITensorHandle * > m_Inputs
const TensorShape & GetShape() const
const TensorInfo & GetTensorInfo(const ITensorHandle *tensorHandle)
float32 helpers