18 unsigned int numActivations = 1;
21 numActivations *= inputInfo.
GetShape()[i];
23 return numActivations;
30 , m_InputShape(info.m_InputTensorInfos[0].GetShape())
31 , m_WeightShape(info.m_InputTensorInfos[1].GetShape())
32 , m_OutputShape(info.m_OutputTensorInfos[0].GetShape())
51 std::unique_ptr<Decoder<float>> inputDecoder = MakeDecoder<float>(
GetTensorInfo(inputs[0]), inputs[0]->Map());
52 std::unique_ptr<Encoder<float>> OutputEncoder = MakeEncoder<float>(
GetTensorInfo(outputs[0]), outputs[0]->Map());
54 std::unique_ptr<Decoder<float>> weightsDecoder = MakeDecoder<float>(
GetTensorInfo(inputs[1]), inputs[1]->Map());
55 std::unique_ptr<Decoder<float>> biasDecoder;
59 biasDecoder = MakeDecoder<float>(
GetTensorInfo(inputs[2]), inputs[2]->Map());
unsigned int GetNumActivations(const TensorInfo &inputInfo)
const TensorShape & GetShape() const
CPU Execution: Reference C++ kernels.
bool m_TransposeWeightMatrix
Enable/disable transpose weight matrix.
Copyright (c) 2021 ARM Limited and Contributors.
LayerDescriptor m_Parameters
#define ARMNN_SCOPED_PROFILING_EVENT(backendId, name)
std::vector< ITensorHandle * > m_Inputs
void Execute() const override
FullyConnectedQueueDescriptor m_Data
bool m_BiasEnabled
Enable/disable bias.
void ExecuteAsync(WorkingMemDescriptor &workingMemDescriptor) override
RefFullyConnectedWorkload(const FullyConnectedQueueDescriptor &descriptor, const WorkloadInfo &info)
std::vector< ITensorHandle * > m_Outputs
std::vector< ITensorHandle * > m_Outputs
Contains information about TensorInfos of a layer.
std::vector< ITensorHandle * > m_Inputs
const TensorInfo & GetTensorInfo(const ITensorHandle *tensorHandle)
float32 helpers
unsigned int GetNumDimensions() const