21 m_Weights = std::make_unique<ScopedCpuTensorHandle>(*(descriptor.
m_Weight));
22 const TensorInfo& weightsInfo = m_Weights->GetTensorInfo();
24 m_WeightsDecoder = MakeDecoder<float>(weightsInfo, m_Weights->Map(
true));
25 m_WeightsShape = weightsInfo.GetShape();
30 m_Biases = std::make_unique<ScopedCpuTensorHandle>(*(descriptor.
m_Bias));
31 const TensorInfo& biasesInfo = m_Biases->GetTensorInfo();
32 m_BiasesDecoder = MakeDecoder<float>(biasesInfo, m_Biases->Map(
true));
43 m_InputDecoder = MakeDecoder<float>(inputInfo);
49 m_OutputShape = outputInfo.
GetShape();
50 m_OutputEncoder = MakeEncoder<float>(outputInfo);
67 m_BiasesDecoder.get());
#define ARMNN_SCOPED_PROFILING_EVENT(backendId, name)
LayerDescriptor m_Parameters
RefTransposeConvolution2dWorkload(const TransposeConvolution2dQueueDescriptor &descriptor, const WorkloadInfo &info)
void TransposeConvolution2dImpl(const TransposeConvolution2dDescriptor &descriptor, const TensorShape &inputShape, Decoder< float > &inputDecoder, const TensorShape &outputShape, Encoder< float > &outputEncoder, const TensorShape &weightsShape, Decoder< float > &weightsDecoder, Decoder< float > *biasesDecoder)
const TransposeConvolution2dQueueDescriptor m_Data
CPU Execution: Reference C++ kernels.
const ConstCpuTensorHandle * m_Bias
const ConstCpuTensorHandle * m_Weight
std::vector< ITensorHandle * > m_Outputs
std::vector< ITensorHandle * > m_Inputs
bool m_BiasEnabled
Enable/disable bias.
void Execute() const override
const TensorShape & GetShape() const
void PostAllocationConfigure() override
const TensorInfo & GetTensorInfo(const ITensorHandle *tensorHandle)
float32 helpers