23.02
|
Go to the documentation of this file.
21 m_Weights = std::make_unique<ScopedTensorHandle>(*(descriptor.
m_Weight));
22 const TensorInfo& weightsInfo = m_Weights->GetTensorInfo();
24 m_WeightsDecoder = MakeDecoder<float>(weightsInfo, m_Weights->Map(
true));
25 m_WeightsShape = weightsInfo.
GetShape();
30 m_Biases = std::make_unique<ScopedTensorHandle>(*(descriptor.
m_Bias));
31 const TensorInfo& biasesInfo = m_Biases->GetTensorInfo();
32 m_BiasesDecoder = MakeDecoder<float>(biasesInfo, m_Biases->Map(
true));
48 std::vector<ITensorHandle*> outputs)
const
55 std::unique_ptr<Decoder<float>> inputDecoder = MakeDecoder<float>(inputInfo, inputs[0]->
Map());
56 std::unique_ptr<Encoder<float>> outputEncoder = MakeEncoder<float>(outputInfo, outputs[0]->
Map());
65 m_BiasesDecoder.get());
const TensorInfo & GetTensorInfo(const ITensorHandle *tensorHandle)
float32 helpers
std::vector< ITensorHandle * > m_Inputs
bool m_BiasEnabled
Enable/disable bias.
TransposeConvolution2dQueueDescriptor m_Data
void Execute() const override
Copyright (c) 2021 ARM Limited and Contributors.
#define ARMNN_SCOPED_PROFILING_EVENT(backendId, name)
void ExecuteAsync(ExecutionData &executionData) override
const TensorShape & GetShape() const
RefTransposeConvolution2dWorkload(const TransposeConvolution2dQueueDescriptor &descriptor, const WorkloadInfo &info)
Contains information about TensorInfos of a layer.
LayerDescriptor m_Parameters
std::vector< ITensorHandle * > m_Outputs
std::vector< ITensorHandle * > m_Outputs
const ConstTensorHandle * m_Bias
void TransposeConvolution2dImpl(const TransposeConvolution2dDescriptor &descriptor, const TensorShape &inputShape, Decoder< float > &inputDecoder, const TensorShape &outputShape, Encoder< float > &outputEncoder, const TensorShape &weightsShape, Decoder< float > &weightsDecoder, Decoder< float > *biasesDecoder)
const ConstTensorHandle * m_Weight
@ CpuRef
CPU Execution: Reference C++ kernels.
std::vector< ITensorHandle * > m_Inputs