35 const TensorInfo& weightsInfo = m_Weight->GetTensorInfo();
38 const int16_t* inputData = GetInputTensorData<int16_t>(0,
m_Data);;
39 const int16_t* weightsData = m_Weight->template GetConstTensor<int16_t>();
42 EthosnRefTransposeConvolutionImpl<armnn::TransposeConvolution2dQueueDescriptor, int16_t, int16_t, int32_t, int64_t>(
44 inputData, inputInfo.GetQuantizationScale(), inputInfo.GetQuantizationOffset(),
45 weightsData, weightsInfo.GetQuantizationScale(), weightsInfo.GetQuantizationOffset(),
47 outputInfo.GetQuantizationScale(), outputInfo.GetQuantizationOffset(), weightsInfo);
50 const int8_t* inputData = GetInputTensorData<int8_t>(0,
m_Data);;
51 const int8_t* weightsData = m_Weight->template GetConstTensor<int8_t>();
52 const int32_t* biasData = m_Data.m_Parameters.m_BiasEnabled ? m_Bias->template GetConstTensor<int32_t>() :
nullptr;
53 const TensorInfo& outputInfo =
GetTensorInfo(m_Data.m_Outputs[0]);
54 EthosnRefTransposeConvolutionImpl<armnn::TransposeConvolution2dQueueDescriptor, int8_t, int8_t, int32_t, int32_t>(
56 inputData, inputInfo.GetQuantizationScale(), inputInfo.GetQuantizationOffset(),
57 weightsData, weightsInfo.GetQuantizationScale(), weightsInfo.GetQuantizationOffset(),
59 outputInfo.GetQuantizationScale(), outputInfo.GetQuantizationOffset(), weightsInfo);
63 const uint8_t* inputData = GetInputTensorData<uint8_t>(0,
m_Data);;
64 const uint8_t* weightsData = m_Weight->template GetConstTensor<uint8_t>();
65 const int32_t* biasData = m_Data.m_Parameters.m_BiasEnabled ? m_Bias->template GetConstTensor<int32_t>() :
nullptr;
66 const TensorInfo& outputInfo =
GetTensorInfo(m_Data.m_Outputs[0]);
67 EthosnRefTransposeConvolutionImpl<armnn::TransposeConvolution2dQueueDescriptor, uint8_t, uint8_t, int32_t, int32_t>(
69 inputData, inputInfo.GetQuantizationScale(), inputInfo.GetQuantizationOffset(),
70 weightsData, weightsInfo.GetQuantizationScale(), weightsInfo.GetQuantizationOffset(),
72 outputInfo.GetQuantizationScale(), outputInfo.GetQuantizationOffset(), weightsInfo);
bool m_BiasEnabled
Enable/disable bias.
LayerDescriptor m_Parameters
bool CheckDataType(DataType type, DataType inputType, DataType weightsType)
TransposeConvolution2dQueueDescriptor m_Data
#define ARMNN_SCOPED_PROFILING_EVENT_ETHOSN(name)
std::vector< ITensorHandle * > m_Outputs
std::vector< ITensorHandle * > m_Inputs
const TensorInfo & GetTensorInfo(const ITensorHandle *tensorHandle)
float32 helpers