// // Copyright © 2017 Arm Ltd and Contributors. All rights reserved. // SPDX-License-Identifier: MIT // #include "Convolution2dLayer.hpp" #include "LayerCloneBase.hpp" #include #include #include #include #include using namespace armnnUtils; namespace armnn { Convolution2dLayer::Convolution2dLayer(const Convolution2dDescriptor& param, const char* name) : LayerWithParameters(1, 1, LayerType::Convolution2d, param, name) { } void Convolution2dLayer::SerializeLayerParameters(ParameterStringifyFunction& fn) const { //using DescriptorType = Parameters; const std::vector& inputShapes = { GetInputSlot(0).GetConnection()->GetTensorInfo().GetShape(), m_Weight->GetTensorInfo().GetShape() }; const TensorShape filterShape = inputShapes[1]; DataLayoutIndexed dataLayoutIndex(m_Param.m_DataLayout); unsigned int filterWidth = filterShape[dataLayoutIndex.GetWidthIndex()]; unsigned int filterHeight = filterShape[dataLayoutIndex.GetHeightIndex()]; unsigned int outChannels = filterShape[0]; fn("OutputChannels",std::to_string(outChannels)); fn("FilterWidth",std::to_string(filterWidth)); fn("FilterHeight",std::to_string(filterHeight)); LayerWithParameters::SerializeLayerParameters(fn); } std::unique_ptr Convolution2dLayer::CreateWorkload(const IWorkloadFactory& factory) const { // on this level constant data should not be released.. ARMNN_ASSERT_MSG(m_Weight != nullptr, "Convolution2dLayer: Weights data should not be null."); ARMNN_SCOPED_PROFILING_EVENT(Compute::Undefined, "Convolution2dLayer_CreateWorkload"); Convolution2dQueueDescriptor descriptor; descriptor.m_Weight = m_Weight.get(); if (m_Param.m_BiasEnabled) { ARMNN_ASSERT_MSG(m_Bias != nullptr, "Convolution2dLayer: Bias data should not be null."); descriptor.m_Bias = m_Bias.get(); } SetAdditionalInfo(descriptor); return factory.CreateWorkload(LayerType::Convolution2d, descriptor, PrepInfoAndDesc(descriptor)); } Convolution2dLayer* Convolution2dLayer::Clone(Graph& graph) const { auto layer = CloneBase(graph, m_Param, GetName()); layer->m_Weight = m_Weight ? m_Weight : nullptr; if (layer->m_Param.m_BiasEnabled) { layer->m_Bias = m_Bias ? m_Bias : nullptr; } return std::move(layer); } std::vector Convolution2dLayer::InferOutputShapes(const std::vector& inputShapes) const { ARMNN_ASSERT(inputShapes.size() == 2); const TensorShape& inputShape = inputShapes[0]; const TensorShape filterShape = inputShapes[1]; // If we support multiple batch dimensions in the future, then this assert will need to change. ARMNN_ASSERT_MSG(inputShape.GetNumDimensions() == 4, "Convolutions will always have 4D input."); ARMNN_ASSERT( m_Param.m_StrideX > 0); ARMNN_ASSERT( m_Param.m_StrideY > 0); DataLayoutIndexed dataLayoutIndex(m_Param.m_DataLayout); unsigned int inWidth = inputShape[dataLayoutIndex.GetWidthIndex()]; unsigned int inHeight = inputShape[dataLayoutIndex.GetHeightIndex()]; unsigned int inBatchSize = inputShape[0]; unsigned int filterWidth = filterShape[dataLayoutIndex.GetWidthIndex()]; unsigned int dilatedFilterWidth = filterWidth + (m_Param.m_DilationX - 1) * (filterWidth - 1); unsigned int readWidth = (inWidth + m_Param.m_PadLeft + m_Param.m_PadRight) - dilatedFilterWidth; unsigned int outWidth = 1 + (readWidth / m_Param.m_StrideX); unsigned int filterHeight = filterShape[dataLayoutIndex.GetHeightIndex()]; unsigned int dilatedFilterHeight = filterHeight + (m_Param.m_DilationY - 1) * (filterHeight - 1); unsigned int readHeight = (inHeight + m_Param.m_PadTop + m_Param.m_PadBottom) - dilatedFilterHeight; unsigned int outHeight = 1 + (readHeight / m_Param.m_StrideY); unsigned int outChannels = filterShape[0]; unsigned int outBatchSize = inBatchSize; TensorShape tensorShape = m_Param.m_DataLayout == armnn::DataLayout::NHWC ? TensorShape( { outBatchSize, outHeight, outWidth, outChannels } ) : TensorShape( { outBatchSize, outChannels, outHeight, outWidth }); return std::vector({ tensorShape }); } void Convolution2dLayer::ValidateTensorShapesFromInputs() { VerifyLayerConnections(1, CHECK_LOCATION()); const TensorShape& outputShape = GetOutputSlot(0).GetTensorInfo().GetShape(); VerifyShapeInferenceType(outputShape, m_ShapeInferenceMethod); // check if we m_Weight data is not nullptr ARMNN_ASSERT_MSG(m_Weight != nullptr, "Convolution2dLayer: Weights data should not be null."); auto inferredShapes = InferOutputShapes({ GetInputSlot(0).GetConnection()->GetTensorInfo().GetShape(), m_Weight->GetTensorInfo().GetShape() }); ARMNN_ASSERT(inferredShapes.size() == 1); ValidateAndCopyShape(outputShape, inferredShapes[0], m_ShapeInferenceMethod, "Convolution2dLayer"); } Layer::ConstantTensors Convolution2dLayer::GetConstantTensorsByRef() { // For API stability DO NOT ALTER order and add new members to the end of vector return {m_Weight, m_Bias}; } ARMNN_NO_DEPRECATE_WARN_BEGIN void Convolution2dLayer::Accept(ILayerVisitor& visitor) const { ManagedConstTensorHandle managedWeight(m_Weight); ConstTensor weightsTensor(managedWeight.GetTensorInfo(), managedWeight.Map()); Optional optionalBiasTensor = EmptyOptional(); ManagedConstTensorHandle managedBias(m_Bias); if (GetParameters().m_BiasEnabled) { ConstTensor biasTensor(managedBias.GetTensorInfo(), managedBias.Map()); optionalBiasTensor = Optional(biasTensor); } visitor.VisitConvolution2dLayer(this, GetParameters(), weightsTensor, optionalBiasTensor, GetName()); } ARMNN_NO_DEPRECATE_WARN_END void Convolution2dLayer::ExecuteStrategy(IStrategy& strategy) const { ManagedConstTensorHandle managedWeight(m_Weight); std::vector constTensors { { managedWeight.GetTensorInfo(), managedWeight.Map() } }; ManagedConstTensorHandle managedBias(m_Bias); if (GetParameters().m_BiasEnabled) { constTensors.emplace_back(ConstTensor(managedBias.GetTensorInfo(), managedBias.Map())); } strategy.ExecuteStrategy(this, GetParameters(), constTensors, GetName()); } } // namespace armnn