From b4dd5cc86d4eb841de670f0f102ede599e0d9c40 Mon Sep 17 00:00:00 2001 From: Keith Davis Date: Thu, 7 Apr 2022 11:32:00 +0100 Subject: IVGCVSW-6124 ConstTensorsAsInput: Conv2d - FrontEnd * Update Front-end and Tools. * Updated Serializer, Deserializer and unit tests to reflect this. * Updated TfLiteDelegate, TfLiteParser and OnnxParser. * Updated Ref. * Fixed resulting Neon / CL tests * Unified optimizers for conv2d ops * Optimizer Fix - Fp32ToBf16 * Partial implementation for ACL backends to fix VTS failures !android-nn-driver:7477 Signed-off-by: Keith Davis Change-Id: I5fb18877f7ee32643e15a9818945356274bb401b --- src/armnnDeserializer/Deserializer.cpp | 73 +++++++++++++++++++++++----------- 1 file changed, 49 insertions(+), 24 deletions(-) (limited to 'src/armnnDeserializer/Deserializer.cpp') diff --git a/src/armnnDeserializer/Deserializer.cpp b/src/armnnDeserializer/Deserializer.cpp index 704b6c35c1..04dde73b20 100644 --- a/src/armnnDeserializer/Deserializer.cpp +++ b/src/armnnDeserializer/Deserializer.cpp @@ -1423,44 +1423,69 @@ void IDeserializer::DeserializerImpl::ParseConvolution2d(GraphPtr graph, unsigne CHECK_LAYERS(graph, 0, layerIndex); auto inputs = GetInputs(graph, layerIndex); CHECK_LOCATION(); - CHECK_VALID_SIZE(inputs.size(), 1); auto outputs = GetOutputs(graph, layerIndex); CHECK_VALID_SIZE(outputs.size(), 1); - auto serializerLayer = graph->layers()->Get(layerIndex)->layer_as_Convolution2dLayer(); + auto flatBufferLayer = graph->layers()->Get(layerIndex)->layer_as_Convolution2dLayer(); + auto layerName = GetLayerName(graph, layerIndex); - auto serializerDescriptor = serializerLayer->descriptor(); + auto flatbufferDescriptor = flatBufferLayer->descriptor(); armnn::Convolution2dDescriptor descriptor; - descriptor.m_PadLeft = serializerDescriptor->padLeft(); - descriptor.m_PadRight = serializerDescriptor->padRight(); - descriptor.m_PadTop = serializerDescriptor->padTop(); - descriptor.m_PadBottom = serializerDescriptor->padBottom(); - descriptor.m_StrideX = serializerDescriptor->strideX(); - descriptor.m_StrideY = serializerDescriptor->strideY();; - descriptor.m_DilationX = serializerDescriptor->dilationX(); - descriptor.m_DilationY = serializerDescriptor->dilationY();; - descriptor.m_BiasEnabled = serializerDescriptor->biasEnabled();; - descriptor.m_DataLayout = ToDataLayout(serializerDescriptor->dataLayout()); + descriptor.m_PadLeft = flatbufferDescriptor->padLeft(); + descriptor.m_PadRight = flatbufferDescriptor->padRight(); + descriptor.m_PadTop = flatbufferDescriptor->padTop(); + descriptor.m_PadBottom = flatbufferDescriptor->padBottom(); + descriptor.m_StrideX = flatbufferDescriptor->strideX(); + descriptor.m_StrideY = flatbufferDescriptor->strideY();; + descriptor.m_DilationX = flatbufferDescriptor->dilationX(); + descriptor.m_DilationY = flatbufferDescriptor->dilationY();; + descriptor.m_BiasEnabled = flatbufferDescriptor->biasEnabled();; + descriptor.m_DataLayout = ToDataLayout(flatbufferDescriptor->dataLayout()); - armnn::ConstTensor weights = ToConstTensor(serializerLayer->weights()); - armnn::ConstTensor biases; + armnn::IConnectableLayer* layer; + std::vector ignoreSlots {}; - armnn::Optional optionalBiases = armnn::EmptyOptional(); - if (descriptor.m_BiasEnabled) + armnn::ConstTensor biasTensor; + // Weights and biases used to be always constant and were stored as members of the layer. This has changed and + // they are now passed as inputs. If they are constant then they will be stored in a ConstantLayer. + if (this->GetFeatureVersions(graph).m_ConstTensorsAsInputs <= 0) + { + // If the model stores weights and biases as members of the layer we have to read them from there + // but add them to their own ConstantLayer for compatibility + CHECK_VALID_SIZE(inputs.size(), 1); + + layer = m_Network->AddConvolution2dLayer(descriptor, + layerName.c_str()); + + armnn::ConstTensor weightsTensor = ToConstTensor(flatBufferLayer->weights()); + auto weightsLayer = m_Network->AddConstantLayer(weightsTensor); + weightsLayer->GetOutputSlot(0).Connect(layer->GetInputSlot(1u)); + weightsLayer->GetOutputSlot(0).SetTensorInfo(weightsTensor.GetInfo()); + ignoreSlots.emplace_back(1u); + + if (descriptor.m_BiasEnabled) + { + biasTensor = ToConstTensor(flatBufferLayer->biases()); + auto biasLayer = m_Network->AddConstantLayer(biasTensor); + biasLayer->GetOutputSlot(0).Connect(layer->GetInputSlot(2u)); + biasLayer->GetOutputSlot(0).SetTensorInfo(biasTensor.GetInfo()); + ignoreSlots.emplace_back(2u); + } + } + else { - biases = ToConstTensor(serializerLayer->biases()); - optionalBiases = armnn::Optional(biases); + layer = m_Network->AddConvolution2dLayer(descriptor, + layerName.c_str()); + uint32_t numInputs = descriptor.GetNumInputs(); + CHECK_VALID_SIZE(inputs.size(), numInputs); } - IConnectableLayer* layer = m_Network->AddConvolution2dLayer(descriptor, - weights, - optionalBiases, - layerName.c_str()); + armnn::TensorInfo outputTensorInfo = ToTensorInfo(outputs[0]); layer->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); - RegisterInputSlots(graph, layerIndex, layer); + RegisterInputSlots(graph, layerIndex, layer, ignoreSlots); RegisterOutputSlots(graph, layerIndex, layer); } -- cgit v1.2.1