From 2e4a24aafc622c00d14ad2dcc684d05dfaacfe33 Mon Sep 17 00:00:00 2001 From: Sadik Armagan Date: Thu, 18 Mar 2021 13:59:40 +0000 Subject: IVGCVSW-5742 'NonConstWeights: Update FullyConnected in android-nn-driver' * Enabled weights and bias as inputs in FULLY_CONNECTED operator. !armnn:5180 Signed-off-by: Sadik Armagan Change-Id: Id325a8bf5be5a772191d27ae89485e992f0c48fa --- ConversionUtils.hpp | 98 +++++++++++++++++++++++++++++++++++++++++------------ 1 file changed, 77 insertions(+), 21 deletions(-) (limited to 'ConversionUtils.hpp') diff --git a/ConversionUtils.hpp b/ConversionUtils.hpp index 3432d9f8..e5f99ed4 100644 --- a/ConversionUtils.hpp +++ b/ConversionUtils.hpp @@ -3034,26 +3034,72 @@ bool ConvertFullyConnected(const HalOperation& operation, const HalModel& model, const armnn::TensorInfo& inputInfo = input.GetTensorInfo(); const armnn::TensorInfo& outputInfo = GetTensorInfoForOperand(*output); - ConstTensorPin weightsPin = DequantizeAndMakeConstTensorPin(operation, model, data, 1); - ConstTensorPin biasPin = ConvertOperationInputToConstTensorPin(operation, 2, model, data); // 1D + LayerInputHandle weightsInput = LayerInputHandle(); + const HalOperand* weightsOperand = GetInputOperand(operation, 1, model); + if (!weightsOperand) + { + return Fail("%s: Could not read weights", __func__); + } + + const armnn::TensorInfo& weightsInfo = GetTensorInfoForOperand(*weightsOperand); + bool constantWeights = IsOperandConstant(*weightsOperand); - if (!weightsPin.IsValid()) + armnn::Optional optionalWeights = armnn::EmptyOptional(); + if (!constantWeights) { - return Fail("%s: Operation has invalid weights", __func__); + weightsInput = ConvertToLayerInputHandle(operation, 1, model, data); + if (!weightsInput.IsValid()) + { + return Fail("%s: Operation has invalid inputs", __func__); + } + } + else + { + ConstTensorPin weightsPin = DequantizeAndMakeConstTensorPin(operation, model, data, 1); + if (!weightsPin.IsValid()) + { + return Fail("%s: Operation has invalid weights", __func__); + } + optionalWeights = armnn::Optional(weightsPin.GetConstTensor()); } - if (!biasPin.IsValid()) + LayerInputHandle biasInput = LayerInputHandle(); + const HalOperand* biasOperand = GetInputOperand(operation, 2, model); + if (!biasOperand) { - return Fail("%s: Operation has invalid bias", __func__); + return Fail("%s: Could not read bias", __func__); } + armnn::TensorInfo biasInfo = GetTensorInfoForOperand(*biasOperand); + bool constantBias = IsOperandConstant(*biasOperand); - armnn::ConstTensor weights = weightsPin.GetConstTensor(); - armnn::ConstTensor bias = biasPin.GetConstTensor(); - armnn::TensorInfo reshapedInfo = inputInfo; + armnn::Optional optionalBias = armnn::EmptyOptional(); + if (!constantBias) + { + biasInput = ConvertToLayerInputHandle(operation, 2, model, data); + if (!biasInput.IsValid()) + { + return Fail("%s: Operation has invalid inputs", __func__); + } + } + else + { + ConstTensorPin biasPin = ConvertOperationInputToConstTensorPin(operation, 2, model, data); // 1D + if (!biasPin.IsValid()) + { + return Fail("%s: Operation has invalid bias", __func__); + } + optionalBias = armnn::Optional(biasPin.GetConstTensor()); + } + if ((constantWeights && !constantBias) || (!constantWeights && constantBias)) + { + return Fail("%s: Non-compatible weights and bias", __func__); + } + + armnn::TensorInfo reshapedInfo = inputInfo; try { - reshapedInfo.SetShape(FlattenFullyConnectedInput(inputInfo.GetShape(), weights.GetInfo().GetShape())); + reshapedInfo.SetShape(FlattenFullyConnectedInput(inputInfo.GetShape(), weightsInfo.GetShape())); } catch (const std::exception& e) { @@ -3061,7 +3107,7 @@ bool ConvertFullyConnected(const HalOperation& operation, const HalModel& model, } // ensuring that the bias value is within 1% of the weights input (small float differences can exist) - SanitizeBiasQuantizationScale(bias.GetInfo(), weights.GetInfo(), reshapedInfo); + SanitizeBiasQuantizationScale(biasInfo, weightsInfo, reshapedInfo); ActivationFn activationFunction; if (!GetInputActivationFunction(operation, 3, activationFunction, model, data)) @@ -3072,12 +3118,13 @@ bool ConvertFullyConnected(const HalOperation& operation, const HalModel& model, armnn::FullyConnectedDescriptor desc; desc.m_TransposeWeightMatrix = true; desc.m_BiasEnabled = true; + desc.m_ConstantWeights = constantWeights; bool isSupported = false; auto validateFunc = [&](const armnn::TensorInfo& outputInfo, bool& isSupported) { if (!VerifyFullyConnectedShapes(reshapedInfo.GetShape(), - weights.GetInfo().GetShape(), + weightsInfo.GetShape(), outputInfo.GetShape(), desc.m_TransposeWeightMatrix)) { @@ -3087,14 +3134,14 @@ bool ConvertFullyConnected(const HalOperation& operation, const HalModel& model, } FORWARD_LAYER_SUPPORT_FUNC(__func__, - IsFullyConnectedSupported, - data.m_Backends, - isSupported, - reshapedInfo, - outputInfo, - weights.GetInfo(), - bias.GetInfo(), - desc); + IsFullyConnectedSupported, + data.m_Backends, + isSupported, + reshapedInfo, + outputInfo, + weightsInfo, + biasInfo, + desc); }; if(!IsDynamicTensor(outputInfo)) @@ -3112,7 +3159,9 @@ bool ConvertFullyConnected(const HalOperation& operation, const HalModel& model, } armnn::IConnectableLayer* startLayer = - data.m_Network->AddFullyConnectedLayer(desc, weights, armnn::Optional(bias)); + data.m_Network->AddFullyConnectedLayer(desc, + optionalWeights, + optionalBias); if (inputInfo.GetNumDimensions() > 2U) { @@ -3130,6 +3179,13 @@ bool ConvertFullyConnected(const HalOperation& operation, const HalModel& model, input.Connect(startLayer->GetInputSlot(0)); } + // connect weights input + if (!desc.m_ConstantWeights) + { + weightsInput.Connect(startLayer->GetInputSlot(1)); + biasInput.Connect(startLayer->GetInputSlot(2)); + } + return SetupAndTrackLayerOutputSlot(operation, 0, *startLayer, model, data, nullptr, validateFunc, activationFunction); } -- cgit v1.2.1