From f03fcf0dd180ba2c87648a524fcca9214e1f979b Mon Sep 17 00:00:00 2001 From: Aron Virginas-Tar Date: Tue, 9 Jul 2019 17:44:24 +0100 Subject: IVGCVSW-3456 Add support for dynamic output shape in ConvertPrelu Signed-off-by: Aron Virginas-Tar Change-Id: I8fc7a716455be3f51b51177f6896a73790a41fc3 --- 1.2/HalPolicy.cpp | 17 +++++++++++++++-- 1 file changed, 15 insertions(+), 2 deletions(-) (limited to '1.2') diff --git a/1.2/HalPolicy.cpp b/1.2/HalPolicy.cpp index fe0cfbdc..b194a57a 100644 --- a/1.2/HalPolicy.cpp +++ b/1.2/HalPolicy.cpp @@ -5,6 +5,8 @@ #include "HalPolicy.hpp" +#include "OutputShapeUtils.hpp" + #include "../1.0/HalPolicy.hpp" #include "../1.1/HalPolicy.hpp" @@ -539,7 +541,13 @@ bool HalPolicy::ConvertPrelu(const Operation& operation, const Model& model, Con const armnn::TensorInfo& inputInfo = input.GetTensorInfo(); const armnn::TensorInfo& alphaInfo = alpha.GetTensorInfo(); - const armnn::TensorInfo& outputInfo = GetTensorInfoForOperand(*output); + + armnn::TensorInfo outputInfo = GetTensorInfoForOperand(*output); + if (outputInfo.GetNumElements() == 0u) + { + ALOGD("Output shape not set, will infer from inputs"); + outputInfo.SetShape(InferPreluOutputShape(inputInfo.GetShape(), alphaInfo.GetShape())); + } if (!IsLayerSupportedForAnyBackend(__func__, armnn::IsPreluSupported, @@ -560,7 +568,12 @@ bool HalPolicy::ConvertPrelu(const Operation& operation, const Model& model, Con BroadcastTensor(input, alpha, layer, *data.m_Network); - return SetupAndTrackLayerOutputSlot(operation, 0, *layer, model, data); + return SetupAndTrackLayerOutputSlot(operation, + 0, + *layer, + model, + data, + armnn::Optional(outputInfo)); } bool HalPolicy::ConvertResize(const Operation& operation, -- cgit v1.2.1