From aed08ac9b985766a6ce442695ff1742defb0a189 Mon Sep 17 00:00:00 2001 From: Kevin May Date: Thu, 12 Dec 2019 16:33:31 +0000 Subject: IVGCVSW-4262 Add Calls to IsReshapeSupported and IsPermuteSupported !armnn:2643 * Add calls before addition of these layers in ConvertConcatenation * Add outputInfo parameter wherever needed for IsReshapeSupported Signed-off-by: Kevin May Change-Id: Ic5d142ea046161960ff2fc137bd261ebb4e6ac0c --- 1.2/HalPolicy.cpp | 7 ++++--- 1 file changed, 4 insertions(+), 3 deletions(-) (limited to '1.2') diff --git a/1.2/HalPolicy.cpp b/1.2/HalPolicy.cpp index 61daeef5..bf4525aa 100644 --- a/1.2/HalPolicy.cpp +++ b/1.2/HalPolicy.cpp @@ -749,6 +749,7 @@ bool HalPolicy::ConvertExpandDims(const Operation& operation, const Model& model data.m_Backends, isSupported, input.GetTensorInfo(), + outputInfo, reshapeDescriptor); if (!isSupported) @@ -1346,7 +1347,7 @@ bool HalPolicy::ConvertMaximum(const Operation& operation, const Model& model, C IConnectableLayer* layer = data.m_Network->AddMaximumLayer(); assert(layer != nullptr); - bool isReshapeSupported = BroadcastTensor(input0, input1, layer, data); + bool isReshapeSupported = BroadcastTensor(input0, input1, outInfo, layer, data); if (!isReshapeSupported) { return false; @@ -1401,7 +1402,7 @@ bool HalPolicy::ConvertMinimum(const Operation& operation, const Model& model, C IConnectableLayer* const layer = data.m_Network->AddMinimumLayer(); assert(layer != nullptr); - bool isReshapeSupported = BroadcastTensor(input0, input1, layer, data); + bool isReshapeSupported = BroadcastTensor(input0, input1, outputInfo, layer, data); if (!isReshapeSupported) { return false; @@ -1564,7 +1565,7 @@ bool HalPolicy::ConvertPrelu(const Operation& operation, const Model& model, Con return Fail("%s: AddPreluLayer failed", __func__); } - bool isReshapeSupported = BroadcastTensor(input, alpha, layer, data); + bool isReshapeSupported = BroadcastTensor(input, alpha, outputInfo, layer, data); if (!isReshapeSupported) { return false; -- cgit v1.2.1