From 2c14db6f5747a279b7bf7dd2244ac3c82b86656c Mon Sep 17 00:00:00 2001 From: Mike Kelly Date: Wed, 15 Mar 2023 15:06:23 +0000 Subject: IVGCVSW-3808 Deprecation notices for old ElementwiseBinary layers * Added Deprecation notices for old ElementwiseBinary layers. Signed-off-by: Mike Kelly Change-Id: I5bd0f186aaed675885d667f47e1e210ee9ec84f8 --- include/armnn/BackendHelper.hpp | 6 ++++++ include/armnn/INetwork.hpp | 6 ++++++ samples/PreImportMemorySample.cpp | 2 ++ shim/sl/canonical/Converter.cpp | 24 ++++++++++++++++++++++ src/armnn/Network.cpp | 12 +++++++++++ src/armnn/Network.hpp | 7 ++++++- src/armnn/test/RuntimeTests.cpp | 18 +++++++++++----- src/armnn/test/TestNameOnlyLayerVisitor.cpp | 13 ++++++++++-- .../test/optimizations/FuseActivationTests.cpp | 11 ++++++++-- src/armnnSerializer/test/SerializerTests.cpp | 16 +++++++++++++++ src/backends/aclCommon/ArmComputeSubgraphUtils.hpp | 8 ++++++++ .../test/AdditionEndToEndTestImpl.hpp | 3 ++- .../backendsCommon/test/EndToEndTestImpl.hpp | 2 ++ .../test/IsLayerSupportedTestImpl.hpp | 12 +++++++++++ .../backendsCommon/test/OptimizationViewsTests.cpp | 2 ++ src/backends/cl/ClLayerSupport.cpp | 12 +++++++++++ src/backends/cl/ClLayerSupport.hpp | 10 +++++++-- 17 files changed, 151 insertions(+), 13 deletions(-) diff --git a/include/armnn/BackendHelper.hpp b/include/armnn/BackendHelper.hpp index 85aabe0dd2..ddf2308da2 100644 --- a/include/armnn/BackendHelper.hpp +++ b/include/armnn/BackendHelper.hpp @@ -42,6 +42,7 @@ public: const ActivationDescriptor& descriptor, Optional reasonIfUnsupported = EmptyOptional()); + ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use IsElementwiseBinarySupported instead", "24.02") bool IsAdditionSupported(const TensorInfo& input0, const TensorInfo& input1, const TensorInfo& output, @@ -156,6 +157,7 @@ public: const Optional& biases, Optional reasonIfUnsupported = EmptyOptional()); + ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use IsElementwiseBinarySupported instead", "24.02") bool IsDivisionSupported(const TensorInfo& input0, const TensorInfo& input1, const TensorInfo& output, @@ -244,6 +246,7 @@ public: const LstmInputParamsInfo& paramsInfo, Optional reasonIfUnsupported = EmptyOptional()); + ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use IsElementwiseBinarySupported instead", "24.02") bool IsMaximumSupported(const TensorInfo& input0, const TensorInfo& input1, const TensorInfo& output, @@ -267,11 +270,13 @@ public: const TensorInfo& output, Optional reasonIfUnsupported = EmptyOptional()); + ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use IsElementwiseBinarySupported instead", "24.02") bool IsMinimumSupported(const TensorInfo& input0, const TensorInfo& input1, const TensorInfo& output, Optional reasonIfUnsupported = EmptyOptional()); + ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use IsElementwiseBinarySupported instead", "24.02") bool IsMultiplicationSupported(const TensorInfo& input0, const TensorInfo& input1, const TensorInfo& output, @@ -400,6 +405,7 @@ public: const StridedSliceDescriptor& descriptor, Optional reasonIfUnsupported = EmptyOptional()); + ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use IsElementwiseBinarySupported instead", "24.02") bool IsSubtractionSupported(const TensorInfo& input0, const TensorInfo& input1, const TensorInfo& output, diff --git a/include/armnn/INetwork.hpp b/include/armnn/INetwork.hpp index 4eac0cfe2a..273753752d 100644 --- a/include/armnn/INetwork.hpp +++ b/include/armnn/INetwork.hpp @@ -481,11 +481,13 @@ public: /// Adds an addition layer to the network. /// @param name - Optional name for the layer. /// @return - Interface for configuring the layer. + ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use AddElementwiseBinaryLayer instead", "24.02") IConnectableLayer* AddAdditionLayer(const char* name = nullptr); /// Adds a multiplication layer to the network. /// @param name - Optional name for the layer. /// @return - Interface for configuring the layer. + ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use AddElementwiseBinaryLayer instead", "24.02") IConnectableLayer* AddMultiplicationLayer(const char* name = nullptr); /// Adds a batch normalization layer to the network. @@ -603,16 +605,19 @@ public: /// Adds a division layer to the network. /// @param name - Optional name for the layer. /// @return - Interface for configuring the layer. + ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use AddElementwiseBinaryLayer instead", "24.02") IConnectableLayer* AddDivisionLayer(const char* name = nullptr); /// Adds a subtraction layer to the network. /// @param name - Optional name for the layer. /// @return - Interface for configuring the layer. + ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use AddElementwiseBinaryLayer instead", "24.02") IConnectableLayer* AddSubtractionLayer(const char* name = nullptr); /// Add a Maximum layer to the network. /// @param name - Optional name for the layer. /// @return - Interface for configuring the layer. + ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use AddElementwiseBinaryLayer instead", "24.02") IConnectableLayer* AddMaximumLayer(const char* name = nullptr); /// Add a Mean layer to the network. @@ -645,6 +650,7 @@ public: /// Add a Minimum layer to the network. /// @param name - Optional name for the layer. /// @return - Interface for configuring the layer. + ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use AddElementwiseBinaryLayer instead", "24.02") IConnectableLayer* AddMinimumLayer(const char* name = nullptr); /// Add Gather layer to the network. diff --git a/samples/PreImportMemorySample.cpp b/samples/PreImportMemorySample.cpp index cf196fd20e..98f386bfdf 100644 --- a/samples/PreImportMemorySample.cpp +++ b/samples/PreImportMemorySample.cpp @@ -31,7 +31,9 @@ int main() armnn::INetworkPtr testNetwork(armnn::INetwork::Create()); auto inputLayer1 = testNetwork->AddInputLayer(0, "input 1 layer"); auto inputLayer2 = testNetwork->AddInputLayer(1, "input 2 layer"); + ARMNN_NO_DEPRECATE_WARN_BEGIN auto addLayer = testNetwork->AddAdditionLayer("add layer"); + ARMNN_NO_DEPRECATE_WARN_END auto outputLayer = testNetwork->AddOutputLayer(2, "output layer"); // Set the tensors in the network. diff --git a/shim/sl/canonical/Converter.cpp b/shim/sl/canonical/Converter.cpp index be052a6faa..90fd71154f 100644 --- a/shim/sl/canonical/Converter.cpp +++ b/shim/sl/canonical/Converter.cpp @@ -209,6 +209,7 @@ bool Converter::ConvertAdd(const Operation& operation, const Model& model, Conve armnn::BackendId setBackend; auto validateFunc = [&](const armnn::TensorInfo& outputInfo, bool& isSupported) { + ARMNN_NO_DEPRECATE_WARN_BEGIN FORWARD_LAYER_SUPPORT_FUNC(__func__, IsAdditionSupported, data.m_Backends, @@ -217,6 +218,7 @@ bool Converter::ConvertAdd(const Operation& operation, const Model& model, Conve inputInfo0, inputInfo1, outputInfo); + ARMNN_NO_DEPRECATE_WARN_END }; if(!IsDynamicTensor(outputInfo)) @@ -233,7 +235,9 @@ bool Converter::ConvertAdd(const Operation& operation, const Model& model, Conve return false; } + ARMNN_NO_DEPRECATE_WARN_BEGIN armnn::IConnectableLayer* const startLayer = data.m_Network->AddAdditionLayer(); + ARMNN_NO_DEPRECATE_WARN_END startLayer->SetBackendId(setBackend); bool isReshapeSupported = BroadcastTensor(input0, input1, startLayer, data); @@ -1536,6 +1540,7 @@ bool Converter::ConvertDiv(const Operation& operation, const Model& model, Conve armnn::BackendId setBackend; auto validateFunc = [&](const armnn::TensorInfo& outputInfo, bool& isSupported) { + ARMNN_NO_DEPRECATE_WARN_BEGIN FORWARD_LAYER_SUPPORT_FUNC(__func__, IsDivisionSupported, data.m_Backends, @@ -1544,6 +1549,7 @@ bool Converter::ConvertDiv(const Operation& operation, const Model& model, Conve input0.GetTensorInfo(), input1.GetTensorInfo(), outputInfo); + ARMNN_NO_DEPRECATE_WARN_END }; if(!IsDynamicTensor(outputInfo)) @@ -1560,7 +1566,9 @@ bool Converter::ConvertDiv(const Operation& operation, const Model& model, Conve return false; } + ARMNN_NO_DEPRECATE_WARN_BEGIN armnn::IConnectableLayer* const startLayer = data.m_Network->AddDivisionLayer(); + ARMNN_NO_DEPRECATE_WARN_END startLayer->SetBackendId(setBackend); bool isReshapeSupported = BroadcastTensor(input0, input1, startLayer, data); @@ -3378,6 +3386,7 @@ bool Converter::ConvertMaximum(const Operation& operation, const Model& model, C armnn::BackendId setBackend; auto validateFunc = [&](const armnn::TensorInfo& outInfo, bool& isSupported) { + ARMNN_NO_DEPRECATE_WARN_BEGIN FORWARD_LAYER_SUPPORT_FUNC(__func__, IsMaximumSupported, data.m_Backends, @@ -3386,6 +3395,7 @@ bool Converter::ConvertMaximum(const Operation& operation, const Model& model, C input0.GetTensorInfo(), input1.GetTensorInfo(), outInfo); + ARMNN_NO_DEPRECATE_WARN_END }; if(IsDynamicTensor(outInfo)) @@ -3402,7 +3412,9 @@ bool Converter::ConvertMaximum(const Operation& operation, const Model& model, C return false; } + ARMNN_NO_DEPRECATE_WARN_BEGIN IConnectableLayer* layer = data.m_Network->AddMaximumLayer(); + ARMNN_NO_DEPRECATE_WARN_END layer->SetBackendId(setBackend); assert(layer != nullptr); bool isReshapeSupported = BroadcastTensor(input0, input1, layer, data); @@ -3524,6 +3536,7 @@ bool Converter::ConvertMinimum(const Operation& operation, const Model& model, C armnn::BackendId setBackend; auto validateFunc = [&](const armnn::TensorInfo& outputInfo, bool& isSupported) { + ARMNN_NO_DEPRECATE_WARN_BEGIN FORWARD_LAYER_SUPPORT_FUNC(__func__, IsMinimumSupported, data.m_Backends, @@ -3532,6 +3545,7 @@ bool Converter::ConvertMinimum(const Operation& operation, const Model& model, C input0.GetTensorInfo(), input1.GetTensorInfo(), outputInfo); + ARMNN_NO_DEPRECATE_WARN_END }; if(IsDynamicTensor(outputInfo)) @@ -3548,7 +3562,9 @@ bool Converter::ConvertMinimum(const Operation& operation, const Model& model, C return false; } + ARMNN_NO_DEPRECATE_WARN_BEGIN IConnectableLayer* const layer = data.m_Network->AddMinimumLayer(); + ARMNN_NO_DEPRECATE_WARN_END layer->SetBackendId(setBackend); assert(layer != nullptr); bool isReshapeSupported = BroadcastTensor(input0, input1, layer, data); @@ -3593,6 +3609,7 @@ bool Converter::ConvertMul(const Operation& operation, const Model& model, Conve armnn::BackendId setBackend; auto validateFunc = [&](const armnn::TensorInfo& outputInfo, bool& isSupported) { + ARMNN_NO_DEPRECATE_WARN_BEGIN FORWARD_LAYER_SUPPORT_FUNC(__func__, IsMultiplicationSupported, data.m_Backends, @@ -3601,6 +3618,7 @@ bool Converter::ConvertMul(const Operation& operation, const Model& model, Conve input0.GetTensorInfo(), input1.GetTensorInfo(), outputInfo); + ARMNN_NO_DEPRECATE_WARN_END }; if(!IsDynamicTensor(outputInfo)) @@ -3617,7 +3635,9 @@ bool Converter::ConvertMul(const Operation& operation, const Model& model, Conve return false; } + ARMNN_NO_DEPRECATE_WARN_BEGIN armnn::IConnectableLayer* const startLayer = data.m_Network->AddMultiplicationLayer(); + ARMNN_NO_DEPRECATE_WARN_END startLayer->SetBackendId(setBackend); bool isReshapeSupported = BroadcastTensor(input0, input1, startLayer, data); @@ -5340,6 +5360,7 @@ bool Converter::ConvertSub(const Operation& operation, const Model& model, Conve armnn::BackendId setBackend; auto validateFunc = [&](const armnn::TensorInfo& outputInfo, bool& isSupported) { + ARMNN_NO_DEPRECATE_WARN_BEGIN FORWARD_LAYER_SUPPORT_FUNC(__func__, IsSubtractionSupported, data.m_Backends, @@ -5348,6 +5369,7 @@ bool Converter::ConvertSub(const Operation& operation, const Model& model, Conve input0.GetTensorInfo(), input1.GetTensorInfo(), outputInfo); + ARMNN_NO_DEPRECATE_WARN_END }; if(IsDynamicTensor(outputInfo)) @@ -5364,7 +5386,9 @@ bool Converter::ConvertSub(const Operation& operation, const Model& model, Conve return false; } + ARMNN_NO_DEPRECATE_WARN_BEGIN armnn::IConnectableLayer* const startLayer = data.m_Network->AddSubtractionLayer(); + ARMNN_NO_DEPRECATE_WARN_END startLayer->SetBackendId(setBackend); bool isReshapeSupported = BroadcastTensor(input0, input1, startLayer, data); diff --git a/src/armnn/Network.cpp b/src/armnn/Network.cpp index 9ebb67b593..4b89daf977 100644 --- a/src/armnn/Network.cpp +++ b/src/armnn/Network.cpp @@ -213,12 +213,16 @@ IConnectableLayer* INetwork::AddMergeLayer(const char* name) IConnectableLayer* INetwork::AddAdditionLayer(const char* name) { + ARMNN_NO_DEPRECATE_WARN_BEGIN return pNetworkImpl->AddAdditionLayer(name); + ARMNN_NO_DEPRECATE_WARN_END } IConnectableLayer* INetwork::AddMultiplicationLayer(const char* name) { + ARMNN_NO_DEPRECATE_WARN_BEGIN return pNetworkImpl->AddMultiplicationLayer(name); + ARMNN_NO_DEPRECATE_WARN_END } IConnectableLayer* INetwork::AddBatchNormalizationLayer(const BatchNormalizationDescriptor& desc, @@ -308,17 +312,23 @@ IConnectableLayer* INetwork::AddLstmLayer(const LstmDescriptor& descriptor, IConnectableLayer* INetwork::AddDivisionLayer(const char* name) { + ARMNN_NO_DEPRECATE_WARN_BEGIN return pNetworkImpl->AddDivisionLayer(name); + ARMNN_NO_DEPRECATE_WARN_END } IConnectableLayer* INetwork::AddSubtractionLayer(const char* name) { + ARMNN_NO_DEPRECATE_WARN_BEGIN return pNetworkImpl->AddSubtractionLayer(name); + ARMNN_NO_DEPRECATE_WARN_END } IConnectableLayer* INetwork::AddMaximumLayer(const char* name) { + ARMNN_NO_DEPRECATE_WARN_BEGIN return pNetworkImpl->AddMaximumLayer(name); + ARMNN_NO_DEPRECATE_WARN_END } IConnectableLayer* INetwork::AddMeanLayer(const MeanDescriptor& meanDescriptor, const char* name) @@ -345,7 +355,9 @@ IConnectableLayer* INetwork::AddStridedSliceLayer(const StridedSliceDescriptor& IConnectableLayer* INetwork::AddMinimumLayer(const char* name) { + ARMNN_NO_DEPRECATE_WARN_BEGIN return pNetworkImpl->AddMinimumLayer(name); + ARMNN_NO_DEPRECATE_WARN_END } IConnectableLayer* INetwork::AddGatherLayer(const GatherDescriptor& descriptor, diff --git a/src/armnn/Network.hpp b/src/armnn/Network.hpp index 03642ce993..c6bf0859f7 100644 --- a/src/armnn/Network.hpp +++ b/src/armnn/Network.hpp @@ -43,7 +43,7 @@ public: IConnectableLayer* AddActivationLayer(const ActivationDescriptor& activationDescriptor, const char* name = nullptr); - + ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use AddElementwiseBinaryLayer instead", "24.02") IConnectableLayer* AddAdditionLayer(const char* name = nullptr); IConnectableLayer* AddArgMinMaxLayer(const ArgMinMaxDescriptor& desc, @@ -93,6 +93,7 @@ public: const ConstTensor& anchors, const char* name = nullptr); + ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use AddElementwiseBinaryLayer instead", "24.02") IConnectableLayer* AddDivisionLayer(const char* name = nullptr); IConnectableLayer* AddElementwiseBinaryLayer(const ElementwiseBinaryDescriptor& elementwiseBinaryDescriptor, @@ -132,12 +133,15 @@ public: const LstmInputParams& params, const char* name = nullptr); + ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use AddElementwiseBinaryLayer instead", "24.02") IConnectableLayer* AddMaximumLayer(const char* name = nullptr); IConnectableLayer* AddMeanLayer(const MeanDescriptor& meanDescriptor, const char* name = nullptr); + ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use AddElementwiseBinaryLayer instead", "24.02") IConnectableLayer* AddMinimumLayer(const char* name = nullptr); + ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use AddElementwiseBinaryLayer instead", "24.02") IConnectableLayer* AddMultiplicationLayer(const char* name = nullptr); IConnectableLayer* AddNormalizationLayer(const NormalizationDescriptor& normalizationDescriptor, @@ -208,6 +212,7 @@ public: IConnectableLayer* AddStridedSliceLayer(const StridedSliceDescriptor& stridedSliceDescriptor, const char* name = nullptr); + ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use AddElementwiseBinaryLayer instead", "24.02") IConnectableLayer* AddSubtractionLayer(const char* name = nullptr); IConnectableLayer* AddSwitchLayer(const char* name = nullptr); diff --git a/src/armnn/test/RuntimeTests.cpp b/src/armnn/test/RuntimeTests.cpp index e0d3a222fe..6b3fe0f211 100644 --- a/src/armnn/test/RuntimeTests.cpp +++ b/src/armnn/test/RuntimeTests.cpp @@ -1,5 +1,5 @@ // -// Copyright © 2017 Arm Ltd and Contributors. All rights reserved. +// Copyright © 2017-2023 Arm Ltd and Contributors. All rights reserved. // SPDX-License-Identifier: MIT // @@ -10,7 +10,6 @@ #include #include #include -#include #include @@ -19,9 +18,6 @@ #include -#include -#include - #ifdef WITH_VALGRIND #include #endif @@ -76,7 +72,9 @@ TEST_CASE("RuntimePreImportInputs") auto inputLayer1 = testNetwork->AddInputLayer(0, "input 1 layer"); auto inputLayer2 = testNetwork->AddInputLayer(1, "input 2 layer"); + ARMNN_NO_DEPRECATE_WARN_BEGIN auto addLayer = testNetwork->AddAdditionLayer("add layer"); + ARMNN_NO_DEPRECATE_WARN_END auto outputLayer = testNetwork->AddOutputLayer(2, "output layer"); TensorInfo tensorInfo{{4}, armnn::DataType::Signed32}; @@ -1306,7 +1304,9 @@ TEST_CASE("RuntimeOptimizeImportOff_LoadNetworkImportOn") auto inputLayer1 = testNetwork->AddInputLayer(0, "input 1 layer"); auto inputLayer2 = testNetwork->AddInputLayer(1, "input 2 layer"); + ARMNN_NO_DEPRECATE_WARN_BEGIN auto addLayer = testNetwork->AddAdditionLayer("add layer"); + ARMNN_NO_DEPRECATE_WARN_END auto outputLayer = testNetwork->AddOutputLayer(2, "output layer"); TensorInfo tensorInfo{{4}, armnn::DataType::Signed32}; @@ -1349,7 +1349,9 @@ TEST_CASE("RuntimeOptimizeExportOff_LoadNetworkExportOn") auto inputLayer1 = testNetwork->AddInputLayer(0, "input 1 layer"); auto inputLayer2 = testNetwork->AddInputLayer(1, "input 2 layer"); + ARMNN_NO_DEPRECATE_WARN_BEGIN auto addLayer = testNetwork->AddAdditionLayer("add layer"); + ARMNN_NO_DEPRECATE_WARN_END auto outputLayer = testNetwork->AddOutputLayer(2, "output layer"); TensorInfo tensorInfo{{4}, armnn::DataType::Signed32}; @@ -1392,7 +1394,9 @@ TEST_CASE("RuntimeOptimizeImportOn_LoadNetworkImportOff") auto inputLayer1 = testNetwork->AddInputLayer(0, "input 1 layer"); auto inputLayer2 = testNetwork->AddInputLayer(1, "input 2 layer"); + ARMNN_NO_DEPRECATE_WARN_BEGIN auto addLayer = testNetwork->AddAdditionLayer("add layer"); + ARMNN_NO_DEPRECATE_WARN_END auto outputLayer = testNetwork->AddOutputLayer(2, "output layer"); TensorInfo tensorInfo{{4}, armnn::DataType::Signed32}; @@ -1435,7 +1439,9 @@ TEST_CASE("RuntimeOptimizeExportOn_LoadNetworkExportOff") auto inputLayer1 = testNetwork->AddInputLayer(0, "input 1 layer"); auto inputLayer2 = testNetwork->AddInputLayer(1, "input 2 layer"); + ARMNN_NO_DEPRECATE_WARN_BEGIN auto addLayer = testNetwork->AddAdditionLayer("add layer"); + ARMNN_NO_DEPRECATE_WARN_END auto outputLayer = testNetwork->AddOutputLayer(2, "output layer"); TensorInfo tensorInfo{{4}, armnn::DataType::Signed32}; @@ -1483,7 +1489,9 @@ TEST_CASE("SyncExecutePreImportInputsHappyPath") auto inputLayer1 = testNetwork->AddInputLayer(0, "input 1 layer"); auto inputLayer2 = testNetwork->AddInputLayer(1, "input 2 layer"); + ARMNN_NO_DEPRECATE_WARN_BEGIN auto addLayer = testNetwork->AddAdditionLayer("add layer"); + ARMNN_NO_DEPRECATE_WARN_END auto outputLayer = testNetwork->AddOutputLayer(2, "output layer"); TensorInfo tensorInfo{ { 4 }, armnn::DataType::Signed32 }; diff --git a/src/armnn/test/TestNameOnlyLayerVisitor.cpp b/src/armnn/test/TestNameOnlyLayerVisitor.cpp index 497c36b079..eb488a5bcb 100644 --- a/src/armnn/test/TestNameOnlyLayerVisitor.cpp +++ b/src/armnn/test/TestNameOnlyLayerVisitor.cpp @@ -1,5 +1,5 @@ // -// Copyright © 2017 Arm Ltd and Contributors. All rights reserved. +// Copyright © 2017,2019-2021,2023 Arm Ltd and Contributors. All rights reserved. // SPDX-License-Identifier: MIT // @@ -34,31 +34,40 @@ TEST_CASE(#testName) \ TEST_SUITE("TestNameOnlyLayerVisitor") { +ARMNN_NO_DEPRECATE_WARN_BEGIN TEST_CASE_CHECK_LAYER_VISITOR_NAME(Addition, CheckAdditionLayerVisitorName) TEST_CASE_CHECK_LAYER_VISITOR_NAME_NULLPTR(Addition, CheckAdditionLayerVisitorNameNullptr) +ARMNN_NO_DEPRECATE_WARN_END TEST_CASE_CHECK_LAYER_VISITOR_NAME(Dequantize, CheckDequantizeLayerVisitorName) TEST_CASE_CHECK_LAYER_VISITOR_NAME_NULLPTR(Dequantize, CheckDequantizeLayerVisitorNameNullptr) +ARMNN_NO_DEPRECATE_WARN_BEGIN TEST_CASE_CHECK_LAYER_VISITOR_NAME(Division, CheckDivisionLayerVisitorName) TEST_CASE_CHECK_LAYER_VISITOR_NAME_NULLPTR(Division, CheckDivisionLayerVisitorNameNullptr) +ARMNN_NO_DEPRECATE_WARN_END TEST_CASE_CHECK_LAYER_VISITOR_NAME(Floor, CheckFloorLayerVisitorName) TEST_CASE_CHECK_LAYER_VISITOR_NAME_NULLPTR(Floor, CheckFloorLayerVisitorNameNullptr) +ARMNN_NO_DEPRECATE_WARN_BEGIN TEST_CASE_CHECK_LAYER_VISITOR_NAME(Maximum, CheckMaximumLayerVisitorName) TEST_CASE_CHECK_LAYER_VISITOR_NAME_NULLPTR(Maximum, CheckMaximumLayerVisitorNameNullptr) +ARMNN_NO_DEPRECATE_WARN_END TEST_CASE_CHECK_LAYER_VISITOR_NAME(Merge, CheckMergeLayerVisitorName) TEST_CASE_CHECK_LAYER_VISITOR_NAME_NULLPTR(Merge, CheckMergeLayerVisitorNameNullptr) +ARMNN_NO_DEPRECATE_WARN_BEGIN TEST_CASE_CHECK_LAYER_VISITOR_NAME(Minimum, CheckMinimumLayerVisitorName) TEST_CASE_CHECK_LAYER_VISITOR_NAME_NULLPTR(Minimum, CheckMinimumLayerVisitorNameNullptr) TEST_CASE_CHECK_LAYER_VISITOR_NAME(Multiplication, CheckMultiplicationLayerVisitorName) TEST_CASE_CHECK_LAYER_VISITOR_NAME_NULLPTR(Multiplication, CheckMultiplicationLayerVisitorNameNullptr) +ARMNN_NO_DEPRECATE_WARN_END TEST_CASE_CHECK_LAYER_VISITOR_NAME(Prelu, CheckPreluLayerVisitorName) TEST_CASE_CHECK_LAYER_VISITOR_NAME_NULLPTR(Prelu, CheckPreluLayerVisitorNameNullptr) TEST_CASE_CHECK_LAYER_VISITOR_NAME(Quantize, CheckQuantizeLayerVisitorName) TEST_CASE_CHECK_LAYER_VISITOR_NAME_NULLPTR(Quantize, CheckQuantizeLayerVisitorNameNullptr) TEST_CASE_CHECK_LAYER_VISITOR_NAME(Rank, CheckRankLayerVisitorName) TEST_CASE_CHECK_LAYER_VISITOR_NAME_NULLPTR(Rank, CheckRankLayerVisitorNameNullptr) +ARMNN_NO_DEPRECATE_WARN_BEGIN TEST_CASE_CHECK_LAYER_VISITOR_NAME(Subtraction, CheckSubtractionLayerVisitorName) TEST_CASE_CHECK_LAYER_VISITOR_NAME_NULLPTR(Subtraction, CheckSubtractionLayerVisitorNameNullptr) +ARMNN_NO_DEPRECATE_WARN_END TEST_CASE_CHECK_LAYER_VISITOR_NAME(Switch, CheckSwitchLayerVisitorName) TEST_CASE_CHECK_LAYER_VISITOR_NAME_NULLPTR(Switch, CheckSwitchLayerVisitorNameNullptr) - } diff --git a/src/armnn/test/optimizations/FuseActivationTests.cpp b/src/armnn/test/optimizations/FuseActivationTests.cpp index 3b8917192d..a82550780a 100644 --- a/src/armnn/test/optimizations/FuseActivationTests.cpp +++ b/src/armnn/test/optimizations/FuseActivationTests.cpp @@ -1,11 +1,10 @@ // -// Copyright © 2020 Arm Ltd and Contributors. All rights reserved. +// Copyright © 2020-2023 Arm Ltd and Contributors. All rights reserved. // SPDX-License-Identifier: MIT // #include "LayersFwd.hpp" -#include #include #include #include @@ -259,7 +258,9 @@ struct MultiplicationTest IgnoreUnused(scale); IgnoreUnused(offset); + ARMNN_NO_DEPRECATE_WARN_BEGIN return network->AddMultiplicationLayer(name); + ARMNN_NO_DEPRECATE_WARN_END } static std::vector AddConstantLayers(INetwork* network, @@ -294,7 +295,9 @@ struct AdditionTest IgnoreUnused(scale); IgnoreUnused(offset); + ARMNN_NO_DEPRECATE_WARN_BEGIN return network->AddAdditionLayer(name); + ARMNN_NO_DEPRECATE_WARN_END } static std::vector AddConstantLayers(INetwork* network, @@ -329,7 +332,9 @@ struct SubtractionTest IgnoreUnused(scale); IgnoreUnused(offset); + ARMNN_NO_DEPRECATE_WARN_BEGIN return network->AddSubtractionLayer(name); + ARMNN_NO_DEPRECATE_WARN_END } static std::vector AddConstantLayers(INetwork* network, @@ -364,7 +369,9 @@ struct DivisionTest IgnoreUnused(scale); IgnoreUnused(offset); + ARMNN_NO_DEPRECATE_WARN_BEGIN return network->AddDivisionLayer(name); + ARMNN_NO_DEPRECATE_WARN_END } static std::vector AddConstantLayers(INetwork* network, diff --git a/src/armnnSerializer/test/SerializerTests.cpp b/src/armnnSerializer/test/SerializerTests.cpp index 6ddc971f36..3998ee730d 100644 --- a/src/armnnSerializer/test/SerializerTests.cpp +++ b/src/armnnSerializer/test/SerializerTests.cpp @@ -32,7 +32,9 @@ TEST_CASE("SerializeAddition") armnn::INetworkPtr network = armnn::INetwork::Create(); armnn::IConnectableLayer* const inputLayer0 = network->AddInputLayer(0); armnn::IConnectableLayer* const inputLayer1 = network->AddInputLayer(1); + ARMNN_NO_DEPRECATE_WARN_BEGIN armnn::IConnectableLayer* const additionLayer = network->AddAdditionLayer(layerName.c_str()); + ARMNN_NO_DEPRECATE_WARN_END armnn::IConnectableLayer* const outputLayer = network->AddOutputLayer(0); inputLayer0->GetOutputSlot(0).Connect(additionLayer->GetInputSlot(0)); @@ -357,7 +359,9 @@ TEST_CASE("SerializeConstant") armnn::INetworkPtr network(armnn::INetwork::Create()); armnn::IConnectableLayer* input = network->AddInputLayer(0); armnn::IConnectableLayer* constant = network->AddConstantLayer(constTensor, layerName.c_str()); + ARMNN_NO_DEPRECATE_WARN_BEGIN armnn::IConnectableLayer* add = network->AddAdditionLayer(); + ARMNN_NO_DEPRECATE_WARN_END armnn::IConnectableLayer* output = network->AddOutputLayer(0); input->GetOutputSlot(0).Connect(add->GetInputSlot(0)); @@ -927,7 +931,9 @@ TEST_CASE("SerializeDivision") armnn::INetworkPtr network = armnn::INetwork::Create(); armnn::IConnectableLayer* const inputLayer0 = network->AddInputLayer(0); armnn::IConnectableLayer* const inputLayer1 = network->AddInputLayer(1); + ARMNN_NO_DEPRECATE_WARN_BEGIN armnn::IConnectableLayer* const divisionLayer = network->AddDivisionLayer(layerName.c_str()); + ARMNN_NO_DEPRECATE_WARN_END armnn::IConnectableLayer* const outputLayer = network->AddOutputLayer(0); inputLayer0->GetOutputSlot(0).Connect(divisionLayer->GetInputSlot(0)); @@ -1627,7 +1633,9 @@ TEST_CASE("SerializeMaximum") armnn::INetworkPtr network = armnn::INetwork::Create(); armnn::IConnectableLayer* const inputLayer0 = network->AddInputLayer(0); armnn::IConnectableLayer* const inputLayer1 = network->AddInputLayer(1); + ARMNN_NO_DEPRECATE_WARN_BEGIN armnn::IConnectableLayer* const maximumLayer = network->AddMaximumLayer(layerName.c_str()); + ARMNN_NO_DEPRECATE_WARN_END armnn::IConnectableLayer* const outputLayer = network->AddOutputLayer(0); inputLayer0->GetOutputSlot(0).Connect(maximumLayer->GetInputSlot(0)); @@ -1852,7 +1860,9 @@ TEST_CASE("SerializeMinimum") armnn::INetworkPtr network = armnn::INetwork::Create(); armnn::IConnectableLayer* const inputLayer0 = network->AddInputLayer(0); armnn::IConnectableLayer* const inputLayer1 = network->AddInputLayer(1); + ARMNN_NO_DEPRECATE_WARN_BEGIN armnn::IConnectableLayer* const minimumLayer = network->AddMinimumLayer(layerName.c_str()); + ARMNN_NO_DEPRECATE_WARN_END armnn::IConnectableLayer* const outputLayer = network->AddOutputLayer(0); inputLayer0->GetOutputSlot(0).Connect(minimumLayer->GetInputSlot(0)); @@ -1878,7 +1888,9 @@ TEST_CASE("SerializeMultiplication") armnn::INetworkPtr network = armnn::INetwork::Create(); armnn::IConnectableLayer* const inputLayer0 = network->AddInputLayer(0); armnn::IConnectableLayer* const inputLayer1 = network->AddInputLayer(1); + ARMNN_NO_DEPRECATE_WARN_BEGIN armnn::IConnectableLayer* const multiplicationLayer = network->AddMultiplicationLayer(layerName.c_str()); + ARMNN_NO_DEPRECATE_WARN_END armnn::IConnectableLayer* const outputLayer = network->AddOutputLayer(0); inputLayer0->GetOutputSlot(0).Connect(multiplicationLayer->GetInputSlot(0)); @@ -2736,7 +2748,9 @@ TEST_CASE("SerializeSubtraction") armnn::INetworkPtr network = armnn::INetwork::Create(); armnn::IConnectableLayer* const inputLayer0 = network->AddInputLayer(0); armnn::IConnectableLayer* const inputLayer1 = network->AddInputLayer(1); + ARMNN_NO_DEPRECATE_WARN_BEGIN armnn::IConnectableLayer* const subtractionLayer = network->AddSubtractionLayer(layerName.c_str()); + ARMNN_NO_DEPRECATE_WARN_END armnn::IConnectableLayer* const outputLayer = network->AddOutputLayer(0); inputLayer0->GetOutputSlot(0).Connect(subtractionLayer->GetInputSlot(0)); @@ -2945,7 +2959,9 @@ TEST_CASE("SerializeDeserializeNonLinearNetwork") armnn::INetworkPtr network(armnn::INetwork::Create()); armnn::IConnectableLayer* input = network->AddInputLayer(0); + ARMNN_NO_DEPRECATE_WARN_BEGIN armnn::IConnectableLayer* add = network->AddAdditionLayer(); + ARMNN_NO_DEPRECATE_WARN_END armnn::IConnectableLayer* constant = network->AddConstantLayer(constTensor, layerName.c_str()); armnn::IConnectableLayer* output = network->AddOutputLayer(0); diff --git a/src/backends/aclCommon/ArmComputeSubgraphUtils.hpp b/src/backends/aclCommon/ArmComputeSubgraphUtils.hpp index 599d3538eb..973d104195 100644 --- a/src/backends/aclCommon/ArmComputeSubgraphUtils.hpp +++ b/src/backends/aclCommon/ArmComputeSubgraphUtils.hpp @@ -77,7 +77,9 @@ LayerType* FuseAdditionLayer(OptimizationViews& optimizationViews, ActivationDescriptor& activationDesc, std::string name) { + ARMNN_NO_DEPRECATE_WARN_BEGIN IConnectableLayer* replacement = optimizationViews.GetINetwork()->AddAdditionLayer(name.c_str()); + ARMNN_NO_DEPRECATE_WARN_END LayerType* replacementLayer = PolymorphicDowncast(replacement); FuseLayer(optimizationViews, @@ -96,7 +98,9 @@ LayerType* FuseSubtractionLayer(OptimizationViews& optimizationViews, ActivationDescriptor& activationDesc, std::string name) { + ARMNN_NO_DEPRECATE_WARN_BEGIN IConnectableLayer* replacement = optimizationViews.GetINetwork()->AddSubtractionLayer(name.c_str()); + ARMNN_NO_DEPRECATE_WARN_END LayerType* replacementLayer = PolymorphicDowncast(replacement); FuseLayer(optimizationViews, @@ -115,7 +119,9 @@ LayerType* FuseDivisionLayer(OptimizationViews& optimizationViews, ActivationDescriptor& activationDesc, std::string name) { + ARMNN_NO_DEPRECATE_WARN_BEGIN IConnectableLayer* replacement = optimizationViews.GetINetwork()->AddDivisionLayer(name.c_str()); + ARMNN_NO_DEPRECATE_WARN_END LayerType* replacementLayer = PolymorphicDowncast(replacement); FuseLayer(optimizationViews, @@ -134,7 +140,9 @@ LayerType* FuseMultiplicationLayer(OptimizationViews& optimizationViews, ActivationDescriptor& activationDesc, std::string name) { + ARMNN_NO_DEPRECATE_WARN_BEGIN IConnectableLayer* replacement = optimizationViews.GetINetwork()->AddMultiplicationLayer(name.c_str()); + ARMNN_NO_DEPRECATE_WARN_END LayerType* replacementLayer = PolymorphicDowncast(replacement); FuseLayer(optimizationViews, diff --git a/src/backends/backendsCommon/test/AdditionEndToEndTestImpl.hpp b/src/backends/backendsCommon/test/AdditionEndToEndTestImpl.hpp index f33521888f..c850c518f5 100644 --- a/src/backends/backendsCommon/test/AdditionEndToEndTestImpl.hpp +++ b/src/backends/backendsCommon/test/AdditionEndToEndTestImpl.hpp @@ -30,8 +30,9 @@ armnn::INetworkPtr CreateAdditionNetwork(const armnn::TensorShape& inputXShape, TensorInfo outputTensorInfo(outputShape, DataType, qScale, qOffset); - + ARMNN_NO_DEPRECATE_WARN_BEGIN IConnectableLayer* addition = network->AddAdditionLayer("addition"); + ARMNN_NO_DEPRECATE_WARN_END IConnectableLayer* inputX = network->AddInputLayer(0, "inputX"); IConnectableLayer* inputY = network->AddInputLayer(1, "inputY"); IConnectableLayer* output = network->AddOutputLayer(0, "output"); diff --git a/src/backends/backendsCommon/test/EndToEndTestImpl.hpp b/src/backends/backendsCommon/test/EndToEndTestImpl.hpp index 795fc13c32..73cef16aad 100644 --- a/src/backends/backendsCommon/test/EndToEndTestImpl.hpp +++ b/src/backends/backendsCommon/test/EndToEndTestImpl.hpp @@ -39,7 +39,9 @@ bool ConstantUsageTest(const std::vector& computeDevice, IConnectableLayer* input = net->AddInputLayer(0); IConnectableLayer* constant = net->AddConstantLayer(ConstTensor(commonTensorInfo, constantData)); + ARMNN_NO_DEPRECATE_WARN_BEGIN IConnectableLayer* add = net->AddAdditionLayer(); + ARMNN_NO_DEPRECATE_WARN_END IConnectableLayer* output = net->AddOutputLayer(0); input->GetOutputSlot(0).Connect(add->GetInputSlot(0)); diff --git a/src/backends/backendsCommon/test/IsLayerSupportedTestImpl.hpp b/src/backends/backendsCommon/test/IsLayerSupportedTestImpl.hpp index 5b95d3cd92..5475762a53 100644 --- a/src/backends/backendsCommon/test/IsLayerSupportedTestImpl.hpp +++ b/src/backends/backendsCommon/test/IsLayerSupportedTestImpl.hpp @@ -622,7 +622,9 @@ struct LayerTypePolicy; // Every entry in the armnn::LayerType enum must be accounted for below. DECLARE_LAYER_POLICY_2_PARAM(Activation) +ARMNN_NO_DEPRECATE_WARN_BEGIN DECLARE_LAYER_POLICY_1_PARAM(Addition) +ARMNN_NO_DEPRECATE_WARN_END DECLARE_LAYER_POLICY_2_PARAM(ArgMinMax) @@ -694,15 +696,21 @@ DECLARE_LAYER_POLICY_2_PARAM(Lstm) DECLARE_LAYER_POLICY_MAP_PARAM(Map, void) +ARMNN_NO_DEPRECATE_WARN_BEGIN DECLARE_LAYER_POLICY_1_PARAM(Maximum) +ARMNN_NO_DEPRECATE_WARN_END DECLARE_LAYER_POLICY_2_PARAM(Mean) DECLARE_LAYER_POLICY_1_PARAM(Merge) +ARMNN_NO_DEPRECATE_WARN_BEGIN DECLARE_LAYER_POLICY_1_PARAM(Minimum) +ARMNN_NO_DEPRECATE_WARN_END +ARMNN_NO_DEPRECATE_WARN_BEGIN DECLARE_LAYER_POLICY_1_PARAM(Multiplication) +ARMNN_NO_DEPRECATE_WARN_END DECLARE_LAYER_POLICY_2_PARAM(Normalization) @@ -726,7 +734,9 @@ DECLARE_LAYER_POLICY_2_PARAM(QLstm) DECLARE_LAYER_POLICY_1_PARAM(QuantizedLstm) +ARMNN_NO_DEPRECATE_WARN_BEGIN DECLARE_LAYER_POLICY_1_PARAM(Division) +ARMNN_NO_DEPRECATE_WARN_END DECLARE_LAYER_POLICY_1_PARAM(Rank) @@ -752,7 +762,9 @@ DECLARE_LAYER_POLICY_EXCEPTION_2_PARAM(StandIn) DECLARE_LAYER_POLICY_2_PARAM(StridedSlice) +ARMNN_NO_DEPRECATE_WARN_BEGIN DECLARE_LAYER_POLICY_1_PARAM(Subtraction) +ARMNN_NO_DEPRECATE_WARN_END DECLARE_LAYER_POLICY_2_PARAM(Reduce) diff --git a/src/backends/backendsCommon/test/OptimizationViewsTests.cpp b/src/backends/backendsCommon/test/OptimizationViewsTests.cpp index ff3217911a..665358b9c6 100644 --- a/src/backends/backendsCommon/test/OptimizationViewsTests.cpp +++ b/src/backends/backendsCommon/test/OptimizationViewsTests.cpp @@ -263,7 +263,9 @@ TEST_CASE("OptimizeViewsValidateDeviceMockBackend") armnn::IConnectableLayer* input = net->AddInputLayer(0, "inLayer0"); armnn::IConnectableLayer* input1 = net->AddInputLayer(1, "inLayer1"); + ARMNN_NO_DEPRECATE_WARN_BEGIN armnn::IConnectableLayer* addition = net->AddAdditionLayer("addLayer"); + ARMNN_NO_DEPRECATE_WARN_END armnn::IConnectableLayer* output = net->AddOutputLayer(0, "outLayer"); diff --git a/src/backends/cl/ClLayerSupport.cpp b/src/backends/cl/ClLayerSupport.cpp index 89bcf9bc01..b63837539e 100644 --- a/src/backends/cl/ClLayerSupport.cpp +++ b/src/backends/cl/ClLayerSupport.cpp @@ -196,7 +196,9 @@ bool ClLayerSupport::IsLayerSupported(const LayerType& type, *(PolymorphicDowncast(&descriptor)), reasonIfUnsupported); case LayerType::Addition: + ARMNN_NO_DEPRECATE_WARN_BEGIN return IsAdditionSupported(infos[0], infos[1], infos[2], reasonIfUnsupported); + ARMNN_NO_DEPRECATE_WARN_END case LayerType::ArgMinMax: return IsArgMinMaxSupported(infos[0], infos[1], @@ -346,7 +348,9 @@ bool ClLayerSupport::IsLayerSupported(const LayerType& type, case LayerType::Dequantize: return IsDequantizeSupported(infos[0], infos[1], reasonIfUnsupported); case LayerType::Division: + ARMNN_NO_DEPRECATE_WARN_BEGIN return IsDivisionSupported(infos[0], infos[1], infos[2], reasonIfUnsupported); + ARMNN_NO_DEPRECATE_WARN_END case LayerType::ElementwiseBinary: { auto desc = *(PolymorphicDowncast(&descriptor)); @@ -474,16 +478,22 @@ bool ClLayerSupport::IsLayerSupported(const LayerType& type, infos[2], reasonIfUnsupported); case LayerType::Maximum: + ARMNN_NO_DEPRECATE_WARN_BEGIN return IsMaximumSupported(infos[0], infos[1], infos[2], reasonIfUnsupported); + ARMNN_NO_DEPRECATE_WARN_END case LayerType::Mean: return IsMeanSupported(infos[0], infos[1], *(PolymorphicDowncast(&descriptor)), reasonIfUnsupported); case LayerType::Minimum: + ARMNN_NO_DEPRECATE_WARN_BEGIN return IsMinimumSupported(infos[0], infos[1], infos[2], reasonIfUnsupported); + ARMNN_NO_DEPRECATE_WARN_END case LayerType::Multiplication: + ARMNN_NO_DEPRECATE_WARN_BEGIN return IsMultiplicationSupported(infos[0], infos[1], infos[2], reasonIfUnsupported); + ARMNN_NO_DEPRECATE_WARN_END case LayerType::Normalization: return IsNormalizationSupported(infos[0], infos[1], @@ -604,7 +614,9 @@ bool ClLayerSupport::IsLayerSupported(const LayerType& type, *(PolymorphicDowncast(&descriptor)), reasonIfUnsupported); case LayerType::Subtraction: + ARMNN_NO_DEPRECATE_WARN_BEGIN return IsSubtractionSupported(infos[0], infos[1], infos[2], reasonIfUnsupported); + ARMNN_NO_DEPRECATE_WARN_END case LayerType::Transpose: return IsTransposeSupported(infos[0], infos[1], diff --git a/src/backends/cl/ClLayerSupport.hpp b/src/backends/cl/ClLayerSupport.hpp index 2d784e3df8..2b3593630a 100644 --- a/src/backends/cl/ClLayerSupport.hpp +++ b/src/backends/cl/ClLayerSupport.hpp @@ -1,5 +1,5 @@ // -// Copyright © 2017 Arm Ltd and Contributors. All rights reserved. +// Copyright © 2017-2023 Arm Ltd and Contributors. All rights reserved. // SPDX-License-Identifier: MIT // #pragma once @@ -30,6 +30,7 @@ public: const ActivationDescriptor& descriptor, Optional reasonIfUnsupported = EmptyOptional()) const override; + ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use IsLayerSupported instead", "24.02") bool IsAdditionSupported(const TensorInfo& input0, const TensorInfo& input1, const TensorInfo& output, @@ -128,13 +129,14 @@ public: const Optional& biases, Optional reason = EmptyOptional()) const override; + ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use IsLayerSupported instead", "24.02") bool IsDivisionSupported(const TensorInfo& input0, const TensorInfo& input1, const TensorInfo& output, Optional reasonIfUnsupported = EmptyOptional()) const override; bool IsElementwiseUnarySupported(const TensorInfo& input, - const TensorInfo& ouput, + const TensorInfo& output, const ElementwiseUnaryDescriptor& descriptor, Optional reasonIfUnsupported = EmptyOptional()) const override; @@ -200,6 +202,7 @@ public: const LstmInputParamsInfo& paramsInfo, Optional reasonIfUnsupported = EmptyOptional()) const override; + ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use IsLayerSupported instead", "24.02") bool IsMaximumSupported(const TensorInfo& input0, const TensorInfo& input1, const TensorInfo& output, @@ -210,11 +213,13 @@ public: const MeanDescriptor& descriptor, Optional reasonIfUnsupported = EmptyOptional()) const override; + ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use IsLayerSupported instead", "24.02") bool IsMinimumSupported(const TensorInfo& input0, const TensorInfo& input1, const TensorInfo& output, Optional reasonIfUnsupported = EmptyOptional()) const override; + ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use IsLayerSupported instead", "24.02") bool IsMultiplicationSupported(const TensorInfo& input0, const TensorInfo& input1, const TensorInfo& output, @@ -325,6 +330,7 @@ public: const StridedSliceDescriptor& descriptor, Optional reasonIfUnsupported = EmptyOptional()) const override; + ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use IsLayerSupported instead", "24.02") bool IsSubtractionSupported(const TensorInfo& input0, const TensorInfo& input1, const TensorInfo& output, -- cgit v1.2.1