aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorMike Kelly <mike.kelly@arm.com>2023-03-15 15:06:23 +0000
committerTeresaARM <teresa.charlinreyes@arm.com>2023-04-03 16:30:45 +0000
commit2c14db6f5747a279b7bf7dd2244ac3c82b86656c (patch)
tree0932c959351ebe03e93d9803b83b3dec2645916b
parent7be50dca00995c290b4eefd75abed70bdbfb3d0f (diff)
downloadarmnn-2c14db6f5747a279b7bf7dd2244ac3c82b86656c.tar.gz
IVGCVSW-3808 Deprecation notices for old ElementwiseBinary layers
* Added Deprecation notices for old ElementwiseBinary layers. Signed-off-by: Mike Kelly <mike.kelly@arm.com> Change-Id: I5bd0f186aaed675885d667f47e1e210ee9ec84f8
-rw-r--r--include/armnn/BackendHelper.hpp6
-rw-r--r--include/armnn/INetwork.hpp6
-rw-r--r--samples/PreImportMemorySample.cpp2
-rw-r--r--shim/sl/canonical/Converter.cpp24
-rw-r--r--src/armnn/Network.cpp12
-rw-r--r--src/armnn/Network.hpp7
-rw-r--r--src/armnn/test/RuntimeTests.cpp18
-rw-r--r--src/armnn/test/TestNameOnlyLayerVisitor.cpp13
-rw-r--r--src/armnn/test/optimizations/FuseActivationTests.cpp11
-rw-r--r--src/armnnSerializer/test/SerializerTests.cpp16
-rw-r--r--src/backends/aclCommon/ArmComputeSubgraphUtils.hpp8
-rw-r--r--src/backends/backendsCommon/test/AdditionEndToEndTestImpl.hpp3
-rw-r--r--src/backends/backendsCommon/test/EndToEndTestImpl.hpp2
-rw-r--r--src/backends/backendsCommon/test/IsLayerSupportedTestImpl.hpp12
-rw-r--r--src/backends/backendsCommon/test/OptimizationViewsTests.cpp2
-rw-r--r--src/backends/cl/ClLayerSupport.cpp12
-rw-r--r--src/backends/cl/ClLayerSupport.hpp10
17 files changed, 151 insertions, 13 deletions
diff --git a/include/armnn/BackendHelper.hpp b/include/armnn/BackendHelper.hpp
index 85aabe0dd2..ddf2308da2 100644
--- a/include/armnn/BackendHelper.hpp
+++ b/include/armnn/BackendHelper.hpp
@@ -42,6 +42,7 @@ public:
const ActivationDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported = EmptyOptional());
+ ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use IsElementwiseBinarySupported instead", "24.02")
bool IsAdditionSupported(const TensorInfo& input0,
const TensorInfo& input1,
const TensorInfo& output,
@@ -156,6 +157,7 @@ public:
const Optional<TensorInfo>& biases,
Optional<std::string&> reasonIfUnsupported = EmptyOptional());
+ ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use IsElementwiseBinarySupported instead", "24.02")
bool IsDivisionSupported(const TensorInfo& input0,
const TensorInfo& input1,
const TensorInfo& output,
@@ -244,6 +246,7 @@ public:
const LstmInputParamsInfo& paramsInfo,
Optional<std::string&> reasonIfUnsupported = EmptyOptional());
+ ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use IsElementwiseBinarySupported instead", "24.02")
bool IsMaximumSupported(const TensorInfo& input0,
const TensorInfo& input1,
const TensorInfo& output,
@@ -267,11 +270,13 @@ public:
const TensorInfo& output,
Optional<std::string&> reasonIfUnsupported = EmptyOptional());
+ ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use IsElementwiseBinarySupported instead", "24.02")
bool IsMinimumSupported(const TensorInfo& input0,
const TensorInfo& input1,
const TensorInfo& output,
Optional<std::string&> reasonIfUnsupported = EmptyOptional());
+ ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use IsElementwiseBinarySupported instead", "24.02")
bool IsMultiplicationSupported(const TensorInfo& input0,
const TensorInfo& input1,
const TensorInfo& output,
@@ -400,6 +405,7 @@ public:
const StridedSliceDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported = EmptyOptional());
+ ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use IsElementwiseBinarySupported instead", "24.02")
bool IsSubtractionSupported(const TensorInfo& input0,
const TensorInfo& input1,
const TensorInfo& output,
diff --git a/include/armnn/INetwork.hpp b/include/armnn/INetwork.hpp
index 4eac0cfe2a..273753752d 100644
--- a/include/armnn/INetwork.hpp
+++ b/include/armnn/INetwork.hpp
@@ -481,11 +481,13 @@ public:
/// Adds an addition layer to the network.
/// @param name - Optional name for the layer.
/// @return - Interface for configuring the layer.
+ ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use AddElementwiseBinaryLayer instead", "24.02")
IConnectableLayer* AddAdditionLayer(const char* name = nullptr);
/// Adds a multiplication layer to the network.
/// @param name - Optional name for the layer.
/// @return - Interface for configuring the layer.
+ ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use AddElementwiseBinaryLayer instead", "24.02")
IConnectableLayer* AddMultiplicationLayer(const char* name = nullptr);
/// Adds a batch normalization layer to the network.
@@ -603,16 +605,19 @@ public:
/// Adds a division layer to the network.
/// @param name - Optional name for the layer.
/// @return - Interface for configuring the layer.
+ ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use AddElementwiseBinaryLayer instead", "24.02")
IConnectableLayer* AddDivisionLayer(const char* name = nullptr);
/// Adds a subtraction layer to the network.
/// @param name - Optional name for the layer.
/// @return - Interface for configuring the layer.
+ ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use AddElementwiseBinaryLayer instead", "24.02")
IConnectableLayer* AddSubtractionLayer(const char* name = nullptr);
/// Add a Maximum layer to the network.
/// @param name - Optional name for the layer.
/// @return - Interface for configuring the layer.
+ ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use AddElementwiseBinaryLayer instead", "24.02")
IConnectableLayer* AddMaximumLayer(const char* name = nullptr);
/// Add a Mean layer to the network.
@@ -645,6 +650,7 @@ public:
/// Add a Minimum layer to the network.
/// @param name - Optional name for the layer.
/// @return - Interface for configuring the layer.
+ ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use AddElementwiseBinaryLayer instead", "24.02")
IConnectableLayer* AddMinimumLayer(const char* name = nullptr);
/// Add Gather layer to the network.
diff --git a/samples/PreImportMemorySample.cpp b/samples/PreImportMemorySample.cpp
index cf196fd20e..98f386bfdf 100644
--- a/samples/PreImportMemorySample.cpp
+++ b/samples/PreImportMemorySample.cpp
@@ -31,7 +31,9 @@ int main()
armnn::INetworkPtr testNetwork(armnn::INetwork::Create());
auto inputLayer1 = testNetwork->AddInputLayer(0, "input 1 layer");
auto inputLayer2 = testNetwork->AddInputLayer(1, "input 2 layer");
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
auto addLayer = testNetwork->AddAdditionLayer("add layer");
+ ARMNN_NO_DEPRECATE_WARN_END
auto outputLayer = testNetwork->AddOutputLayer(2, "output layer");
// Set the tensors in the network.
diff --git a/shim/sl/canonical/Converter.cpp b/shim/sl/canonical/Converter.cpp
index be052a6faa..90fd71154f 100644
--- a/shim/sl/canonical/Converter.cpp
+++ b/shim/sl/canonical/Converter.cpp
@@ -209,6 +209,7 @@ bool Converter::ConvertAdd(const Operation& operation, const Model& model, Conve
armnn::BackendId setBackend;
auto validateFunc = [&](const armnn::TensorInfo& outputInfo, bool& isSupported)
{
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
FORWARD_LAYER_SUPPORT_FUNC(__func__,
IsAdditionSupported,
data.m_Backends,
@@ -217,6 +218,7 @@ bool Converter::ConvertAdd(const Operation& operation, const Model& model, Conve
inputInfo0,
inputInfo1,
outputInfo);
+ ARMNN_NO_DEPRECATE_WARN_END
};
if(!IsDynamicTensor(outputInfo))
@@ -233,7 +235,9 @@ bool Converter::ConvertAdd(const Operation& operation, const Model& model, Conve
return false;
}
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
armnn::IConnectableLayer* const startLayer = data.m_Network->AddAdditionLayer();
+ ARMNN_NO_DEPRECATE_WARN_END
startLayer->SetBackendId(setBackend);
bool isReshapeSupported = BroadcastTensor(input0, input1, startLayer, data);
@@ -1536,6 +1540,7 @@ bool Converter::ConvertDiv(const Operation& operation, const Model& model, Conve
armnn::BackendId setBackend;
auto validateFunc = [&](const armnn::TensorInfo& outputInfo, bool& isSupported)
{
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
FORWARD_LAYER_SUPPORT_FUNC(__func__,
IsDivisionSupported,
data.m_Backends,
@@ -1544,6 +1549,7 @@ bool Converter::ConvertDiv(const Operation& operation, const Model& model, Conve
input0.GetTensorInfo(),
input1.GetTensorInfo(),
outputInfo);
+ ARMNN_NO_DEPRECATE_WARN_END
};
if(!IsDynamicTensor(outputInfo))
@@ -1560,7 +1566,9 @@ bool Converter::ConvertDiv(const Operation& operation, const Model& model, Conve
return false;
}
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
armnn::IConnectableLayer* const startLayer = data.m_Network->AddDivisionLayer();
+ ARMNN_NO_DEPRECATE_WARN_END
startLayer->SetBackendId(setBackend);
bool isReshapeSupported = BroadcastTensor(input0, input1, startLayer, data);
@@ -3378,6 +3386,7 @@ bool Converter::ConvertMaximum(const Operation& operation, const Model& model, C
armnn::BackendId setBackend;
auto validateFunc = [&](const armnn::TensorInfo& outInfo, bool& isSupported)
{
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
FORWARD_LAYER_SUPPORT_FUNC(__func__,
IsMaximumSupported,
data.m_Backends,
@@ -3386,6 +3395,7 @@ bool Converter::ConvertMaximum(const Operation& operation, const Model& model, C
input0.GetTensorInfo(),
input1.GetTensorInfo(),
outInfo);
+ ARMNN_NO_DEPRECATE_WARN_END
};
if(IsDynamicTensor(outInfo))
@@ -3402,7 +3412,9 @@ bool Converter::ConvertMaximum(const Operation& operation, const Model& model, C
return false;
}
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
IConnectableLayer* layer = data.m_Network->AddMaximumLayer();
+ ARMNN_NO_DEPRECATE_WARN_END
layer->SetBackendId(setBackend);
assert(layer != nullptr);
bool isReshapeSupported = BroadcastTensor(input0, input1, layer, data);
@@ -3524,6 +3536,7 @@ bool Converter::ConvertMinimum(const Operation& operation, const Model& model, C
armnn::BackendId setBackend;
auto validateFunc = [&](const armnn::TensorInfo& outputInfo, bool& isSupported)
{
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
FORWARD_LAYER_SUPPORT_FUNC(__func__,
IsMinimumSupported,
data.m_Backends,
@@ -3532,6 +3545,7 @@ bool Converter::ConvertMinimum(const Operation& operation, const Model& model, C
input0.GetTensorInfo(),
input1.GetTensorInfo(),
outputInfo);
+ ARMNN_NO_DEPRECATE_WARN_END
};
if(IsDynamicTensor(outputInfo))
@@ -3548,7 +3562,9 @@ bool Converter::ConvertMinimum(const Operation& operation, const Model& model, C
return false;
}
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
IConnectableLayer* const layer = data.m_Network->AddMinimumLayer();
+ ARMNN_NO_DEPRECATE_WARN_END
layer->SetBackendId(setBackend);
assert(layer != nullptr);
bool isReshapeSupported = BroadcastTensor(input0, input1, layer, data);
@@ -3593,6 +3609,7 @@ bool Converter::ConvertMul(const Operation& operation, const Model& model, Conve
armnn::BackendId setBackend;
auto validateFunc = [&](const armnn::TensorInfo& outputInfo, bool& isSupported)
{
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
FORWARD_LAYER_SUPPORT_FUNC(__func__,
IsMultiplicationSupported,
data.m_Backends,
@@ -3601,6 +3618,7 @@ bool Converter::ConvertMul(const Operation& operation, const Model& model, Conve
input0.GetTensorInfo(),
input1.GetTensorInfo(),
outputInfo);
+ ARMNN_NO_DEPRECATE_WARN_END
};
if(!IsDynamicTensor(outputInfo))
@@ -3617,7 +3635,9 @@ bool Converter::ConvertMul(const Operation& operation, const Model& model, Conve
return false;
}
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
armnn::IConnectableLayer* const startLayer = data.m_Network->AddMultiplicationLayer();
+ ARMNN_NO_DEPRECATE_WARN_END
startLayer->SetBackendId(setBackend);
bool isReshapeSupported = BroadcastTensor(input0, input1, startLayer, data);
@@ -5340,6 +5360,7 @@ bool Converter::ConvertSub(const Operation& operation, const Model& model, Conve
armnn::BackendId setBackend;
auto validateFunc = [&](const armnn::TensorInfo& outputInfo, bool& isSupported)
{
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
FORWARD_LAYER_SUPPORT_FUNC(__func__,
IsSubtractionSupported,
data.m_Backends,
@@ -5348,6 +5369,7 @@ bool Converter::ConvertSub(const Operation& operation, const Model& model, Conve
input0.GetTensorInfo(),
input1.GetTensorInfo(),
outputInfo);
+ ARMNN_NO_DEPRECATE_WARN_END
};
if(IsDynamicTensor(outputInfo))
@@ -5364,7 +5386,9 @@ bool Converter::ConvertSub(const Operation& operation, const Model& model, Conve
return false;
}
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
armnn::IConnectableLayer* const startLayer = data.m_Network->AddSubtractionLayer();
+ ARMNN_NO_DEPRECATE_WARN_END
startLayer->SetBackendId(setBackend);
bool isReshapeSupported = BroadcastTensor(input0, input1, startLayer, data);
diff --git a/src/armnn/Network.cpp b/src/armnn/Network.cpp
index 9ebb67b593..4b89daf977 100644
--- a/src/armnn/Network.cpp
+++ b/src/armnn/Network.cpp
@@ -213,12 +213,16 @@ IConnectableLayer* INetwork::AddMergeLayer(const char* name)
IConnectableLayer* INetwork::AddAdditionLayer(const char* name)
{
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
return pNetworkImpl->AddAdditionLayer(name);
+ ARMNN_NO_DEPRECATE_WARN_END
}
IConnectableLayer* INetwork::AddMultiplicationLayer(const char* name)
{
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
return pNetworkImpl->AddMultiplicationLayer(name);
+ ARMNN_NO_DEPRECATE_WARN_END
}
IConnectableLayer* INetwork::AddBatchNormalizationLayer(const BatchNormalizationDescriptor& desc,
@@ -308,17 +312,23 @@ IConnectableLayer* INetwork::AddLstmLayer(const LstmDescriptor& descriptor,
IConnectableLayer* INetwork::AddDivisionLayer(const char* name)
{
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
return pNetworkImpl->AddDivisionLayer(name);
+ ARMNN_NO_DEPRECATE_WARN_END
}
IConnectableLayer* INetwork::AddSubtractionLayer(const char* name)
{
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
return pNetworkImpl->AddSubtractionLayer(name);
+ ARMNN_NO_DEPRECATE_WARN_END
}
IConnectableLayer* INetwork::AddMaximumLayer(const char* name)
{
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
return pNetworkImpl->AddMaximumLayer(name);
+ ARMNN_NO_DEPRECATE_WARN_END
}
IConnectableLayer* INetwork::AddMeanLayer(const MeanDescriptor& meanDescriptor, const char* name)
@@ -345,7 +355,9 @@ IConnectableLayer* INetwork::AddStridedSliceLayer(const StridedSliceDescriptor&
IConnectableLayer* INetwork::AddMinimumLayer(const char* name)
{
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
return pNetworkImpl->AddMinimumLayer(name);
+ ARMNN_NO_DEPRECATE_WARN_END
}
IConnectableLayer* INetwork::AddGatherLayer(const GatherDescriptor& descriptor,
diff --git a/src/armnn/Network.hpp b/src/armnn/Network.hpp
index 03642ce993..c6bf0859f7 100644
--- a/src/armnn/Network.hpp
+++ b/src/armnn/Network.hpp
@@ -43,7 +43,7 @@ public:
IConnectableLayer* AddActivationLayer(const ActivationDescriptor& activationDescriptor,
const char* name = nullptr);
-
+ ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use AddElementwiseBinaryLayer instead", "24.02")
IConnectableLayer* AddAdditionLayer(const char* name = nullptr);
IConnectableLayer* AddArgMinMaxLayer(const ArgMinMaxDescriptor& desc,
@@ -93,6 +93,7 @@ public:
const ConstTensor& anchors,
const char* name = nullptr);
+ ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use AddElementwiseBinaryLayer instead", "24.02")
IConnectableLayer* AddDivisionLayer(const char* name = nullptr);
IConnectableLayer* AddElementwiseBinaryLayer(const ElementwiseBinaryDescriptor& elementwiseBinaryDescriptor,
@@ -132,12 +133,15 @@ public:
const LstmInputParams& params,
const char* name = nullptr);
+ ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use AddElementwiseBinaryLayer instead", "24.02")
IConnectableLayer* AddMaximumLayer(const char* name = nullptr);
IConnectableLayer* AddMeanLayer(const MeanDescriptor& meanDescriptor, const char* name = nullptr);
+ ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use AddElementwiseBinaryLayer instead", "24.02")
IConnectableLayer* AddMinimumLayer(const char* name = nullptr);
+ ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use AddElementwiseBinaryLayer instead", "24.02")
IConnectableLayer* AddMultiplicationLayer(const char* name = nullptr);
IConnectableLayer* AddNormalizationLayer(const NormalizationDescriptor& normalizationDescriptor,
@@ -208,6 +212,7 @@ public:
IConnectableLayer* AddStridedSliceLayer(const StridedSliceDescriptor& stridedSliceDescriptor,
const char* name = nullptr);
+ ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use AddElementwiseBinaryLayer instead", "24.02")
IConnectableLayer* AddSubtractionLayer(const char* name = nullptr);
IConnectableLayer* AddSwitchLayer(const char* name = nullptr);
diff --git a/src/armnn/test/RuntimeTests.cpp b/src/armnn/test/RuntimeTests.cpp
index e0d3a222fe..6b3fe0f211 100644
--- a/src/armnn/test/RuntimeTests.cpp
+++ b/src/armnn/test/RuntimeTests.cpp
@@ -1,5 +1,5 @@
//
-// Copyright © 2017 Arm Ltd and Contributors. All rights reserved.
+// Copyright © 2017-2023 Arm Ltd and Contributors. All rights reserved.
// SPDX-License-Identifier: MIT
//
@@ -10,7 +10,6 @@
#include <armnn/Descriptors.hpp>
#include <armnn/IRuntime.hpp>
#include <armnn/INetwork.hpp>
-#include <armnn/TypesUtils.hpp>
#include <armnn/profiling/ArmNNProfiling.hpp>
@@ -19,9 +18,6 @@
#include <test/ProfilingTestUtils.hpp>
-#include <HeapProfiling.hpp>
-#include <LeakChecking.hpp>
-
#ifdef WITH_VALGRIND
#include <valgrind/memcheck.h>
#endif
@@ -76,7 +72,9 @@ TEST_CASE("RuntimePreImportInputs")
auto inputLayer1 = testNetwork->AddInputLayer(0, "input 1 layer");
auto inputLayer2 = testNetwork->AddInputLayer(1, "input 2 layer");
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
auto addLayer = testNetwork->AddAdditionLayer("add layer");
+ ARMNN_NO_DEPRECATE_WARN_END
auto outputLayer = testNetwork->AddOutputLayer(2, "output layer");
TensorInfo tensorInfo{{4}, armnn::DataType::Signed32};
@@ -1306,7 +1304,9 @@ TEST_CASE("RuntimeOptimizeImportOff_LoadNetworkImportOn")
auto inputLayer1 = testNetwork->AddInputLayer(0, "input 1 layer");
auto inputLayer2 = testNetwork->AddInputLayer(1, "input 2 layer");
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
auto addLayer = testNetwork->AddAdditionLayer("add layer");
+ ARMNN_NO_DEPRECATE_WARN_END
auto outputLayer = testNetwork->AddOutputLayer(2, "output layer");
TensorInfo tensorInfo{{4}, armnn::DataType::Signed32};
@@ -1349,7 +1349,9 @@ TEST_CASE("RuntimeOptimizeExportOff_LoadNetworkExportOn")
auto inputLayer1 = testNetwork->AddInputLayer(0, "input 1 layer");
auto inputLayer2 = testNetwork->AddInputLayer(1, "input 2 layer");
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
auto addLayer = testNetwork->AddAdditionLayer("add layer");
+ ARMNN_NO_DEPRECATE_WARN_END
auto outputLayer = testNetwork->AddOutputLayer(2, "output layer");
TensorInfo tensorInfo{{4}, armnn::DataType::Signed32};
@@ -1392,7 +1394,9 @@ TEST_CASE("RuntimeOptimizeImportOn_LoadNetworkImportOff")
auto inputLayer1 = testNetwork->AddInputLayer(0, "input 1 layer");
auto inputLayer2 = testNetwork->AddInputLayer(1, "input 2 layer");
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
auto addLayer = testNetwork->AddAdditionLayer("add layer");
+ ARMNN_NO_DEPRECATE_WARN_END
auto outputLayer = testNetwork->AddOutputLayer(2, "output layer");
TensorInfo tensorInfo{{4}, armnn::DataType::Signed32};
@@ -1435,7 +1439,9 @@ TEST_CASE("RuntimeOptimizeExportOn_LoadNetworkExportOff")
auto inputLayer1 = testNetwork->AddInputLayer(0, "input 1 layer");
auto inputLayer2 = testNetwork->AddInputLayer(1, "input 2 layer");
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
auto addLayer = testNetwork->AddAdditionLayer("add layer");
+ ARMNN_NO_DEPRECATE_WARN_END
auto outputLayer = testNetwork->AddOutputLayer(2, "output layer");
TensorInfo tensorInfo{{4}, armnn::DataType::Signed32};
@@ -1483,7 +1489,9 @@ TEST_CASE("SyncExecutePreImportInputsHappyPath")
auto inputLayer1 = testNetwork->AddInputLayer(0, "input 1 layer");
auto inputLayer2 = testNetwork->AddInputLayer(1, "input 2 layer");
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
auto addLayer = testNetwork->AddAdditionLayer("add layer");
+ ARMNN_NO_DEPRECATE_WARN_END
auto outputLayer = testNetwork->AddOutputLayer(2, "output layer");
TensorInfo tensorInfo{ { 4 }, armnn::DataType::Signed32 };
diff --git a/src/armnn/test/TestNameOnlyLayerVisitor.cpp b/src/armnn/test/TestNameOnlyLayerVisitor.cpp
index 497c36b079..eb488a5bcb 100644
--- a/src/armnn/test/TestNameOnlyLayerVisitor.cpp
+++ b/src/armnn/test/TestNameOnlyLayerVisitor.cpp
@@ -1,5 +1,5 @@
//
-// Copyright © 2017 Arm Ltd and Contributors. All rights reserved.
+// Copyright © 2017,2019-2021,2023 Arm Ltd and Contributors. All rights reserved.
// SPDX-License-Identifier: MIT
//
@@ -34,31 +34,40 @@ TEST_CASE(#testName) \
TEST_SUITE("TestNameOnlyLayerVisitor")
{
+ARMNN_NO_DEPRECATE_WARN_BEGIN
TEST_CASE_CHECK_LAYER_VISITOR_NAME(Addition, CheckAdditionLayerVisitorName)
TEST_CASE_CHECK_LAYER_VISITOR_NAME_NULLPTR(Addition, CheckAdditionLayerVisitorNameNullptr)
+ARMNN_NO_DEPRECATE_WARN_END
TEST_CASE_CHECK_LAYER_VISITOR_NAME(Dequantize, CheckDequantizeLayerVisitorName)
TEST_CASE_CHECK_LAYER_VISITOR_NAME_NULLPTR(Dequantize, CheckDequantizeLayerVisitorNameNullptr)
+ARMNN_NO_DEPRECATE_WARN_BEGIN
TEST_CASE_CHECK_LAYER_VISITOR_NAME(Division, CheckDivisionLayerVisitorName)
TEST_CASE_CHECK_LAYER_VISITOR_NAME_NULLPTR(Division, CheckDivisionLayerVisitorNameNullptr)
+ARMNN_NO_DEPRECATE_WARN_END
TEST_CASE_CHECK_LAYER_VISITOR_NAME(Floor, CheckFloorLayerVisitorName)
TEST_CASE_CHECK_LAYER_VISITOR_NAME_NULLPTR(Floor, CheckFloorLayerVisitorNameNullptr)
+ARMNN_NO_DEPRECATE_WARN_BEGIN
TEST_CASE_CHECK_LAYER_VISITOR_NAME(Maximum, CheckMaximumLayerVisitorName)
TEST_CASE_CHECK_LAYER_VISITOR_NAME_NULLPTR(Maximum, CheckMaximumLayerVisitorNameNullptr)
+ARMNN_NO_DEPRECATE_WARN_END
TEST_CASE_CHECK_LAYER_VISITOR_NAME(Merge, CheckMergeLayerVisitorName)
TEST_CASE_CHECK_LAYER_VISITOR_NAME_NULLPTR(Merge, CheckMergeLayerVisitorNameNullptr)
+ARMNN_NO_DEPRECATE_WARN_BEGIN
TEST_CASE_CHECK_LAYER_VISITOR_NAME(Minimum, CheckMinimumLayerVisitorName)
TEST_CASE_CHECK_LAYER_VISITOR_NAME_NULLPTR(Minimum, CheckMinimumLayerVisitorNameNullptr)
TEST_CASE_CHECK_LAYER_VISITOR_NAME(Multiplication, CheckMultiplicationLayerVisitorName)
TEST_CASE_CHECK_LAYER_VISITOR_NAME_NULLPTR(Multiplication, CheckMultiplicationLayerVisitorNameNullptr)
+ARMNN_NO_DEPRECATE_WARN_END
TEST_CASE_CHECK_LAYER_VISITOR_NAME(Prelu, CheckPreluLayerVisitorName)
TEST_CASE_CHECK_LAYER_VISITOR_NAME_NULLPTR(Prelu, CheckPreluLayerVisitorNameNullptr)
TEST_CASE_CHECK_LAYER_VISITOR_NAME(Quantize, CheckQuantizeLayerVisitorName)
TEST_CASE_CHECK_LAYER_VISITOR_NAME_NULLPTR(Quantize, CheckQuantizeLayerVisitorNameNullptr)
TEST_CASE_CHECK_LAYER_VISITOR_NAME(Rank, CheckRankLayerVisitorName)
TEST_CASE_CHECK_LAYER_VISITOR_NAME_NULLPTR(Rank, CheckRankLayerVisitorNameNullptr)
+ARMNN_NO_DEPRECATE_WARN_BEGIN
TEST_CASE_CHECK_LAYER_VISITOR_NAME(Subtraction, CheckSubtractionLayerVisitorName)
TEST_CASE_CHECK_LAYER_VISITOR_NAME_NULLPTR(Subtraction, CheckSubtractionLayerVisitorNameNullptr)
+ARMNN_NO_DEPRECATE_WARN_END
TEST_CASE_CHECK_LAYER_VISITOR_NAME(Switch, CheckSwitchLayerVisitorName)
TEST_CASE_CHECK_LAYER_VISITOR_NAME_NULLPTR(Switch, CheckSwitchLayerVisitorNameNullptr)
-
}
diff --git a/src/armnn/test/optimizations/FuseActivationTests.cpp b/src/armnn/test/optimizations/FuseActivationTests.cpp
index 3b8917192d..a82550780a 100644
--- a/src/armnn/test/optimizations/FuseActivationTests.cpp
+++ b/src/armnn/test/optimizations/FuseActivationTests.cpp
@@ -1,11 +1,10 @@
//
-// Copyright © 2020 Arm Ltd and Contributors. All rights reserved.
+// Copyright © 2020-2023 Arm Ltd and Contributors. All rights reserved.
// SPDX-License-Identifier: MIT
//
#include "LayersFwd.hpp"
-#include <Network.hpp>
#include <ResolveType.hpp>
#include <armnn/INetwork.hpp>
#include <GraphUtils.hpp>
@@ -259,7 +258,9 @@ struct MultiplicationTest
IgnoreUnused(scale);
IgnoreUnused(offset);
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
return network->AddMultiplicationLayer(name);
+ ARMNN_NO_DEPRECATE_WARN_END
}
static std::vector<IConnectableLayer*> AddConstantLayers(INetwork* network,
@@ -294,7 +295,9 @@ struct AdditionTest
IgnoreUnused(scale);
IgnoreUnused(offset);
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
return network->AddAdditionLayer(name);
+ ARMNN_NO_DEPRECATE_WARN_END
}
static std::vector<IConnectableLayer*> AddConstantLayers(INetwork* network,
@@ -329,7 +332,9 @@ struct SubtractionTest
IgnoreUnused(scale);
IgnoreUnused(offset);
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
return network->AddSubtractionLayer(name);
+ ARMNN_NO_DEPRECATE_WARN_END
}
static std::vector<IConnectableLayer*> AddConstantLayers(INetwork* network,
@@ -364,7 +369,9 @@ struct DivisionTest
IgnoreUnused(scale);
IgnoreUnused(offset);
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
return network->AddDivisionLayer(name);
+ ARMNN_NO_DEPRECATE_WARN_END
}
static std::vector<IConnectableLayer*> AddConstantLayers(INetwork* network,
diff --git a/src/armnnSerializer/test/SerializerTests.cpp b/src/armnnSerializer/test/SerializerTests.cpp
index 6ddc971f36..3998ee730d 100644
--- a/src/armnnSerializer/test/SerializerTests.cpp
+++ b/src/armnnSerializer/test/SerializerTests.cpp
@@ -32,7 +32,9 @@ TEST_CASE("SerializeAddition")
armnn::INetworkPtr network = armnn::INetwork::Create();
armnn::IConnectableLayer* const inputLayer0 = network->AddInputLayer(0);
armnn::IConnectableLayer* const inputLayer1 = network->AddInputLayer(1);
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
armnn::IConnectableLayer* const additionLayer = network->AddAdditionLayer(layerName.c_str());
+ ARMNN_NO_DEPRECATE_WARN_END
armnn::IConnectableLayer* const outputLayer = network->AddOutputLayer(0);
inputLayer0->GetOutputSlot(0).Connect(additionLayer->GetInputSlot(0));
@@ -357,7 +359,9 @@ TEST_CASE("SerializeConstant")
armnn::INetworkPtr network(armnn::INetwork::Create());
armnn::IConnectableLayer* input = network->AddInputLayer(0);
armnn::IConnectableLayer* constant = network->AddConstantLayer(constTensor, layerName.c_str());
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
armnn::IConnectableLayer* add = network->AddAdditionLayer();
+ ARMNN_NO_DEPRECATE_WARN_END
armnn::IConnectableLayer* output = network->AddOutputLayer(0);
input->GetOutputSlot(0).Connect(add->GetInputSlot(0));
@@ -927,7 +931,9 @@ TEST_CASE("SerializeDivision")
armnn::INetworkPtr network = armnn::INetwork::Create();
armnn::IConnectableLayer* const inputLayer0 = network->AddInputLayer(0);
armnn::IConnectableLayer* const inputLayer1 = network->AddInputLayer(1);
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
armnn::IConnectableLayer* const divisionLayer = network->AddDivisionLayer(layerName.c_str());
+ ARMNN_NO_DEPRECATE_WARN_END
armnn::IConnectableLayer* const outputLayer = network->AddOutputLayer(0);
inputLayer0->GetOutputSlot(0).Connect(divisionLayer->GetInputSlot(0));
@@ -1627,7 +1633,9 @@ TEST_CASE("SerializeMaximum")
armnn::INetworkPtr network = armnn::INetwork::Create();
armnn::IConnectableLayer* const inputLayer0 = network->AddInputLayer(0);
armnn::IConnectableLayer* const inputLayer1 = network->AddInputLayer(1);
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
armnn::IConnectableLayer* const maximumLayer = network->AddMaximumLayer(layerName.c_str());
+ ARMNN_NO_DEPRECATE_WARN_END
armnn::IConnectableLayer* const outputLayer = network->AddOutputLayer(0);
inputLayer0->GetOutputSlot(0).Connect(maximumLayer->GetInputSlot(0));
@@ -1852,7 +1860,9 @@ TEST_CASE("SerializeMinimum")
armnn::INetworkPtr network = armnn::INetwork::Create();
armnn::IConnectableLayer* const inputLayer0 = network->AddInputLayer(0);
armnn::IConnectableLayer* const inputLayer1 = network->AddInputLayer(1);
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
armnn::IConnectableLayer* const minimumLayer = network->AddMinimumLayer(layerName.c_str());
+ ARMNN_NO_DEPRECATE_WARN_END
armnn::IConnectableLayer* const outputLayer = network->AddOutputLayer(0);
inputLayer0->GetOutputSlot(0).Connect(minimumLayer->GetInputSlot(0));
@@ -1878,7 +1888,9 @@ TEST_CASE("SerializeMultiplication")
armnn::INetworkPtr network = armnn::INetwork::Create();
armnn::IConnectableLayer* const inputLayer0 = network->AddInputLayer(0);
armnn::IConnectableLayer* const inputLayer1 = network->AddInputLayer(1);
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
armnn::IConnectableLayer* const multiplicationLayer = network->AddMultiplicationLayer(layerName.c_str());
+ ARMNN_NO_DEPRECATE_WARN_END
armnn::IConnectableLayer* const outputLayer = network->AddOutputLayer(0);
inputLayer0->GetOutputSlot(0).Connect(multiplicationLayer->GetInputSlot(0));
@@ -2736,7 +2748,9 @@ TEST_CASE("SerializeSubtraction")
armnn::INetworkPtr network = armnn::INetwork::Create();
armnn::IConnectableLayer* const inputLayer0 = network->AddInputLayer(0);
armnn::IConnectableLayer* const inputLayer1 = network->AddInputLayer(1);
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
armnn::IConnectableLayer* const subtractionLayer = network->AddSubtractionLayer(layerName.c_str());
+ ARMNN_NO_DEPRECATE_WARN_END
armnn::IConnectableLayer* const outputLayer = network->AddOutputLayer(0);
inputLayer0->GetOutputSlot(0).Connect(subtractionLayer->GetInputSlot(0));
@@ -2945,7 +2959,9 @@ TEST_CASE("SerializeDeserializeNonLinearNetwork")
armnn::INetworkPtr network(armnn::INetwork::Create());
armnn::IConnectableLayer* input = network->AddInputLayer(0);
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
armnn::IConnectableLayer* add = network->AddAdditionLayer();
+ ARMNN_NO_DEPRECATE_WARN_END
armnn::IConnectableLayer* constant = network->AddConstantLayer(constTensor, layerName.c_str());
armnn::IConnectableLayer* output = network->AddOutputLayer(0);
diff --git a/src/backends/aclCommon/ArmComputeSubgraphUtils.hpp b/src/backends/aclCommon/ArmComputeSubgraphUtils.hpp
index 599d3538eb..973d104195 100644
--- a/src/backends/aclCommon/ArmComputeSubgraphUtils.hpp
+++ b/src/backends/aclCommon/ArmComputeSubgraphUtils.hpp
@@ -77,7 +77,9 @@ LayerType* FuseAdditionLayer(OptimizationViews& optimizationViews,
ActivationDescriptor& activationDesc,
std::string name)
{
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
IConnectableLayer* replacement = optimizationViews.GetINetwork()->AddAdditionLayer(name.c_str());
+ ARMNN_NO_DEPRECATE_WARN_END
LayerType* replacementLayer = PolymorphicDowncast<LayerType*>(replacement);
FuseLayer(optimizationViews,
@@ -96,7 +98,9 @@ LayerType* FuseSubtractionLayer(OptimizationViews& optimizationViews,
ActivationDescriptor& activationDesc,
std::string name)
{
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
IConnectableLayer* replacement = optimizationViews.GetINetwork()->AddSubtractionLayer(name.c_str());
+ ARMNN_NO_DEPRECATE_WARN_END
LayerType* replacementLayer = PolymorphicDowncast<LayerType*>(replacement);
FuseLayer(optimizationViews,
@@ -115,7 +119,9 @@ LayerType* FuseDivisionLayer(OptimizationViews& optimizationViews,
ActivationDescriptor& activationDesc,
std::string name)
{
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
IConnectableLayer* replacement = optimizationViews.GetINetwork()->AddDivisionLayer(name.c_str());
+ ARMNN_NO_DEPRECATE_WARN_END
LayerType* replacementLayer = PolymorphicDowncast<LayerType*>(replacement);
FuseLayer(optimizationViews,
@@ -134,7 +140,9 @@ LayerType* FuseMultiplicationLayer(OptimizationViews& optimizationViews,
ActivationDescriptor& activationDesc,
std::string name)
{
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
IConnectableLayer* replacement = optimizationViews.GetINetwork()->AddMultiplicationLayer(name.c_str());
+ ARMNN_NO_DEPRECATE_WARN_END
LayerType* replacementLayer = PolymorphicDowncast<LayerType*>(replacement);
FuseLayer(optimizationViews,
diff --git a/src/backends/backendsCommon/test/AdditionEndToEndTestImpl.hpp b/src/backends/backendsCommon/test/AdditionEndToEndTestImpl.hpp
index f33521888f..c850c518f5 100644
--- a/src/backends/backendsCommon/test/AdditionEndToEndTestImpl.hpp
+++ b/src/backends/backendsCommon/test/AdditionEndToEndTestImpl.hpp
@@ -30,8 +30,9 @@ armnn::INetworkPtr CreateAdditionNetwork(const armnn::TensorShape& inputXShape,
TensorInfo outputTensorInfo(outputShape, DataType, qScale, qOffset);
-
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
IConnectableLayer* addition = network->AddAdditionLayer("addition");
+ ARMNN_NO_DEPRECATE_WARN_END
IConnectableLayer* inputX = network->AddInputLayer(0, "inputX");
IConnectableLayer* inputY = network->AddInputLayer(1, "inputY");
IConnectableLayer* output = network->AddOutputLayer(0, "output");
diff --git a/src/backends/backendsCommon/test/EndToEndTestImpl.hpp b/src/backends/backendsCommon/test/EndToEndTestImpl.hpp
index 795fc13c32..73cef16aad 100644
--- a/src/backends/backendsCommon/test/EndToEndTestImpl.hpp
+++ b/src/backends/backendsCommon/test/EndToEndTestImpl.hpp
@@ -39,7 +39,9 @@ bool ConstantUsageTest(const std::vector<BackendId>& computeDevice,
IConnectableLayer* input = net->AddInputLayer(0);
IConnectableLayer* constant = net->AddConstantLayer(ConstTensor(commonTensorInfo, constantData));
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
IConnectableLayer* add = net->AddAdditionLayer();
+ ARMNN_NO_DEPRECATE_WARN_END
IConnectableLayer* output = net->AddOutputLayer(0);
input->GetOutputSlot(0).Connect(add->GetInputSlot(0));
diff --git a/src/backends/backendsCommon/test/IsLayerSupportedTestImpl.hpp b/src/backends/backendsCommon/test/IsLayerSupportedTestImpl.hpp
index 5b95d3cd92..5475762a53 100644
--- a/src/backends/backendsCommon/test/IsLayerSupportedTestImpl.hpp
+++ b/src/backends/backendsCommon/test/IsLayerSupportedTestImpl.hpp
@@ -622,7 +622,9 @@ struct LayerTypePolicy;
// Every entry in the armnn::LayerType enum must be accounted for below.
DECLARE_LAYER_POLICY_2_PARAM(Activation)
+ARMNN_NO_DEPRECATE_WARN_BEGIN
DECLARE_LAYER_POLICY_1_PARAM(Addition)
+ARMNN_NO_DEPRECATE_WARN_END
DECLARE_LAYER_POLICY_2_PARAM(ArgMinMax)
@@ -694,15 +696,21 @@ DECLARE_LAYER_POLICY_2_PARAM(Lstm)
DECLARE_LAYER_POLICY_MAP_PARAM(Map, void)
+ARMNN_NO_DEPRECATE_WARN_BEGIN
DECLARE_LAYER_POLICY_1_PARAM(Maximum)
+ARMNN_NO_DEPRECATE_WARN_END
DECLARE_LAYER_POLICY_2_PARAM(Mean)
DECLARE_LAYER_POLICY_1_PARAM(Merge)
+ARMNN_NO_DEPRECATE_WARN_BEGIN
DECLARE_LAYER_POLICY_1_PARAM(Minimum)
+ARMNN_NO_DEPRECATE_WARN_END
+ARMNN_NO_DEPRECATE_WARN_BEGIN
DECLARE_LAYER_POLICY_1_PARAM(Multiplication)
+ARMNN_NO_DEPRECATE_WARN_END
DECLARE_LAYER_POLICY_2_PARAM(Normalization)
@@ -726,7 +734,9 @@ DECLARE_LAYER_POLICY_2_PARAM(QLstm)
DECLARE_LAYER_POLICY_1_PARAM(QuantizedLstm)
+ARMNN_NO_DEPRECATE_WARN_BEGIN
DECLARE_LAYER_POLICY_1_PARAM(Division)
+ARMNN_NO_DEPRECATE_WARN_END
DECLARE_LAYER_POLICY_1_PARAM(Rank)
@@ -752,7 +762,9 @@ DECLARE_LAYER_POLICY_EXCEPTION_2_PARAM(StandIn)
DECLARE_LAYER_POLICY_2_PARAM(StridedSlice)
+ARMNN_NO_DEPRECATE_WARN_BEGIN
DECLARE_LAYER_POLICY_1_PARAM(Subtraction)
+ARMNN_NO_DEPRECATE_WARN_END
DECLARE_LAYER_POLICY_2_PARAM(Reduce)
diff --git a/src/backends/backendsCommon/test/OptimizationViewsTests.cpp b/src/backends/backendsCommon/test/OptimizationViewsTests.cpp
index ff3217911a..665358b9c6 100644
--- a/src/backends/backendsCommon/test/OptimizationViewsTests.cpp
+++ b/src/backends/backendsCommon/test/OptimizationViewsTests.cpp
@@ -263,7 +263,9 @@ TEST_CASE("OptimizeViewsValidateDeviceMockBackend")
armnn::IConnectableLayer* input = net->AddInputLayer(0, "inLayer0");
armnn::IConnectableLayer* input1 = net->AddInputLayer(1, "inLayer1");
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
armnn::IConnectableLayer* addition = net->AddAdditionLayer("addLayer");
+ ARMNN_NO_DEPRECATE_WARN_END
armnn::IConnectableLayer* output = net->AddOutputLayer(0, "outLayer");
diff --git a/src/backends/cl/ClLayerSupport.cpp b/src/backends/cl/ClLayerSupport.cpp
index 89bcf9bc01..b63837539e 100644
--- a/src/backends/cl/ClLayerSupport.cpp
+++ b/src/backends/cl/ClLayerSupport.cpp
@@ -196,7 +196,9 @@ bool ClLayerSupport::IsLayerSupported(const LayerType& type,
*(PolymorphicDowncast<const ActivationDescriptor*>(&descriptor)),
reasonIfUnsupported);
case LayerType::Addition:
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
return IsAdditionSupported(infos[0], infos[1], infos[2], reasonIfUnsupported);
+ ARMNN_NO_DEPRECATE_WARN_END
case LayerType::ArgMinMax:
return IsArgMinMaxSupported(infos[0],
infos[1],
@@ -346,7 +348,9 @@ bool ClLayerSupport::IsLayerSupported(const LayerType& type,
case LayerType::Dequantize:
return IsDequantizeSupported(infos[0], infos[1], reasonIfUnsupported);
case LayerType::Division:
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
return IsDivisionSupported(infos[0], infos[1], infos[2], reasonIfUnsupported);
+ ARMNN_NO_DEPRECATE_WARN_END
case LayerType::ElementwiseBinary:
{
auto desc = *(PolymorphicDowncast<const ElementwiseBinaryDescriptor *>(&descriptor));
@@ -474,16 +478,22 @@ bool ClLayerSupport::IsLayerSupported(const LayerType& type,
infos[2],
reasonIfUnsupported);
case LayerType::Maximum:
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
return IsMaximumSupported(infos[0], infos[1], infos[2], reasonIfUnsupported);
+ ARMNN_NO_DEPRECATE_WARN_END
case LayerType::Mean:
return IsMeanSupported(infos[0],
infos[1],
*(PolymorphicDowncast<const MeanDescriptor*>(&descriptor)),
reasonIfUnsupported);
case LayerType::Minimum:
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
return IsMinimumSupported(infos[0], infos[1], infos[2], reasonIfUnsupported);
+ ARMNN_NO_DEPRECATE_WARN_END
case LayerType::Multiplication:
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
return IsMultiplicationSupported(infos[0], infos[1], infos[2], reasonIfUnsupported);
+ ARMNN_NO_DEPRECATE_WARN_END
case LayerType::Normalization:
return IsNormalizationSupported(infos[0],
infos[1],
@@ -604,7 +614,9 @@ bool ClLayerSupport::IsLayerSupported(const LayerType& type,
*(PolymorphicDowncast<const StridedSliceDescriptor*>(&descriptor)),
reasonIfUnsupported);
case LayerType::Subtraction:
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
return IsSubtractionSupported(infos[0], infos[1], infos[2], reasonIfUnsupported);
+ ARMNN_NO_DEPRECATE_WARN_END
case LayerType::Transpose:
return IsTransposeSupported(infos[0],
infos[1],
diff --git a/src/backends/cl/ClLayerSupport.hpp b/src/backends/cl/ClLayerSupport.hpp
index 2d784e3df8..2b3593630a 100644
--- a/src/backends/cl/ClLayerSupport.hpp
+++ b/src/backends/cl/ClLayerSupport.hpp
@@ -1,5 +1,5 @@
//
-// Copyright © 2017 Arm Ltd and Contributors. All rights reserved.
+// Copyright © 2017-2023 Arm Ltd and Contributors. All rights reserved.
// SPDX-License-Identifier: MIT
//
#pragma once
@@ -30,6 +30,7 @@ public:
const ActivationDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override;
+ ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use IsLayerSupported instead", "24.02")
bool IsAdditionSupported(const TensorInfo& input0,
const TensorInfo& input1,
const TensorInfo& output,
@@ -128,13 +129,14 @@ public:
const Optional<TensorInfo>& biases,
Optional<std::string&> reason = EmptyOptional()) const override;
+ ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use IsLayerSupported instead", "24.02")
bool IsDivisionSupported(const TensorInfo& input0,
const TensorInfo& input1,
const TensorInfo& output,
Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override;
bool IsElementwiseUnarySupported(const TensorInfo& input,
- const TensorInfo& ouput,
+ const TensorInfo& output,
const ElementwiseUnaryDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override;
@@ -200,6 +202,7 @@ public:
const LstmInputParamsInfo& paramsInfo,
Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override;
+ ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use IsLayerSupported instead", "24.02")
bool IsMaximumSupported(const TensorInfo& input0,
const TensorInfo& input1,
const TensorInfo& output,
@@ -210,11 +213,13 @@ public:
const MeanDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override;
+ ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use IsLayerSupported instead", "24.02")
bool IsMinimumSupported(const TensorInfo& input0,
const TensorInfo& input1,
const TensorInfo& output,
Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override;
+ ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use IsLayerSupported instead", "24.02")
bool IsMultiplicationSupported(const TensorInfo& input0,
const TensorInfo& input1,
const TensorInfo& output,
@@ -325,6 +330,7 @@ public:
const StridedSliceDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported = EmptyOptional()) const override;
+ ARMNN_DEPRECATED_MSG_REMOVAL_DATE("Use IsLayerSupported instead", "24.02")
bool IsSubtractionSupported(const TensorInfo& input0,
const TensorInfo& input1,
const TensorInfo& output,