aboutsummaryrefslogtreecommitdiff
path: root/src/backends/backendsCommon/test/layerTests/SoftmaxTestImpl.cpp
diff options
context:
space:
mode:
Diffstat (limited to 'src/backends/backendsCommon/test/layerTests/SoftmaxTestImpl.cpp')
-rw-r--r--src/backends/backendsCommon/test/layerTests/SoftmaxTestImpl.cpp9
1 files changed, 7 insertions, 2 deletions
diff --git a/src/backends/backendsCommon/test/layerTests/SoftmaxTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/SoftmaxTestImpl.cpp
index fa7ab90b16..1ffb34f932 100644
--- a/src/backends/backendsCommon/test/layerTests/SoftmaxTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/SoftmaxTestImpl.cpp
@@ -1,5 +1,5 @@
//
-// Copyright © 2017 Arm Ltd. All rights reserved.
+// Copyright © 2017 Arm Ltd and Contributors. All rights reserved.
// SPDX-License-Identifier: MIT
//
@@ -87,8 +87,10 @@ LayerTestResult<T, n> SimpleSoftmaxBaseTestImpl(
// Each row is independently softmax'd.
auto input = MakeTensor<T, n>(inputTensorInfo, armnnUtils::QuantizedVector<T>(inputData, qScale, qOffset));
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
std::unique_ptr<armnn::ITensorHandle> inputHandle = workloadFactory.CreateTensorHandle(inputTensorInfo);
std::unique_ptr<armnn::ITensorHandle> outputHandle = workloadFactory.CreateTensorHandle(outputTensorInfo);
+ ARMNN_NO_DEPRECATE_WARN_END
armnn::SoftmaxQueueDescriptor data;
data.m_Parameters.m_Beta = beta;
@@ -258,8 +260,10 @@ LayerTestResult<T, 2> CompareSoftmaxTestImpl(
LayerTestResult<T, 2> ret(outputTensorInfo);
auto input = MakeRandomTensor<T, 2>(inputTensorInfo, 0xF00D, 0.0f, 1.0f);
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
std::unique_ptr<armnn::ITensorHandle> inputHandle = workloadFactory.CreateTensorHandle(inputTensorInfo);
std::unique_ptr<armnn::ITensorHandle> outputHandle = workloadFactory.CreateTensorHandle(outputTensorInfo);
+ ARMNN_NO_DEPRECATE_WARN_END
armnn::SoftmaxQueueDescriptor data;
data.m_Parameters.m_Beta = beta;
@@ -268,9 +272,10 @@ LayerTestResult<T, 2> CompareSoftmaxTestImpl(
AddInputToWorkload(data, info, inputTensorInfo, inputHandle.get());
AddOutputToWorkload(data, info, outputTensorInfo, outputHandle.get());
+ ARMNN_NO_DEPRECATE_WARN_BEGIN
std::unique_ptr<armnn::ITensorHandle> outputHandleRef = refWorkloadFactory.CreateTensorHandle(outputTensorInfo);
std::unique_ptr<armnn::ITensorHandle> inputHandleRef = refWorkloadFactory.CreateTensorHandle(inputTensorInfo);
-
+ ARMNN_NO_DEPRECATE_WARN_END
armnn::SoftmaxQueueDescriptor refData = data;
armnn::WorkloadInfo refInfo = info;