aboutsummaryrefslogtreecommitdiff
path: root/src/armnn/test
diff options
context:
space:
mode:
Diffstat (limited to 'src/armnn/test')
-rw-r--r--src/armnn/test/CreateWorkload.hpp4
-rw-r--r--src/armnn/test/DebugCallbackTest.cpp2
-rw-r--r--src/armnn/test/EndToEndTest.cpp2
-rw-r--r--src/armnn/test/OptimizerTests.cpp6
-rw-r--r--src/armnn/test/OptionalTest.cpp6
-rw-r--r--src/armnn/test/ProfilerTests.cpp3
-rw-r--r--src/armnn/test/QuantizerTest.cpp112
-rw-r--r--src/armnn/test/RuntimeTests.cpp4
-rw-r--r--src/armnn/test/TensorHandleStrategyTest.cpp23
9 files changed, 81 insertions, 81 deletions
diff --git a/src/armnn/test/CreateWorkload.hpp b/src/armnn/test/CreateWorkload.hpp
index 4782c432a2..72ad9d45ef 100644
--- a/src/armnn/test/CreateWorkload.hpp
+++ b/src/armnn/test/CreateWorkload.hpp
@@ -11,6 +11,7 @@
#include <ResolveType.hpp>
#include <armnnUtils/DataLayoutIndexed.hpp>
+#include <armnn/utility/IgnoreUnused.hpp>
#include <backendsCommon/WorkloadData.hpp>
#include <backendsCommon/WorkloadFactory.hpp>
@@ -18,7 +19,6 @@
#include <boost/test/unit_test.hpp>
#include <boost/cast.hpp>
-#include <boost/core/ignore_unused.hpp>
#include <utility>
@@ -1298,7 +1298,7 @@ std::pair<armnn::IOptimizedNetworkPtr, std::unique_ptr<PreCompiledWorkload>> Cre
armnn::Graph& graph,
bool biasEnabled = false)
{
- boost::ignore_unused(graph);
+ IgnoreUnused(graph);
// To create a PreCompiled layer, create a network and Optimize it.
armnn::Network net;
diff --git a/src/armnn/test/DebugCallbackTest.cpp b/src/armnn/test/DebugCallbackTest.cpp
index bd8bdd543c..c89da83a89 100644
--- a/src/armnn/test/DebugCallbackTest.cpp
+++ b/src/armnn/test/DebugCallbackTest.cpp
@@ -60,7 +60,7 @@ BOOST_AUTO_TEST_CASE(RuntimeRegisterDebugCallback)
std::vector<unsigned int> slotIndexes;
auto mockCallback = [&](LayerGuid guid, unsigned int slotIndex, ITensorHandle* tensor)
{
- boost::ignore_unused(guid);
+ IgnoreUnused(guid);
slotIndexes.push_back(slotIndex);
tensorShapes.push_back(tensor->GetShape());
callCount++;
diff --git a/src/armnn/test/EndToEndTest.cpp b/src/armnn/test/EndToEndTest.cpp
index df84be4277..a8192a6480 100644
--- a/src/armnn/test/EndToEndTest.cpp
+++ b/src/armnn/test/EndToEndTest.cpp
@@ -6,8 +6,8 @@
#include <armnn/Descriptors.hpp>
#include <armnn/IRuntime.hpp>
#include <armnn/INetwork.hpp>
+#include <armnn/utility/IgnoreUnused.hpp>
-#include <boost/core/ignore_unused.hpp>
#include <boost/test/unit_test.hpp>
#include <set>
diff --git a/src/armnn/test/OptimizerTests.cpp b/src/armnn/test/OptimizerTests.cpp
index 0ca4fc4764..56032adc33 100644
--- a/src/armnn/test/OptimizerTests.cpp
+++ b/src/armnn/test/OptimizerTests.cpp
@@ -691,7 +691,7 @@ BOOST_AUTO_TEST_CASE(BackendHintTest)
LayerBindingId id,
const char* name = nullptr) override
{
- boost::ignore_unused(id, name);
+ IgnoreUnused(id, name);
auto inputLayer = boost::polymorphic_downcast<const InputLayer*>(layer);
BOOST_TEST((inputLayer->GetBackendId() == "MockBackend"));
}
@@ -700,7 +700,7 @@ BOOST_AUTO_TEST_CASE(BackendHintTest)
LayerBindingId id,
const char* name = nullptr) override
{
- boost::ignore_unused(id, name);
+ IgnoreUnused(id, name);
auto outputLayer = boost::polymorphic_downcast<const OutputLayer*>(layer);
BOOST_TEST((outputLayer->GetBackendId() == "MockBackend"));
}
@@ -709,7 +709,7 @@ BOOST_AUTO_TEST_CASE(BackendHintTest)
const ActivationDescriptor& activationDescriptor,
const char* name = nullptr) override
{
- boost::ignore_unused(activationDescriptor, name);
+ IgnoreUnused(activationDescriptor, name);
auto activation = boost::polymorphic_downcast<const ActivationLayer*>(layer);
BOOST_TEST((activation->GetBackendId() == "CustomBackend"));
}
diff --git a/src/armnn/test/OptionalTest.cpp b/src/armnn/test/OptionalTest.cpp
index fd136439c8..73c96431fb 100644
--- a/src/armnn/test/OptionalTest.cpp
+++ b/src/armnn/test/OptionalTest.cpp
@@ -7,19 +7,19 @@
#include <armnn/Optional.hpp>
#include <string>
-#include <boost/core/ignore_unused.hpp>
+#include <armnn/utility/IgnoreUnused.hpp>
namespace
{
void PassStringRef(armnn::Optional<std::string&> value)
{
- boost::ignore_unused(value);
+ armnn::IgnoreUnused(value);
}
void PassStringRefWithDefault(armnn::Optional<std::string&> value = armnn::EmptyOptional())
{
- boost::ignore_unused(value);
+ armnn::IgnoreUnused(value);
}
} // namespace <anonymous>
diff --git a/src/armnn/test/ProfilerTests.cpp b/src/armnn/test/ProfilerTests.cpp
index a052862bdd..9376fa4cea 100644
--- a/src/armnn/test/ProfilerTests.cpp
+++ b/src/armnn/test/ProfilerTests.cpp
@@ -5,6 +5,7 @@
#include <armnn/IRuntime.hpp>
#include <armnn/TypesUtils.hpp>
+#include <armnn/utility/IgnoreUnused.hpp>
#include <boost/test/unit_test.hpp>
#include <boost/test/tools/output_test_stream.hpp>
@@ -309,7 +310,7 @@ BOOST_AUTO_TEST_CASE(ProfilerJsonPrinter)
profiler->Print(json);
std::string output = buffer.str();
- boost::ignore_unused(output);
+ armnn::IgnoreUnused(output);
// Disable profiling here to not print out anything on stdout.
profiler->EnableProfiling(false);
diff --git a/src/armnn/test/QuantizerTest.cpp b/src/armnn/test/QuantizerTest.cpp
index 2dc054af07..ef9b2da782 100644
--- a/src/armnn/test/QuantizerTest.cpp
+++ b/src/armnn/test/QuantizerTest.cpp
@@ -3,15 +3,6 @@
// SPDX-License-Identifier: MIT
//
-#include <armnn/INetwork.hpp>
-#include <armnn/LayerVisitorBase.hpp>
-#include <armnn/Tensor.hpp>
-#include <armnn/Types.hpp>
-
-#include <armnnQuantizer/INetworkQuantizer.hpp>
-
-#include <QuantizeHelper.hpp>
-
#include "../Graph.hpp"
#include "../Network.hpp"
#include "../NetworkQuantizerUtils.hpp"
@@ -19,7 +10,14 @@
#include "../RangeTracker.hpp"
#include "../../armnnQuantizer/CommandLineProcessor.hpp"
-#include <boost/core/ignore_unused.hpp>
+#include <armnn/INetwork.hpp>
+#include <armnn/LayerVisitorBase.hpp>
+#include <armnn/Tensor.hpp>
+#include <armnn/Types.hpp>
+#include <armnn/utility/IgnoreUnused.hpp>
+#include <armnnQuantizer/INetworkQuantizer.hpp>
+#include <QuantizeHelper.hpp>
+
#include <boost/test/unit_test.hpp>
#include <unordered_map>
@@ -58,7 +56,7 @@ public:
LayerBindingId id,
const char* name = nullptr) override
{
- boost::ignore_unused(id, name);
+ IgnoreUnused(id, name);
const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo();
BOOST_TEST(m_InputShape == info.GetShape());
// Based off current default [-15.0f, 15.0f]
@@ -72,7 +70,7 @@ public:
LayerBindingId id,
const char* name = nullptr) override
{
- boost::ignore_unused(id, name);
+ IgnoreUnused(id, name);
const TensorInfo& info = layer->GetInputSlot(0).GetConnection()->GetTensorInfo();
BOOST_TEST(m_OutputShape == info.GetShape());
}
@@ -116,7 +114,7 @@ protected:
const OffsetScalePair& params,
DataType dataType = DataType::QAsymmU8)
{
- boost::ignore_unused(dataType);
+ IgnoreUnused(dataType);
TestQuantizationParamsImpl(info, dataType, params.first, params.second);
}
@@ -212,7 +210,7 @@ public:
void VisitAdditionLayer(const IConnectableLayer* layer,
const char* name = nullptr) override
{
- boost::ignore_unused(name);
+ IgnoreUnused(name);
TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo();
// Based off default static range [-20.0f, 20.0f]
@@ -282,7 +280,7 @@ public:
const ActivationDescriptor& descriptor,
const char* name = nullptr) override
{
- boost::ignore_unused(descriptor, name);
+ IgnoreUnused(descriptor, name);
TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo();
@@ -385,7 +383,7 @@ BOOST_AUTO_TEST_CASE(InputOutputLayerDynamicQuant)
LayerBindingId id,
const char* name = nullptr) override
{
- boost::ignore_unused(id, name);
+ IgnoreUnused(id, name);
const TensorInfo& info = layer->GetInputSlot(0).GetConnection()->GetTensorInfo();
BOOST_CHECK_MESSAGE(info.GetDataType() == m_DataType,
std::string(armnn::GetDataTypeName(info.GetDataType()))
@@ -543,7 +541,7 @@ BOOST_AUTO_TEST_CASE(QuantizeBoundedReluActivation)
const ActivationDescriptor& descriptor,
const char* name = nullptr) override
{
- boost::ignore_unused(descriptor, name);
+ IgnoreUnused(descriptor, name);
TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo();
// Based off default static range [0.0f, 3.5f]
@@ -599,7 +597,7 @@ BOOST_AUTO_TEST_CASE(QuantizeTanHActivation)
const ActivationDescriptor& descriptor,
const char* name = nullptr) override
{
- boost::ignore_unused(descriptor, name);
+ IgnoreUnused(descriptor, name);
TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo();
// Based off default static range [-1.0f, 1.0f]
@@ -654,7 +652,7 @@ public:
const ActivationDescriptor& descriptor,
const char* name = nullptr) override
{
- boost::ignore_unused(descriptor, name);
+ IgnoreUnused(descriptor, name);
TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo();
// Based off default static range [-5.0f, 15.0f]
@@ -725,7 +723,7 @@ BOOST_AUTO_TEST_CASE(QuantizeELuActivation)
const ActivationDescriptor& descriptor,
const char* name = nullptr) override
{
- boost::ignore_unused(descriptor, name);
+ IgnoreUnused(descriptor, name);
TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo();
// Based off default static range [-15.0f, 15.0f]
@@ -779,7 +777,7 @@ BOOST_AUTO_TEST_CASE(QuantizeHardSwishActivation)
const ActivationDescriptor& descriptor,
const char* name = nullptr) override
{
- boost::ignore_unused(descriptor, name);
+ IgnoreUnused(descriptor, name);
TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo();
// Based off default static range [-15.0f, 15.0f]
@@ -839,7 +837,7 @@ BOOST_AUTO_TEST_CASE(QuantizeBatchNorm)
const ConstTensor& gamma,
const char* name = nullptr) override
{
- boost::ignore_unused(desc, name);
+ IgnoreUnused(desc, name);
TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo();
// Based off default static range [-15.0f, 15.0f]
@@ -924,7 +922,7 @@ BOOST_AUTO_TEST_CASE(QuantizeDepthToSpace)
const DepthToSpaceDescriptor& desc,
const char* name = nullptr)
{
- boost::ignore_unused(desc, name);
+ IgnoreUnused(desc, name);
const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo();
const OffsetScalePair qAsymmU8Params{ 30.0f / g_AsymmU8QuantizationBase, 128 };
@@ -1116,7 +1114,7 @@ void ValidateFullyConnectedLayer(const bool biasEnabled)
const Optional<ConstTensor>& biases,
const char* name = nullptr) override
{
- boost::ignore_unused(desc, name);
+ IgnoreUnused(desc, name);
TestQuantizationOnLayersWithBiases(layer, weights, biases);
}
};
@@ -1173,7 +1171,7 @@ void TestQuantizeConvolution2d(bool useBiases)
const Optional<ConstTensor>& biases,
const char *name = nullptr) override
{
- boost::ignore_unused(convolution2dDescriptor, name);
+ IgnoreUnused(convolution2dDescriptor, name);
TestQuantizationOnLayersWithBiases(layer, weights, biases);
}
};
@@ -1259,7 +1257,7 @@ void TestQuantizeDepthwiseConvolution2d(bool useBiases)
const Optional<ConstTensor>& biases,
const char *name = nullptr) override
{
- boost::ignore_unused(convolution2dDescriptor, name);
+ IgnoreUnused(convolution2dDescriptor, name);
TestQuantizationOnLayersWithBiases(layer, weights, biases);
}
};
@@ -1343,7 +1341,7 @@ BOOST_AUTO_TEST_CASE(QuantizeInstanceNormalization)
const InstanceNormalizationDescriptor& descriptor,
const char* name = nullptr)
{
- boost::ignore_unused(descriptor, name);
+ IgnoreUnused(descriptor, name);
const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo();
const OffsetScalePair qAsymmU8Params{ 30.0f / g_AsymmU8QuantizationBase, 128 };
@@ -1411,7 +1409,7 @@ BOOST_AUTO_TEST_CASE(QuantizeLogSoftmax)
const SoftmaxDescriptor& descriptor,
const char* name = nullptr) override
{
- boost::ignore_unused(descriptor, name);
+ IgnoreUnused(descriptor, name);
TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo();
const OffsetScalePair qAsymmU8Params{ 30.0f / g_AsymmU8QuantizationBase, 128 };
@@ -1503,7 +1501,7 @@ BOOST_AUTO_TEST_CASE(QuantizeSoftmax)
const SoftmaxDescriptor& descriptor,
const char* name = nullptr) override
{
- boost::ignore_unused(descriptor, name);
+ IgnoreUnused(descriptor, name);
TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo();
// Based off default static range [0.0f, 1.0f]
@@ -1636,7 +1634,7 @@ BOOST_AUTO_TEST_CASE(QuantizePermute)
const PermuteDescriptor& desc,
const char* name = nullptr) override
{
- boost::ignore_unused(desc, name);
+ IgnoreUnused(desc, name);
CheckForwardedQuantizationSettings(layer);
}
};
@@ -1691,7 +1689,7 @@ BOOST_AUTO_TEST_CASE(QuantizeSpaceToBatch)
const SpaceToBatchNdDescriptor& spaceToBatchNdDescriptor,
const char* name = nullptr) override
{
- boost::ignore_unused(spaceToBatchNdDescriptor, name);
+ IgnoreUnused(spaceToBatchNdDescriptor, name);
CheckForwardedQuantizationSettings(layer);
}
};
@@ -1804,7 +1802,7 @@ BOOST_AUTO_TEST_CASE(QuantizePooling2d)
const Pooling2dDescriptor& desc,
const char* name = nullptr) override
{
- boost::ignore_unused(desc, name);
+ IgnoreUnused(desc, name);
CheckForwardedQuantizationSettings(layer);
}
};
@@ -1873,7 +1871,7 @@ BOOST_AUTO_TEST_CASE(QuantizeConstant)
const ConstTensor& input,
const char* name = nullptr) override
{
- boost::ignore_unused(input, name);
+ IgnoreUnused(input, name);
TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo();
// Based off the range of values in the const tensor used for the test: [-2.0f, 6.0f]
@@ -1946,20 +1944,20 @@ BOOST_AUTO_TEST_CASE(QuantizeArgMinMax)
LayerBindingId id,
const char* name = nullptr) override
{
- boost::ignore_unused(layer, id, name);
+ IgnoreUnused(layer, id, name);
}
void VisitOutputLayer(const IConnectableLayer* layer,
LayerBindingId id,
const char* name = nullptr) override
{
- boost::ignore_unused(layer, id, name);
+ IgnoreUnused(layer, id, name);
}
void VisitArgMinMaxLayer(const IConnectableLayer* layer,
const ArgMinMaxDescriptor& argMinMaxDescriptor,
const char* name = nullptr) override
{
- boost::ignore_unused(argMinMaxDescriptor, name);
+ IgnoreUnused(argMinMaxDescriptor, name);
TensorInfo outputInfo = layer->GetOutputSlot(0).GetTensorInfo();
TestQuantizationParams(outputInfo,
@@ -2034,7 +2032,7 @@ BOOST_AUTO_TEST_CASE(QuantizeComparison)
const ComparisonDescriptor& descriptor,
const char* name = nullptr) override
{
- boost::ignore_unused(descriptor, name);
+ IgnoreUnused(descriptor, name);
TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo();
const OffsetScalePair qAsymmU8Params{ 30.0f / g_AsymmU8QuantizationBase, 128 };
@@ -2106,19 +2104,19 @@ BOOST_AUTO_TEST_CASE(QuantizeConcat)
LayerBindingId id,
const char* name = nullptr) override
{
- boost::ignore_unused(layer, id, name);
+ IgnoreUnused(layer, id, name);
}
void VisitOutputLayer(const IConnectableLayer* layer,
LayerBindingId id,
const char* name = nullptr) override
{
- boost::ignore_unused(layer, id, name);
+ IgnoreUnused(layer, id, name);
}
void VisitConcatLayer(const IConnectableLayer* layer,
const OriginsDescriptor& originsDescriptor,
const char* name = nullptr) override
{
- boost::ignore_unused(originsDescriptor, name);
+ IgnoreUnused(originsDescriptor, name);
TensorInfo outputInfo = layer->GetOutputSlot(0).GetTensorInfo();
TestQuantizationParams(
outputInfo, {60.8f / g_AsymmU8QuantizationBase, 65},
@@ -2214,7 +2212,7 @@ BOOST_AUTO_TEST_CASE(QuantizeReshape)
const ReshapeDescriptor& reshapeDescriptor,
const char* name = nullptr) override
{
- boost::ignore_unused(reshapeDescriptor, name);
+ IgnoreUnused(reshapeDescriptor, name);
CheckForwardedQuantizationSettings(layer);
}
};
@@ -2269,7 +2267,7 @@ BOOST_AUTO_TEST_CASE(QuantizeSplitter)
const SplitterDescriptor& desc,
const char* name = nullptr)
{
- boost::ignore_unused(desc, name);
+ IgnoreUnused(desc, name);
CheckForwardedQuantizationSettings(layer);
}
};
@@ -2325,7 +2323,7 @@ BOOST_AUTO_TEST_CASE(QuantizeResize)
const ResizeDescriptor& resizeDescriptor,
const char* name = nullptr) override
{
- boost::ignore_unused(resizeDescriptor, name);
+ IgnoreUnused(resizeDescriptor, name);
CheckForwardedQuantizationSettings(layer);
}
};
@@ -2382,7 +2380,7 @@ BOOST_AUTO_TEST_CASE(QuantizeStridedSlice)
const StridedSliceDescriptor& desc,
const char* name = nullptr)
{
- boost::ignore_unused(desc, name);
+ IgnoreUnused(desc, name);
CheckForwardedQuantizationSettings(layer);
}
};
@@ -2437,7 +2435,7 @@ BOOST_AUTO_TEST_CASE(QuantizeBatchToSpace)
const BatchToSpaceNdDescriptor& batchToSpaceNdDescriptor,
const char* name = nullptr) override
{
- boost::ignore_unused(batchToSpaceNdDescriptor, name);
+ IgnoreUnused(batchToSpaceNdDescriptor, name);
CheckForwardedQuantizationSettings(layer);
}
};
@@ -2499,7 +2497,7 @@ BOOST_AUTO_TEST_CASE(QuantizePrelu)
LayerBindingId id,
const char* name = nullptr) override
{
- boost::ignore_unused(id, name);
+ IgnoreUnused(id, name);
const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo();
switch (id)
@@ -2526,7 +2524,7 @@ BOOST_AUTO_TEST_CASE(QuantizePrelu)
LayerBindingId id,
const char* name = nullptr) override
{
- boost::ignore_unused(id, name);
+ IgnoreUnused(id, name);
const TensorInfo& info = layer->GetInputSlot(0).GetConnection()->GetTensorInfo();
BOOST_TEST(m_OutputShape == info.GetShape());
}
@@ -2534,7 +2532,7 @@ BOOST_AUTO_TEST_CASE(QuantizePrelu)
void VisitPreluLayer(const IConnectableLayer* layer,
const char* name = nullptr) override
{
- boost::ignore_unused(name);
+ IgnoreUnused(name);
const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo();
TestQuantizationParams(info,
{ 30.0f / g_AsymmU8QuantizationBase, 128 }, // QASymmU8
@@ -2617,7 +2615,7 @@ void TestQuantizeTransposeConvolution2d(bool useBiases)
const Optional<ConstTensor>& biases,
const char *name = nullptr) override
{
- boost::ignore_unused(descriptor, name);
+ IgnoreUnused(descriptor, name);
TestQuantizationOnLayersWithBiases(layer, weights, biases);
}
};
@@ -2704,20 +2702,20 @@ BOOST_AUTO_TEST_CASE(QuantizeStack)
LayerBindingId id,
const char* name = nullptr) override
{
- boost::ignore_unused(layer, id, name);
+ IgnoreUnused(layer, id, name);
}
void VisitOutputLayer(const IConnectableLayer* layer,
LayerBindingId id,
const char* name = nullptr) override
{
- boost::ignore_unused(layer, id, name);
+ IgnoreUnused(layer, id, name);
}
void VisitStackLayer(const IConnectableLayer* layer,
const StackDescriptor& descriptor,
const char* name = nullptr) override
{
- boost::ignore_unused(descriptor, name);
+ IgnoreUnused(descriptor, name);
TensorInfo outputInfo = layer->GetOutputSlot(0).GetTensorInfo();
TestQuantizationParams(outputInfo,
@@ -2784,7 +2782,7 @@ BOOST_AUTO_TEST_CASE(QuantizeSlice)
const SliceDescriptor& desc,
const char* name = nullptr)
{
- boost::ignore_unused(desc, name);
+ IgnoreUnused(desc, name);
const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo();
const OffsetScalePair qAsymmU8Params{ 30.0f / g_AsymmU8QuantizationBase, 128 };
@@ -2876,7 +2874,7 @@ public:
LayerBindingId id,
const char* name = nullptr) override
{
- boost::ignore_unused(id, name);
+ IgnoreUnused(id, name);
const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo();
BOOST_TEST(GetDataTypeName(info.GetDataType()) == GetDataTypeName(m_DataType));
BOOST_TEST(m_InputShape == info.GetShape());
@@ -2886,7 +2884,7 @@ public:
LayerBindingId id,
const char* name = nullptr) override
{
- boost::ignore_unused(id, name);
+ IgnoreUnused(id, name);
const TensorInfo& info = layer->GetInputSlot(0).GetConnection()->GetTensorInfo();
BOOST_TEST(GetDataTypeName(info.GetDataType()) == GetDataTypeName(m_DataType));
BOOST_TEST(m_OutputShape == info.GetShape());
@@ -2895,14 +2893,14 @@ public:
void VisitQuantizeLayer(const IConnectableLayer* layer,
const char* name = nullptr) override
{
- boost::ignore_unused(layer, name);
+ IgnoreUnused(layer, name);
m_VisitedQuantizeLayer = true;
}
void VisitDequantizeLayer(const IConnectableLayer* layer,
const char* name = nullptr) override
{
- boost::ignore_unused(layer, name);
+ IgnoreUnused(layer, name);
m_VisitedDequantizeLayer = true;
}
diff --git a/src/armnn/test/RuntimeTests.cpp b/src/armnn/test/RuntimeTests.cpp
index d9ed18bdd5..e3cbe03c62 100644
--- a/src/armnn/test/RuntimeTests.cpp
+++ b/src/armnn/test/RuntimeTests.cpp
@@ -158,8 +158,8 @@ BOOST_AUTO_TEST_CASE(RuntimeMemoryLeak)
// These are needed because VALGRIND_COUNT_LEAKS is a macro that assigns to the parameters
// so they are assigned to, but still considered unused, causing a warning.
- boost::ignore_unused(dubious);
- boost::ignore_unused(suppressed);
+ IgnoreUnused(dubious);
+ IgnoreUnused(suppressed);
}
#endif // WITH_VALGRIND
diff --git a/src/armnn/test/TensorHandleStrategyTest.cpp b/src/armnn/test/TensorHandleStrategyTest.cpp
index 3e59c0b604..976e58eb50 100644
--- a/src/armnn/test/TensorHandleStrategyTest.cpp
+++ b/src/armnn/test/TensorHandleStrategyTest.cpp
@@ -16,10 +16,11 @@
#include <Network.hpp>
+#include <armnn/utility/IgnoreUnused.hpp>
+
#include <vector>
#include <string>
-#include <boost/core/ignore_unused.hpp>
using namespace armnn;
@@ -44,20 +45,20 @@ public:
TensorShape const& subTensorShape,
unsigned int const* subTensorOrigin) const override
{
- boost::ignore_unused(parent, subTensorShape, subTensorOrigin);
+ IgnoreUnused(parent, subTensorShape, subTensorOrigin);
return nullptr;
}
std::unique_ptr<ITensorHandle> CreateTensorHandle(const TensorInfo& tensorInfo) const override
{
- boost::ignore_unused(tensorInfo);
+ IgnoreUnused(tensorInfo);
return nullptr;
}
std::unique_ptr<ITensorHandle> CreateTensorHandle(const TensorInfo& tensorInfo,
DataLayout dataLayout) const override
{
- boost::ignore_unused(tensorInfo, dataLayout);
+ IgnoreUnused(tensorInfo, dataLayout);
return nullptr;
}
@@ -85,20 +86,20 @@ public:
TensorShape const& subTensorShape,
unsigned int const* subTensorOrigin) const override
{
- boost::ignore_unused(parent, subTensorShape, subTensorOrigin);
+ IgnoreUnused(parent, subTensorShape, subTensorOrigin);
return nullptr;
}
std::unique_ptr<ITensorHandle> CreateTensorHandle(const TensorInfo& tensorInfo) const override
{
- boost::ignore_unused(tensorInfo);
+ IgnoreUnused(tensorInfo);
return nullptr;
}
std::unique_ptr<ITensorHandle> CreateTensorHandle(const TensorInfo& tensorInfo,
DataLayout dataLayout) const override
{
- boost::ignore_unused(tensorInfo, dataLayout);
+ IgnoreUnused(tensorInfo, dataLayout);
return nullptr;
}
@@ -123,7 +124,7 @@ public:
IWorkloadFactoryPtr CreateWorkloadFactory(const IMemoryManagerSharedPtr& memoryManager = nullptr) const override
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
return IWorkloadFactoryPtr{};
}
@@ -164,7 +165,7 @@ public:
IWorkloadFactoryPtr CreateWorkloadFactory(const IMemoryManagerSharedPtr& memoryManager = nullptr) const override
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
return IWorkloadFactoryPtr{};
}
@@ -202,7 +203,7 @@ public:
IWorkloadFactoryPtr CreateWorkloadFactory(const IMemoryManagerSharedPtr& memoryManager = nullptr) const override
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
return IWorkloadFactoryPtr{};
}
@@ -239,7 +240,7 @@ public:
IWorkloadFactoryPtr CreateWorkloadFactory(const IMemoryManagerSharedPtr& memoryManager = nullptr) const override
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
return IWorkloadFactoryPtr{};
}