aboutsummaryrefslogtreecommitdiff
path: root/src/backends
diff options
context:
space:
mode:
Diffstat (limited to 'src/backends')
-rw-r--r--src/backends/backendsCommon/CpuTensorHandle.cpp1
-rw-r--r--src/backends/backendsCommon/LayerSupportBase.cpp2
-rw-r--r--src/backends/backendsCommon/MakeWorkloadHelper.hpp6
-rw-r--r--src/backends/backendsCommon/WorkloadUtils.hpp4
-rw-r--r--src/backends/backendsCommon/test/BackendProfilingTests.cpp7
-rw-r--r--src/backends/backendsCommon/test/IsLayerSupportedTestImpl.hpp12
-rw-r--r--src/backends/backendsCommon/test/MockBackend.cpp2
-rw-r--r--src/backends/backendsCommon/test/TestDynamicBackend.cpp4
-rw-r--r--src/backends/backendsCommon/test/layerTests/ActivationTestImpl.cpp12
-rw-r--r--src/backends/backendsCommon/test/layerTests/AdditionTestImpl.cpp8
-rw-r--r--src/backends/backendsCommon/test/layerTests/ArgMinMaxTestImpl.cpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/BatchNormalizationTestImpl.cpp8
-rw-r--r--src/backends/backendsCommon/test/layerTests/BatchToSpaceNdTestImpl.hpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/ComparisonTestImpl.cpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/ConcatTestImpl.cpp14
-rw-r--r--src/backends/backendsCommon/test/layerTests/ConstantTestImpl.cpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/Conv2dTestImpl.cpp10
-rw-r--r--src/backends/backendsCommon/test/layerTests/ConvertFp16ToFp32TestImpl.cpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/ConvertFp32ToFp16TestImpl.cpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/DebugTestImpl.cpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/DepthToSpaceTestImpl.cpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/DequantizeTestImpl.cpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/DivisionTestImpl.cpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/FakeQuantizationTestImpl.cpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/FloorTestImpl.cpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/FullyConnectedTestImpl.cpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/GatherTestImpl.cpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/InstanceNormalizationTestImpl.cpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/L2NormalizationTestImpl.cpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/LogSoftmaxTestImpl.cpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/LstmTestImpl.cpp10
-rw-r--r--src/backends/backendsCommon/test/layerTests/MaximumTestImpl.cpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/MeanTestImpl.hpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/MinimumTestImpl.cpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/MultiplicationTestImpl.cpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/NormalizationTestImpl.cpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/PadTestImpl.cpp6
-rw-r--r--src/backends/backendsCommon/test/layerTests/PermuteTestImpl.hpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/Pooling2dTestImpl.cpp6
-rw-r--r--src/backends/backendsCommon/test/layerTests/PreluTestImpl.hpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/QuantizeTestImpl.cpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/ReshapeTestImpl.cpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/ResizeTestImpl.cpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/SliceTestImpl.cpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/SoftmaxTestImpl.cpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/SpaceToBatchNdTestImpl.cpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/SpaceToDepthTestImpl.cpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/SplitterTestImpl.cpp4
-rw-r--r--src/backends/backendsCommon/test/layerTests/StackTestImpl.cpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/StridedSliceTestImpl.cpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/TransposeConvolution2dTestImpl.cpp2
-rw-r--r--src/backends/backendsCommon/test/layerTests/TransposeTestImpl.hpp2
-rw-r--r--src/backends/cl/ClContextControl.cpp5
-rw-r--r--src/backends/cl/ClLayerSupport.cpp12
-rw-r--r--src/backends/cl/ClWorkloadFactory.cpp14
-rw-r--r--src/backends/cl/OpenClTimer.cpp5
-rw-r--r--src/backends/cl/test/ClRuntimeTests.cpp6
-rw-r--r--src/backends/cl/test/Fp16SupportTest.cpp2
-rw-r--r--src/backends/neon/NeonLayerSupport.cpp29
-rw-r--r--src/backends/neon/NeonTensorHandleFactory.cpp2
-rw-r--r--src/backends/neon/NeonWorkloadFactory.cpp10
-rw-r--r--src/backends/neon/NeonWorkloadFactory.hpp3
-rw-r--r--src/backends/neon/test/NeonLayerTests.cpp2
-rw-r--r--src/backends/reference/RefLayerSupport.cpp71
-rw-r--r--src/backends/reference/RefTensorHandleFactory.cpp6
-rw-r--r--src/backends/reference/RefWorkloadFactory.cpp12
-rw-r--r--src/backends/reference/RefWorkloadFactory.hpp9
-rw-r--r--src/backends/reference/test/RefWorkloadFactoryHelper.hpp2
-rw-r--r--src/backends/reference/workloads/ArgMinMax.cpp2
-rw-r--r--src/backends/reference/workloads/BaseIterator.hpp5
-rw-r--r--src/backends/reference/workloads/Dequantize.cpp5
-rw-r--r--src/backends/reference/workloads/DetectionPostProcess.cpp2
-rw-r--r--src/backends/reference/workloads/Gather.cpp4
-rw-r--r--src/backends/reference/workloads/LogSoftmax.cpp4
-rw-r--r--src/backends/reference/workloads/RefStridedSliceWorkload.cpp2
-rw-r--r--src/backends/reference/workloads/Slice.cpp5
76 files changed, 202 insertions, 203 deletions
diff --git a/src/backends/backendsCommon/CpuTensorHandle.cpp b/src/backends/backendsCommon/CpuTensorHandle.cpp
index de83048340..65e6c47179 100644
--- a/src/backends/backendsCommon/CpuTensorHandle.cpp
+++ b/src/backends/backendsCommon/CpuTensorHandle.cpp
@@ -3,6 +3,7 @@
// SPDX-License-Identifier: MIT
//
#include <armnn/Exceptions.hpp>
+#include <armnn/utility/IgnoreUnused.hpp>
#include <backendsCommon/CpuTensorHandle.hpp>
diff --git a/src/backends/backendsCommon/LayerSupportBase.cpp b/src/backends/backendsCommon/LayerSupportBase.cpp
index 127913447c..e8ef46ecd8 100644
--- a/src/backends/backendsCommon/LayerSupportBase.cpp
+++ b/src/backends/backendsCommon/LayerSupportBase.cpp
@@ -10,7 +10,7 @@
#include <backendsCommon/LayerSupportBase.hpp>
-#include <boost/core/ignore_unused.hpp>
+#include <armnn/utility/IgnoreUnused.hpp>
namespace
{
diff --git a/src/backends/backendsCommon/MakeWorkloadHelper.hpp b/src/backends/backendsCommon/MakeWorkloadHelper.hpp
index 250a10a61c..7ef140e453 100644
--- a/src/backends/backendsCommon/MakeWorkloadHelper.hpp
+++ b/src/backends/backendsCommon/MakeWorkloadHelper.hpp
@@ -31,9 +31,9 @@ struct MakeWorkloadForType<NullWorkload>
const WorkloadInfo& info,
Args&&... args)
{
- boost::ignore_unused(descriptor);
- boost::ignore_unused(info);
- boost::ignore_unused(args...);
+ IgnoreUnused(descriptor);
+ IgnoreUnused(info);
+ IgnoreUnused(args...);
return nullptr;
}
};
diff --git a/src/backends/backendsCommon/WorkloadUtils.hpp b/src/backends/backendsCommon/WorkloadUtils.hpp
index 92ef2d224f..66056db4ca 100644
--- a/src/backends/backendsCommon/WorkloadUtils.hpp
+++ b/src/backends/backendsCommon/WorkloadUtils.hpp
@@ -55,11 +55,11 @@ void CopyTensorContentsGeneric(const ITensorHandle* srcTensor, ITensorHandle* ds
TensorShape srcStrides = srcTensor->GetStrides();
const TensorShape& srcShape = srcTensor->GetShape();
const auto srcSize = srcTensor->GetStrides()[0] * srcShape[0];
- boost::ignore_unused(srcSize); // Only used for asserts
+ IgnoreUnused(srcSize); // Only used for asserts
TensorShape dstStrides = dstTensor->GetStrides();
const TensorShape& dstShape = dstTensor->GetShape();
const auto dstSize = dstTensor->GetStrides()[0] * dstShape[0];
- boost::ignore_unused(dstSize); // Only used for asserts
+ IgnoreUnused(dstSize); // Only used for asserts
size_t srcDepth = 1;
size_t srcBatches = 1;
diff --git a/src/backends/backendsCommon/test/BackendProfilingTests.cpp b/src/backends/backendsCommon/test/BackendProfilingTests.cpp
index 455533699d..b9e0e45c0a 100644
--- a/src/backends/backendsCommon/test/BackendProfilingTests.cpp
+++ b/src/backends/backendsCommon/test/BackendProfilingTests.cpp
@@ -14,6 +14,7 @@
#include "ProfilingUtils.hpp"
#include "RequestCounterDirectoryCommandHandler.hpp"
+#include <armnn/utility/IgnoreUnused.hpp>
#include <armnn/BackendId.hpp>
#include <armnn/Logging.hpp>
#include <armnn/profiling/ISendTimelinePacket.hpp>
@@ -56,7 +57,7 @@ public:
/// Create and write a CounterDirectoryPacket from the parameters to the buffer.
virtual void SendCounterDirectoryPacket(const ICounterDirectory& counterDirectory)
{
- boost::ignore_unused(counterDirectory);
+ armnn::IgnoreUnused(counterDirectory);
}
/// Create and write a PeriodicCounterCapturePacket from the parameters to the buffer.
@@ -69,8 +70,8 @@ public:
virtual void SendPeriodicCounterSelectionPacket(uint32_t capturePeriod,
const std::vector<uint16_t>& selectedCounterIds)
{
- boost::ignore_unused(capturePeriod);
- boost::ignore_unused(selectedCounterIds);
+ armnn::IgnoreUnused(capturePeriod);
+ armnn::IgnoreUnused(selectedCounterIds);
}
std::vector<Timestamp> GetTimestamps()
diff --git a/src/backends/backendsCommon/test/IsLayerSupportedTestImpl.hpp b/src/backends/backendsCommon/test/IsLayerSupportedTestImpl.hpp
index 395a63d6e6..15608ccdd8 100644
--- a/src/backends/backendsCommon/test/IsLayerSupportedTestImpl.hpp
+++ b/src/backends/backendsCommon/test/IsLayerSupportedTestImpl.hpp
@@ -8,7 +8,7 @@
#include <backendsCommon/WorkloadFactory.hpp>
-#include <boost/core/ignore_unused.hpp>
+#include <armnn/utility/IgnoreUnused.hpp>
namespace
{
@@ -414,7 +414,7 @@ struct LayerTypePolicy<armnn::LayerType::name, DataType> \
static std::unique_ptr<armnn::IWorkload> MakeDummyWorkload(armnn::IWorkloadFactory *factory, \
unsigned int nIn, unsigned int nOut) \
{ \
- boost::ignore_unused(factory, nIn, nOut); \
+ IgnoreUnused(factory, nIn, nOut); \
return std::unique_ptr<armnn::IWorkload>(); \
} \
};
@@ -559,7 +559,7 @@ unsigned int GetNumOutputs(const armnn::Layer& layer)
template<>
unsigned int GetNumInputs<armnn::LayerType::Concat>(const armnn::Layer& layer)
{
- boost::ignore_unused(layer);
+ IgnoreUnused(layer);
return 2;
}
@@ -613,7 +613,7 @@ bool IsLayerSupportedTest(FactoryType *factory, Tag<Type>)
}
catch(const armnn::InvalidArgumentException& e)
{
- boost::ignore_unused(e);
+ IgnoreUnused(e);
// This is ok since we throw InvalidArgumentException when creating the dummy workload.
return true;
}
@@ -644,12 +644,12 @@ bool IsLayerSupportedTest(FactoryType *factory, Tag<Type>)
// InvalidArgumentException or UnimplementedException.
catch(const armnn::InvalidArgumentException& e)
{
- boost::ignore_unused(e);
+ IgnoreUnused(e);
return true;
}
catch(const armnn::UnimplementedException& e)
{
- boost::ignore_unused(e);
+ IgnoreUnused(e);
return true;
}
catch(const std::exception& e)
diff --git a/src/backends/backendsCommon/test/MockBackend.cpp b/src/backends/backendsCommon/test/MockBackend.cpp
index b2388cf45a..8d40117741 100644
--- a/src/backends/backendsCommon/test/MockBackend.cpp
+++ b/src/backends/backendsCommon/test/MockBackend.cpp
@@ -108,7 +108,7 @@ IBackendInternal::IBackendContextPtr MockBackend::CreateBackendContext(const IRu
IBackendInternal::IBackendProfilingContextPtr MockBackend::CreateBackendProfilingContext(
const IRuntime::CreationOptions& options, IBackendProfilingPtr& backendProfiling)
{
- boost::ignore_unused(options);
+ IgnoreUnused(options);
std::shared_ptr<armnn::MockBackendProfilingContext> context =
std::make_shared<MockBackendProfilingContext>(backendProfiling);
MockBackendProfilingService::Instance().SetProfilingContextPtr(context);
diff --git a/src/backends/backendsCommon/test/TestDynamicBackend.cpp b/src/backends/backendsCommon/test/TestDynamicBackend.cpp
index cbfe09377e..5018b4459d 100644
--- a/src/backends/backendsCommon/test/TestDynamicBackend.cpp
+++ b/src/backends/backendsCommon/test/TestDynamicBackend.cpp
@@ -7,7 +7,7 @@
#include <armnn/backends/IBackendInternal.hpp>
-#include <boost/core/ignore_unused.hpp>
+#include <armnn/utility/IgnoreUnused.hpp>
constexpr const char* TestDynamicBackendId()
{
@@ -65,7 +65,7 @@ public:
}
IWorkloadFactoryPtr CreateWorkloadFactory(const IMemoryManagerSharedPtr& memoryManager) const override
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
return IWorkloadFactoryPtr{};
}
ILayerSupportSharedPtr GetLayerSupport() const override
diff --git a/src/backends/backendsCommon/test/layerTests/ActivationTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/ActivationTestImpl.cpp
index 6993b9e9b1..319434e093 100644
--- a/src/backends/backendsCommon/test/layerTests/ActivationTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/ActivationTestImpl.cpp
@@ -36,7 +36,7 @@ LayerTestResult<T, 4> BoundedReLuTestCommon(
unsigned int inputChannels,
unsigned int inputBatchSize)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
unsigned int outputWidth = inputWidth;
unsigned int outputHeight = inputHeight;
unsigned int outputChannels = inputChannels;
@@ -245,7 +245,7 @@ boost::multi_array<float, 4> BoundedReLuRandomInputTest(
float upperBound,
const armnn::ActivationDescriptor& activationDescriptor)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
const armnn::TensorInfo inputTensorInfo = BoundedReLuRandomInputTestTraits::GetInputTensorInfo();
const armnn::TensorInfo outputTensorInfo = BoundedReLuRandomInputTestTraits::GetOutputTensorInfo();
@@ -310,7 +310,7 @@ LayerTestResult<T,4> ConstantLinearActivationTestCommon(
float qScale = 0.0f,
int32_t qOffset = 0)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
unsigned int inputHeight = 20;
unsigned int inputWidth = 17;
unsigned int inputChannels = 3;
@@ -402,7 +402,7 @@ LayerTestResult<T, 4> SimpleActivationTest(
int32_t outOffset,
const std::vector<float>& outputExpectedData)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
constexpr static unsigned int inputWidth = 16u;
constexpr static unsigned int inputHeight = 1u;
constexpr static unsigned int inputChannels = 1u;
@@ -793,7 +793,7 @@ LayerTestResult<float, 5> SqrtNNTest(
armnn::IWorkloadFactory& workloadFactory,
const armnn::IBackendInternal::IMemoryManagerSharedPtr& memoryManager)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
const int inputDataSize = 120;
std::vector<float> inputData(inputDataSize);
@@ -1148,7 +1148,7 @@ LayerTestResult<T,4> CompareActivationTestImpl(
float qScale = 0.0f,
int32_t qOffset = 0)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
unsigned int width = 17;
unsigned int height = 29;
unsigned int channels = 2;
diff --git a/src/backends/backendsCommon/test/layerTests/AdditionTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/AdditionTestImpl.cpp
index 82dc59b66b..bfe0282dc4 100644
--- a/src/backends/backendsCommon/test/layerTests/AdditionTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/AdditionTestImpl.cpp
@@ -165,7 +165,7 @@ LayerTestResult<T, 4> AdditionBroadcastTestImpl(
float qScale,
int32_t qOffset)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
armnn::TensorInfo inputTensorInfo1 = armnn::TensorInfo({1, 3, 2, 1}, ArmnnType);
armnn::TensorInfo inputTensorInfo2 = armnn::TensorInfo({1, 1, 2, 3}, ArmnnType);
armnn::TensorInfo outputTensorInfo = armnn::TensorInfo({1, 3, 2, 3}, ArmnnType);
@@ -248,7 +248,7 @@ LayerTestResult<T, 4> AdditionBroadcast1ElementTestImpl(
float qScale,
int32_t qOffset)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
armnn::TensorInfo inputTensorInfo1 = armnn::TensorInfo({1, 3, 2, 3}, ArmnnType);
armnn::TensorInfo inputTensorInfo2 = armnn::TensorInfo({1, 1, 1, 1}, ArmnnType);
armnn::TensorInfo outputTensorInfo = armnn::TensorInfo({1, 3, 2, 3}, ArmnnType);
@@ -455,7 +455,7 @@ LayerTestResult<float, 4> AdditionAfterMaxPoolTest(
armnn::IWorkloadFactory& workloadFactory,
const armnn::IBackendInternal::IMemoryManagerSharedPtr& memoryManager)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
// Create Initial Tensor
// 1, 2, 3
@@ -563,7 +563,7 @@ LayerTestResult<float,4> CompareAdditionTest(
const armnn::IBackendInternal::IMemoryManagerSharedPtr& memoryManager,
armnn::IWorkloadFactory& refWorkloadFactory)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
unsigned int batchSize = 4;
unsigned int channels = 1;
unsigned int height = 2;
diff --git a/src/backends/backendsCommon/test/layerTests/ArgMinMaxTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/ArgMinMaxTestImpl.cpp
index 7bfccd64d2..20dcef5dd4 100644
--- a/src/backends/backendsCommon/test/layerTests/ArgMinMaxTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/ArgMinMaxTestImpl.cpp
@@ -26,7 +26,7 @@ LayerTestResult<int32_t, 3> ArgMinMaxTestCommon(
const std::vector<int32_t>& outputData,
int axis = 3)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
auto inputTensor = MakeTensor<T, 4>(inputTensorInfo, ConvertToDataType<ArmnnType>(inputData, inputTensorInfo));
LayerTestResult<int32_t, 3> result(outputTensorInfo);
diff --git a/src/backends/backendsCommon/test/layerTests/BatchNormalizationTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/BatchNormalizationTestImpl.cpp
index f64b06d84b..48f7257a2e 100644
--- a/src/backends/backendsCommon/test/layerTests/BatchNormalizationTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/BatchNormalizationTestImpl.cpp
@@ -8,7 +8,7 @@
#include <QuantizeHelper.hpp>
#include <ResolveType.hpp>
-
+#include <armnn/utility/IgnoreUnused.hpp>
#include <armnnUtils/DataLayoutIndexed.hpp>
#include <backendsCommon/CpuTensorHandle.hpp>
@@ -36,7 +36,7 @@ LayerTestResult<T, 4> BatchNormTestImpl(
int32_t qOffset,
armnn::DataLayout dataLayout)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
armnn::TensorInfo inputTensorInfo(inputOutputTensorShape, ArmnnType);
armnn::TensorInfo outputTensorInfo(inputOutputTensorShape, ArmnnType);
@@ -115,7 +115,7 @@ LayerTestResult<T,4> BatchNormTestNhwcImpl(
float qScale,
int32_t qOffset)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
const unsigned int width = 2;
const unsigned int height = 3;
@@ -589,7 +589,7 @@ LayerTestResult<float,4> CompareBatchNormTest(
const armnn::IBackendInternal::IMemoryManagerSharedPtr& memoryManager,
armnn::IWorkloadFactory& refWorkloadFactory)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
const unsigned int width = 2;
const unsigned int height = 3;
const unsigned int channels = 5;
diff --git a/src/backends/backendsCommon/test/layerTests/BatchToSpaceNdTestImpl.hpp b/src/backends/backendsCommon/test/layerTests/BatchToSpaceNdTestImpl.hpp
index 12413667b9..2ba3a0c68c 100644
--- a/src/backends/backendsCommon/test/layerTests/BatchToSpaceNdTestImpl.hpp
+++ b/src/backends/backendsCommon/test/layerTests/BatchToSpaceNdTestImpl.hpp
@@ -39,7 +39,7 @@ LayerTestResult<T, OutputDim> BatchToSpaceNdHelper(
float scale = 1.0f,
int32_t offset = 0)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
armnn::TensorInfo inputTensorInfo(InputDim, inputShape, ArmnnType);
armnn::TensorInfo outputTensorInfo(OutputDim, outputShape, ArmnnType);
diff --git a/src/backends/backendsCommon/test/layerTests/ComparisonTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/ComparisonTestImpl.cpp
index 9f9944d6db..2156b0ee9e 100644
--- a/src/backends/backendsCommon/test/layerTests/ComparisonTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/ComparisonTestImpl.cpp
@@ -43,7 +43,7 @@ LayerTestResult<uint8_t, NumDims> ComparisonTestImpl(
float outQuantScale,
int outQuantOffset)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
BOOST_ASSERT(shape0.GetNumDimensions() == NumDims);
armnn::TensorInfo inputTensorInfo0(shape0, ArmnnInType, quantScale0, quantOffset0);
diff --git a/src/backends/backendsCommon/test/layerTests/ConcatTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/ConcatTestImpl.cpp
index 9d590e3feb..f6f4b09f6a 100644
--- a/src/backends/backendsCommon/test/layerTests/ConcatTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/ConcatTestImpl.cpp
@@ -126,7 +126,7 @@ template<typename T> void PermuteTensorData(
const T * inputData,
std::vector<T>& outputData)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
BOOST_ASSERT_MSG(inputData != nullptr, "inputData must not be null");
if (inputData == nullptr)
{
@@ -178,7 +178,7 @@ template<typename T> void PermuteInputsForConcat(
unsigned int & concatDim,
TensorInfo & outputTensorInfo)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
BOOST_ASSERT_MSG(inputTensorInfos.size() > 1,
"Expecting more than one tensor to be concatenated here");
@@ -1918,7 +1918,7 @@ LayerTestResult<T, 3> ConcatDifferentInputOutputQParamTest(
const IBackendInternal::IMemoryManagerSharedPtr& memoryManager,
bool useSubtensor)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
// Defines the tensor descriptors.
TensorInfo outputTensorInfo({ 3, 6, 3 }, ArmnnType);
@@ -2073,7 +2073,7 @@ LayerTestResult<float,3> ConcatTest(
IWorkloadFactory& workloadFactory,
const IBackendInternal::IMemoryManagerSharedPtr& memoryManager)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
unsigned int outputWidth = 3;
unsigned int outputHeight = 6;
@@ -2346,7 +2346,7 @@ LayerTestResult<uint8_t, 3> ConcatUint8DifferentQParamsTest(
IWorkloadFactory& workloadFactory,
const IBackendInternal::IMemoryManagerSharedPtr& memoryManager)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
unsigned int outputWidth = 3;
unsigned int outputHeight = 6;
@@ -2491,7 +2491,7 @@ LayerTestResult<uint8_t, 3> ConcatUint8Test(
IWorkloadFactory& workloadFactory,
const IBackendInternal::IMemoryManagerSharedPtr& memoryManager)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
unsigned int outputWidth = 3;
unsigned int outputHeight = 6;
@@ -2629,7 +2629,7 @@ LayerTestResult<uint16_t, 3> ConcatUint16Test(
IWorkloadFactory& workloadFactory,
const IBackendInternal::IMemoryManagerSharedPtr& memoryManager)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
unsigned int outputWidth = 3;
unsigned int outputHeight = 6;
diff --git a/src/backends/backendsCommon/test/layerTests/ConstantTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/ConstantTestImpl.cpp
index d11004c069..7a8aac4686 100644
--- a/src/backends/backendsCommon/test/layerTests/ConstantTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/ConstantTestImpl.cpp
@@ -28,7 +28,7 @@ LayerTestResult<T, 4> ConstantTestImpl(
float qScale,
int32_t qOffset)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
constexpr unsigned int inputWidth = 3;
constexpr unsigned int inputHeight = 4;
constexpr unsigned int inputChannels = 3;
diff --git a/src/backends/backendsCommon/test/layerTests/Conv2dTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/Conv2dTestImpl.cpp
index 669398fb54..89cdd96e37 100644
--- a/src/backends/backendsCommon/test/layerTests/Conv2dTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/Conv2dTestImpl.cpp
@@ -8,7 +8,7 @@
#include <QuantizeHelper.hpp>
#include <armnnUtils/TensorUtils.hpp>
-
+#include <armnn/utility/IgnoreUnused.hpp>
#include <armnnUtils/DataLayoutIndexed.hpp>
#include <armnnUtils/Permute.hpp>
@@ -217,7 +217,7 @@ LayerTestResult<T, 4> SimpleConvolution2dTestImpl(
uint32_t dilationX = 1,
uint32_t dilationY = 1)
{
- boost::ignore_unused(memoryManager);
+ armnn::IgnoreUnused(memoryManager);
unsigned int inputHeight = boost::numeric_cast<unsigned int>(originalInput.shape()[2]);
unsigned int inputWidth = boost::numeric_cast<unsigned int>(originalInput.shape()[3]);
unsigned int inputChannels = boost::numeric_cast<unsigned int>(originalInput.shape()[1]);
@@ -381,7 +381,7 @@ LayerTestResult<T, 4> SimpleConvolution2dNhwcTestImpl(
uint32_t strideX = 1,
uint32_t strideY = 1)
{
- boost::ignore_unused(qScale, qOffset);
+ armnn::IgnoreUnused(qScale, qOffset);
unsigned int inputNum = boost::numeric_cast<unsigned int>(input.shape()[0]);
unsigned int inputChannels = boost::numeric_cast<unsigned int>(input.shape()[3]);
unsigned int inputHeight = boost::numeric_cast<unsigned int>(input.shape()[1]);
@@ -587,7 +587,7 @@ LayerTestResult<T, 4> SimpleConvolution2d3x3NhwcTestCommon(
bool biasEnabled,
armnn::DataLayout dataLayout)
{
- boost::ignore_unused(biasEnabled);
+ armnn::IgnoreUnused(biasEnabled);
// Use common single-batch 5x5 image.
armnn::TensorInfo inputDesc({1, 3, 4, 1}, ArmnnType);
@@ -640,7 +640,7 @@ LayerTestResult<T, 4> SimpleConvolution2d3x3Stride2x2TestCommon(
bool biasEnabled,
const armnn::DataLayout& dataLayout)
{
- boost::ignore_unused(biasEnabled);
+ armnn::IgnoreUnused(biasEnabled);
// Input is a single-batch, 1 channel, 5x5 image.
armnn::TensorInfo inputDesc({1, 5, 5, 1}, ArmnnType);
diff --git a/src/backends/backendsCommon/test/layerTests/ConvertFp16ToFp32TestImpl.cpp b/src/backends/backendsCommon/test/layerTests/ConvertFp16ToFp32TestImpl.cpp
index cea6efb549..8b3bbd82e2 100644
--- a/src/backends/backendsCommon/test/layerTests/ConvertFp16ToFp32TestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/ConvertFp16ToFp32TestImpl.cpp
@@ -17,7 +17,7 @@ LayerTestResult<float, 4> SimpleConvertFp16ToFp32Test(
armnn::IWorkloadFactory& workloadFactory,
const armnn::IBackendInternal::IMemoryManagerSharedPtr& memoryManager)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
using namespace half_float::literal;
const armnn::TensorInfo inputTensorInfo({1, 3, 2, 3}, armnn::DataType::Float16);
diff --git a/src/backends/backendsCommon/test/layerTests/ConvertFp32ToFp16TestImpl.cpp b/src/backends/backendsCommon/test/layerTests/ConvertFp32ToFp16TestImpl.cpp
index 9f4eecac2e..1e604719e7 100644
--- a/src/backends/backendsCommon/test/layerTests/ConvertFp32ToFp16TestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/ConvertFp32ToFp16TestImpl.cpp
@@ -15,7 +15,7 @@ LayerTestResult<armnn::Half, 4> SimpleConvertFp32ToFp16Test(
armnn::IWorkloadFactory& workloadFactory,
const armnn::IBackendInternal::IMemoryManagerSharedPtr& memoryManager)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
using namespace half_float::literal;
const armnn::TensorInfo inputTensorInfo({1, 3, 2, 3}, armnn::DataType::Float32);
diff --git a/src/backends/backendsCommon/test/layerTests/DebugTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/DebugTestImpl.cpp
index 92c5d92fcc..149779b9ef 100644
--- a/src/backends/backendsCommon/test/layerTests/DebugTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/DebugTestImpl.cpp
@@ -30,7 +30,7 @@ LayerTestResult<T, Dim> DebugTestImpl(
const float qScale = 1.0f,
const int32_t qOffset = 0)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
if(armnn::IsQuantizedType<T>())
{
inputTensorInfo.SetQuantizationScale(qScale);
diff --git a/src/backends/backendsCommon/test/layerTests/DepthToSpaceTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/DepthToSpaceTestImpl.cpp
index 4ddfb3089f..4d4a6bc156 100644
--- a/src/backends/backendsCommon/test/layerTests/DepthToSpaceTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/DepthToSpaceTestImpl.cpp
@@ -29,7 +29,7 @@ LayerTestResult<T, 4> DepthToSpaceTestImpl(
const float qScale = 1.0f,
const int32_t qOffset = 0)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
if (descriptor.m_Parameters.m_DataLayout == armnn::DataLayout::NCHW)
{
PermuteTensorNhwcToNchw<float>(inputInfo, inputData);
diff --git a/src/backends/backendsCommon/test/layerTests/DequantizeTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/DequantizeTestImpl.cpp
index 91d56bb492..0a4bdb8124 100644
--- a/src/backends/backendsCommon/test/layerTests/DequantizeTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/DequantizeTestImpl.cpp
@@ -26,7 +26,7 @@ LayerTestResult<T1, Dim> DequantizeTestImpl(
const std::vector<T1>& expectedOutputData,
armnn::DequantizeQueueDescriptor descriptor)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
boost::multi_array<T, Dim> input = MakeTensor<T, Dim>(inputTensorInfo, inputData);
LayerTestResult<T1, Dim> ret(outputTensorInfo);
diff --git a/src/backends/backendsCommon/test/layerTests/DivisionTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/DivisionTestImpl.cpp
index 223beb49e8..2359f777b8 100644
--- a/src/backends/backendsCommon/test/layerTests/DivisionTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/DivisionTestImpl.cpp
@@ -20,7 +20,7 @@ LayerTestResult<float, 4> DivisionByZeroTest(
armnn::IWorkloadFactory& workloadFactory,
const armnn::IBackendInternal::IMemoryManagerSharedPtr& memoryManager)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
const unsigned int width = 2u;
const unsigned int height = 2u;
const unsigned int channelCount = 2u;
diff --git a/src/backends/backendsCommon/test/layerTests/FakeQuantizationTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/FakeQuantizationTestImpl.cpp
index d905bde7a0..905f97b968 100644
--- a/src/backends/backendsCommon/test/layerTests/FakeQuantizationTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/FakeQuantizationTestImpl.cpp
@@ -17,7 +17,7 @@ LayerTestResult<float, 2> FakeQuantizationTest(
armnn::IWorkloadFactory& workloadFactory,
const armnn::IBackendInternal::IMemoryManagerSharedPtr& memoryManager)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
constexpr unsigned int width = 2;
constexpr unsigned int height = 3;
diff --git a/src/backends/backendsCommon/test/layerTests/FloorTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/FloorTestImpl.cpp
index ebad7fc91c..444809f69b 100644
--- a/src/backends/backendsCommon/test/layerTests/FloorTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/FloorTestImpl.cpp
@@ -16,7 +16,7 @@ LayerTestResult<T, 4> SimpleFloorTest(
armnn::IWorkloadFactory& workloadFactory,
const armnn::IBackendInternal::IMemoryManagerSharedPtr& memoryManager)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
armnn::TensorInfo inputTensorInfo({1, 3, 2, 3}, ArmnnType);
inputTensorInfo.SetQuantizationScale(0.1f);
diff --git a/src/backends/backendsCommon/test/layerTests/FullyConnectedTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/FullyConnectedTestImpl.cpp
index 7c6122e5fa..43bcfb1d76 100644
--- a/src/backends/backendsCommon/test/layerTests/FullyConnectedTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/FullyConnectedTestImpl.cpp
@@ -34,7 +34,7 @@ LayerTestResult<T, 2> SimpleFullyConnectedTestImpl(
bool biasEnabled,
bool transposeWeights)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
std::unique_ptr<armnn::ITensorHandle> inputHandle = workloadFactory.CreateTensorHandle(inputTensorInfo);
std::unique_ptr<armnn::ITensorHandle> outputHandle = workloadFactory.CreateTensorHandle(outputTensorInfo);
diff --git a/src/backends/backendsCommon/test/layerTests/GatherTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/GatherTestImpl.cpp
index 68410559f7..47adb22fe7 100644
--- a/src/backends/backendsCommon/test/layerTests/GatherTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/GatherTestImpl.cpp
@@ -31,7 +31,7 @@ LayerTestResult<T, OutputDim> GatherTestImpl(
const std::vector<int32_t>& indicesData,
const std::vector<T>& outputData)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
auto params = MakeTensor<T, ParamsDim>(paramsInfo, paramsData);
auto indices = MakeTensor<int32_t, IndicesDim>(indicesInfo, indicesData);
diff --git a/src/backends/backendsCommon/test/layerTests/InstanceNormalizationTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/InstanceNormalizationTestImpl.cpp
index ae28bc03f3..a13198b2f9 100644
--- a/src/backends/backendsCommon/test/layerTests/InstanceNormalizationTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/InstanceNormalizationTestImpl.cpp
@@ -34,7 +34,7 @@ LayerTestResult<T, 4> InstanceNormTestImpl(
float qScale = 0.0f,
int32_t qOffset = 0)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
auto inputTensor = MakeTensor<T, 4>(inputTensorInfo,
armnnUtils::QuantizedVector<T>(inputValues, qScale, qOffset));
diff --git a/src/backends/backendsCommon/test/layerTests/L2NormalizationTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/L2NormalizationTestImpl.cpp
index e500a126f6..4d98e236d3 100644
--- a/src/backends/backendsCommon/test/layerTests/L2NormalizationTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/L2NormalizationTestImpl.cpp
@@ -33,7 +33,7 @@ LayerTestResult<T, 4> L2NormalizationTestImpl(
const armnn::DataLayout layout,
float epsilon = 1e-12f)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
const armnn::TensorInfo inputTensorInfo(inputOutputTensorShape, ArmnnType, scale, offset);
const armnn::TensorInfo outputTensorInfo(inputOutputTensorShape, ArmnnType, outScale, outOffset);
diff --git a/src/backends/backendsCommon/test/layerTests/LogSoftmaxTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/LogSoftmaxTestImpl.cpp
index 392983c311..208bed24a5 100644
--- a/src/backends/backendsCommon/test/layerTests/LogSoftmaxTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/LogSoftmaxTestImpl.cpp
@@ -36,7 +36,7 @@ LayerTestResult<T, NumDims> LogSoftmaxTestImpl(
float qScale = 1.0f,
int32_t qOffset = 0)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
LayerTestResult<T, NumDims> result(outputInfo);
result.outputExpected =
MakeTensor<T, NumDims>(outputInfo, armnnUtils::QuantizedVector<T>(expectedOutputValues, qScale, qOffset));
diff --git a/src/backends/backendsCommon/test/layerTests/LstmTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/LstmTestImpl.cpp
index b12df8a2b9..50ef5c9758 100644
--- a/src/backends/backendsCommon/test/layerTests/LstmTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/LstmTestImpl.cpp
@@ -142,7 +142,7 @@ LstmNoCifgNoPeepholeNoProjectionTestImpl(
int32_t qOffset = 0,
armnn::DataType constantDataType = armnn::DataType::Float32)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
unsigned int batchSize = boost::numeric_cast<unsigned int>(input.shape()[0]);
unsigned int inputSize = boost::numeric_cast<unsigned int>(input.shape()[1]);
unsigned int outputSize = boost::numeric_cast<unsigned int>(outputExpected.shape()[1]);
@@ -345,7 +345,7 @@ LstmLayerNoCifgWithPeepholeWithProjectionTestImpl(armnn::IWorkloadFactory& workl
int32_t qOffset = 0,
armnn::DataType constantDataType = armnn::DataType::Float32)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
unsigned int batchSize = 2;
unsigned int outputSize = 16;
unsigned int inputSize = 5;
@@ -1060,7 +1060,7 @@ LayerTestResult<T, 2> LstmLayerWithCifgWithPeepholeNoProjectionTestImpl(
int32_t qOffset = 0,
armnn::DataType constantDataType = armnn::DataType::Float32)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
bool cifgEnabled = true;
bool peepholeEnabled = true;
bool projectionEnabled = false;
@@ -1285,7 +1285,7 @@ LstmLayerNoCifgWithPeepholeWithProjectionWithLayerNormTestImpl(armnn::IWorkloadF
int32_t qOffset = 0,
armnn::DataType constantDataType = armnn::DataType::Float32)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
unsigned int batchSize = 2;
unsigned int outputSize = 3;
unsigned int inputSize = 5;
@@ -1552,7 +1552,7 @@ LayerTestResult<uint8_t, 2> QuantizedLstmTestImpl(
const boost::multi_array<uint8_t, 2>& input,
const boost::multi_array<uint8_t, 2>& outputExpected)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
auto numBatches = boost::numeric_cast<unsigned int>(input.shape()[0]);
auto inputSize = boost::numeric_cast<unsigned int>(input.shape()[1]);
auto outputSize = boost::numeric_cast<unsigned int>(outputExpected.shape()[1]);
diff --git a/src/backends/backendsCommon/test/layerTests/MaximumTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/MaximumTestImpl.cpp
index 5147cffddb..0e66d9fecd 100644
--- a/src/backends/backendsCommon/test/layerTests/MaximumTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/MaximumTestImpl.cpp
@@ -19,7 +19,7 @@ std::unique_ptr<armnn::IWorkload> CreateWorkload<armnn::MaximumQueueDescriptor>(
LayerTestResult<float, 4> MaximumSimpleTest(armnn::IWorkloadFactory& workloadFactory,
const armnn::IBackendInternal::IMemoryManagerSharedPtr& memoryManager)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
const unsigned int width = 2u;
const unsigned int height = 2u;
const unsigned int channelCount = 2u;
diff --git a/src/backends/backendsCommon/test/layerTests/MeanTestImpl.hpp b/src/backends/backendsCommon/test/layerTests/MeanTestImpl.hpp
index b8eae1c93e..cd7b22ea42 100644
--- a/src/backends/backendsCommon/test/layerTests/MeanTestImpl.hpp
+++ b/src/backends/backendsCommon/test/layerTests/MeanTestImpl.hpp
@@ -28,7 +28,7 @@ LayerTestResult<T, OutputDim> MeanTestHelper(
float scale = 1.0f,
int32_t offset = 0)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
armnn::TensorInfo inputTensorInfo(InputDim, inputShape, ArmnnType);
armnn::TensorInfo outputTensorInfo(OutputDim, outputShape, ArmnnType);
diff --git a/src/backends/backendsCommon/test/layerTests/MinimumTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/MinimumTestImpl.cpp
index a0a4029115..ae54746ad8 100644
--- a/src/backends/backendsCommon/test/layerTests/MinimumTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/MinimumTestImpl.cpp
@@ -20,7 +20,7 @@ LayerTestResult<float, 4> MinimumBroadcast1ElementTest1(
armnn::IWorkloadFactory& workloadFactory,
const armnn::IBackendInternal::IMemoryManagerSharedPtr& memoryManager)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
unsigned int shape0[] = { 1, 2, 2, 2 };
unsigned int shape1[] = { 1, 1, 1, 1 };
diff --git a/src/backends/backendsCommon/test/layerTests/MultiplicationTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/MultiplicationTestImpl.cpp
index d32e0cf89b..1a9cf5b173 100644
--- a/src/backends/backendsCommon/test/layerTests/MultiplicationTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/MultiplicationTestImpl.cpp
@@ -401,7 +401,7 @@ LayerTestResult<float,4> CompareMultiplicationTest(
const armnn::IBackendInternal::IMemoryManagerSharedPtr& memoryManager,
armnn::IWorkloadFactory& refWorkloadFactory)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
const unsigned int width = 16;
const unsigned int height = 32;
const unsigned int channelCount = 2;
diff --git a/src/backends/backendsCommon/test/layerTests/NormalizationTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/NormalizationTestImpl.cpp
index ef828555fa..ef3a45b391 100644
--- a/src/backends/backendsCommon/test/layerTests/NormalizationTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/NormalizationTestImpl.cpp
@@ -24,7 +24,7 @@ LayerTestResult<float,4> SimpleNormalizationTestImpl(
armnn::NormalizationAlgorithmChannel normChannel,
armnn::NormalizationAlgorithmMethod normMethod)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
const unsigned int inputHeight = 2;
const unsigned int inputWidth = 2;
const unsigned int inputChannels = 1;
diff --git a/src/backends/backendsCommon/test/layerTests/PadTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/PadTestImpl.cpp
index 9239c665eb..69c651b5cd 100644
--- a/src/backends/backendsCommon/test/layerTests/PadTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/PadTestImpl.cpp
@@ -24,7 +24,7 @@ LayerTestResult<T, 2> Pad2dTestCommon(
int32_t qOffset,
const float customPaddingValue)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
const armnn::TensorShape inputShape{ 3, 3 };
const armnn::TensorShape outputShape{ 7, 7 };
@@ -96,7 +96,7 @@ LayerTestResult<T, 3> Pad3dTestCommon(
float qScale,
int32_t qOffset)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
const armnn::TensorShape inputShape{ 2, 2, 2 };
const armnn::TensorShape outputShape{ 3, 5, 6 };
@@ -180,7 +180,7 @@ LayerTestResult<T, 4> Pad4dTestCommon(
float qScale,
int32_t qOffset)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
const armnn::TensorShape inputShape{ 2, 2, 3, 2 };
const armnn::TensorShape outputShape{ 4, 5, 7, 4 };
diff --git a/src/backends/backendsCommon/test/layerTests/PermuteTestImpl.hpp b/src/backends/backendsCommon/test/layerTests/PermuteTestImpl.hpp
index 9460592888..71e15334e7 100644
--- a/src/backends/backendsCommon/test/layerTests/PermuteTestImpl.hpp
+++ b/src/backends/backendsCommon/test/layerTests/PermuteTestImpl.hpp
@@ -25,7 +25,7 @@ LayerTestResult<T, 4> SimplePermuteTestImpl(
const std::vector<T>& inputData,
const std::vector<T>& outputExpectedData)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
auto input = MakeTensor<T, 4>(inputTensorInfo, inputData);
LayerTestResult<T, 4> ret(outputTensorInfo);
diff --git a/src/backends/backendsCommon/test/layerTests/Pooling2dTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/Pooling2dTestImpl.cpp
index b58e9826b8..89e46fbdb1 100644
--- a/src/backends/backendsCommon/test/layerTests/Pooling2dTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/Pooling2dTestImpl.cpp
@@ -14,6 +14,8 @@
#include <armnnUtils/DataLayoutIndexed.hpp>
#include <armnnUtils/Permute.hpp>
+#include <armnn/utility/IgnoreUnused.hpp>
+
#include <backendsCommon/WorkloadInfo.hpp>
#include <backendsCommon/test/TensorCopyUtils.hpp>
@@ -38,7 +40,7 @@ LayerTestResult<T, 4> SimplePooling2dTestImpl(
const boost::multi_array<T, 4>& input,
const boost::multi_array<T, 4>& outputExpected)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
const armnn::DataLayout dataLayout = descriptor.m_DataLayout;
const armnnUtils::DataLayoutIndexed dimensionIndices = dataLayout;
auto heightIndex = dimensionIndices.GetHeightIndex();
@@ -740,7 +742,7 @@ LayerTestResult<T, 4> ComparePooling2dTestCommon(
float qScale = 1.0f,
int32_t qOffset = 0)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
const unsigned int inputWidth = 16;
const unsigned int inputHeight = 32;
const unsigned int channelCount = 2;
diff --git a/src/backends/backendsCommon/test/layerTests/PreluTestImpl.hpp b/src/backends/backendsCommon/test/layerTests/PreluTestImpl.hpp
index 7138b46f88..3b6c2d8412 100644
--- a/src/backends/backendsCommon/test/layerTests/PreluTestImpl.hpp
+++ b/src/backends/backendsCommon/test/layerTests/PreluTestImpl.hpp
@@ -24,7 +24,7 @@ LayerTestResult<T, 4> PreluTest(
armnn::IWorkloadFactory& workloadFactory,
const armnn::IBackendInternal::IMemoryManagerSharedPtr& memoryManager)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
armnn::TensorInfo inputTensorInfo ({ 1, 2, 2, 3 }, ArmnnType);
armnn::TensorInfo alphaTensorInfo ({ 1, 1, 1, 3 }, ArmnnType);
diff --git a/src/backends/backendsCommon/test/layerTests/QuantizeTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/QuantizeTestImpl.cpp
index e8996d4a51..673bfef678 100644
--- a/src/backends/backendsCommon/test/layerTests/QuantizeTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/QuantizeTestImpl.cpp
@@ -29,7 +29,7 @@ LayerTestResult<T, Dim> QuantizeTestImpl(
const std::vector<T>& expectedOutputData,
armnn::QuantizeQueueDescriptor descriptor)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
boost::multi_array<float, Dim> input = MakeTensor<float, Dim>(inputTensorInfo, inputData);
LayerTestResult<T, Dim> ret(outputTensorInfo);
diff --git a/src/backends/backendsCommon/test/layerTests/ReshapeTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/ReshapeTestImpl.cpp
index 894ece65a5..5ed947d8c3 100644
--- a/src/backends/backendsCommon/test/layerTests/ReshapeTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/ReshapeTestImpl.cpp
@@ -23,7 +23,7 @@ LayerTestResult<T, NumDims> SimpleReshapeTestImpl(
const std::vector<T>& inputData,
const std::vector<T>& outputExpectedData)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
auto input = MakeTensor<T, NumDims>(inputTensorInfo, inputData);
LayerTestResult<T, NumDims> ret(outputTensorInfo);
diff --git a/src/backends/backendsCommon/test/layerTests/ResizeTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/ResizeTestImpl.cpp
index a6e0ca19a5..e95f18b7a5 100644
--- a/src/backends/backendsCommon/test/layerTests/ResizeTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/ResizeTestImpl.cpp
@@ -73,7 +73,7 @@ LayerTestResult<T, NumDims> ResizeTestImpl(
const armnn::IBackendInternal::IMemoryManagerSharedPtr& memoryManager,
const ResizeTestParams& params)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
armnn::TensorInfo inputInfo(params.m_InputShape, ArmnnType);
armnn::TensorInfo outputInfo(params.m_OutputShape, ArmnnType);
diff --git a/src/backends/backendsCommon/test/layerTests/SliceTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/SliceTestImpl.cpp
index 09f15c022e..df3b6238c5 100644
--- a/src/backends/backendsCommon/test/layerTests/SliceTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/SliceTestImpl.cpp
@@ -29,7 +29,7 @@ LayerTestResult<T, NumDims> SliceTestImpl(
const float qScale = 1.0f,
const int qOffset = 0)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
if(armnn::IsQuantizedType<T>())
{
inputInfo.SetQuantizationScale(qScale);
diff --git a/src/backends/backendsCommon/test/layerTests/SoftmaxTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/SoftmaxTestImpl.cpp
index 044589b9d2..772ae2ccc7 100644
--- a/src/backends/backendsCommon/test/layerTests/SoftmaxTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/SoftmaxTestImpl.cpp
@@ -65,7 +65,7 @@ LayerTestResult<T, n> SimpleSoftmaxBaseTestImpl(
const std::vector<float>& inputData,
int axis = 1)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
using std::exp;
const float qScale = 1.f / 256.f;
diff --git a/src/backends/backendsCommon/test/layerTests/SpaceToBatchNdTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/SpaceToBatchNdTestImpl.cpp
index ed35413e1e..d1bc2a950f 100644
--- a/src/backends/backendsCommon/test/layerTests/SpaceToBatchNdTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/SpaceToBatchNdTestImpl.cpp
@@ -31,7 +31,7 @@ LayerTestResult<T, 4> SpaceToBatchNdTestImpl(
const float qScale = 1.0f,
const int32_t qOffset = 0)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
const armnn::PermutationVector NCHWToNHWC = {0, 3, 1, 2};
if (descriptor.m_Parameters.m_DataLayout == armnn::DataLayout::NHWC)
{
diff --git a/src/backends/backendsCommon/test/layerTests/SpaceToDepthTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/SpaceToDepthTestImpl.cpp
index bf2f48ca61..c6a5bbed0e 100644
--- a/src/backends/backendsCommon/test/layerTests/SpaceToDepthTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/SpaceToDepthTestImpl.cpp
@@ -31,7 +31,7 @@ LayerTestResult<T, 4> SpaceToDepthTestImpl(
const float qScale = 1.0f,
const int32_t qOffset = 0)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
const armnn::PermutationVector NHWCToNCHW = {0, 2, 3, 1};
if (descriptor.m_Parameters.m_DataLayout == armnn::DataLayout::NCHW)
diff --git a/src/backends/backendsCommon/test/layerTests/SplitterTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/SplitterTestImpl.cpp
index f55aca1baa..88b18b9732 100644
--- a/src/backends/backendsCommon/test/layerTests/SplitterTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/SplitterTestImpl.cpp
@@ -24,7 +24,7 @@ std::vector<LayerTestResult<T,3>> SplitterTestCommon(
float qScale = 0.0f,
int32_t qOffset = 0)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
unsigned int inputWidth = 5;
unsigned int inputHeight = 6;
unsigned int inputChannels = 3;
@@ -257,7 +257,7 @@ LayerTestResult<T, 3> CopyViaSplitterTestImpl(
const armnn::IBackendInternal::IMemoryManagerSharedPtr& memoryManager,
float qScale, int32_t qOffset)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
const armnn::TensorInfo tensorInfo({ 3, 6, 5 }, ArmnnType, qScale, qOffset);
auto input = MakeTensor<T, 3>(
tensorInfo,
diff --git a/src/backends/backendsCommon/test/layerTests/StackTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/StackTestImpl.cpp
index 45dff96ea7..eeaa846972 100644
--- a/src/backends/backendsCommon/test/layerTests/StackTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/StackTestImpl.cpp
@@ -30,7 +30,7 @@ LayerTestResult<T, outputDimLength> StackTestHelper(
const std::vector<std::vector<T>>& inputData,
const std::vector<T>& outputExpectedData)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
unsigned int numInputs = static_cast<unsigned int>(inputData.size());
std::vector<boost::multi_array<T, outputDimLength-1>> inputs;
for (unsigned int i = 0; i < numInputs; ++i)
diff --git a/src/backends/backendsCommon/test/layerTests/StridedSliceTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/StridedSliceTestImpl.cpp
index 63a95b1008..b857a1b23b 100644
--- a/src/backends/backendsCommon/test/layerTests/StridedSliceTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/StridedSliceTestImpl.cpp
@@ -29,7 +29,7 @@ LayerTestResult<T, OutDim> StridedSliceTestImpl(
const float qScale = 1.0f,
const int32_t qOffset = 0)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
if(armnn::IsQuantizedType<T>())
{
inputTensorInfo.SetQuantizationScale(qScale);
diff --git a/src/backends/backendsCommon/test/layerTests/TransposeConvolution2dTestImpl.cpp b/src/backends/backendsCommon/test/layerTests/TransposeConvolution2dTestImpl.cpp
index 378ec46bd1..07f52584ca 100644
--- a/src/backends/backendsCommon/test/layerTests/TransposeConvolution2dTestImpl.cpp
+++ b/src/backends/backendsCommon/test/layerTests/TransposeConvolution2dTestImpl.cpp
@@ -51,7 +51,7 @@ void TransposeConvolution2dTestImpl(armnn::IWorkloadFactory& workloadFactory,
const TensorData<T>& weights,
const armnn::Optional<TensorData<BT>>& biases)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
using namespace armnn;
VerifyInputTensorData(input, "input");
diff --git a/src/backends/backendsCommon/test/layerTests/TransposeTestImpl.hpp b/src/backends/backendsCommon/test/layerTests/TransposeTestImpl.hpp
index 3949dcc142..0e0f317a3e 100644
--- a/src/backends/backendsCommon/test/layerTests/TransposeTestImpl.hpp
+++ b/src/backends/backendsCommon/test/layerTests/TransposeTestImpl.hpp
@@ -25,7 +25,7 @@ LayerTestResult<T, 4> SimpleTransposeTestImpl(
const std::vector<T>& inputData,
const std::vector<T>& outputExpectedData)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
auto input = MakeTensor<T, 4>(inputTensorInfo, inputData);
LayerTestResult<T, 4> ret(outputTensorInfo);
diff --git a/src/backends/cl/ClContextControl.cpp b/src/backends/cl/ClContextControl.cpp
index 72c8e9fe45..f307133085 100644
--- a/src/backends/cl/ClContextControl.cpp
+++ b/src/backends/cl/ClContextControl.cpp
@@ -9,13 +9,14 @@
#include <LeakChecking.hpp>
+#include <armnn/utility/IgnoreUnused.hpp>
+
#include <arm_compute/core/CL/CLKernelLibrary.h>
#include <arm_compute/runtime/CL/CLScheduler.h>
#include <boost/assert.hpp>
#include <boost/format.hpp>
#include <boost/polymorphic_cast.hpp>
-#include <boost/core/ignore_unused.hpp>
namespace cl
{
@@ -33,7 +34,7 @@ ClContextControl::ClContextControl(arm_compute::CLTuner *tuner,
, m_ProfilingEnabled(profilingEnabled)
{
// Ignore m_ProfilingEnabled if unused to avoid compiling problems when ArmCompute is disabled.
- boost::ignore_unused(m_ProfilingEnabled);
+ IgnoreUnused(m_ProfilingEnabled);
try
{
diff --git a/src/backends/cl/ClLayerSupport.cpp b/src/backends/cl/ClLayerSupport.cpp
index d3ac98655a..cdb93d7218 100644
--- a/src/backends/cl/ClLayerSupport.cpp
+++ b/src/backends/cl/ClLayerSupport.cpp
@@ -6,14 +6,13 @@
#include "ClLayerSupport.hpp"
#include "ClBackendId.hpp"
+#include <armnn/utility/IgnoreUnused.hpp>
#include <armnn/Descriptors.hpp>
#include <armnn/BackendRegistry.hpp>
#include <InternalTypes.hpp>
#include <LayerSupportCommon.hpp>
-#include <boost/core/ignore_unused.hpp>
-
#if defined(ARMCOMPUTECL_ENABLED)
#include <aclCommon/ArmComputeUtils.hpp>
#include <aclCommon/ArmComputeTensorUtils.hpp>
@@ -63,7 +62,6 @@
#include "workloads/ClTransposeWorkload.hpp"
#endif
-using namespace boost;
namespace armnn
{
@@ -93,7 +91,7 @@ bool IsMatchingStride(uint32_t actualStride)
template<typename ... Args>
bool IsClBackendSupported(Optional<std::string&> reasonIfUnsupported, Args... args)
{
- boost::ignore_unused(reasonIfUnsupported, (args)...);
+ IgnoreUnused(reasonIfUnsupported, (args)...);
#if defined(ARMCOMPUTECL_ENABLED)
return true;
#else
@@ -649,7 +647,7 @@ bool ClLayerSupport::IsReshapeSupported(const TensorInfo& input,
const ReshapeDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
{
- ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
FORWARD_WORKLOAD_VALIDATE_FUNC(ClReshapeWorkloadValidate, reasonIfUnsupported, input, output);
}
@@ -728,7 +726,7 @@ bool ClLayerSupport::IsSplitterSupported(const TensorInfo& input,
const ViewsDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
{
- ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
return IsSupportedForDataTypeCl(reasonIfUnsupported,
input.GetDataType(),
&TrueFunc<>,
@@ -756,7 +754,7 @@ bool ClLayerSupport::IsSplitterSupported(const TensorInfo& input,
*splitAxis.begin());
}
#endif
- boost::ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
for (auto output : outputs)
{
if (!input.IsTypeSpaceMatch(output)) // Cannot use sub-tensors if the types are not same space
diff --git a/src/backends/cl/ClWorkloadFactory.cpp b/src/backends/cl/ClWorkloadFactory.cpp
index 21c26296af..e7e4fa7e1b 100644
--- a/src/backends/cl/ClWorkloadFactory.cpp
+++ b/src/backends/cl/ClWorkloadFactory.cpp
@@ -9,6 +9,7 @@
#include <armnn/Exceptions.hpp>
#include <armnn/Utils.hpp>
+#include <armnn/utility/IgnoreUnused.hpp>
#include <backendsCommon/CpuTensorHandle.hpp>
#include <backendsCommon/MakeWorkloadHelper.hpp>
@@ -23,7 +24,6 @@
#include <arm_compute/runtime/CL/CLBufferAllocator.h>
#include <arm_compute/runtime/CL/CLScheduler.h>
-#include <boost/core/ignore_unused.hpp>
#include <boost/polymorphic_cast.hpp>
#include <boost/format.hpp>
@@ -85,7 +85,7 @@ ClWorkloadFactory::ClWorkloadFactory(const std::shared_ptr<ClMemoryManager>& mem
std::unique_ptr<ITensorHandle> ClWorkloadFactory::CreateTensorHandle(const TensorInfo& tensorInfo,
const bool IsMemoryManaged) const
{
- boost::ignore_unused(IsMemoryManaged);
+ IgnoreUnused(IsMemoryManaged);
std::unique_ptr<ClTensorHandle> tensorHandle = std::make_unique<ClTensorHandle>(tensorInfo);
tensorHandle->SetMemoryGroup(m_MemoryManager->GetInterLayerMemoryGroup());
@@ -96,7 +96,7 @@ std::unique_ptr<ITensorHandle> ClWorkloadFactory::CreateTensorHandle(const Tenso
DataLayout dataLayout,
const bool IsMemoryManaged) const
{
- boost::ignore_unused(IsMemoryManaged);
+ IgnoreUnused(IsMemoryManaged);
std::unique_ptr<ClTensorHandle> tensorHandle = std::make_unique<ClTensorHandle>(tensorInfo, dataLayout);
tensorHandle->SetMemoryGroup(m_MemoryManager->GetInterLayerMemoryGroup());
@@ -131,7 +131,7 @@ std::unique_ptr<ITensorHandle> ClWorkloadFactory::CreateSubTensorHandle(ITensorH
std::unique_ptr<IWorkload> ClWorkloadFactory::CreateAbs(const AbsQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- boost::ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
ElementwiseUnaryQueueDescriptor elementwiseUnaryDescriptor;
elementwiseUnaryDescriptor.m_Parameters = ElementwiseUnaryDescriptor(UnaryOperation::Abs);
@@ -279,7 +279,7 @@ std::unique_ptr<IWorkload> ClWorkloadFactory::CreateElementwiseUnary(const Eleme
std::unique_ptr<IWorkload> ClWorkloadFactory::CreateEqual(const EqualQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- boost::ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
ComparisonQueueDescriptor comparisonDescriptor;
comparisonDescriptor.m_Parameters = ComparisonDescriptor(ComparisonOperation::Equal);
@@ -308,7 +308,7 @@ std::unique_ptr<IWorkload> ClWorkloadFactory::CreateGather(const GatherQueueDesc
std::unique_ptr<IWorkload> ClWorkloadFactory::CreateGreater(const GreaterQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- boost::ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
ComparisonQueueDescriptor comparisonDescriptor;
comparisonDescriptor.m_Parameters = ComparisonDescriptor(ComparisonOperation::Greater);
@@ -477,7 +477,7 @@ std::unique_ptr<IWorkload> ClWorkloadFactory::CreateResizeBilinear(const ResizeB
std::unique_ptr<IWorkload> ClWorkloadFactory::CreateRsqrt(const RsqrtQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- boost::ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
ElementwiseUnaryQueueDescriptor elementwiseUnaryDescriptor;
elementwiseUnaryDescriptor.m_Parameters = ElementwiseUnaryDescriptor(UnaryOperation::Rsqrt);
diff --git a/src/backends/cl/OpenClTimer.cpp b/src/backends/cl/OpenClTimer.cpp
index ee3c114ba0..5f106993ec 100644
--- a/src/backends/cl/OpenClTimer.cpp
+++ b/src/backends/cl/OpenClTimer.cpp
@@ -5,10 +5,11 @@
#include "OpenClTimer.hpp"
+#include <armnn/utility/IgnoreUnused.hpp>
+
#include <string>
#include <sstream>
-#include <boost/core/ignore_unused.hpp>
namespace armnn
{
@@ -31,7 +32,7 @@ void OpenClTimer::Start()
const cl_event * event_wait_list,
cl_event * event)
{
- boost::ignore_unused(event);
+ IgnoreUnused(event);
cl_int retVal = 0;
// Get the name of the kernel
diff --git a/src/backends/cl/test/ClRuntimeTests.cpp b/src/backends/cl/test/ClRuntimeTests.cpp
index 9aa36173d0..a0d7963aa8 100644
--- a/src/backends/cl/test/ClRuntimeTests.cpp
+++ b/src/backends/cl/test/ClRuntimeTests.cpp
@@ -9,8 +9,8 @@
#include <backendsCommon/test/RuntimeTestImpl.hpp>
#include <test/ProfilingTestUtils.hpp>
+#include <armnn/utility/IgnoreUnused.hpp>
-#include <boost/core/ignore_unused.hpp>
#include <boost/test/unit_test.hpp>
#ifdef WITH_VALGRIND
@@ -144,8 +144,8 @@ BOOST_AUTO_TEST_CASE(RuntimeMemoryUsage)
// These are needed because VALGRIND_COUNT_LEAKS is a macro that assigns to the parameters
// so they are assigned to, but still considered unused, causing a warning.
- boost::ignore_unused(dubious);
- boost::ignore_unused(suppressed);
+ IgnoreUnused(dubious);
+ IgnoreUnused(suppressed);
}
#endif
diff --git a/src/backends/cl/test/Fp16SupportTest.cpp b/src/backends/cl/test/Fp16SupportTest.cpp
index ee5163f668..b7d274fdca 100644
--- a/src/backends/cl/test/Fp16SupportTest.cpp
+++ b/src/backends/cl/test/Fp16SupportTest.cpp
@@ -11,8 +11,8 @@
#include <Graph.hpp>
#include <Optimizer.hpp>
#include <backendsCommon/CpuTensorHandle.hpp>
+#include <armnn/utility/IgnoreUnused.hpp>
-#include <boost/core/ignore_unused.hpp>
#include <boost/test/unit_test.hpp>
#include <set>
diff --git a/src/backends/neon/NeonLayerSupport.cpp b/src/backends/neon/NeonLayerSupport.cpp
index 7e58dabe93..78776124ec 100644
--- a/src/backends/neon/NeonLayerSupport.cpp
+++ b/src/backends/neon/NeonLayerSupport.cpp
@@ -13,8 +13,7 @@
#include <InternalTypes.hpp>
#include <LayerSupportCommon.hpp>
-
-#include <boost/core/ignore_unused.hpp>
+#include <armnn/utility/IgnoreUnused.hpp>
#if defined(ARMCOMPUTENEON_ENABLED)
#include <aclCommon/ArmComputeUtils.hpp>
@@ -62,8 +61,6 @@
#include "workloads/NeonTransposeWorkload.hpp"
#endif
-using namespace boost;
-
namespace armnn
{
@@ -73,7 +70,7 @@ namespace
template< typename ... Args>
bool IsNeonBackendSupported(Optional<std::string&> reasonIfUnsupported, Args... args)
{
- boost::ignore_unused(reasonIfUnsupported, (args)...);
+ IgnoreUnused(reasonIfUnsupported, (args)...);
#if defined(ARMCOMPUTENEON_ENABLED)
return true;
#else
@@ -134,7 +131,7 @@ bool NeonLayerSupport::IsActivationSupported(const TensorInfo& input,
const ActivationDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
{
- ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
FORWARD_WORKLOAD_VALIDATE_FUNC(NeonActivationWorkloadValidate,
reasonIfUnsupported,
input,
@@ -268,9 +265,9 @@ bool NeonLayerSupport::IsConvertFp16ToFp32Supported(const TensorInfo& input,
const TensorInfo& output,
Optional<std::string&> reasonIfUnsupported) const
{
- ignore_unused(input);
- ignore_unused(output);
- ignore_unused(reasonIfUnsupported);
+ armnn::IgnoreUnused(input);
+ armnn::IgnoreUnused(output);
+ armnn::IgnoreUnused(reasonIfUnsupported);
return true;
}
@@ -278,9 +275,9 @@ bool NeonLayerSupport::IsConvertFp32ToFp16Supported(const TensorInfo& input,
const TensorInfo& output,
Optional<std::string&> reasonIfUnsupported) const
{
- ignore_unused(input);
- ignore_unused(output);
- ignore_unused(reasonIfUnsupported);
+ armnn::IgnoreUnused(input);
+ armnn::IgnoreUnused(output);
+ armnn::IgnoreUnused(reasonIfUnsupported);
return true;
}
@@ -381,7 +378,7 @@ bool NeonLayerSupport::IsFloorSupported(const TensorInfo& input,
const TensorInfo& output,
Optional<std::string&> reasonIfUnsupported) const
{
- ignore_unused(output);
+ armnn::IgnoreUnused(output);
return IsNeonBackendSupported(reasonIfUnsupported) &&
IsSupportedForDataTypeGeneric(reasonIfUnsupported,
input.GetDataType(),
@@ -622,7 +619,7 @@ bool NeonLayerSupport::IsReshapeSupported(const TensorInfo& input,
const ReshapeDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
{
- ignore_unused(descriptor);
+ armnn::IgnoreUnused(descriptor);
FORWARD_WORKLOAD_VALIDATE_FUNC(NeonReshapeWorkloadValidate,
reasonIfUnsupported,
input,
@@ -712,7 +709,7 @@ bool NeonLayerSupport::IsSplitterSupported(const TensorInfo& input,
const ViewsDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
{
- ignore_unused(descriptor);
+ armnn::IgnoreUnused(descriptor);
return IsSupportedForDataTypeNeon(reasonIfUnsupported,
input.GetDataType(),
&TrueFunc<>,
@@ -740,7 +737,7 @@ bool NeonLayerSupport::IsSplitterSupported(const TensorInfo& input,
*splitAxis.begin());
}
#endif
- boost::ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
for (auto output : outputs)
{
if (!input.IsTypeSpaceMatch(output)) // Cannot use sub-tensors if the types are not same space
diff --git a/src/backends/neon/NeonTensorHandleFactory.cpp b/src/backends/neon/NeonTensorHandleFactory.cpp
index d5fef4ea95..26b14af144 100644
--- a/src/backends/neon/NeonTensorHandleFactory.cpp
+++ b/src/backends/neon/NeonTensorHandleFactory.cpp
@@ -6,7 +6,7 @@
#include "NeonTensorHandleFactory.hpp"
#include "NeonTensorHandle.hpp"
-#include <boost/core/ignore_unused.hpp>
+#include <armnn/utility/IgnoreUnused.hpp>
namespace armnn
{
diff --git a/src/backends/neon/NeonWorkloadFactory.cpp b/src/backends/neon/NeonWorkloadFactory.cpp
index 939590cad6..cf9999f5d2 100644
--- a/src/backends/neon/NeonWorkloadFactory.cpp
+++ b/src/backends/neon/NeonWorkloadFactory.cpp
@@ -10,6 +10,7 @@
#include <Layer.hpp>
#include <armnn/Utils.hpp>
+#include <armnn/utility/IgnoreUnused.hpp>
#include <backendsCommon/CpuTensorHandle.hpp>
#include <backendsCommon/MakeWorkloadHelper.hpp>
@@ -19,7 +20,6 @@
#include <neon/workloads/NeonWorkloadUtils.hpp>
#include <neon/workloads/NeonWorkloads.hpp>
-#include <boost/core/ignore_unused.hpp>
#include <boost/polymorphic_cast.hpp>
namespace armnn
@@ -98,7 +98,7 @@ std::unique_ptr<ITensorHandle> NeonWorkloadFactory::CreateTensorHandle(const Ten
std::unique_ptr<IWorkload> NeonWorkloadFactory::CreateAbs(const AbsQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- boost::ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
ElementwiseUnaryQueueDescriptor elementwiseUnaryDescriptor;
elementwiseUnaryDescriptor.m_Parameters = ElementwiseUnaryDescriptor(UnaryOperation::Abs);
@@ -245,7 +245,7 @@ std::unique_ptr<IWorkload> NeonWorkloadFactory::CreateElementwiseUnary(const Ele
std::unique_ptr<IWorkload> NeonWorkloadFactory::CreateEqual(const EqualQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- boost::ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
ComparisonQueueDescriptor comparisonDescriptor;
comparisonDescriptor.m_Parameters = ComparisonDescriptor(ComparisonOperation::Equal);
@@ -275,7 +275,7 @@ std::unique_ptr<IWorkload> NeonWorkloadFactory::CreateGather(const armnn::Gather
std::unique_ptr<IWorkload> NeonWorkloadFactory::CreateGreater(const GreaterQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- boost::ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
ComparisonQueueDescriptor comparisonDescriptor;
comparisonDescriptor.m_Parameters = ComparisonDescriptor(ComparisonOperation::Greater);
@@ -446,7 +446,7 @@ std::unique_ptr<armnn::IWorkload> NeonWorkloadFactory::CreateResizeBilinear(
std::unique_ptr<IWorkload> NeonWorkloadFactory::CreateRsqrt(const RsqrtQueueDescriptor &descriptor,
const WorkloadInfo &info) const
{
- boost::ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
ElementwiseUnaryQueueDescriptor elementwiseUnaryDescriptor;
elementwiseUnaryDescriptor.m_Parameters = ElementwiseUnaryDescriptor(UnaryOperation::Rsqrt);
diff --git a/src/backends/neon/NeonWorkloadFactory.hpp b/src/backends/neon/NeonWorkloadFactory.hpp
index bc4107dbb0..f122792203 100644
--- a/src/backends/neon/NeonWorkloadFactory.hpp
+++ b/src/backends/neon/NeonWorkloadFactory.hpp
@@ -8,8 +8,7 @@
#include <backendsCommon/WorkloadFactoryBase.hpp>
#include <aclCommon/BaseMemoryManager.hpp>
-
-#include <boost/core/ignore_unused.hpp>
+#include <armnn/utility/IgnoreUnused.hpp>
namespace armnn
{
diff --git a/src/backends/neon/test/NeonLayerTests.cpp b/src/backends/neon/test/NeonLayerTests.cpp
index 4dc9641a74..d3ca675679 100644
--- a/src/backends/neon/test/NeonLayerTests.cpp
+++ b/src/backends/neon/test/NeonLayerTests.cpp
@@ -219,7 +219,7 @@ DepthwiseConvolution2dDescriptor MakeDepthwiseConv2dDesc(uint32_t strideX, uint3
uint32_t depthMultiplier = 1, uint32_t padLeft = 0, uint32_t padRight = 0,
uint32_t padTop = 0, uint32_t padBottom = 0)
{
- boost::ignore_unused(depthMultiplier);
+ IgnoreUnused(depthMultiplier);
DepthwiseConvolution2dDescriptor desc;
diff --git a/src/backends/reference/RefLayerSupport.cpp b/src/backends/reference/RefLayerSupport.cpp
index 7d5c3b509e..bd2e7289d8 100644
--- a/src/backends/reference/RefLayerSupport.cpp
+++ b/src/backends/reference/RefLayerSupport.cpp
@@ -8,13 +8,12 @@
#include <armnn/TypesUtils.hpp>
#include <armnn/Types.hpp>
#include <armnn/Descriptors.hpp>
+#include <armnn/utility/IgnoreUnused.hpp>
#include <LayerSupportCommon.hpp>
-
#include <backendsCommon/LayerSupportRules.hpp>
#include <boost/cast.hpp>
-#include <boost/core/ignore_unused.hpp>
#include <vector>
#include <array>
@@ -178,7 +177,7 @@ bool RefLayerSupport::IsArgMinMaxSupported(const armnn::TensorInfo &input, const
const armnn::ArgMinMaxDescriptor &descriptor,
armnn::Optional<std::string &> reasonIfUnsupported) const
{
- ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
std::array<DataType, 4> supportedTypes =
{
@@ -207,7 +206,7 @@ bool RefLayerSupport::IsBatchNormalizationSupported(const TensorInfo& input,
const BatchNormalizationDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
{
- ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
std::array<DataType, 4> supportedTypes =
{
@@ -248,7 +247,7 @@ bool RefLayerSupport::IsBatchToSpaceNdSupported(const TensorInfo& input,
const BatchToSpaceNdDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
{
- ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
bool supported = true;
@@ -297,7 +296,7 @@ bool RefLayerSupport::IsComparisonSupported(const TensorInfo& input0,
const ComparisonDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
{
- boost::ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
std::array<DataType, 4> supportedInputTypes =
{
@@ -325,7 +324,7 @@ bool RefLayerSupport::IsConcatSupported(const std::vector<const TensorInfo*> inp
const ConcatDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
{
- ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
bool supported = true;
std::array<DataType,5> supportedTypes =
@@ -475,7 +474,7 @@ bool RefLayerSupport::IsConvolution2dSupported(const TensorInfo& input,
supported &= CheckSupportRule(TypeAnyOf(biases.value(), biasesSupportedTypes), reasonIfUnsupported,
"Reference Convolution2d: biases is not a supported type.");
}
- ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
return supported;
}
@@ -514,7 +513,7 @@ bool RefLayerSupport::IsDepthToSpaceSupported(const TensorInfo& input,
const DepthToSpaceDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
{
- ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
bool supported = true;
std::array<DataType,4> supportedTypes =
@@ -602,7 +601,7 @@ bool RefLayerSupport::IsDepthwiseConvolutionSupported(const TensorInfo& input,
supported &= CheckSupportRule(TypeAnyOf(biases.value(), biasesSupportedTypes), reasonIfUnsupported,
"Reference DepthwiseConvolution2d: biases is not a supported type.");
}
- ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
return supported;
@@ -655,7 +654,7 @@ bool RefLayerSupport::IsDetectionPostProcessSupported(const TensorInfo& boxEncod
const DetectionPostProcessDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
{
- boost::ignore_unused(anchors, detectionBoxes, detectionClasses, detectionScores, numDetections, descriptor);
+ IgnoreUnused(anchors, detectionBoxes, detectionClasses, detectionScores, numDetections, descriptor);
bool supported = true;
@@ -725,7 +724,7 @@ bool RefLayerSupport::IsElementwiseUnarySupported(const TensorInfo& input,
const ElementwiseUnaryDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
{
- boost::ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
std::array<DataType, 4> supportedTypes =
{
@@ -769,7 +768,7 @@ bool RefLayerSupport::IsFakeQuantizationSupported(const TensorInfo& input,
const FakeQuantizationDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
{
- ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
bool supported = true;
std::array<DataType,1> supportedTypes =
@@ -787,7 +786,7 @@ bool RefLayerSupport::IsFloorSupported(const TensorInfo& input,
const TensorInfo& output,
Optional<std::string&> reasonIfUnsupported) const
{
- ignore_unused(output);
+ IgnoreUnused(output);
bool supported = true;
std::array<DataType,3> supportedTypes =
@@ -916,7 +915,7 @@ bool RefLayerSupport::IsInstanceNormalizationSupported(const TensorInfo& input,
const InstanceNormalizationDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
{
- ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
// Define supported types
std::array<DataType, 4> supportedTypes =
{
@@ -947,7 +946,7 @@ bool RefLayerSupport::IsL2NormalizationSupported(const TensorInfo& input,
const L2NormalizationDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
{
- ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
// Define supported types
std::array<DataType, 4> supportedTypes =
{
@@ -980,7 +979,7 @@ bool RefLayerSupport::IsLogSoftmaxSupported(const TensorInfo& input,
const LogSoftmaxDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
{
- ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
std::array<DataType, 2> supportedTypes =
{
@@ -1012,8 +1011,8 @@ bool RefLayerSupport::IsLstmSupported(const TensorInfo& input,
const LstmInputParamsInfo& paramsInfo,
Optional<std::string&> reasonIfUnsupported) const
{
- ignore_unused(descriptor);
- ignore_unused(paramsInfo);
+ IgnoreUnused(descriptor);
+ IgnoreUnused(paramsInfo);
bool supported = true;
@@ -1319,7 +1318,7 @@ bool RefLayerSupport::IsNormalizationSupported(const TensorInfo& input,
const NormalizationDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
{
- ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
// Define supported types
std::array<DataType, 4> supportedTypes =
@@ -1356,7 +1355,7 @@ bool RefLayerSupport::IsPadSupported(const TensorInfo& input,
const PadDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
{
- ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
bool supported = true;
// Define supported output and inputs types.
@@ -1385,7 +1384,7 @@ bool RefLayerSupport::IsPermuteSupported(const TensorInfo& input,
const PermuteDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
{
- ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
bool supported = true;
// Define supported output and inputs types.
@@ -1414,7 +1413,7 @@ bool RefLayerSupport::IsPooling2dSupported(const TensorInfo& input,
const Pooling2dDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
{
- ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
bool supported = true;
// Define supported output and inputs types.
@@ -1479,8 +1478,8 @@ bool RefLayerSupport::IsReshapeSupported(const TensorInfo& input,
const ReshapeDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
{
- ignore_unused(output);
- ignore_unused(descriptor);
+ IgnoreUnused(output);
+ IgnoreUnused(descriptor);
// Define supported output types.
std::array<DataType,7> supportedOutputTypes =
{
@@ -1526,7 +1525,7 @@ bool RefLayerSupport::IsResizeSupported(const TensorInfo& input,
const ResizeDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
{
- boost::ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
bool supported = true;
std::array<DataType,5> supportedTypes =
{
@@ -1564,7 +1563,7 @@ bool RefLayerSupport::IsSliceSupported(const TensorInfo& input,
const SliceDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
{
- boost::ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
bool supported = true;
std::array<DataType, 3> supportedTypes =
@@ -1591,7 +1590,7 @@ bool RefLayerSupport::IsSoftmaxSupported(const TensorInfo& input,
const SoftmaxDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
{
- boost::ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
bool supported = true;
std::array<DataType,6> supportedTypes =
{
@@ -1620,7 +1619,7 @@ bool RefLayerSupport::IsSpaceToBatchNdSupported(const TensorInfo& input,
const SpaceToBatchNdDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
{
- boost::ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
bool supported = true;
std::array<DataType,4> supportedTypes =
{
@@ -1648,7 +1647,7 @@ bool RefLayerSupport::IsSpaceToDepthSupported(const TensorInfo& input,
Optional<std::string&> reasonIfUnsupported) const
{
- ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
bool supported = true;
std::array<DataType,4> supportedTypes =
@@ -1675,7 +1674,7 @@ bool RefLayerSupport::IsSplitterSupported(const TensorInfo& input,
const ViewsDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
{
- ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
bool supported = true;
std::array<DataType,4> supportedTypes =
{
@@ -1696,7 +1695,7 @@ bool RefLayerSupport::IsSplitterSupported(const TensorInfo& input,
const ViewsDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
{
- ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
bool supported = true;
std::array<DataType,4> supportedTypes =
{
@@ -1725,7 +1724,7 @@ bool RefLayerSupport::IsStackSupported(const std::vector<const TensorInfo*>& inp
const StackDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
{
- ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
bool supported = true;
std::array<DataType,4> supportedTypes =
@@ -1756,7 +1755,7 @@ bool RefLayerSupport::IsStridedSliceSupported(const TensorInfo& input,
const StridedSliceDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
{
- ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
bool supported = true;
std::array<DataType,3> supportedTypes =
@@ -1853,7 +1852,7 @@ bool RefLayerSupport::IsTransposeConvolution2dSupported(const TensorInfo& input,
const Optional<TensorInfo>& biases,
Optional<std::string&> reasonIfUnsupported) const
{
- boost::ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
bool supported = true;
std::array<DataType,4> supportedTypes =
@@ -1919,7 +1918,7 @@ bool RefLayerSupport::IsTransposeSupported(const TensorInfo& input,
const TransposeDescriptor& descriptor,
Optional<std::string&> reasonIfUnsupported) const
{
- ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
bool supported = true;
// Define supported output and inputs types.
diff --git a/src/backends/reference/RefTensorHandleFactory.cpp b/src/backends/reference/RefTensorHandleFactory.cpp
index c97a779cb3..d687c78b17 100644
--- a/src/backends/reference/RefTensorHandleFactory.cpp
+++ b/src/backends/reference/RefTensorHandleFactory.cpp
@@ -6,7 +6,7 @@
#include "RefTensorHandleFactory.hpp"
#include "RefTensorHandle.hpp"
-#include <boost/core/ignore_unused.hpp>
+#include <armnn/utility/IgnoreUnused.hpp>
namespace armnn
{
@@ -23,7 +23,7 @@ std::unique_ptr<ITensorHandle> RefTensorHandleFactory::CreateSubTensorHandle(ITe
TensorShape const& subTensorShape,
unsigned int const* subTensorOrigin) const
{
- boost::ignore_unused(parent, subTensorShape, subTensorOrigin);
+ IgnoreUnused(parent, subTensorShape, subTensorOrigin);
return nullptr;
}
@@ -35,7 +35,7 @@ std::unique_ptr<ITensorHandle> RefTensorHandleFactory::CreateTensorHandle(const
std::unique_ptr<ITensorHandle> RefTensorHandleFactory::CreateTensorHandle(const TensorInfo& tensorInfo,
DataLayout dataLayout) const
{
- boost::ignore_unused(dataLayout);
+ IgnoreUnused(dataLayout);
return std::make_unique<RefTensorHandle>(tensorInfo, m_MemoryManager, m_ImportFlags);
}
diff --git a/src/backends/reference/RefWorkloadFactory.cpp b/src/backends/reference/RefWorkloadFactory.cpp
index 2a415bfbf0..52d71df936 100644
--- a/src/backends/reference/RefWorkloadFactory.cpp
+++ b/src/backends/reference/RefWorkloadFactory.cpp
@@ -102,7 +102,7 @@ std::unique_ptr<ITensorHandle> RefWorkloadFactory::CreateTensorHandle(const Tens
{
// For Ref it is okay to make the TensorHandle memory managed as it can also store a pointer
// to unmanaged memory. This also ensures memory alignment.
- boost::ignore_unused(isMemoryManaged);
+ IgnoreUnused(isMemoryManaged);
return std::make_unique<RefTensorHandle>(tensorInfo, m_MemoryManager);
}
@@ -112,14 +112,14 @@ std::unique_ptr<ITensorHandle> RefWorkloadFactory::CreateTensorHandle(const Tens
{
// For Ref it is okay to make the TensorHandle memory managed as it can also store a pointer
// to unmanaged memory. This also ensures memory alignment.
- boost::ignore_unused(isMemoryManaged, dataLayout);
+ IgnoreUnused(isMemoryManaged, dataLayout);
return std::make_unique<RefTensorHandle>(tensorInfo, m_MemoryManager);
}
std::unique_ptr<IWorkload> RefWorkloadFactory::CreateAbs(const AbsQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- boost::ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
ElementwiseUnaryQueueDescriptor elementwiseUnaryDescriptor;
elementwiseUnaryDescriptor.m_Parameters.m_Operation = UnaryOperation::Abs;
@@ -267,7 +267,7 @@ std::unique_ptr<IWorkload> RefWorkloadFactory::CreateElementwiseUnary(const Elem
std::unique_ptr<IWorkload> RefWorkloadFactory::CreateEqual(const EqualQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- boost::ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
ComparisonQueueDescriptor comparisonDescriptor;
comparisonDescriptor.m_Parameters.m_Operation = ComparisonOperation::Equal;
@@ -303,7 +303,7 @@ std::unique_ptr<IWorkload> RefWorkloadFactory::CreateGather(const GatherQueueDes
std::unique_ptr<IWorkload> RefWorkloadFactory::CreateGreater(const GreaterQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- boost::ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
ComparisonQueueDescriptor comparisonDescriptor;
comparisonDescriptor.m_Parameters.m_Operation = ComparisonOperation::Greater;
@@ -506,7 +506,7 @@ std::unique_ptr<IWorkload> RefWorkloadFactory::CreateResizeBilinear(const Resize
std::unique_ptr<IWorkload> RefWorkloadFactory::CreateRsqrt(const RsqrtQueueDescriptor& descriptor,
const WorkloadInfo& info) const
{
- boost::ignore_unused(descriptor);
+ IgnoreUnused(descriptor);
ElementwiseUnaryQueueDescriptor elementwiseUnaryDescriptor;
elementwiseUnaryDescriptor.m_Parameters.m_Operation = UnaryOperation::Rsqrt;
diff --git a/src/backends/reference/RefWorkloadFactory.hpp b/src/backends/reference/RefWorkloadFactory.hpp
index 030ce6f03d..b64479e207 100644
--- a/src/backends/reference/RefWorkloadFactory.hpp
+++ b/src/backends/reference/RefWorkloadFactory.hpp
@@ -4,12 +4,11 @@
//
#pragma once
-#include <armnn/Optional.hpp>
-#include <backendsCommon/WorkloadFactory.hpp>
-
#include "RefMemoryManager.hpp"
-#include <boost/core/ignore_unused.hpp>
+#include <armnn/Optional.hpp>
+#include <backendsCommon/WorkloadFactory.hpp>
+#include <armnn/utility/IgnoreUnused.hpp>
namespace armnn
@@ -48,7 +47,7 @@ public:
TensorShape const& subTensorShape,
unsigned int const* subTensorOrigin) const override
{
- boost::ignore_unused(parent, subTensorShape, subTensorOrigin);
+ IgnoreUnused(parent, subTensorShape, subTensorOrigin);
return nullptr;
}
diff --git a/src/backends/reference/test/RefWorkloadFactoryHelper.hpp b/src/backends/reference/test/RefWorkloadFactoryHelper.hpp
index 10e5b9fa28..30d2037a32 100644
--- a/src/backends/reference/test/RefWorkloadFactoryHelper.hpp
+++ b/src/backends/reference/test/RefWorkloadFactoryHelper.hpp
@@ -25,7 +25,7 @@ struct WorkloadFactoryHelper<armnn::RefWorkloadFactory>
static armnn::RefWorkloadFactory GetFactory(
const armnn::IBackendInternal::IMemoryManagerSharedPtr& memoryManager = nullptr)
{
- boost::ignore_unused(memoryManager);
+ IgnoreUnused(memoryManager);
return armnn::RefWorkloadFactory();
}
};
diff --git a/src/backends/reference/workloads/ArgMinMax.cpp b/src/backends/reference/workloads/ArgMinMax.cpp
index db85b958e9..637aa17013 100644
--- a/src/backends/reference/workloads/ArgMinMax.cpp
+++ b/src/backends/reference/workloads/ArgMinMax.cpp
@@ -15,7 +15,7 @@ namespace armnn
void ArgMinMax(Decoder<float>& in, int32_t* out, const TensorInfo& inputTensorInfo,
const TensorInfo& outputTensorInfo, ArgMinMaxFunction function, int axis)
{
- boost::ignore_unused(outputTensorInfo);
+ IgnoreUnused(outputTensorInfo);
unsigned int uAxis = armnnUtils::GetUnsignedAxis(inputTensorInfo.GetNumDimensions(), axis);
diff --git a/src/backends/reference/workloads/BaseIterator.hpp b/src/backends/reference/workloads/BaseIterator.hpp
index 3f0144670f..5cae5bda83 100644
--- a/src/backends/reference/workloads/BaseIterator.hpp
+++ b/src/backends/reference/workloads/BaseIterator.hpp
@@ -5,14 +5,13 @@
#pragma once
-
+#include <armnn/utility/IgnoreUnused.hpp>
#include <armnn/TypesUtils.hpp>
#include <armnnUtils/FloatingPointConverter.hpp>
#include <ResolveType.hpp>
#include <boost/assert.hpp>
-#include <boost/core/ignore_unused.hpp>
namespace armnn
{
@@ -107,7 +106,7 @@ public:
TypedIterator& SetIndex(unsigned int index, unsigned int axisIndex = 0) override
{
- boost::ignore_unused(axisIndex);
+ IgnoreUnused(axisIndex);
BOOST_ASSERT(m_Iterator);
m_Iterator = m_Start + index;
return *this;
diff --git a/src/backends/reference/workloads/Dequantize.cpp b/src/backends/reference/workloads/Dequantize.cpp
index 4025e8d7fa..63c0405efe 100644
--- a/src/backends/reference/workloads/Dequantize.cpp
+++ b/src/backends/reference/workloads/Dequantize.cpp
@@ -5,7 +5,8 @@
#include "Dequantize.hpp"
-#include <boost/core/ignore_unused.hpp>
+#include <armnn/utility/IgnoreUnused.hpp>
+
namespace armnn
{
@@ -14,7 +15,7 @@ void Dequantize(Decoder<float>& inputDecoder,
const TensorInfo& inputInfo,
const TensorInfo& outputInfo)
{
- boost::ignore_unused(outputInfo);
+ IgnoreUnused(outputInfo);
BOOST_ASSERT(inputInfo.GetNumElements() == outputInfo.GetNumElements());
for (unsigned int i = 0; i < inputInfo.GetNumElements(); i++)
{
diff --git a/src/backends/reference/workloads/DetectionPostProcess.cpp b/src/backends/reference/workloads/DetectionPostProcess.cpp
index 96e57803a1..57cf01e4a1 100644
--- a/src/backends/reference/workloads/DetectionPostProcess.cpp
+++ b/src/backends/reference/workloads/DetectionPostProcess.cpp
@@ -154,7 +154,7 @@ void DetectionPostProcess(const TensorInfo& boxEncodingsInfo,
float* detectionScores,
float* numDetections)
{
- boost::ignore_unused(anchorsInfo, detectionClassesInfo, detectionScoresInfo, numDetectionsInfo);
+ IgnoreUnused(anchorsInfo, detectionClassesInfo, detectionScoresInfo, numDetectionsInfo);
// Transform center-size format which is (ycenter, xcenter, height, width) to box-corner format,
// which represents the lower left corner and the upper right corner (ymin, xmin, ymax, xmax)
diff --git a/src/backends/reference/workloads/Gather.cpp b/src/backends/reference/workloads/Gather.cpp
index 5416855f48..4cf3a142a0 100644
--- a/src/backends/reference/workloads/Gather.cpp
+++ b/src/backends/reference/workloads/Gather.cpp
@@ -8,8 +8,8 @@
#include "RefWorkloadUtils.hpp"
#include <backendsCommon/WorkloadData.hpp>
+#include <armnn/utility/IgnoreUnused.hpp>
-#include <boost/core/ignore_unused.hpp>
#include <boost/numeric/conversion/cast.hpp>
namespace armnn
@@ -22,7 +22,7 @@ void Gather(const TensorInfo& paramsInfo,
const int32_t* indices,
Encoder<float>& output)
{
- boost::ignore_unused(outputInfo);
+ IgnoreUnused(outputInfo);
const TensorShape& paramsShape = paramsInfo.GetShape();
unsigned int paramsProduct = 1;
diff --git a/src/backends/reference/workloads/LogSoftmax.cpp b/src/backends/reference/workloads/LogSoftmax.cpp
index ddf5674fb8..103d62a8df 100644
--- a/src/backends/reference/workloads/LogSoftmax.cpp
+++ b/src/backends/reference/workloads/LogSoftmax.cpp
@@ -6,11 +6,11 @@
#include "LogSoftmax.hpp"
#include <armnnUtils/TensorUtils.hpp>
+#include <armnn/utility/IgnoreUnused.hpp>
#include <cmath>
#include <boost/assert.hpp>
-#include <boost/core/ignore_unused.hpp>
#include <boost/numeric/conversion/cast.hpp>
namespace
@@ -37,7 +37,7 @@ void LogSoftmax(Decoder<float>& input,
bool axisIsValid = ValidateAxis(descriptor.m_Axis, numDimensions);
BOOST_ASSERT_MSG(axisIsValid,
"Axis index is not in range [-numDimensions, numDimensions).");
- boost::ignore_unused(axisIsValid);
+ IgnoreUnused(axisIsValid);
unsigned int uAxis = descriptor.m_Axis < 0 ?
numDimensions - boost::numeric_cast<unsigned int>(std::abs(descriptor.m_Axis)) :
diff --git a/src/backends/reference/workloads/RefStridedSliceWorkload.cpp b/src/backends/reference/workloads/RefStridedSliceWorkload.cpp
index 8bb1670a48..bfd3c284ae 100644
--- a/src/backends/reference/workloads/RefStridedSliceWorkload.cpp
+++ b/src/backends/reference/workloads/RefStridedSliceWorkload.cpp
@@ -28,7 +28,7 @@ void RefStridedSliceWorkload::Execute() const
DataType outputDataType = outputInfo.GetDataType();
BOOST_ASSERT(inputDataType == outputDataType);
- boost::ignore_unused(outputDataType);
+ IgnoreUnused(outputDataType);
StridedSlice(inputInfo,
m_Data.m_Parameters,
diff --git a/src/backends/reference/workloads/Slice.cpp b/src/backends/reference/workloads/Slice.cpp
index c7ca3b156e..0223cdc56a 100644
--- a/src/backends/reference/workloads/Slice.cpp
+++ b/src/backends/reference/workloads/Slice.cpp
@@ -5,8 +5,9 @@
#include "Slice.hpp"
+#include <armnn/utility/IgnoreUnused.hpp>
+
#include <boost/assert.hpp>
-#include <boost/core/ignore_unused.hpp>
#include <boost/numeric/conversion/cast.hpp>
namespace armnn
@@ -72,7 +73,7 @@ void Slice(const TensorInfo& inputInfo,
const unsigned char* input = reinterpret_cast<const unsigned char*>(inputData);
unsigned char* output = reinterpret_cast<unsigned char*>(outputData);
- boost::ignore_unused(dim0);
+ IgnoreUnused(dim0);
for (unsigned int idx0 = begin0; idx0 < begin0 + size0; ++idx0)
{
for (unsigned int idx1 = begin1; idx1 < begin1 + size1; ++idx1)