diff options
author | Jan Eilers <jan.eilers@arm.com> | 2020-03-09 12:13:48 +0000 |
---|---|---|
committer | Jan Eilers <jan.eilers@arm.com> | 2020-03-10 10:07:43 +0000 |
commit | 8eb256065f0e75ecf8e427d56955e2bac117c2d7 (patch) | |
tree | 1387fb4ea4a741475449d78be63d601f9d84b6e5 /src/backends/reference | |
parent | 8832522f47b701f5f042069e7bf8deae9b75d449 (diff) | |
download | armnn-8eb256065f0e75ecf8e427d56955e2bac117c2d7.tar.gz |
IVGCVSW-4482 Remove boost::ignore_unused
!referencetests:229377
Signed-off-by: Jan Eilers <jan.eilers@arm.com>
Change-Id: Ia9b360b4a057fe7bbce5b268092627c09a0dba82
Diffstat (limited to 'src/backends/reference')
-rw-r--r-- | src/backends/reference/RefLayerSupport.cpp | 71 | ||||
-rw-r--r-- | src/backends/reference/RefTensorHandleFactory.cpp | 6 | ||||
-rw-r--r-- | src/backends/reference/RefWorkloadFactory.cpp | 12 | ||||
-rw-r--r-- | src/backends/reference/RefWorkloadFactory.hpp | 9 | ||||
-rw-r--r-- | src/backends/reference/test/RefWorkloadFactoryHelper.hpp | 2 | ||||
-rw-r--r-- | src/backends/reference/workloads/ArgMinMax.cpp | 2 | ||||
-rw-r--r-- | src/backends/reference/workloads/BaseIterator.hpp | 5 | ||||
-rw-r--r-- | src/backends/reference/workloads/Dequantize.cpp | 5 | ||||
-rw-r--r-- | src/backends/reference/workloads/DetectionPostProcess.cpp | 2 | ||||
-rw-r--r-- | src/backends/reference/workloads/Gather.cpp | 4 | ||||
-rw-r--r-- | src/backends/reference/workloads/LogSoftmax.cpp | 4 | ||||
-rw-r--r-- | src/backends/reference/workloads/RefStridedSliceWorkload.cpp | 2 | ||||
-rw-r--r-- | src/backends/reference/workloads/Slice.cpp | 5 |
13 files changed, 64 insertions, 65 deletions
diff --git a/src/backends/reference/RefLayerSupport.cpp b/src/backends/reference/RefLayerSupport.cpp index 7d5c3b509e..bd2e7289d8 100644 --- a/src/backends/reference/RefLayerSupport.cpp +++ b/src/backends/reference/RefLayerSupport.cpp @@ -8,13 +8,12 @@ #include <armnn/TypesUtils.hpp> #include <armnn/Types.hpp> #include <armnn/Descriptors.hpp> +#include <armnn/utility/IgnoreUnused.hpp> #include <LayerSupportCommon.hpp> - #include <backendsCommon/LayerSupportRules.hpp> #include <boost/cast.hpp> -#include <boost/core/ignore_unused.hpp> #include <vector> #include <array> @@ -178,7 +177,7 @@ bool RefLayerSupport::IsArgMinMaxSupported(const armnn::TensorInfo &input, const const armnn::ArgMinMaxDescriptor &descriptor, armnn::Optional<std::string &> reasonIfUnsupported) const { - ignore_unused(descriptor); + IgnoreUnused(descriptor); std::array<DataType, 4> supportedTypes = { @@ -207,7 +206,7 @@ bool RefLayerSupport::IsBatchNormalizationSupported(const TensorInfo& input, const BatchNormalizationDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported) const { - ignore_unused(descriptor); + IgnoreUnused(descriptor); std::array<DataType, 4> supportedTypes = { @@ -248,7 +247,7 @@ bool RefLayerSupport::IsBatchToSpaceNdSupported(const TensorInfo& input, const BatchToSpaceNdDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported) const { - ignore_unused(descriptor); + IgnoreUnused(descriptor); bool supported = true; @@ -297,7 +296,7 @@ bool RefLayerSupport::IsComparisonSupported(const TensorInfo& input0, const ComparisonDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported) const { - boost::ignore_unused(descriptor); + IgnoreUnused(descriptor); std::array<DataType, 4> supportedInputTypes = { @@ -325,7 +324,7 @@ bool RefLayerSupport::IsConcatSupported(const std::vector<const TensorInfo*> inp const ConcatDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported) const { - ignore_unused(descriptor); + IgnoreUnused(descriptor); bool supported = true; std::array<DataType,5> supportedTypes = @@ -475,7 +474,7 @@ bool RefLayerSupport::IsConvolution2dSupported(const TensorInfo& input, supported &= CheckSupportRule(TypeAnyOf(biases.value(), biasesSupportedTypes), reasonIfUnsupported, "Reference Convolution2d: biases is not a supported type."); } - ignore_unused(descriptor); + IgnoreUnused(descriptor); return supported; } @@ -514,7 +513,7 @@ bool RefLayerSupport::IsDepthToSpaceSupported(const TensorInfo& input, const DepthToSpaceDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported) const { - ignore_unused(descriptor); + IgnoreUnused(descriptor); bool supported = true; std::array<DataType,4> supportedTypes = @@ -602,7 +601,7 @@ bool RefLayerSupport::IsDepthwiseConvolutionSupported(const TensorInfo& input, supported &= CheckSupportRule(TypeAnyOf(biases.value(), biasesSupportedTypes), reasonIfUnsupported, "Reference DepthwiseConvolution2d: biases is not a supported type."); } - ignore_unused(descriptor); + IgnoreUnused(descriptor); return supported; @@ -655,7 +654,7 @@ bool RefLayerSupport::IsDetectionPostProcessSupported(const TensorInfo& boxEncod const DetectionPostProcessDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported) const { - boost::ignore_unused(anchors, detectionBoxes, detectionClasses, detectionScores, numDetections, descriptor); + IgnoreUnused(anchors, detectionBoxes, detectionClasses, detectionScores, numDetections, descriptor); bool supported = true; @@ -725,7 +724,7 @@ bool RefLayerSupport::IsElementwiseUnarySupported(const TensorInfo& input, const ElementwiseUnaryDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported) const { - boost::ignore_unused(descriptor); + IgnoreUnused(descriptor); std::array<DataType, 4> supportedTypes = { @@ -769,7 +768,7 @@ bool RefLayerSupport::IsFakeQuantizationSupported(const TensorInfo& input, const FakeQuantizationDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported) const { - ignore_unused(descriptor); + IgnoreUnused(descriptor); bool supported = true; std::array<DataType,1> supportedTypes = @@ -787,7 +786,7 @@ bool RefLayerSupport::IsFloorSupported(const TensorInfo& input, const TensorInfo& output, Optional<std::string&> reasonIfUnsupported) const { - ignore_unused(output); + IgnoreUnused(output); bool supported = true; std::array<DataType,3> supportedTypes = @@ -916,7 +915,7 @@ bool RefLayerSupport::IsInstanceNormalizationSupported(const TensorInfo& input, const InstanceNormalizationDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported) const { - ignore_unused(descriptor); + IgnoreUnused(descriptor); // Define supported types std::array<DataType, 4> supportedTypes = { @@ -947,7 +946,7 @@ bool RefLayerSupport::IsL2NormalizationSupported(const TensorInfo& input, const L2NormalizationDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported) const { - ignore_unused(descriptor); + IgnoreUnused(descriptor); // Define supported types std::array<DataType, 4> supportedTypes = { @@ -980,7 +979,7 @@ bool RefLayerSupport::IsLogSoftmaxSupported(const TensorInfo& input, const LogSoftmaxDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported) const { - ignore_unused(descriptor); + IgnoreUnused(descriptor); std::array<DataType, 2> supportedTypes = { @@ -1012,8 +1011,8 @@ bool RefLayerSupport::IsLstmSupported(const TensorInfo& input, const LstmInputParamsInfo& paramsInfo, Optional<std::string&> reasonIfUnsupported) const { - ignore_unused(descriptor); - ignore_unused(paramsInfo); + IgnoreUnused(descriptor); + IgnoreUnused(paramsInfo); bool supported = true; @@ -1319,7 +1318,7 @@ bool RefLayerSupport::IsNormalizationSupported(const TensorInfo& input, const NormalizationDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported) const { - ignore_unused(descriptor); + IgnoreUnused(descriptor); // Define supported types std::array<DataType, 4> supportedTypes = @@ -1356,7 +1355,7 @@ bool RefLayerSupport::IsPadSupported(const TensorInfo& input, const PadDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported) const { - ignore_unused(descriptor); + IgnoreUnused(descriptor); bool supported = true; // Define supported output and inputs types. @@ -1385,7 +1384,7 @@ bool RefLayerSupport::IsPermuteSupported(const TensorInfo& input, const PermuteDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported) const { - ignore_unused(descriptor); + IgnoreUnused(descriptor); bool supported = true; // Define supported output and inputs types. @@ -1414,7 +1413,7 @@ bool RefLayerSupport::IsPooling2dSupported(const TensorInfo& input, const Pooling2dDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported) const { - ignore_unused(descriptor); + IgnoreUnused(descriptor); bool supported = true; // Define supported output and inputs types. @@ -1479,8 +1478,8 @@ bool RefLayerSupport::IsReshapeSupported(const TensorInfo& input, const ReshapeDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported) const { - ignore_unused(output); - ignore_unused(descriptor); + IgnoreUnused(output); + IgnoreUnused(descriptor); // Define supported output types. std::array<DataType,7> supportedOutputTypes = { @@ -1526,7 +1525,7 @@ bool RefLayerSupport::IsResizeSupported(const TensorInfo& input, const ResizeDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported) const { - boost::ignore_unused(descriptor); + IgnoreUnused(descriptor); bool supported = true; std::array<DataType,5> supportedTypes = { @@ -1564,7 +1563,7 @@ bool RefLayerSupport::IsSliceSupported(const TensorInfo& input, const SliceDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported) const { - boost::ignore_unused(descriptor); + IgnoreUnused(descriptor); bool supported = true; std::array<DataType, 3> supportedTypes = @@ -1591,7 +1590,7 @@ bool RefLayerSupport::IsSoftmaxSupported(const TensorInfo& input, const SoftmaxDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported) const { - boost::ignore_unused(descriptor); + IgnoreUnused(descriptor); bool supported = true; std::array<DataType,6> supportedTypes = { @@ -1620,7 +1619,7 @@ bool RefLayerSupport::IsSpaceToBatchNdSupported(const TensorInfo& input, const SpaceToBatchNdDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported) const { - boost::ignore_unused(descriptor); + IgnoreUnused(descriptor); bool supported = true; std::array<DataType,4> supportedTypes = { @@ -1648,7 +1647,7 @@ bool RefLayerSupport::IsSpaceToDepthSupported(const TensorInfo& input, Optional<std::string&> reasonIfUnsupported) const { - ignore_unused(descriptor); + IgnoreUnused(descriptor); bool supported = true; std::array<DataType,4> supportedTypes = @@ -1675,7 +1674,7 @@ bool RefLayerSupport::IsSplitterSupported(const TensorInfo& input, const ViewsDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported) const { - ignore_unused(descriptor); + IgnoreUnused(descriptor); bool supported = true; std::array<DataType,4> supportedTypes = { @@ -1696,7 +1695,7 @@ bool RefLayerSupport::IsSplitterSupported(const TensorInfo& input, const ViewsDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported) const { - ignore_unused(descriptor); + IgnoreUnused(descriptor); bool supported = true; std::array<DataType,4> supportedTypes = { @@ -1725,7 +1724,7 @@ bool RefLayerSupport::IsStackSupported(const std::vector<const TensorInfo*>& inp const StackDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported) const { - ignore_unused(descriptor); + IgnoreUnused(descriptor); bool supported = true; std::array<DataType,4> supportedTypes = @@ -1756,7 +1755,7 @@ bool RefLayerSupport::IsStridedSliceSupported(const TensorInfo& input, const StridedSliceDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported) const { - ignore_unused(descriptor); + IgnoreUnused(descriptor); bool supported = true; std::array<DataType,3> supportedTypes = @@ -1853,7 +1852,7 @@ bool RefLayerSupport::IsTransposeConvolution2dSupported(const TensorInfo& input, const Optional<TensorInfo>& biases, Optional<std::string&> reasonIfUnsupported) const { - boost::ignore_unused(descriptor); + IgnoreUnused(descriptor); bool supported = true; std::array<DataType,4> supportedTypes = @@ -1919,7 +1918,7 @@ bool RefLayerSupport::IsTransposeSupported(const TensorInfo& input, const TransposeDescriptor& descriptor, Optional<std::string&> reasonIfUnsupported) const { - ignore_unused(descriptor); + IgnoreUnused(descriptor); bool supported = true; // Define supported output and inputs types. diff --git a/src/backends/reference/RefTensorHandleFactory.cpp b/src/backends/reference/RefTensorHandleFactory.cpp index c97a779cb3..d687c78b17 100644 --- a/src/backends/reference/RefTensorHandleFactory.cpp +++ b/src/backends/reference/RefTensorHandleFactory.cpp @@ -6,7 +6,7 @@ #include "RefTensorHandleFactory.hpp" #include "RefTensorHandle.hpp" -#include <boost/core/ignore_unused.hpp> +#include <armnn/utility/IgnoreUnused.hpp> namespace armnn { @@ -23,7 +23,7 @@ std::unique_ptr<ITensorHandle> RefTensorHandleFactory::CreateSubTensorHandle(ITe TensorShape const& subTensorShape, unsigned int const* subTensorOrigin) const { - boost::ignore_unused(parent, subTensorShape, subTensorOrigin); + IgnoreUnused(parent, subTensorShape, subTensorOrigin); return nullptr; } @@ -35,7 +35,7 @@ std::unique_ptr<ITensorHandle> RefTensorHandleFactory::CreateTensorHandle(const std::unique_ptr<ITensorHandle> RefTensorHandleFactory::CreateTensorHandle(const TensorInfo& tensorInfo, DataLayout dataLayout) const { - boost::ignore_unused(dataLayout); + IgnoreUnused(dataLayout); return std::make_unique<RefTensorHandle>(tensorInfo, m_MemoryManager, m_ImportFlags); } diff --git a/src/backends/reference/RefWorkloadFactory.cpp b/src/backends/reference/RefWorkloadFactory.cpp index 2a415bfbf0..52d71df936 100644 --- a/src/backends/reference/RefWorkloadFactory.cpp +++ b/src/backends/reference/RefWorkloadFactory.cpp @@ -102,7 +102,7 @@ std::unique_ptr<ITensorHandle> RefWorkloadFactory::CreateTensorHandle(const Tens { // For Ref it is okay to make the TensorHandle memory managed as it can also store a pointer // to unmanaged memory. This also ensures memory alignment. - boost::ignore_unused(isMemoryManaged); + IgnoreUnused(isMemoryManaged); return std::make_unique<RefTensorHandle>(tensorInfo, m_MemoryManager); } @@ -112,14 +112,14 @@ std::unique_ptr<ITensorHandle> RefWorkloadFactory::CreateTensorHandle(const Tens { // For Ref it is okay to make the TensorHandle memory managed as it can also store a pointer // to unmanaged memory. This also ensures memory alignment. - boost::ignore_unused(isMemoryManaged, dataLayout); + IgnoreUnused(isMemoryManaged, dataLayout); return std::make_unique<RefTensorHandle>(tensorInfo, m_MemoryManager); } std::unique_ptr<IWorkload> RefWorkloadFactory::CreateAbs(const AbsQueueDescriptor& descriptor, const WorkloadInfo& info) const { - boost::ignore_unused(descriptor); + IgnoreUnused(descriptor); ElementwiseUnaryQueueDescriptor elementwiseUnaryDescriptor; elementwiseUnaryDescriptor.m_Parameters.m_Operation = UnaryOperation::Abs; @@ -267,7 +267,7 @@ std::unique_ptr<IWorkload> RefWorkloadFactory::CreateElementwiseUnary(const Elem std::unique_ptr<IWorkload> RefWorkloadFactory::CreateEqual(const EqualQueueDescriptor& descriptor, const WorkloadInfo& info) const { - boost::ignore_unused(descriptor); + IgnoreUnused(descriptor); ComparisonQueueDescriptor comparisonDescriptor; comparisonDescriptor.m_Parameters.m_Operation = ComparisonOperation::Equal; @@ -303,7 +303,7 @@ std::unique_ptr<IWorkload> RefWorkloadFactory::CreateGather(const GatherQueueDes std::unique_ptr<IWorkload> RefWorkloadFactory::CreateGreater(const GreaterQueueDescriptor& descriptor, const WorkloadInfo& info) const { - boost::ignore_unused(descriptor); + IgnoreUnused(descriptor); ComparisonQueueDescriptor comparisonDescriptor; comparisonDescriptor.m_Parameters.m_Operation = ComparisonOperation::Greater; @@ -506,7 +506,7 @@ std::unique_ptr<IWorkload> RefWorkloadFactory::CreateResizeBilinear(const Resize std::unique_ptr<IWorkload> RefWorkloadFactory::CreateRsqrt(const RsqrtQueueDescriptor& descriptor, const WorkloadInfo& info) const { - boost::ignore_unused(descriptor); + IgnoreUnused(descriptor); ElementwiseUnaryQueueDescriptor elementwiseUnaryDescriptor; elementwiseUnaryDescriptor.m_Parameters.m_Operation = UnaryOperation::Rsqrt; diff --git a/src/backends/reference/RefWorkloadFactory.hpp b/src/backends/reference/RefWorkloadFactory.hpp index 030ce6f03d..b64479e207 100644 --- a/src/backends/reference/RefWorkloadFactory.hpp +++ b/src/backends/reference/RefWorkloadFactory.hpp @@ -4,12 +4,11 @@ // #pragma once -#include <armnn/Optional.hpp> -#include <backendsCommon/WorkloadFactory.hpp> - #include "RefMemoryManager.hpp" -#include <boost/core/ignore_unused.hpp> +#include <armnn/Optional.hpp> +#include <backendsCommon/WorkloadFactory.hpp> +#include <armnn/utility/IgnoreUnused.hpp> namespace armnn @@ -48,7 +47,7 @@ public: TensorShape const& subTensorShape, unsigned int const* subTensorOrigin) const override { - boost::ignore_unused(parent, subTensorShape, subTensorOrigin); + IgnoreUnused(parent, subTensorShape, subTensorOrigin); return nullptr; } diff --git a/src/backends/reference/test/RefWorkloadFactoryHelper.hpp b/src/backends/reference/test/RefWorkloadFactoryHelper.hpp index 10e5b9fa28..30d2037a32 100644 --- a/src/backends/reference/test/RefWorkloadFactoryHelper.hpp +++ b/src/backends/reference/test/RefWorkloadFactoryHelper.hpp @@ -25,7 +25,7 @@ struct WorkloadFactoryHelper<armnn::RefWorkloadFactory> static armnn::RefWorkloadFactory GetFactory( const armnn::IBackendInternal::IMemoryManagerSharedPtr& memoryManager = nullptr) { - boost::ignore_unused(memoryManager); + IgnoreUnused(memoryManager); return armnn::RefWorkloadFactory(); } }; diff --git a/src/backends/reference/workloads/ArgMinMax.cpp b/src/backends/reference/workloads/ArgMinMax.cpp index db85b958e9..637aa17013 100644 --- a/src/backends/reference/workloads/ArgMinMax.cpp +++ b/src/backends/reference/workloads/ArgMinMax.cpp @@ -15,7 +15,7 @@ namespace armnn void ArgMinMax(Decoder<float>& in, int32_t* out, const TensorInfo& inputTensorInfo, const TensorInfo& outputTensorInfo, ArgMinMaxFunction function, int axis) { - boost::ignore_unused(outputTensorInfo); + IgnoreUnused(outputTensorInfo); unsigned int uAxis = armnnUtils::GetUnsignedAxis(inputTensorInfo.GetNumDimensions(), axis); diff --git a/src/backends/reference/workloads/BaseIterator.hpp b/src/backends/reference/workloads/BaseIterator.hpp index 3f0144670f..5cae5bda83 100644 --- a/src/backends/reference/workloads/BaseIterator.hpp +++ b/src/backends/reference/workloads/BaseIterator.hpp @@ -5,14 +5,13 @@ #pragma once - +#include <armnn/utility/IgnoreUnused.hpp> #include <armnn/TypesUtils.hpp> #include <armnnUtils/FloatingPointConverter.hpp> #include <ResolveType.hpp> #include <boost/assert.hpp> -#include <boost/core/ignore_unused.hpp> namespace armnn { @@ -107,7 +106,7 @@ public: TypedIterator& SetIndex(unsigned int index, unsigned int axisIndex = 0) override { - boost::ignore_unused(axisIndex); + IgnoreUnused(axisIndex); BOOST_ASSERT(m_Iterator); m_Iterator = m_Start + index; return *this; diff --git a/src/backends/reference/workloads/Dequantize.cpp b/src/backends/reference/workloads/Dequantize.cpp index 4025e8d7fa..63c0405efe 100644 --- a/src/backends/reference/workloads/Dequantize.cpp +++ b/src/backends/reference/workloads/Dequantize.cpp @@ -5,7 +5,8 @@ #include "Dequantize.hpp" -#include <boost/core/ignore_unused.hpp> +#include <armnn/utility/IgnoreUnused.hpp> + namespace armnn { @@ -14,7 +15,7 @@ void Dequantize(Decoder<float>& inputDecoder, const TensorInfo& inputInfo, const TensorInfo& outputInfo) { - boost::ignore_unused(outputInfo); + IgnoreUnused(outputInfo); BOOST_ASSERT(inputInfo.GetNumElements() == outputInfo.GetNumElements()); for (unsigned int i = 0; i < inputInfo.GetNumElements(); i++) { diff --git a/src/backends/reference/workloads/DetectionPostProcess.cpp b/src/backends/reference/workloads/DetectionPostProcess.cpp index 96e57803a1..57cf01e4a1 100644 --- a/src/backends/reference/workloads/DetectionPostProcess.cpp +++ b/src/backends/reference/workloads/DetectionPostProcess.cpp @@ -154,7 +154,7 @@ void DetectionPostProcess(const TensorInfo& boxEncodingsInfo, float* detectionScores, float* numDetections) { - boost::ignore_unused(anchorsInfo, detectionClassesInfo, detectionScoresInfo, numDetectionsInfo); + IgnoreUnused(anchorsInfo, detectionClassesInfo, detectionScoresInfo, numDetectionsInfo); // Transform center-size format which is (ycenter, xcenter, height, width) to box-corner format, // which represents the lower left corner and the upper right corner (ymin, xmin, ymax, xmax) diff --git a/src/backends/reference/workloads/Gather.cpp b/src/backends/reference/workloads/Gather.cpp index 5416855f48..4cf3a142a0 100644 --- a/src/backends/reference/workloads/Gather.cpp +++ b/src/backends/reference/workloads/Gather.cpp @@ -8,8 +8,8 @@ #include "RefWorkloadUtils.hpp" #include <backendsCommon/WorkloadData.hpp> +#include <armnn/utility/IgnoreUnused.hpp> -#include <boost/core/ignore_unused.hpp> #include <boost/numeric/conversion/cast.hpp> namespace armnn @@ -22,7 +22,7 @@ void Gather(const TensorInfo& paramsInfo, const int32_t* indices, Encoder<float>& output) { - boost::ignore_unused(outputInfo); + IgnoreUnused(outputInfo); const TensorShape& paramsShape = paramsInfo.GetShape(); unsigned int paramsProduct = 1; diff --git a/src/backends/reference/workloads/LogSoftmax.cpp b/src/backends/reference/workloads/LogSoftmax.cpp index ddf5674fb8..103d62a8df 100644 --- a/src/backends/reference/workloads/LogSoftmax.cpp +++ b/src/backends/reference/workloads/LogSoftmax.cpp @@ -6,11 +6,11 @@ #include "LogSoftmax.hpp" #include <armnnUtils/TensorUtils.hpp> +#include <armnn/utility/IgnoreUnused.hpp> #include <cmath> #include <boost/assert.hpp> -#include <boost/core/ignore_unused.hpp> #include <boost/numeric/conversion/cast.hpp> namespace @@ -37,7 +37,7 @@ void LogSoftmax(Decoder<float>& input, bool axisIsValid = ValidateAxis(descriptor.m_Axis, numDimensions); BOOST_ASSERT_MSG(axisIsValid, "Axis index is not in range [-numDimensions, numDimensions)."); - boost::ignore_unused(axisIsValid); + IgnoreUnused(axisIsValid); unsigned int uAxis = descriptor.m_Axis < 0 ? numDimensions - boost::numeric_cast<unsigned int>(std::abs(descriptor.m_Axis)) : diff --git a/src/backends/reference/workloads/RefStridedSliceWorkload.cpp b/src/backends/reference/workloads/RefStridedSliceWorkload.cpp index 8bb1670a48..bfd3c284ae 100644 --- a/src/backends/reference/workloads/RefStridedSliceWorkload.cpp +++ b/src/backends/reference/workloads/RefStridedSliceWorkload.cpp @@ -28,7 +28,7 @@ void RefStridedSliceWorkload::Execute() const DataType outputDataType = outputInfo.GetDataType(); BOOST_ASSERT(inputDataType == outputDataType); - boost::ignore_unused(outputDataType); + IgnoreUnused(outputDataType); StridedSlice(inputInfo, m_Data.m_Parameters, diff --git a/src/backends/reference/workloads/Slice.cpp b/src/backends/reference/workloads/Slice.cpp index c7ca3b156e..0223cdc56a 100644 --- a/src/backends/reference/workloads/Slice.cpp +++ b/src/backends/reference/workloads/Slice.cpp @@ -5,8 +5,9 @@ #include "Slice.hpp" +#include <armnn/utility/IgnoreUnused.hpp> + #include <boost/assert.hpp> -#include <boost/core/ignore_unused.hpp> #include <boost/numeric/conversion/cast.hpp> namespace armnn @@ -72,7 +73,7 @@ void Slice(const TensorInfo& inputInfo, const unsigned char* input = reinterpret_cast<const unsigned char*>(inputData); unsigned char* output = reinterpret_cast<unsigned char*>(outputData); - boost::ignore_unused(dim0); + IgnoreUnused(dim0); for (unsigned int idx0 = begin0; idx0 < begin0 + size0; ++idx0) { for (unsigned int idx1 = begin1; idx1 < begin1 + size1; ++idx1) |