From 8eb256065f0e75ecf8e427d56955e2bac117c2d7 Mon Sep 17 00:00:00 2001 From: Jan Eilers Date: Mon, 9 Mar 2020 12:13:48 +0000 Subject: IVGCVSW-4482 Remove boost::ignore_unused !referencetests:229377 Signed-off-by: Jan Eilers Change-Id: Ia9b360b4a057fe7bbce5b268092627c09a0dba82 --- src/armnn/test/CreateWorkload.hpp | 4 +- src/armnn/test/DebugCallbackTest.cpp | 2 +- src/armnn/test/EndToEndTest.cpp | 2 +- src/armnn/test/OptimizerTests.cpp | 6 +- src/armnn/test/OptionalTest.cpp | 6 +- src/armnn/test/ProfilerTests.cpp | 3 +- src/armnn/test/QuantizerTest.cpp | 112 ++++++++++++++-------------- src/armnn/test/RuntimeTests.cpp | 4 +- src/armnn/test/TensorHandleStrategyTest.cpp | 23 +++--- 9 files changed, 81 insertions(+), 81 deletions(-) (limited to 'src/armnn/test') diff --git a/src/armnn/test/CreateWorkload.hpp b/src/armnn/test/CreateWorkload.hpp index 4782c432a2..72ad9d45ef 100644 --- a/src/armnn/test/CreateWorkload.hpp +++ b/src/armnn/test/CreateWorkload.hpp @@ -11,6 +11,7 @@ #include #include +#include #include #include @@ -18,7 +19,6 @@ #include #include -#include #include @@ -1298,7 +1298,7 @@ std::pair> Cre armnn::Graph& graph, bool biasEnabled = false) { - boost::ignore_unused(graph); + IgnoreUnused(graph); // To create a PreCompiled layer, create a network and Optimize it. armnn::Network net; diff --git a/src/armnn/test/DebugCallbackTest.cpp b/src/armnn/test/DebugCallbackTest.cpp index bd8bdd543c..c89da83a89 100644 --- a/src/armnn/test/DebugCallbackTest.cpp +++ b/src/armnn/test/DebugCallbackTest.cpp @@ -60,7 +60,7 @@ BOOST_AUTO_TEST_CASE(RuntimeRegisterDebugCallback) std::vector slotIndexes; auto mockCallback = [&](LayerGuid guid, unsigned int slotIndex, ITensorHandle* tensor) { - boost::ignore_unused(guid); + IgnoreUnused(guid); slotIndexes.push_back(slotIndex); tensorShapes.push_back(tensor->GetShape()); callCount++; diff --git a/src/armnn/test/EndToEndTest.cpp b/src/armnn/test/EndToEndTest.cpp index df84be4277..a8192a6480 100644 --- a/src/armnn/test/EndToEndTest.cpp +++ b/src/armnn/test/EndToEndTest.cpp @@ -6,8 +6,8 @@ #include #include #include +#include -#include #include #include diff --git a/src/armnn/test/OptimizerTests.cpp b/src/armnn/test/OptimizerTests.cpp index 0ca4fc4764..56032adc33 100644 --- a/src/armnn/test/OptimizerTests.cpp +++ b/src/armnn/test/OptimizerTests.cpp @@ -691,7 +691,7 @@ BOOST_AUTO_TEST_CASE(BackendHintTest) LayerBindingId id, const char* name = nullptr) override { - boost::ignore_unused(id, name); + IgnoreUnused(id, name); auto inputLayer = boost::polymorphic_downcast(layer); BOOST_TEST((inputLayer->GetBackendId() == "MockBackend")); } @@ -700,7 +700,7 @@ BOOST_AUTO_TEST_CASE(BackendHintTest) LayerBindingId id, const char* name = nullptr) override { - boost::ignore_unused(id, name); + IgnoreUnused(id, name); auto outputLayer = boost::polymorphic_downcast(layer); BOOST_TEST((outputLayer->GetBackendId() == "MockBackend")); } @@ -709,7 +709,7 @@ BOOST_AUTO_TEST_CASE(BackendHintTest) const ActivationDescriptor& activationDescriptor, const char* name = nullptr) override { - boost::ignore_unused(activationDescriptor, name); + IgnoreUnused(activationDescriptor, name); auto activation = boost::polymorphic_downcast(layer); BOOST_TEST((activation->GetBackendId() == "CustomBackend")); } diff --git a/src/armnn/test/OptionalTest.cpp b/src/armnn/test/OptionalTest.cpp index fd136439c8..73c96431fb 100644 --- a/src/armnn/test/OptionalTest.cpp +++ b/src/armnn/test/OptionalTest.cpp @@ -7,19 +7,19 @@ #include #include -#include +#include namespace { void PassStringRef(armnn::Optional value) { - boost::ignore_unused(value); + armnn::IgnoreUnused(value); } void PassStringRefWithDefault(armnn::Optional value = armnn::EmptyOptional()) { - boost::ignore_unused(value); + armnn::IgnoreUnused(value); } } // namespace diff --git a/src/armnn/test/ProfilerTests.cpp b/src/armnn/test/ProfilerTests.cpp index a052862bdd..9376fa4cea 100644 --- a/src/armnn/test/ProfilerTests.cpp +++ b/src/armnn/test/ProfilerTests.cpp @@ -5,6 +5,7 @@ #include #include +#include #include #include @@ -309,7 +310,7 @@ BOOST_AUTO_TEST_CASE(ProfilerJsonPrinter) profiler->Print(json); std::string output = buffer.str(); - boost::ignore_unused(output); + armnn::IgnoreUnused(output); // Disable profiling here to not print out anything on stdout. profiler->EnableProfiling(false); diff --git a/src/armnn/test/QuantizerTest.cpp b/src/armnn/test/QuantizerTest.cpp index 2dc054af07..ef9b2da782 100644 --- a/src/armnn/test/QuantizerTest.cpp +++ b/src/armnn/test/QuantizerTest.cpp @@ -3,15 +3,6 @@ // SPDX-License-Identifier: MIT // -#include -#include -#include -#include - -#include - -#include - #include "../Graph.hpp" #include "../Network.hpp" #include "../NetworkQuantizerUtils.hpp" @@ -19,7 +10,14 @@ #include "../RangeTracker.hpp" #include "../../armnnQuantizer/CommandLineProcessor.hpp" -#include +#include +#include +#include +#include +#include +#include +#include + #include #include @@ -58,7 +56,7 @@ public: LayerBindingId id, const char* name = nullptr) override { - boost::ignore_unused(id, name); + IgnoreUnused(id, name); const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo(); BOOST_TEST(m_InputShape == info.GetShape()); // Based off current default [-15.0f, 15.0f] @@ -72,7 +70,7 @@ public: LayerBindingId id, const char* name = nullptr) override { - boost::ignore_unused(id, name); + IgnoreUnused(id, name); const TensorInfo& info = layer->GetInputSlot(0).GetConnection()->GetTensorInfo(); BOOST_TEST(m_OutputShape == info.GetShape()); } @@ -116,7 +114,7 @@ protected: const OffsetScalePair& params, DataType dataType = DataType::QAsymmU8) { - boost::ignore_unused(dataType); + IgnoreUnused(dataType); TestQuantizationParamsImpl(info, dataType, params.first, params.second); } @@ -212,7 +210,7 @@ public: void VisitAdditionLayer(const IConnectableLayer* layer, const char* name = nullptr) override { - boost::ignore_unused(name); + IgnoreUnused(name); TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); // Based off default static range [-20.0f, 20.0f] @@ -282,7 +280,7 @@ public: const ActivationDescriptor& descriptor, const char* name = nullptr) override { - boost::ignore_unused(descriptor, name); + IgnoreUnused(descriptor, name); TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); @@ -385,7 +383,7 @@ BOOST_AUTO_TEST_CASE(InputOutputLayerDynamicQuant) LayerBindingId id, const char* name = nullptr) override { - boost::ignore_unused(id, name); + IgnoreUnused(id, name); const TensorInfo& info = layer->GetInputSlot(0).GetConnection()->GetTensorInfo(); BOOST_CHECK_MESSAGE(info.GetDataType() == m_DataType, std::string(armnn::GetDataTypeName(info.GetDataType())) @@ -543,7 +541,7 @@ BOOST_AUTO_TEST_CASE(QuantizeBoundedReluActivation) const ActivationDescriptor& descriptor, const char* name = nullptr) override { - boost::ignore_unused(descriptor, name); + IgnoreUnused(descriptor, name); TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); // Based off default static range [0.0f, 3.5f] @@ -599,7 +597,7 @@ BOOST_AUTO_TEST_CASE(QuantizeTanHActivation) const ActivationDescriptor& descriptor, const char* name = nullptr) override { - boost::ignore_unused(descriptor, name); + IgnoreUnused(descriptor, name); TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); // Based off default static range [-1.0f, 1.0f] @@ -654,7 +652,7 @@ public: const ActivationDescriptor& descriptor, const char* name = nullptr) override { - boost::ignore_unused(descriptor, name); + IgnoreUnused(descriptor, name); TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); // Based off default static range [-5.0f, 15.0f] @@ -725,7 +723,7 @@ BOOST_AUTO_TEST_CASE(QuantizeELuActivation) const ActivationDescriptor& descriptor, const char* name = nullptr) override { - boost::ignore_unused(descriptor, name); + IgnoreUnused(descriptor, name); TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); // Based off default static range [-15.0f, 15.0f] @@ -779,7 +777,7 @@ BOOST_AUTO_TEST_CASE(QuantizeHardSwishActivation) const ActivationDescriptor& descriptor, const char* name = nullptr) override { - boost::ignore_unused(descriptor, name); + IgnoreUnused(descriptor, name); TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); // Based off default static range [-15.0f, 15.0f] @@ -839,7 +837,7 @@ BOOST_AUTO_TEST_CASE(QuantizeBatchNorm) const ConstTensor& gamma, const char* name = nullptr) override { - boost::ignore_unused(desc, name); + IgnoreUnused(desc, name); TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); // Based off default static range [-15.0f, 15.0f] @@ -924,7 +922,7 @@ BOOST_AUTO_TEST_CASE(QuantizeDepthToSpace) const DepthToSpaceDescriptor& desc, const char* name = nullptr) { - boost::ignore_unused(desc, name); + IgnoreUnused(desc, name); const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo(); const OffsetScalePair qAsymmU8Params{ 30.0f / g_AsymmU8QuantizationBase, 128 }; @@ -1116,7 +1114,7 @@ void ValidateFullyConnectedLayer(const bool biasEnabled) const Optional& biases, const char* name = nullptr) override { - boost::ignore_unused(desc, name); + IgnoreUnused(desc, name); TestQuantizationOnLayersWithBiases(layer, weights, biases); } }; @@ -1173,7 +1171,7 @@ void TestQuantizeConvolution2d(bool useBiases) const Optional& biases, const char *name = nullptr) override { - boost::ignore_unused(convolution2dDescriptor, name); + IgnoreUnused(convolution2dDescriptor, name); TestQuantizationOnLayersWithBiases(layer, weights, biases); } }; @@ -1259,7 +1257,7 @@ void TestQuantizeDepthwiseConvolution2d(bool useBiases) const Optional& biases, const char *name = nullptr) override { - boost::ignore_unused(convolution2dDescriptor, name); + IgnoreUnused(convolution2dDescriptor, name); TestQuantizationOnLayersWithBiases(layer, weights, biases); } }; @@ -1343,7 +1341,7 @@ BOOST_AUTO_TEST_CASE(QuantizeInstanceNormalization) const InstanceNormalizationDescriptor& descriptor, const char* name = nullptr) { - boost::ignore_unused(descriptor, name); + IgnoreUnused(descriptor, name); const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo(); const OffsetScalePair qAsymmU8Params{ 30.0f / g_AsymmU8QuantizationBase, 128 }; @@ -1411,7 +1409,7 @@ BOOST_AUTO_TEST_CASE(QuantizeLogSoftmax) const SoftmaxDescriptor& descriptor, const char* name = nullptr) override { - boost::ignore_unused(descriptor, name); + IgnoreUnused(descriptor, name); TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); const OffsetScalePair qAsymmU8Params{ 30.0f / g_AsymmU8QuantizationBase, 128 }; @@ -1503,7 +1501,7 @@ BOOST_AUTO_TEST_CASE(QuantizeSoftmax) const SoftmaxDescriptor& descriptor, const char* name = nullptr) override { - boost::ignore_unused(descriptor, name); + IgnoreUnused(descriptor, name); TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); // Based off default static range [0.0f, 1.0f] @@ -1636,7 +1634,7 @@ BOOST_AUTO_TEST_CASE(QuantizePermute) const PermuteDescriptor& desc, const char* name = nullptr) override { - boost::ignore_unused(desc, name); + IgnoreUnused(desc, name); CheckForwardedQuantizationSettings(layer); } }; @@ -1691,7 +1689,7 @@ BOOST_AUTO_TEST_CASE(QuantizeSpaceToBatch) const SpaceToBatchNdDescriptor& spaceToBatchNdDescriptor, const char* name = nullptr) override { - boost::ignore_unused(spaceToBatchNdDescriptor, name); + IgnoreUnused(spaceToBatchNdDescriptor, name); CheckForwardedQuantizationSettings(layer); } }; @@ -1804,7 +1802,7 @@ BOOST_AUTO_TEST_CASE(QuantizePooling2d) const Pooling2dDescriptor& desc, const char* name = nullptr) override { - boost::ignore_unused(desc, name); + IgnoreUnused(desc, name); CheckForwardedQuantizationSettings(layer); } }; @@ -1873,7 +1871,7 @@ BOOST_AUTO_TEST_CASE(QuantizeConstant) const ConstTensor& input, const char* name = nullptr) override { - boost::ignore_unused(input, name); + IgnoreUnused(input, name); TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); // Based off the range of values in the const tensor used for the test: [-2.0f, 6.0f] @@ -1946,20 +1944,20 @@ BOOST_AUTO_TEST_CASE(QuantizeArgMinMax) LayerBindingId id, const char* name = nullptr) override { - boost::ignore_unused(layer, id, name); + IgnoreUnused(layer, id, name); } void VisitOutputLayer(const IConnectableLayer* layer, LayerBindingId id, const char* name = nullptr) override { - boost::ignore_unused(layer, id, name); + IgnoreUnused(layer, id, name); } void VisitArgMinMaxLayer(const IConnectableLayer* layer, const ArgMinMaxDescriptor& argMinMaxDescriptor, const char* name = nullptr) override { - boost::ignore_unused(argMinMaxDescriptor, name); + IgnoreUnused(argMinMaxDescriptor, name); TensorInfo outputInfo = layer->GetOutputSlot(0).GetTensorInfo(); TestQuantizationParams(outputInfo, @@ -2034,7 +2032,7 @@ BOOST_AUTO_TEST_CASE(QuantizeComparison) const ComparisonDescriptor& descriptor, const char* name = nullptr) override { - boost::ignore_unused(descriptor, name); + IgnoreUnused(descriptor, name); TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); const OffsetScalePair qAsymmU8Params{ 30.0f / g_AsymmU8QuantizationBase, 128 }; @@ -2106,19 +2104,19 @@ BOOST_AUTO_TEST_CASE(QuantizeConcat) LayerBindingId id, const char* name = nullptr) override { - boost::ignore_unused(layer, id, name); + IgnoreUnused(layer, id, name); } void VisitOutputLayer(const IConnectableLayer* layer, LayerBindingId id, const char* name = nullptr) override { - boost::ignore_unused(layer, id, name); + IgnoreUnused(layer, id, name); } void VisitConcatLayer(const IConnectableLayer* layer, const OriginsDescriptor& originsDescriptor, const char* name = nullptr) override { - boost::ignore_unused(originsDescriptor, name); + IgnoreUnused(originsDescriptor, name); TensorInfo outputInfo = layer->GetOutputSlot(0).GetTensorInfo(); TestQuantizationParams( outputInfo, {60.8f / g_AsymmU8QuantizationBase, 65}, @@ -2214,7 +2212,7 @@ BOOST_AUTO_TEST_CASE(QuantizeReshape) const ReshapeDescriptor& reshapeDescriptor, const char* name = nullptr) override { - boost::ignore_unused(reshapeDescriptor, name); + IgnoreUnused(reshapeDescriptor, name); CheckForwardedQuantizationSettings(layer); } }; @@ -2269,7 +2267,7 @@ BOOST_AUTO_TEST_CASE(QuantizeSplitter) const SplitterDescriptor& desc, const char* name = nullptr) { - boost::ignore_unused(desc, name); + IgnoreUnused(desc, name); CheckForwardedQuantizationSettings(layer); } }; @@ -2325,7 +2323,7 @@ BOOST_AUTO_TEST_CASE(QuantizeResize) const ResizeDescriptor& resizeDescriptor, const char* name = nullptr) override { - boost::ignore_unused(resizeDescriptor, name); + IgnoreUnused(resizeDescriptor, name); CheckForwardedQuantizationSettings(layer); } }; @@ -2382,7 +2380,7 @@ BOOST_AUTO_TEST_CASE(QuantizeStridedSlice) const StridedSliceDescriptor& desc, const char* name = nullptr) { - boost::ignore_unused(desc, name); + IgnoreUnused(desc, name); CheckForwardedQuantizationSettings(layer); } }; @@ -2437,7 +2435,7 @@ BOOST_AUTO_TEST_CASE(QuantizeBatchToSpace) const BatchToSpaceNdDescriptor& batchToSpaceNdDescriptor, const char* name = nullptr) override { - boost::ignore_unused(batchToSpaceNdDescriptor, name); + IgnoreUnused(batchToSpaceNdDescriptor, name); CheckForwardedQuantizationSettings(layer); } }; @@ -2499,7 +2497,7 @@ BOOST_AUTO_TEST_CASE(QuantizePrelu) LayerBindingId id, const char* name = nullptr) override { - boost::ignore_unused(id, name); + IgnoreUnused(id, name); const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo(); switch (id) @@ -2526,7 +2524,7 @@ BOOST_AUTO_TEST_CASE(QuantizePrelu) LayerBindingId id, const char* name = nullptr) override { - boost::ignore_unused(id, name); + IgnoreUnused(id, name); const TensorInfo& info = layer->GetInputSlot(0).GetConnection()->GetTensorInfo(); BOOST_TEST(m_OutputShape == info.GetShape()); } @@ -2534,7 +2532,7 @@ BOOST_AUTO_TEST_CASE(QuantizePrelu) void VisitPreluLayer(const IConnectableLayer* layer, const char* name = nullptr) override { - boost::ignore_unused(name); + IgnoreUnused(name); const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo(); TestQuantizationParams(info, { 30.0f / g_AsymmU8QuantizationBase, 128 }, // QASymmU8 @@ -2617,7 +2615,7 @@ void TestQuantizeTransposeConvolution2d(bool useBiases) const Optional& biases, const char *name = nullptr) override { - boost::ignore_unused(descriptor, name); + IgnoreUnused(descriptor, name); TestQuantizationOnLayersWithBiases(layer, weights, biases); } }; @@ -2704,20 +2702,20 @@ BOOST_AUTO_TEST_CASE(QuantizeStack) LayerBindingId id, const char* name = nullptr) override { - boost::ignore_unused(layer, id, name); + IgnoreUnused(layer, id, name); } void VisitOutputLayer(const IConnectableLayer* layer, LayerBindingId id, const char* name = nullptr) override { - boost::ignore_unused(layer, id, name); + IgnoreUnused(layer, id, name); } void VisitStackLayer(const IConnectableLayer* layer, const StackDescriptor& descriptor, const char* name = nullptr) override { - boost::ignore_unused(descriptor, name); + IgnoreUnused(descriptor, name); TensorInfo outputInfo = layer->GetOutputSlot(0).GetTensorInfo(); TestQuantizationParams(outputInfo, @@ -2784,7 +2782,7 @@ BOOST_AUTO_TEST_CASE(QuantizeSlice) const SliceDescriptor& desc, const char* name = nullptr) { - boost::ignore_unused(desc, name); + IgnoreUnused(desc, name); const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo(); const OffsetScalePair qAsymmU8Params{ 30.0f / g_AsymmU8QuantizationBase, 128 }; @@ -2876,7 +2874,7 @@ public: LayerBindingId id, const char* name = nullptr) override { - boost::ignore_unused(id, name); + IgnoreUnused(id, name); const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo(); BOOST_TEST(GetDataTypeName(info.GetDataType()) == GetDataTypeName(m_DataType)); BOOST_TEST(m_InputShape == info.GetShape()); @@ -2886,7 +2884,7 @@ public: LayerBindingId id, const char* name = nullptr) override { - boost::ignore_unused(id, name); + IgnoreUnused(id, name); const TensorInfo& info = layer->GetInputSlot(0).GetConnection()->GetTensorInfo(); BOOST_TEST(GetDataTypeName(info.GetDataType()) == GetDataTypeName(m_DataType)); BOOST_TEST(m_OutputShape == info.GetShape()); @@ -2895,14 +2893,14 @@ public: void VisitQuantizeLayer(const IConnectableLayer* layer, const char* name = nullptr) override { - boost::ignore_unused(layer, name); + IgnoreUnused(layer, name); m_VisitedQuantizeLayer = true; } void VisitDequantizeLayer(const IConnectableLayer* layer, const char* name = nullptr) override { - boost::ignore_unused(layer, name); + IgnoreUnused(layer, name); m_VisitedDequantizeLayer = true; } diff --git a/src/armnn/test/RuntimeTests.cpp b/src/armnn/test/RuntimeTests.cpp index d9ed18bdd5..e3cbe03c62 100644 --- a/src/armnn/test/RuntimeTests.cpp +++ b/src/armnn/test/RuntimeTests.cpp @@ -158,8 +158,8 @@ BOOST_AUTO_TEST_CASE(RuntimeMemoryLeak) // These are needed because VALGRIND_COUNT_LEAKS is a macro that assigns to the parameters // so they are assigned to, but still considered unused, causing a warning. - boost::ignore_unused(dubious); - boost::ignore_unused(suppressed); + IgnoreUnused(dubious); + IgnoreUnused(suppressed); } #endif // WITH_VALGRIND diff --git a/src/armnn/test/TensorHandleStrategyTest.cpp b/src/armnn/test/TensorHandleStrategyTest.cpp index 3e59c0b604..976e58eb50 100644 --- a/src/armnn/test/TensorHandleStrategyTest.cpp +++ b/src/armnn/test/TensorHandleStrategyTest.cpp @@ -16,10 +16,11 @@ #include +#include + #include #include -#include using namespace armnn; @@ -44,20 +45,20 @@ public: TensorShape const& subTensorShape, unsigned int const* subTensorOrigin) const override { - boost::ignore_unused(parent, subTensorShape, subTensorOrigin); + IgnoreUnused(parent, subTensorShape, subTensorOrigin); return nullptr; } std::unique_ptr CreateTensorHandle(const TensorInfo& tensorInfo) const override { - boost::ignore_unused(tensorInfo); + IgnoreUnused(tensorInfo); return nullptr; } std::unique_ptr CreateTensorHandle(const TensorInfo& tensorInfo, DataLayout dataLayout) const override { - boost::ignore_unused(tensorInfo, dataLayout); + IgnoreUnused(tensorInfo, dataLayout); return nullptr; } @@ -85,20 +86,20 @@ public: TensorShape const& subTensorShape, unsigned int const* subTensorOrigin) const override { - boost::ignore_unused(parent, subTensorShape, subTensorOrigin); + IgnoreUnused(parent, subTensorShape, subTensorOrigin); return nullptr; } std::unique_ptr CreateTensorHandle(const TensorInfo& tensorInfo) const override { - boost::ignore_unused(tensorInfo); + IgnoreUnused(tensorInfo); return nullptr; } std::unique_ptr CreateTensorHandle(const TensorInfo& tensorInfo, DataLayout dataLayout) const override { - boost::ignore_unused(tensorInfo, dataLayout); + IgnoreUnused(tensorInfo, dataLayout); return nullptr; } @@ -123,7 +124,7 @@ public: IWorkloadFactoryPtr CreateWorkloadFactory(const IMemoryManagerSharedPtr& memoryManager = nullptr) const override { - boost::ignore_unused(memoryManager); + IgnoreUnused(memoryManager); return IWorkloadFactoryPtr{}; } @@ -164,7 +165,7 @@ public: IWorkloadFactoryPtr CreateWorkloadFactory(const IMemoryManagerSharedPtr& memoryManager = nullptr) const override { - boost::ignore_unused(memoryManager); + IgnoreUnused(memoryManager); return IWorkloadFactoryPtr{}; } @@ -202,7 +203,7 @@ public: IWorkloadFactoryPtr CreateWorkloadFactory(const IMemoryManagerSharedPtr& memoryManager = nullptr) const override { - boost::ignore_unused(memoryManager); + IgnoreUnused(memoryManager); return IWorkloadFactoryPtr{}; } @@ -239,7 +240,7 @@ public: IWorkloadFactoryPtr CreateWorkloadFactory(const IMemoryManagerSharedPtr& memoryManager = nullptr) const override { - boost::ignore_unused(memoryManager); + IgnoreUnused(memoryManager); return IWorkloadFactoryPtr{}; } -- cgit v1.2.1