From 859f9ced8360fcfe6daedf6c8f6613005893f814 Mon Sep 17 00:00:00 2001 From: Derek Lamberti Date: Tue, 10 Dec 2019 22:05:21 +0000 Subject: IVGCVSW-4246 Clean build of Visitors with -Wextra Change-Id: Icb1b35ff55fa22103777853e6f49fc282d61750d Signed-off-by: Derek Lamberti --- src/armnn/DynamicQuantizationVisitor.cpp | 4 ++ src/armnn/OverrideInputRangeVisitor.cpp | 2 + src/armnn/StaticRangeVisitor.cpp | 6 +++ src/armnn/test/QuantizerTest.cpp | 66 +++++++++++++++++++++++++++++--- 4 files changed, 72 insertions(+), 6 deletions(-) (limited to 'src/armnn') diff --git a/src/armnn/DynamicQuantizationVisitor.cpp b/src/armnn/DynamicQuantizationVisitor.cpp index d4e0c9006c..ba87c6d335 100644 --- a/src/armnn/DynamicQuantizationVisitor.cpp +++ b/src/armnn/DynamicQuantizationVisitor.cpp @@ -84,6 +84,7 @@ void DynamicQuantizationVisitor::VisitNonCalibratedLayers() { void DynamicQuantizationVisitor::VisitAdditionLayer(const IConnectableLayer* layer, const char* name) { + boost::ignore_unused(name); SetRange(layer, 0, -20.f, 20.f); AddToCalibratedLayers(layer); } @@ -138,6 +139,7 @@ void DynamicQuantizationVisitor::VisitActivationLayer(const IConnectableLayer* l const ActivationDescriptor& activationDescriptor, const char* name) { + boost::ignore_unused(name, activationDescriptor); switch (activationDescriptor.m_Function) { // Range is 0, 15 for Abs, Linear, ReLu and Soft ReLu @@ -246,6 +248,8 @@ void DynamicQuantizationVisitor::VisitConcatLayer(const IConnectableLayer* layer const ConcatDescriptor& originsDescriptor, const char* name) { + boost::ignore_unused(name); + boost::ignore_unused(originsDescriptor); float min = std::numeric_limits::max(); float max = std::numeric_limits::lowest(); for (unsigned int i = 0; i < layer->GetNumInputSlots(); ++i) diff --git a/src/armnn/OverrideInputRangeVisitor.cpp b/src/armnn/OverrideInputRangeVisitor.cpp index 058e630c79..d047c5bbe8 100644 --- a/src/armnn/OverrideInputRangeVisitor.cpp +++ b/src/armnn/OverrideInputRangeVisitor.cpp @@ -8,6 +8,7 @@ #include "Layer.hpp" #include +#include namespace armnn { @@ -22,6 +23,7 @@ OverrideInputRangeVisitor::OverrideInputRangeVisitor(RangeTracker& ranges, void OverrideInputRangeVisitor::VisitInputLayer(const IConnectableLayer* layer, LayerBindingId id, const char* name) { + boost::ignore_unused(name); if (m_LayerId == id) { m_Ranges.SetRange(layer, 0, m_MinMaxRange.first, m_MinMaxRange.second); diff --git a/src/armnn/StaticRangeVisitor.cpp b/src/armnn/StaticRangeVisitor.cpp index 94f0a01997..81428c1f90 100644 --- a/src/armnn/StaticRangeVisitor.cpp +++ b/src/armnn/StaticRangeVisitor.cpp @@ -31,6 +31,7 @@ void StaticRangeVisitor::ForwardParentParameters(const IConnectableLayer* layer) void StaticRangeVisitor::VisitAdditionLayer(const IConnectableLayer* layer, const char* name) { + boost::ignore_unused(name); SetRange(layer, 0, -20.f, 20.f); } @@ -81,6 +82,7 @@ void StaticRangeVisitor::VisitActivationLayer(const IConnectableLayer* layer, const ActivationDescriptor& activationDescriptor, const char* name) { + boost::ignore_unused(name); switch (activationDescriptor.m_Function) { // Range is 0, 15 for Abs, Linear, ReLu and Soft ReLu @@ -141,6 +143,7 @@ void StaticRangeVisitor::VisitPooling2dLayer(const IConnectableLayer* layer, const char* name) { boost::ignore_unused(pooling2dDescriptor); + boost::ignore_unused(name); ForwardParentParameters(layer); } @@ -149,6 +152,7 @@ void StaticRangeVisitor::VisitSoftmaxLayer(const IConnectableLayer* layer, const char* name) { boost::ignore_unused(softmaxDescriptor); + boost::ignore_unused(name); SetRange(layer, 0, 0.f, 1.f); } @@ -156,6 +160,8 @@ void StaticRangeVisitor::VisitConcatLayer(const IConnectableLayer* layer, const OriginsDescriptor& originsDescriptor, const char* name) { + boost::ignore_unused(originsDescriptor); + boost::ignore_unused(name); float min = std::numeric_limits::max(); float max = std::numeric_limits::lowest(); for (unsigned int i = 0; i < layer->GetNumInputSlots(); ++i) diff --git a/src/armnn/test/QuantizerTest.cpp b/src/armnn/test/QuantizerTest.cpp index 37b3bfa852..e147a84eb6 100644 --- a/src/armnn/test/QuantizerTest.cpp +++ b/src/armnn/test/QuantizerTest.cpp @@ -19,6 +19,7 @@ #include "../RangeTracker.hpp" #include "../../armnnQuantizer/CommandLineProcessor.hpp" +#include #include #include @@ -55,6 +56,7 @@ public: LayerBindingId id, const char* name = nullptr) override { + boost::ignore_unused(id, name); const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo(); BOOST_TEST(m_InputShape == info.GetShape()); // Based off current default [-15.0f, 15.0f] @@ -67,6 +69,7 @@ public: LayerBindingId id, const char* name = nullptr) override { + boost::ignore_unused(id, name); const TensorInfo& info = layer->GetInputSlot(0).GetConnection()->GetTensorInfo(); BOOST_TEST(m_OutputShape == info.GetShape()); } @@ -105,6 +108,7 @@ protected: const OffsetScalePair& params, DataType dataType = DataType::QuantisedAsymm8) { + boost::ignore_unused(dataType); TestQuantizationParamsImpl(info, DataType::QuantisedAsymm8, params.first, params.second); } @@ -191,6 +195,7 @@ public: void VisitAdditionLayer(const IConnectableLayer* layer, const char* name = nullptr) override { + boost::ignore_unused(name); TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); // Based off default static range [-20.0f, 20.0f] @@ -253,6 +258,8 @@ public: const ActivationDescriptor& descriptor, const char* name = nullptr) override { + boost::ignore_unused(descriptor, name); + TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); // Based off default static range [0.0f, 15.0f] @@ -353,6 +360,7 @@ BOOST_AUTO_TEST_CASE(InputOutputLayerDynamicQuant) LayerBindingId id, const char* name = nullptr) override { + boost::ignore_unused(id, name); const TensorInfo& info = layer->GetInputSlot(0).GetConnection()->GetTensorInfo(); BOOST_CHECK_MESSAGE(info.GetDataType() == m_DataType, std::string(armnn::GetDataTypeName(info.GetDataType())) @@ -489,6 +497,7 @@ BOOST_AUTO_TEST_CASE(QuantizeBoundedReluActivation) const ActivationDescriptor& descriptor, const char* name = nullptr) override { + boost::ignore_unused(descriptor, name); TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); // Based off default static range [0.0f, 3.5f] @@ -538,6 +547,7 @@ BOOST_AUTO_TEST_CASE(QuantizeTanHActivation) const ActivationDescriptor& descriptor, const char* name = nullptr) override { + boost::ignore_unused(descriptor, name); TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); // Based off default static range [-1.0f, 1.0f] @@ -586,6 +596,7 @@ public: const ActivationDescriptor& descriptor, const char* name = nullptr) override { + boost::ignore_unused(descriptor, name); TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); // Based off default static range [-5.0f, 15.0f] @@ -652,6 +663,7 @@ BOOST_AUTO_TEST_CASE(QuantizeBatchNorm) const ConstTensor& gamma, const char* name = nullptr) override { + boost::ignore_unused(desc, name); TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); // Based off default static range [-15.0f, 15.0f] @@ -730,6 +742,7 @@ BOOST_AUTO_TEST_CASE(QuantizeDepthToSpace) const DepthToSpaceDescriptor& desc, const char* name = nullptr) { + boost::ignore_unused(desc, name); const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo(); const OffsetScalePair qAsymm8Params{ 30.0f / g_Asymm8QuantizationBase, 128 }; @@ -914,6 +927,7 @@ void ValidateFullyConnectedLayer(const bool biasEnabled) const Optional& biases, const char* name = nullptr) override { + boost::ignore_unused(desc, name); TestQuantizationOnLayersWithBiases(layer, weights, biases); } }; @@ -965,6 +979,7 @@ void TestQuantizeConvolution2d(bool useBiases) const Optional& biases, const char *name = nullptr) override { + boost::ignore_unused(convolution2dDescriptor, name); TestQuantizationOnLayersWithBiases(layer, weights, biases); } }; @@ -1045,6 +1060,7 @@ void TestQuantizeDepthwiseConvolution2d(bool useBiases) const Optional& biases, const char *name = nullptr) override { + boost::ignore_unused(convolution2dDescriptor, name); TestQuantizationOnLayersWithBiases(layer, weights, biases); } }; @@ -1123,6 +1139,7 @@ BOOST_AUTO_TEST_CASE(QuantizeInstanceNormalization) const InstanceNormalizationDescriptor& descriptor, const char* name = nullptr) { + boost::ignore_unused(descriptor, name); const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo(); const OffsetScalePair qAsymm8Params{ 30.0f / g_Asymm8QuantizationBase, 128 }; @@ -1183,6 +1200,7 @@ BOOST_AUTO_TEST_CASE(QuantizeLogSoftmax) const SoftmaxDescriptor& descriptor, const char* name = nullptr) override { + boost::ignore_unused(descriptor, name); TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); const OffsetScalePair qAsymm8Params{ 30.0f / g_Asymm8QuantizationBase, 128 }; @@ -1267,6 +1285,7 @@ BOOST_AUTO_TEST_CASE(QuantizeSoftmax) const SoftmaxDescriptor& descriptor, const char* name = nullptr) override { + boost::ignore_unused(descriptor, name); TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); // Based off default static range [0.0f, 1.0f] @@ -1388,6 +1407,7 @@ BOOST_AUTO_TEST_CASE(QuantizePermute) const PermuteDescriptor& desc, const char* name = nullptr) override { + boost::ignore_unused(desc, name); CheckForwardedQuantizationSettings(layer); } }; @@ -1437,6 +1457,7 @@ BOOST_AUTO_TEST_CASE(QuantizeSpaceToBatch) const SpaceToBatchNdDescriptor& spaceToBatchNdDescriptor, const char* name = nullptr) override { + boost::ignore_unused(spaceToBatchNdDescriptor, name); CheckForwardedQuantizationSettings(layer); } }; @@ -1538,6 +1559,7 @@ BOOST_AUTO_TEST_CASE(QuantizePooling2d) const Pooling2dDescriptor& desc, const char* name = nullptr) override { + boost::ignore_unused(desc, name); CheckForwardedQuantizationSettings(layer); } }; @@ -1601,6 +1623,7 @@ BOOST_AUTO_TEST_CASE(QuantizeConstant) const ConstTensor& input, const char* name = nullptr) override { + boost::ignore_unused(input, name); TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); // Based off the range of values in the const tensor used for the test: [-2.0f, 6.0f] @@ -1667,6 +1690,7 @@ BOOST_AUTO_TEST_CASE(QuantizeAbs) void VisitAbsLayer(const IConnectableLayer *layer, const char *name = nullptr) override { + boost::ignore_unused(name); TensorInfo outputInfo = layer->GetOutputSlot(0).GetTensorInfo(); TestQuantizationParams(outputInfo, @@ -1720,16 +1744,21 @@ BOOST_AUTO_TEST_CASE(QuantizeArgMinMax) void VisitInputLayer(const IConnectableLayer* layer, LayerBindingId id, const char* name = nullptr) override - {} + { + boost::ignore_unused(layer, id, name); + } void VisitOutputLayer(const IConnectableLayer* layer, LayerBindingId id, const char* name = nullptr) override - {} + { + boost::ignore_unused(layer, id, name); + } void VisitArgMinMaxLayer(const IConnectableLayer* layer, const ArgMinMaxDescriptor& argMinMaxDescriptor, const char* name = nullptr) override { + boost::ignore_unused(argMinMaxDescriptor, name); TensorInfo outputInfo = layer->GetOutputSlot(0).GetTensorInfo(); TestQuantizationParams(outputInfo, @@ -1798,6 +1827,7 @@ BOOST_AUTO_TEST_CASE(QuantizeComparison) const ComparisonDescriptor& descriptor, const char* name = nullptr) override { + boost::ignore_unused(descriptor, name); TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); const OffsetScalePair qAsymm8Params{ 30.0f / g_Asymm8QuantizationBase, 128 }; @@ -1860,15 +1890,20 @@ BOOST_AUTO_TEST_CASE(QuantizeConcat) void VisitInputLayer(const IConnectableLayer* layer, LayerBindingId id, const char* name = nullptr) override - {} + { + boost::ignore_unused(layer, id, name); + } void VisitOutputLayer(const IConnectableLayer* layer, LayerBindingId id, const char* name = nullptr) override - {} + { + boost::ignore_unused(layer, id, name); + } void VisitConcatLayer(const IConnectableLayer* layer, const OriginsDescriptor& originsDescriptor, const char* name = nullptr) override { + boost::ignore_unused(originsDescriptor, name); TensorInfo outputInfo = layer->GetOutputSlot(0).GetTensorInfo(); TestQuantizationParams( @@ -1964,6 +1999,7 @@ BOOST_AUTO_TEST_CASE(QuantizeReshape) const ReshapeDescriptor& reshapeDescriptor, const char* name = nullptr) override { + boost::ignore_unused(reshapeDescriptor, name); CheckForwardedQuantizationSettings(layer); } }; @@ -2013,6 +2049,7 @@ BOOST_AUTO_TEST_CASE(QuantizeSplitter) const SplitterDescriptor& desc, const char* name = nullptr) { + boost::ignore_unused(desc, name); CheckForwardedQuantizationSettings(layer); } }; @@ -2063,6 +2100,7 @@ BOOST_AUTO_TEST_CASE(QuantizeResize) const ResizeDescriptor& resizeDescriptor, const char* name = nullptr) override { + boost::ignore_unused(resizeDescriptor, name); CheckForwardedQuantizationSettings(layer); } }; @@ -2114,6 +2152,7 @@ BOOST_AUTO_TEST_CASE(QuantizeStridedSlice) const StridedSliceDescriptor& desc, const char* name = nullptr) { + boost::ignore_unused(desc, name); CheckForwardedQuantizationSettings(layer); } }; @@ -2163,6 +2202,7 @@ BOOST_AUTO_TEST_CASE(QuantizeBatchToSpace) const BatchToSpaceNdDescriptor& batchToSpaceNdDescriptor, const char* name = nullptr) override { + boost::ignore_unused(batchToSpaceNdDescriptor, name); CheckForwardedQuantizationSettings(layer); } }; @@ -2219,6 +2259,7 @@ BOOST_AUTO_TEST_CASE(QuantizePrelu) LayerBindingId id, const char* name = nullptr) override { + boost::ignore_unused(id, name); const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo(); switch (id) @@ -2244,6 +2285,7 @@ BOOST_AUTO_TEST_CASE(QuantizePrelu) LayerBindingId id, const char* name = nullptr) override { + boost::ignore_unused(id, name); const TensorInfo& info = layer->GetInputSlot(0).GetConnection()->GetTensorInfo(); BOOST_TEST(m_OutputShape == info.GetShape()); } @@ -2251,6 +2293,7 @@ BOOST_AUTO_TEST_CASE(QuantizePrelu) void VisitPreluLayer(const IConnectableLayer* layer, const char* name = nullptr) override { + boost::ignore_unused(name); const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo(); TestQuantizationParams(info, { 30.0f / g_Asymm8QuantizationBase, 128 }, // QASymm8 @@ -2327,6 +2370,7 @@ void TestQuantizeTransposeConvolution2d(bool useBiases) const Optional& biases, const char *name = nullptr) override { + boost::ignore_unused(descriptor, name); TestQuantizationOnLayersWithBiases(layer, weights, biases); } }; @@ -2406,16 +2450,21 @@ BOOST_AUTO_TEST_CASE(QuantizeStack) void VisitInputLayer(const IConnectableLayer* layer, LayerBindingId id, const char* name = nullptr) override - {} + { + boost::ignore_unused(layer, id, name); + } void VisitOutputLayer(const IConnectableLayer* layer, LayerBindingId id, const char* name = nullptr) override - {} + { + boost::ignore_unused(layer, id, name); + } void VisitStackLayer(const IConnectableLayer* layer, const StackDescriptor& descriptor, const char* name = nullptr) override { + boost::ignore_unused(descriptor, name); TensorInfo outputInfo = layer->GetOutputSlot(0).GetTensorInfo(); TestQuantizationParams(outputInfo, @@ -2476,6 +2525,7 @@ BOOST_AUTO_TEST_CASE(QuantizeSlice) const SliceDescriptor& desc, const char* name = nullptr) { + boost::ignore_unused(desc, name); const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo(); const OffsetScalePair qAsymm8Params{ 30.0f / g_Asymm8QuantizationBase, 128 }; @@ -2559,6 +2609,7 @@ public: LayerBindingId id, const char* name = nullptr) override { + boost::ignore_unused(id, name); const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo(); BOOST_TEST(GetDataTypeName(info.GetDataType()) == GetDataTypeName(m_DataType)); BOOST_TEST(m_InputShape == info.GetShape()); @@ -2568,6 +2619,7 @@ public: LayerBindingId id, const char* name = nullptr) override { + boost::ignore_unused(id, name); const TensorInfo& info = layer->GetInputSlot(0).GetConnection()->GetTensorInfo(); BOOST_TEST(GetDataTypeName(info.GetDataType()) == GetDataTypeName(m_DataType)); BOOST_TEST(m_OutputShape == info.GetShape()); @@ -2576,12 +2628,14 @@ public: void VisitQuantizeLayer(const IConnectableLayer* layer, const char* name = nullptr) override { + boost::ignore_unused(layer, name); m_VisitedQuantizeLayer = true; } void VisitDequantizeLayer(const IConnectableLayer* layer, const char* name = nullptr) override { + boost::ignore_unused(layer, name); m_VisitedDequantizeLayer = true; } -- cgit v1.2.1