diff options
Diffstat (limited to 'src/armnn/test/QuantizerTest.cpp')
-rw-r--r-- | src/armnn/test/QuantizerTest.cpp | 66 |
1 files changed, 60 insertions, 6 deletions
diff --git a/src/armnn/test/QuantizerTest.cpp b/src/armnn/test/QuantizerTest.cpp index 37b3bfa852..e147a84eb6 100644 --- a/src/armnn/test/QuantizerTest.cpp +++ b/src/armnn/test/QuantizerTest.cpp @@ -19,6 +19,7 @@ #include "../RangeTracker.hpp" #include "../../armnnQuantizer/CommandLineProcessor.hpp" +#include <boost/core/ignore_unused.hpp> #include <boost/test/unit_test.hpp> #include <unordered_map> @@ -55,6 +56,7 @@ public: LayerBindingId id, const char* name = nullptr) override { + boost::ignore_unused(id, name); const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo(); BOOST_TEST(m_InputShape == info.GetShape()); // Based off current default [-15.0f, 15.0f] @@ -67,6 +69,7 @@ public: LayerBindingId id, const char* name = nullptr) override { + boost::ignore_unused(id, name); const TensorInfo& info = layer->GetInputSlot(0).GetConnection()->GetTensorInfo(); BOOST_TEST(m_OutputShape == info.GetShape()); } @@ -105,6 +108,7 @@ protected: const OffsetScalePair& params, DataType dataType = DataType::QuantisedAsymm8) { + boost::ignore_unused(dataType); TestQuantizationParamsImpl(info, DataType::QuantisedAsymm8, params.first, params.second); } @@ -191,6 +195,7 @@ public: void VisitAdditionLayer(const IConnectableLayer* layer, const char* name = nullptr) override { + boost::ignore_unused(name); TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); // Based off default static range [-20.0f, 20.0f] @@ -253,6 +258,8 @@ public: const ActivationDescriptor& descriptor, const char* name = nullptr) override { + boost::ignore_unused(descriptor, name); + TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); // Based off default static range [0.0f, 15.0f] @@ -353,6 +360,7 @@ BOOST_AUTO_TEST_CASE(InputOutputLayerDynamicQuant) LayerBindingId id, const char* name = nullptr) override { + boost::ignore_unused(id, name); const TensorInfo& info = layer->GetInputSlot(0).GetConnection()->GetTensorInfo(); BOOST_CHECK_MESSAGE(info.GetDataType() == m_DataType, std::string(armnn::GetDataTypeName(info.GetDataType())) @@ -489,6 +497,7 @@ BOOST_AUTO_TEST_CASE(QuantizeBoundedReluActivation) const ActivationDescriptor& descriptor, const char* name = nullptr) override { + boost::ignore_unused(descriptor, name); TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); // Based off default static range [0.0f, 3.5f] @@ -538,6 +547,7 @@ BOOST_AUTO_TEST_CASE(QuantizeTanHActivation) const ActivationDescriptor& descriptor, const char* name = nullptr) override { + boost::ignore_unused(descriptor, name); TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); // Based off default static range [-1.0f, 1.0f] @@ -586,6 +596,7 @@ public: const ActivationDescriptor& descriptor, const char* name = nullptr) override { + boost::ignore_unused(descriptor, name); TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); // Based off default static range [-5.0f, 15.0f] @@ -652,6 +663,7 @@ BOOST_AUTO_TEST_CASE(QuantizeBatchNorm) const ConstTensor& gamma, const char* name = nullptr) override { + boost::ignore_unused(desc, name); TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); // Based off default static range [-15.0f, 15.0f] @@ -730,6 +742,7 @@ BOOST_AUTO_TEST_CASE(QuantizeDepthToSpace) const DepthToSpaceDescriptor& desc, const char* name = nullptr) { + boost::ignore_unused(desc, name); const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo(); const OffsetScalePair qAsymm8Params{ 30.0f / g_Asymm8QuantizationBase, 128 }; @@ -914,6 +927,7 @@ void ValidateFullyConnectedLayer(const bool biasEnabled) const Optional<ConstTensor>& biases, const char* name = nullptr) override { + boost::ignore_unused(desc, name); TestQuantizationOnLayersWithBiases(layer, weights, biases); } }; @@ -965,6 +979,7 @@ void TestQuantizeConvolution2d(bool useBiases) const Optional<ConstTensor>& biases, const char *name = nullptr) override { + boost::ignore_unused(convolution2dDescriptor, name); TestQuantizationOnLayersWithBiases(layer, weights, biases); } }; @@ -1045,6 +1060,7 @@ void TestQuantizeDepthwiseConvolution2d(bool useBiases) const Optional<ConstTensor>& biases, const char *name = nullptr) override { + boost::ignore_unused(convolution2dDescriptor, name); TestQuantizationOnLayersWithBiases(layer, weights, biases); } }; @@ -1123,6 +1139,7 @@ BOOST_AUTO_TEST_CASE(QuantizeInstanceNormalization) const InstanceNormalizationDescriptor& descriptor, const char* name = nullptr) { + boost::ignore_unused(descriptor, name); const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo(); const OffsetScalePair qAsymm8Params{ 30.0f / g_Asymm8QuantizationBase, 128 }; @@ -1183,6 +1200,7 @@ BOOST_AUTO_TEST_CASE(QuantizeLogSoftmax) const SoftmaxDescriptor& descriptor, const char* name = nullptr) override { + boost::ignore_unused(descriptor, name); TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); const OffsetScalePair qAsymm8Params{ 30.0f / g_Asymm8QuantizationBase, 128 }; @@ -1267,6 +1285,7 @@ BOOST_AUTO_TEST_CASE(QuantizeSoftmax) const SoftmaxDescriptor& descriptor, const char* name = nullptr) override { + boost::ignore_unused(descriptor, name); TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); // Based off default static range [0.0f, 1.0f] @@ -1388,6 +1407,7 @@ BOOST_AUTO_TEST_CASE(QuantizePermute) const PermuteDescriptor& desc, const char* name = nullptr) override { + boost::ignore_unused(desc, name); CheckForwardedQuantizationSettings(layer); } }; @@ -1437,6 +1457,7 @@ BOOST_AUTO_TEST_CASE(QuantizeSpaceToBatch) const SpaceToBatchNdDescriptor& spaceToBatchNdDescriptor, const char* name = nullptr) override { + boost::ignore_unused(spaceToBatchNdDescriptor, name); CheckForwardedQuantizationSettings(layer); } }; @@ -1538,6 +1559,7 @@ BOOST_AUTO_TEST_CASE(QuantizePooling2d) const Pooling2dDescriptor& desc, const char* name = nullptr) override { + boost::ignore_unused(desc, name); CheckForwardedQuantizationSettings(layer); } }; @@ -1601,6 +1623,7 @@ BOOST_AUTO_TEST_CASE(QuantizeConstant) const ConstTensor& input, const char* name = nullptr) override { + boost::ignore_unused(input, name); TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); // Based off the range of values in the const tensor used for the test: [-2.0f, 6.0f] @@ -1667,6 +1690,7 @@ BOOST_AUTO_TEST_CASE(QuantizeAbs) void VisitAbsLayer(const IConnectableLayer *layer, const char *name = nullptr) override { + boost::ignore_unused(name); TensorInfo outputInfo = layer->GetOutputSlot(0).GetTensorInfo(); TestQuantizationParams(outputInfo, @@ -1720,16 +1744,21 @@ BOOST_AUTO_TEST_CASE(QuantizeArgMinMax) void VisitInputLayer(const IConnectableLayer* layer, LayerBindingId id, const char* name = nullptr) override - {} + { + boost::ignore_unused(layer, id, name); + } void VisitOutputLayer(const IConnectableLayer* layer, LayerBindingId id, const char* name = nullptr) override - {} + { + boost::ignore_unused(layer, id, name); + } void VisitArgMinMaxLayer(const IConnectableLayer* layer, const ArgMinMaxDescriptor& argMinMaxDescriptor, const char* name = nullptr) override { + boost::ignore_unused(argMinMaxDescriptor, name); TensorInfo outputInfo = layer->GetOutputSlot(0).GetTensorInfo(); TestQuantizationParams(outputInfo, @@ -1798,6 +1827,7 @@ BOOST_AUTO_TEST_CASE(QuantizeComparison) const ComparisonDescriptor& descriptor, const char* name = nullptr) override { + boost::ignore_unused(descriptor, name); TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo(); const OffsetScalePair qAsymm8Params{ 30.0f / g_Asymm8QuantizationBase, 128 }; @@ -1860,15 +1890,20 @@ BOOST_AUTO_TEST_CASE(QuantizeConcat) void VisitInputLayer(const IConnectableLayer* layer, LayerBindingId id, const char* name = nullptr) override - {} + { + boost::ignore_unused(layer, id, name); + } void VisitOutputLayer(const IConnectableLayer* layer, LayerBindingId id, const char* name = nullptr) override - {} + { + boost::ignore_unused(layer, id, name); + } void VisitConcatLayer(const IConnectableLayer* layer, const OriginsDescriptor& originsDescriptor, const char* name = nullptr) override { + boost::ignore_unused(originsDescriptor, name); TensorInfo outputInfo = layer->GetOutputSlot(0).GetTensorInfo(); TestQuantizationParams( @@ -1964,6 +1999,7 @@ BOOST_AUTO_TEST_CASE(QuantizeReshape) const ReshapeDescriptor& reshapeDescriptor, const char* name = nullptr) override { + boost::ignore_unused(reshapeDescriptor, name); CheckForwardedQuantizationSettings(layer); } }; @@ -2013,6 +2049,7 @@ BOOST_AUTO_TEST_CASE(QuantizeSplitter) const SplitterDescriptor& desc, const char* name = nullptr) { + boost::ignore_unused(desc, name); CheckForwardedQuantizationSettings(layer); } }; @@ -2063,6 +2100,7 @@ BOOST_AUTO_TEST_CASE(QuantizeResize) const ResizeDescriptor& resizeDescriptor, const char* name = nullptr) override { + boost::ignore_unused(resizeDescriptor, name); CheckForwardedQuantizationSettings(layer); } }; @@ -2114,6 +2152,7 @@ BOOST_AUTO_TEST_CASE(QuantizeStridedSlice) const StridedSliceDescriptor& desc, const char* name = nullptr) { + boost::ignore_unused(desc, name); CheckForwardedQuantizationSettings(layer); } }; @@ -2163,6 +2202,7 @@ BOOST_AUTO_TEST_CASE(QuantizeBatchToSpace) const BatchToSpaceNdDescriptor& batchToSpaceNdDescriptor, const char* name = nullptr) override { + boost::ignore_unused(batchToSpaceNdDescriptor, name); CheckForwardedQuantizationSettings(layer); } }; @@ -2219,6 +2259,7 @@ BOOST_AUTO_TEST_CASE(QuantizePrelu) LayerBindingId id, const char* name = nullptr) override { + boost::ignore_unused(id, name); const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo(); switch (id) @@ -2244,6 +2285,7 @@ BOOST_AUTO_TEST_CASE(QuantizePrelu) LayerBindingId id, const char* name = nullptr) override { + boost::ignore_unused(id, name); const TensorInfo& info = layer->GetInputSlot(0).GetConnection()->GetTensorInfo(); BOOST_TEST(m_OutputShape == info.GetShape()); } @@ -2251,6 +2293,7 @@ BOOST_AUTO_TEST_CASE(QuantizePrelu) void VisitPreluLayer(const IConnectableLayer* layer, const char* name = nullptr) override { + boost::ignore_unused(name); const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo(); TestQuantizationParams(info, { 30.0f / g_Asymm8QuantizationBase, 128 }, // QASymm8 @@ -2327,6 +2370,7 @@ void TestQuantizeTransposeConvolution2d(bool useBiases) const Optional<ConstTensor>& biases, const char *name = nullptr) override { + boost::ignore_unused(descriptor, name); TestQuantizationOnLayersWithBiases(layer, weights, biases); } }; @@ -2406,16 +2450,21 @@ BOOST_AUTO_TEST_CASE(QuantizeStack) void VisitInputLayer(const IConnectableLayer* layer, LayerBindingId id, const char* name = nullptr) override - {} + { + boost::ignore_unused(layer, id, name); + } void VisitOutputLayer(const IConnectableLayer* layer, LayerBindingId id, const char* name = nullptr) override - {} + { + boost::ignore_unused(layer, id, name); + } void VisitStackLayer(const IConnectableLayer* layer, const StackDescriptor& descriptor, const char* name = nullptr) override { + boost::ignore_unused(descriptor, name); TensorInfo outputInfo = layer->GetOutputSlot(0).GetTensorInfo(); TestQuantizationParams(outputInfo, @@ -2476,6 +2525,7 @@ BOOST_AUTO_TEST_CASE(QuantizeSlice) const SliceDescriptor& desc, const char* name = nullptr) { + boost::ignore_unused(desc, name); const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo(); const OffsetScalePair qAsymm8Params{ 30.0f / g_Asymm8QuantizationBase, 128 }; @@ -2559,6 +2609,7 @@ public: LayerBindingId id, const char* name = nullptr) override { + boost::ignore_unused(id, name); const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo(); BOOST_TEST(GetDataTypeName(info.GetDataType()) == GetDataTypeName(m_DataType)); BOOST_TEST(m_InputShape == info.GetShape()); @@ -2568,6 +2619,7 @@ public: LayerBindingId id, const char* name = nullptr) override { + boost::ignore_unused(id, name); const TensorInfo& info = layer->GetInputSlot(0).GetConnection()->GetTensorInfo(); BOOST_TEST(GetDataTypeName(info.GetDataType()) == GetDataTypeName(m_DataType)); BOOST_TEST(m_OutputShape == info.GetShape()); @@ -2576,12 +2628,14 @@ public: void VisitQuantizeLayer(const IConnectableLayer* layer, const char* name = nullptr) override { + boost::ignore_unused(layer, name); m_VisitedQuantizeLayer = true; } void VisitDequantizeLayer(const IConnectableLayer* layer, const char* name = nullptr) override { + boost::ignore_unused(layer, name); m_VisitedDequantizeLayer = true; } |