aboutsummaryrefslogtreecommitdiff
path: root/src/armnn/test/QuantizerTest.cpp
diff options
context:
space:
mode:
Diffstat (limited to 'src/armnn/test/QuantizerTest.cpp')
-rw-r--r--src/armnn/test/QuantizerTest.cpp66
1 files changed, 60 insertions, 6 deletions
diff --git a/src/armnn/test/QuantizerTest.cpp b/src/armnn/test/QuantizerTest.cpp
index 37b3bfa852..e147a84eb6 100644
--- a/src/armnn/test/QuantizerTest.cpp
+++ b/src/armnn/test/QuantizerTest.cpp
@@ -19,6 +19,7 @@
#include "../RangeTracker.hpp"
#include "../../armnnQuantizer/CommandLineProcessor.hpp"
+#include <boost/core/ignore_unused.hpp>
#include <boost/test/unit_test.hpp>
#include <unordered_map>
@@ -55,6 +56,7 @@ public:
LayerBindingId id,
const char* name = nullptr) override
{
+ boost::ignore_unused(id, name);
const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo();
BOOST_TEST(m_InputShape == info.GetShape());
// Based off current default [-15.0f, 15.0f]
@@ -67,6 +69,7 @@ public:
LayerBindingId id,
const char* name = nullptr) override
{
+ boost::ignore_unused(id, name);
const TensorInfo& info = layer->GetInputSlot(0).GetConnection()->GetTensorInfo();
BOOST_TEST(m_OutputShape == info.GetShape());
}
@@ -105,6 +108,7 @@ protected:
const OffsetScalePair& params,
DataType dataType = DataType::QuantisedAsymm8)
{
+ boost::ignore_unused(dataType);
TestQuantizationParamsImpl(info, DataType::QuantisedAsymm8, params.first, params.second);
}
@@ -191,6 +195,7 @@ public:
void VisitAdditionLayer(const IConnectableLayer* layer,
const char* name = nullptr) override
{
+ boost::ignore_unused(name);
TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo();
// Based off default static range [-20.0f, 20.0f]
@@ -253,6 +258,8 @@ public:
const ActivationDescriptor& descriptor,
const char* name = nullptr) override
{
+ boost::ignore_unused(descriptor, name);
+
TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo();
// Based off default static range [0.0f, 15.0f]
@@ -353,6 +360,7 @@ BOOST_AUTO_TEST_CASE(InputOutputLayerDynamicQuant)
LayerBindingId id,
const char* name = nullptr) override
{
+ boost::ignore_unused(id, name);
const TensorInfo& info = layer->GetInputSlot(0).GetConnection()->GetTensorInfo();
BOOST_CHECK_MESSAGE(info.GetDataType() == m_DataType,
std::string(armnn::GetDataTypeName(info.GetDataType()))
@@ -489,6 +497,7 @@ BOOST_AUTO_TEST_CASE(QuantizeBoundedReluActivation)
const ActivationDescriptor& descriptor,
const char* name = nullptr) override
{
+ boost::ignore_unused(descriptor, name);
TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo();
// Based off default static range [0.0f, 3.5f]
@@ -538,6 +547,7 @@ BOOST_AUTO_TEST_CASE(QuantizeTanHActivation)
const ActivationDescriptor& descriptor,
const char* name = nullptr) override
{
+ boost::ignore_unused(descriptor, name);
TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo();
// Based off default static range [-1.0f, 1.0f]
@@ -586,6 +596,7 @@ public:
const ActivationDescriptor& descriptor,
const char* name = nullptr) override
{
+ boost::ignore_unused(descriptor, name);
TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo();
// Based off default static range [-5.0f, 15.0f]
@@ -652,6 +663,7 @@ BOOST_AUTO_TEST_CASE(QuantizeBatchNorm)
const ConstTensor& gamma,
const char* name = nullptr) override
{
+ boost::ignore_unused(desc, name);
TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo();
// Based off default static range [-15.0f, 15.0f]
@@ -730,6 +742,7 @@ BOOST_AUTO_TEST_CASE(QuantizeDepthToSpace)
const DepthToSpaceDescriptor& desc,
const char* name = nullptr)
{
+ boost::ignore_unused(desc, name);
const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo();
const OffsetScalePair qAsymm8Params{ 30.0f / g_Asymm8QuantizationBase, 128 };
@@ -914,6 +927,7 @@ void ValidateFullyConnectedLayer(const bool biasEnabled)
const Optional<ConstTensor>& biases,
const char* name = nullptr) override
{
+ boost::ignore_unused(desc, name);
TestQuantizationOnLayersWithBiases(layer, weights, biases);
}
};
@@ -965,6 +979,7 @@ void TestQuantizeConvolution2d(bool useBiases)
const Optional<ConstTensor>& biases,
const char *name = nullptr) override
{
+ boost::ignore_unused(convolution2dDescriptor, name);
TestQuantizationOnLayersWithBiases(layer, weights, biases);
}
};
@@ -1045,6 +1060,7 @@ void TestQuantizeDepthwiseConvolution2d(bool useBiases)
const Optional<ConstTensor>& biases,
const char *name = nullptr) override
{
+ boost::ignore_unused(convolution2dDescriptor, name);
TestQuantizationOnLayersWithBiases(layer, weights, biases);
}
};
@@ -1123,6 +1139,7 @@ BOOST_AUTO_TEST_CASE(QuantizeInstanceNormalization)
const InstanceNormalizationDescriptor& descriptor,
const char* name = nullptr)
{
+ boost::ignore_unused(descriptor, name);
const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo();
const OffsetScalePair qAsymm8Params{ 30.0f / g_Asymm8QuantizationBase, 128 };
@@ -1183,6 +1200,7 @@ BOOST_AUTO_TEST_CASE(QuantizeLogSoftmax)
const SoftmaxDescriptor& descriptor,
const char* name = nullptr) override
{
+ boost::ignore_unused(descriptor, name);
TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo();
const OffsetScalePair qAsymm8Params{ 30.0f / g_Asymm8QuantizationBase, 128 };
@@ -1267,6 +1285,7 @@ BOOST_AUTO_TEST_CASE(QuantizeSoftmax)
const SoftmaxDescriptor& descriptor,
const char* name = nullptr) override
{
+ boost::ignore_unused(descriptor, name);
TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo();
// Based off default static range [0.0f, 1.0f]
@@ -1388,6 +1407,7 @@ BOOST_AUTO_TEST_CASE(QuantizePermute)
const PermuteDescriptor& desc,
const char* name = nullptr) override
{
+ boost::ignore_unused(desc, name);
CheckForwardedQuantizationSettings(layer);
}
};
@@ -1437,6 +1457,7 @@ BOOST_AUTO_TEST_CASE(QuantizeSpaceToBatch)
const SpaceToBatchNdDescriptor& spaceToBatchNdDescriptor,
const char* name = nullptr) override
{
+ boost::ignore_unused(spaceToBatchNdDescriptor, name);
CheckForwardedQuantizationSettings(layer);
}
};
@@ -1538,6 +1559,7 @@ BOOST_AUTO_TEST_CASE(QuantizePooling2d)
const Pooling2dDescriptor& desc,
const char* name = nullptr) override
{
+ boost::ignore_unused(desc, name);
CheckForwardedQuantizationSettings(layer);
}
};
@@ -1601,6 +1623,7 @@ BOOST_AUTO_TEST_CASE(QuantizeConstant)
const ConstTensor& input,
const char* name = nullptr) override
{
+ boost::ignore_unused(input, name);
TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo();
// Based off the range of values in the const tensor used for the test: [-2.0f, 6.0f]
@@ -1667,6 +1690,7 @@ BOOST_AUTO_TEST_CASE(QuantizeAbs)
void VisitAbsLayer(const IConnectableLayer *layer,
const char *name = nullptr) override
{
+ boost::ignore_unused(name);
TensorInfo outputInfo = layer->GetOutputSlot(0).GetTensorInfo();
TestQuantizationParams(outputInfo,
@@ -1720,16 +1744,21 @@ BOOST_AUTO_TEST_CASE(QuantizeArgMinMax)
void VisitInputLayer(const IConnectableLayer* layer,
LayerBindingId id,
const char* name = nullptr) override
- {}
+ {
+ boost::ignore_unused(layer, id, name);
+ }
void VisitOutputLayer(const IConnectableLayer* layer,
LayerBindingId id,
const char* name = nullptr) override
- {}
+ {
+ boost::ignore_unused(layer, id, name);
+ }
void VisitArgMinMaxLayer(const IConnectableLayer* layer,
const ArgMinMaxDescriptor& argMinMaxDescriptor,
const char* name = nullptr) override
{
+ boost::ignore_unused(argMinMaxDescriptor, name);
TensorInfo outputInfo = layer->GetOutputSlot(0).GetTensorInfo();
TestQuantizationParams(outputInfo,
@@ -1798,6 +1827,7 @@ BOOST_AUTO_TEST_CASE(QuantizeComparison)
const ComparisonDescriptor& descriptor,
const char* name = nullptr) override
{
+ boost::ignore_unused(descriptor, name);
TensorInfo info = layer->GetOutputSlot(0).GetTensorInfo();
const OffsetScalePair qAsymm8Params{ 30.0f / g_Asymm8QuantizationBase, 128 };
@@ -1860,15 +1890,20 @@ BOOST_AUTO_TEST_CASE(QuantizeConcat)
void VisitInputLayer(const IConnectableLayer* layer,
LayerBindingId id,
const char* name = nullptr) override
- {}
+ {
+ boost::ignore_unused(layer, id, name);
+ }
void VisitOutputLayer(const IConnectableLayer* layer,
LayerBindingId id,
const char* name = nullptr) override
- {}
+ {
+ boost::ignore_unused(layer, id, name);
+ }
void VisitConcatLayer(const IConnectableLayer* layer,
const OriginsDescriptor& originsDescriptor,
const char* name = nullptr) override
{
+ boost::ignore_unused(originsDescriptor, name);
TensorInfo outputInfo = layer->GetOutputSlot(0).GetTensorInfo();
TestQuantizationParams(
@@ -1964,6 +1999,7 @@ BOOST_AUTO_TEST_CASE(QuantizeReshape)
const ReshapeDescriptor& reshapeDescriptor,
const char* name = nullptr) override
{
+ boost::ignore_unused(reshapeDescriptor, name);
CheckForwardedQuantizationSettings(layer);
}
};
@@ -2013,6 +2049,7 @@ BOOST_AUTO_TEST_CASE(QuantizeSplitter)
const SplitterDescriptor& desc,
const char* name = nullptr)
{
+ boost::ignore_unused(desc, name);
CheckForwardedQuantizationSettings(layer);
}
};
@@ -2063,6 +2100,7 @@ BOOST_AUTO_TEST_CASE(QuantizeResize)
const ResizeDescriptor& resizeDescriptor,
const char* name = nullptr) override
{
+ boost::ignore_unused(resizeDescriptor, name);
CheckForwardedQuantizationSettings(layer);
}
};
@@ -2114,6 +2152,7 @@ BOOST_AUTO_TEST_CASE(QuantizeStridedSlice)
const StridedSliceDescriptor& desc,
const char* name = nullptr)
{
+ boost::ignore_unused(desc, name);
CheckForwardedQuantizationSettings(layer);
}
};
@@ -2163,6 +2202,7 @@ BOOST_AUTO_TEST_CASE(QuantizeBatchToSpace)
const BatchToSpaceNdDescriptor& batchToSpaceNdDescriptor,
const char* name = nullptr) override
{
+ boost::ignore_unused(batchToSpaceNdDescriptor, name);
CheckForwardedQuantizationSettings(layer);
}
};
@@ -2219,6 +2259,7 @@ BOOST_AUTO_TEST_CASE(QuantizePrelu)
LayerBindingId id,
const char* name = nullptr) override
{
+ boost::ignore_unused(id, name);
const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo();
switch (id)
@@ -2244,6 +2285,7 @@ BOOST_AUTO_TEST_CASE(QuantizePrelu)
LayerBindingId id,
const char* name = nullptr) override
{
+ boost::ignore_unused(id, name);
const TensorInfo& info = layer->GetInputSlot(0).GetConnection()->GetTensorInfo();
BOOST_TEST(m_OutputShape == info.GetShape());
}
@@ -2251,6 +2293,7 @@ BOOST_AUTO_TEST_CASE(QuantizePrelu)
void VisitPreluLayer(const IConnectableLayer* layer,
const char* name = nullptr) override
{
+ boost::ignore_unused(name);
const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo();
TestQuantizationParams(info,
{ 30.0f / g_Asymm8QuantizationBase, 128 }, // QASymm8
@@ -2327,6 +2370,7 @@ void TestQuantizeTransposeConvolution2d(bool useBiases)
const Optional<ConstTensor>& biases,
const char *name = nullptr) override
{
+ boost::ignore_unused(descriptor, name);
TestQuantizationOnLayersWithBiases(layer, weights, biases);
}
};
@@ -2406,16 +2450,21 @@ BOOST_AUTO_TEST_CASE(QuantizeStack)
void VisitInputLayer(const IConnectableLayer* layer,
LayerBindingId id,
const char* name = nullptr) override
- {}
+ {
+ boost::ignore_unused(layer, id, name);
+ }
void VisitOutputLayer(const IConnectableLayer* layer,
LayerBindingId id,
const char* name = nullptr) override
- {}
+ {
+ boost::ignore_unused(layer, id, name);
+ }
void VisitStackLayer(const IConnectableLayer* layer,
const StackDescriptor& descriptor,
const char* name = nullptr) override
{
+ boost::ignore_unused(descriptor, name);
TensorInfo outputInfo = layer->GetOutputSlot(0).GetTensorInfo();
TestQuantizationParams(outputInfo,
@@ -2476,6 +2525,7 @@ BOOST_AUTO_TEST_CASE(QuantizeSlice)
const SliceDescriptor& desc,
const char* name = nullptr)
{
+ boost::ignore_unused(desc, name);
const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo();
const OffsetScalePair qAsymm8Params{ 30.0f / g_Asymm8QuantizationBase, 128 };
@@ -2559,6 +2609,7 @@ public:
LayerBindingId id,
const char* name = nullptr) override
{
+ boost::ignore_unused(id, name);
const TensorInfo& info = layer->GetOutputSlot(0).GetTensorInfo();
BOOST_TEST(GetDataTypeName(info.GetDataType()) == GetDataTypeName(m_DataType));
BOOST_TEST(m_InputShape == info.GetShape());
@@ -2568,6 +2619,7 @@ public:
LayerBindingId id,
const char* name = nullptr) override
{
+ boost::ignore_unused(id, name);
const TensorInfo& info = layer->GetInputSlot(0).GetConnection()->GetTensorInfo();
BOOST_TEST(GetDataTypeName(info.GetDataType()) == GetDataTypeName(m_DataType));
BOOST_TEST(m_OutputShape == info.GetShape());
@@ -2576,12 +2628,14 @@ public:
void VisitQuantizeLayer(const IConnectableLayer* layer,
const char* name = nullptr) override
{
+ boost::ignore_unused(layer, name);
m_VisitedQuantizeLayer = true;
}
void VisitDequantizeLayer(const IConnectableLayer* layer,
const char* name = nullptr) override
{
+ boost::ignore_unused(layer, name);
m_VisitedDequantizeLayer = true;
}