aboutsummaryrefslogtreecommitdiff
path: root/src/armnnTfLiteParser
diff options
context:
space:
mode:
authorDerek Lamberti <derek.lamberti@arm.com>2020-01-10 17:14:08 +0000
committerKevin May <kevin.may@arm.com>2020-01-13 18:18:12 +0000
commitf90c56d72de4848a2dc5844a97458aaf09df07c2 (patch)
tree71f1c6f16a4687286614f5526ed70938a611b27d /src/armnnTfLiteParser
parent842e0dbd40114e19bf26916fefe06c869dbe416d (diff)
downloadarmnn-f90c56d72de4848a2dc5844a97458aaf09df07c2.tar.gz
Rename quantized data types to remove ambiguity for signed/unsigned payloads
!android-nn-driver:2572 Change-Id: I8fe52ceb09987b3d05c539409510f535165455cc Signed-off-by: Derek Lamberti <derek.lamberti@arm.com>
Diffstat (limited to 'src/armnnTfLiteParser')
-rw-r--r--src/armnnTfLiteParser/TfLiteParser.cpp6
-rw-r--r--src/armnnTfLiteParser/test/Addition.cpp2
-rw-r--r--src/armnnTfLiteParser/test/AvgPool2D.cpp6
-rw-r--r--src/armnnTfLiteParser/test/Concatenation.cpp16
-rw-r--r--src/armnnTfLiteParser/test/Constant.cpp2
-rw-r--r--src/armnnTfLiteParser/test/Conv2D.cpp8
-rw-r--r--src/armnnTfLiteParser/test/DepthwiseConvolution2D.cpp6
-rw-r--r--src/armnnTfLiteParser/test/Dequantize.cpp4
-rw-r--r--src/armnnTfLiteParser/test/DetectionPostProcess.cpp6
-rw-r--r--src/armnnTfLiteParser/test/FullyConnected.cpp6
-rw-r--r--src/armnnTfLiteParser/test/MaxPool2D.cpp6
-rw-r--r--src/armnnTfLiteParser/test/Reshape.cpp8
-rw-r--r--src/armnnTfLiteParser/test/Softmax.cpp2
-rw-r--r--src/armnnTfLiteParser/test/Split.cpp8
-rw-r--r--src/armnnTfLiteParser/test/Squeeze.cpp4
-rw-r--r--src/armnnTfLiteParser/test/Sub.cpp2
-rw-r--r--src/armnnTfLiteParser/test/TransposeConv.cpp2
-rw-r--r--src/armnnTfLiteParser/test/Unpack.cpp4
-rw-r--r--src/armnnTfLiteParser/test/Unsupported.cpp2
19 files changed, 50 insertions, 50 deletions
diff --git a/src/armnnTfLiteParser/TfLiteParser.cpp b/src/armnnTfLiteParser/TfLiteParser.cpp
index 9c7dda8aec..22d65645a3 100644
--- a/src/armnnTfLiteParser/TfLiteParser.cpp
+++ b/src/armnnTfLiteParser/TfLiteParser.cpp
@@ -309,7 +309,7 @@ armnn::TensorInfo ToTensorInfo(TfLiteParser::TensorRawPtr tensorPtr, const std::
switch (tensorPtr->type)
{
case tflite::TensorType_UINT8:
- type = armnn::DataType::QuantisedAsymm8;
+ type = armnn::DataType::QAsymmU8;
break;
case tflite::TensorType_FLOAT32:
type = armnn::DataType::Float32;
@@ -318,7 +318,7 @@ armnn::TensorInfo ToTensorInfo(TfLiteParser::TensorRawPtr tensorPtr, const std::
type = armnn::DataType::QSymmS8;
break;
case tflite::TensorType_INT16:
- type = armnn::DataType::QuantisedSymm16;
+ type = armnn::DataType::QSymmS16;
break;
case tflite::TensorType_INT32:
type = armnn::DataType::Signed32;
@@ -2818,7 +2818,7 @@ TfLiteParser::CreateConstTensor(TensorRawPtr tensorPtr,
tensorPtr,
tensorInfo,
permutationVector);
- case armnn::DataType::QuantisedAsymm8:
+ case armnn::DataType::QAsymmU8:
return CreateConstTensorAndStoreData<uint8_t>(bufferPtr,
tensorPtr,
tensorInfo,
diff --git a/src/armnnTfLiteParser/test/Addition.cpp b/src/armnnTfLiteParser/test/Addition.cpp
index 94389d3134..deeb707a2f 100644
--- a/src/armnnTfLiteParser/test/Addition.cpp
+++ b/src/armnnTfLiteParser/test/Addition.cpp
@@ -97,7 +97,7 @@ struct SimpleAddFixture : AddFixture
BOOST_FIXTURE_TEST_CASE(SimpleAdd, SimpleAddFixture)
{
- RunTest<2, armnn::DataType::QuantisedAsymm8>(
+ RunTest<2, armnn::DataType::QAsymmU8>(
0,
{{"inputTensor1", { 0, 1, 2, 3 }},
{"inputTensor2", { 4, 5, 6, 7 }}},
diff --git a/src/armnnTfLiteParser/test/AvgPool2D.cpp b/src/armnnTfLiteParser/test/AvgPool2D.cpp
index a39c088d44..a56e7e7362 100644
--- a/src/armnnTfLiteParser/test/AvgPool2D.cpp
+++ b/src/armnnTfLiteParser/test/AvgPool2D.cpp
@@ -98,7 +98,7 @@ struct AvgPoolLiteFixture2DOutput : AvgPool2DFixture
BOOST_FIXTURE_TEST_CASE(AvgPoolLite1DOutput, AvgPoolLiteFixtureUint1DOutput)
{
- RunTest<4, armnn::DataType::QuantisedAsymm8>(0, {2, 3, 5, 2 }, { 3 });
+ RunTest<4, armnn::DataType::QAsymmU8>(0, {2, 3, 5, 2 }, { 3 });
}
BOOST_FIXTURE_TEST_CASE(AvgPoolLiteFloat1DOutput, AvgPoolLiteFixtureFloat1DOutput)
@@ -108,13 +108,13 @@ BOOST_FIXTURE_TEST_CASE(AvgPoolLiteFloat1DOutput, AvgPoolLiteFixtureFloat1DOutpu
BOOST_FIXTURE_TEST_CASE(AvgPoolLite2DOutput, AvgPoolLiteFixture2DOutput)
{
- RunTest<4, armnn::DataType::QuantisedAsymm8>(
+ RunTest<4, armnn::DataType::QAsymmU8>(
0, { 1, 2, 2, 3, 5, 6, 7, 8, 3, 2, 1, 0, 1, 2, 3, 4 }, { 4, 5, 2, 2 });
}
BOOST_FIXTURE_TEST_CASE(IncorrectDataTypeError, AvgPoolLiteFixtureFloat1DOutput)
{
- BOOST_CHECK_THROW((RunTest<4, armnn::DataType::QuantisedAsymm8>(0, {2, 3, 5, 2 }, { 3 })), armnn::Exception);
+ BOOST_CHECK_THROW((RunTest<4, armnn::DataType::QAsymmU8>(0, {2, 3, 5, 2 }, { 3 })), armnn::Exception);
}
BOOST_AUTO_TEST_SUITE_END()
diff --git a/src/armnnTfLiteParser/test/Concatenation.cpp b/src/armnnTfLiteParser/test/Concatenation.cpp
index d3d571f174..8e31a3edb0 100644
--- a/src/armnnTfLiteParser/test/Concatenation.cpp
+++ b/src/armnnTfLiteParser/test/Concatenation.cpp
@@ -100,7 +100,7 @@ struct ConcatenationFixtureNegativeDim : ConcatenationFixture
BOOST_FIXTURE_TEST_CASE(ParseConcatenationNegativeDim, ConcatenationFixtureNegativeDim)
{
- RunTest<4, armnn::DataType::QuantisedAsymm8>(
+ RunTest<4, armnn::DataType::QAsymmU8>(
0,
{{"inputTensor1", { 0, 1, 2, 3 }},
{"inputTensor2", { 4, 5, 6, 7 }}},
@@ -114,7 +114,7 @@ struct ConcatenationFixtureNCHW : ConcatenationFixture
BOOST_FIXTURE_TEST_CASE(ParseConcatenationNCHW, ConcatenationFixtureNCHW)
{
- RunTest<4, armnn::DataType::QuantisedAsymm8>(
+ RunTest<4, armnn::DataType::QAsymmU8>(
0,
{{"inputTensor1", { 0, 1, 2, 3 }},
{"inputTensor2", { 4, 5, 6, 7 }}},
@@ -128,7 +128,7 @@ struct ConcatenationFixtureNHWC : ConcatenationFixture
BOOST_FIXTURE_TEST_CASE(ParseConcatenationNHWC, ConcatenationFixtureNHWC)
{
- RunTest<4, armnn::DataType::QuantisedAsymm8>(
+ RunTest<4, armnn::DataType::QAsymmU8>(
0,
{{"inputTensor1", { 0, 1, 2, 3 }},
{"inputTensor2", { 4, 5, 6, 7 }}},
@@ -142,7 +142,7 @@ struct ConcatenationFixtureDim1 : ConcatenationFixture
BOOST_FIXTURE_TEST_CASE(ParseConcatenationDim1, ConcatenationFixtureDim1)
{
- RunTest<4, armnn::DataType::QuantisedAsymm8>(
+ RunTest<4, armnn::DataType::QAsymmU8>(
0,
{ { "inputTensor1", { 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11,
12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23 } },
@@ -161,7 +161,7 @@ struct ConcatenationFixtureDim3 : ConcatenationFixture
BOOST_FIXTURE_TEST_CASE(ParseConcatenationDim3, ConcatenationFixtureDim3)
{
- RunTest<4, armnn::DataType::QuantisedAsymm8>(
+ RunTest<4, armnn::DataType::QAsymmU8>(
0,
{ { "inputTensor1", { 0, 1, 2, 3,
4, 5, 6, 7,
@@ -196,7 +196,7 @@ struct ConcatenationFixture3DDim0 : ConcatenationFixture
BOOST_FIXTURE_TEST_CASE(ParseConcatenation3DDim0, ConcatenationFixture3DDim0)
{
- RunTest<3, armnn::DataType::QuantisedAsymm8>(
+ RunTest<3, armnn::DataType::QAsymmU8>(
0,
{ { "inputTensor1", { 0, 1, 2, 3, 4, 5 } },
{ "inputTensor2", { 6, 7, 8, 9, 10, 11,
@@ -213,7 +213,7 @@ struct ConcatenationFixture3DDim1 : ConcatenationFixture
BOOST_FIXTURE_TEST_CASE(ParseConcatenation3DDim1, ConcatenationFixture3DDim1)
{
- RunTest<3, armnn::DataType::QuantisedAsymm8>(
+ RunTest<3, armnn::DataType::QAsymmU8>(
0,
{ { "inputTensor1", { 0, 1, 2, 3, 4, 5 } },
{ "inputTensor2", { 6, 7, 8, 9, 10, 11,
@@ -230,7 +230,7 @@ struct ConcatenationFixture3DDim2 : ConcatenationFixture
BOOST_FIXTURE_TEST_CASE(ParseConcatenation3DDim2, ConcatenationFixture3DDim2)
{
- RunTest<3, armnn::DataType::QuantisedAsymm8>(
+ RunTest<3, armnn::DataType::QAsymmU8>(
0,
{ { "inputTensor1", { 0, 1, 2,
3, 4, 5 } },
diff --git a/src/armnnTfLiteParser/test/Constant.cpp b/src/armnnTfLiteParser/test/Constant.cpp
index 356e1b769d..cc89223469 100644
--- a/src/armnnTfLiteParser/test/Constant.cpp
+++ b/src/armnnTfLiteParser/test/Constant.cpp
@@ -103,7 +103,7 @@ struct SimpleConstantAddFixture : ConstantAddFixture
BOOST_FIXTURE_TEST_CASE(SimpleConstantAdd, SimpleConstantAddFixture)
{
- RunTest<2, armnn::DataType::QuantisedAsymm8>(
+ RunTest<2, armnn::DataType::QAsymmU8>(
0,
{{"InputTensor", { 0, 1, 2, 3 }}},
{{"OutputTensor", { 4, 6, 8, 10 }}}
diff --git a/src/armnnTfLiteParser/test/Conv2D.cpp b/src/armnnTfLiteParser/test/Conv2D.cpp
index 38c6675ddb..2eae5f5a1a 100644
--- a/src/armnnTfLiteParser/test/Conv2D.cpp
+++ b/src/armnnTfLiteParser/test/Conv2D.cpp
@@ -89,7 +89,7 @@ struct SimpleConv2DFixture : public ParserFlatbuffersFixture
BOOST_FIXTURE_TEST_CASE( ParseSimpleConv2D, SimpleConv2DFixture )
{
- RunTest<4, armnn::DataType::QuantisedAsymm8>(
+ RunTest<4, armnn::DataType::QAsymmU8>(
0,
{
1, 2, 3,
@@ -219,7 +219,7 @@ struct SimpleConv2DWithBiasesFixture : Conv2DWithBiasesFixture
BOOST_FIXTURE_TEST_CASE( ParseConv2DWithBias, SimpleConv2DWithBiasesFixture )
{
- RunTest<4, armnn::DataType::QuantisedAsymm8>(
+ RunTest<4, armnn::DataType::QAsymmU8>(
0,
{
1, 2,
@@ -290,7 +290,7 @@ BOOST_FIXTURE_TEST_CASE( ParseConv2DAndReluWithBias, ReluConv2DWithBiasesFixture
uint8_t outZero = 20;
uint8_t fz = 4; // filter zero point
- RunTest<4, armnn::DataType::QuantisedAsymm8>(
+ RunTest<4, armnn::DataType::QAsymmU8>(
0,
{
1, 2,
@@ -331,7 +331,7 @@ BOOST_FIXTURE_TEST_CASE( ParseConv2DAndRelu6WithBias, Relu6Conv2DWithBiasesFixtu
{
uint8_t relu6Min = 6 / 2; // divide by output scale
- RunTest<4, armnn::DataType::QuantisedAsymm8>(
+ RunTest<4, armnn::DataType::QAsymmU8>(
0,
{
1, 2,
diff --git a/src/armnnTfLiteParser/test/DepthwiseConvolution2D.cpp b/src/armnnTfLiteParser/test/DepthwiseConvolution2D.cpp
index c0767801b3..2bf08fa79f 100644
--- a/src/armnnTfLiteParser/test/DepthwiseConvolution2D.cpp
+++ b/src/armnnTfLiteParser/test/DepthwiseConvolution2D.cpp
@@ -133,7 +133,7 @@ struct DepthwiseConvolution2dSameFixture : DepthwiseConvolution2dFixture
BOOST_FIXTURE_TEST_CASE(ParseDepthwiseConv2DSame, DepthwiseConvolution2dSameFixture)
{
- RunTest<4, armnn::DataType::QuantisedAsymm8>(
+ RunTest<4, armnn::DataType::QAsymmU8>(
0,
{ 0, 1, 2,
3, 4, 5,
@@ -160,7 +160,7 @@ struct DepthwiseConvolution2dValidFixture : DepthwiseConvolution2dFixture
BOOST_FIXTURE_TEST_CASE(ParseDepthwiseConv2DValid, DepthwiseConvolution2dValidFixture)
{
- RunTest<4, armnn::DataType::QuantisedAsymm8>(
+ RunTest<4, armnn::DataType::QAsymmU8>(
0,
{ 0, 1, 2,
3, 4, 5,
@@ -185,7 +185,7 @@ struct DepthwiseConvolution2dSameBiasFixture : DepthwiseConvolution2dFixture
BOOST_FIXTURE_TEST_CASE(ParseDepthwiseConv2DSameBias, DepthwiseConvolution2dSameBiasFixture)
{
- RunTest<4, armnn::DataType::QuantisedAsymm8>(
+ RunTest<4, armnn::DataType::QAsymmU8>(
0,
{ 0, 1, 2,
3, 4, 5,
diff --git a/src/armnnTfLiteParser/test/Dequantize.cpp b/src/armnnTfLiteParser/test/Dequantize.cpp
index 2f98c07a66..79dfe2e26a 100644
--- a/src/armnnTfLiteParser/test/Dequantize.cpp
+++ b/src/armnnTfLiteParser/test/Dequantize.cpp
@@ -82,7 +82,7 @@ BOOST_AUTO_TEST_SUITE(TensorflowLiteParser)
BOOST_FIXTURE_TEST_CASE(SimpleDequantizeQAsymm8, SimpleDequantizeFixtureQAsymm8)
{
- RunTest<2, armnn::DataType::QuantisedAsymm8 , armnn::DataType::Float32>(
+ RunTest<2, armnn::DataType::QAsymmU8 , armnn::DataType::Float32>(
0,
{{"inputTensor", { 0u, 1u, 5u, 100u, 200u, 255u }}},
{{"outputTensor", { 0.0f, 1.5f, 7.5f, 150.0f, 300.0f, 382.5f }}});
@@ -97,7 +97,7 @@ BOOST_AUTO_TEST_SUITE(TensorflowLiteParser)
BOOST_FIXTURE_TEST_CASE(SimpleDequantizeQsymm16, SimpleDequantizeFixtureQSymm16)
{
- RunTest<2, armnn::DataType::QuantisedSymm16 , armnn::DataType::Float32>(
+ RunTest<2, armnn::DataType::QSymmS16 , armnn::DataType::Float32>(
0,
{{"inputTensor", { 0, 1, 5, 32767, -1, -32768 }}},
{{"outputTensor", { 0.0f, 1.5f, 7.5f, 49150.5f, -1.5f,-49152.0f }}});
diff --git a/src/armnnTfLiteParser/test/DetectionPostProcess.cpp b/src/armnnTfLiteParser/test/DetectionPostProcess.cpp
index 1ec87f97d5..f12b2b94d6 100644
--- a/src/armnnTfLiteParser/test/DetectionPostProcess.cpp
+++ b/src/armnnTfLiteParser/test/DetectionPostProcess.cpp
@@ -220,7 +220,7 @@ BOOST_FIXTURE_TEST_CASE( ParseDetectionPostProcess, ParseDetectionPostProcessCus
{ "num_detections", numDetections}
};
- RunTest<armnn::DataType::QuantisedAsymm8, armnn::DataType::Float32>(0, input, output);
+ RunTest<armnn::DataType::QAsymmU8, armnn::DataType::Float32>(0, input, output);
}
BOOST_FIXTURE_TEST_CASE(DetectionPostProcessGraphStructureTest, ParseDetectionPostProcessCustomOptions)
@@ -288,8 +288,8 @@ BOOST_FIXTURE_TEST_CASE(DetectionPostProcessGraphStructureTest, ParseDetectionPo
BOOST_TEST(CheckNumberOfOutputSlot(numDetectionsLayer, 0));
// Check the connections
- armnn::TensorInfo boxEncodingTensor(armnn::TensorShape({ 1, 6, 4 }), armnn::DataType::QuantisedAsymm8, 1, 1);
- armnn::TensorInfo scoresTensor(armnn::TensorShape({ 1, 6, 3 }), armnn::DataType::QuantisedAsymm8,
+ armnn::TensorInfo boxEncodingTensor(armnn::TensorShape({ 1, 6, 4 }), armnn::DataType::QAsymmU8, 1, 1);
+ armnn::TensorInfo scoresTensor(armnn::TensorShape({ 1, 6, 3 }), armnn::DataType::QAsymmU8,
0.00999999978f, 0);
armnn::TensorInfo detectionBoxesTensor(armnn::TensorShape({ 1, 3, 4 }), armnn::DataType::Float32, 0, 0);
diff --git a/src/armnnTfLiteParser/test/FullyConnected.cpp b/src/armnnTfLiteParser/test/FullyConnected.cpp
index 54d7bcb1dc..d1223d5af2 100644
--- a/src/armnnTfLiteParser/test/FullyConnected.cpp
+++ b/src/armnnTfLiteParser/test/FullyConnected.cpp
@@ -125,7 +125,7 @@ struct FullyConnectedWithNoBiasFixture : FullyConnectedFixture
BOOST_FIXTURE_TEST_CASE(FullyConnectedWithNoBias, FullyConnectedWithNoBiasFixture)
{
- RunTest<2, armnn::DataType::QuantisedAsymm8>(
+ RunTest<2, armnn::DataType::QAsymmU8>(
0,
{ 10, 20, 30, 40 },
{ 400/2 });
@@ -145,7 +145,7 @@ struct FullyConnectedWithBiasFixture : FullyConnectedFixture
BOOST_FIXTURE_TEST_CASE(ParseFullyConnectedWithBias, FullyConnectedWithBiasFixture)
{
- RunTest<2, armnn::DataType::QuantisedAsymm8>(
+ RunTest<2, armnn::DataType::QAsymmU8>(
0,
{ 10, 20, 30, 40 },
{ (400+10)/2 });
@@ -165,7 +165,7 @@ struct FullyConnectedWithBiasMultipleOutputsFixture : FullyConnectedFixture
BOOST_FIXTURE_TEST_CASE(FullyConnectedWithBiasMultipleOutputs, FullyConnectedWithBiasMultipleOutputsFixture)
{
- RunTest<2, armnn::DataType::QuantisedAsymm8>(
+ RunTest<2, armnn::DataType::QAsymmU8>(
0,
{ 1, 2, 3, 4, 10, 20, 30, 40 },
{ (40+10)/2, (400+10)/2 });
diff --git a/src/armnnTfLiteParser/test/MaxPool2D.cpp b/src/armnnTfLiteParser/test/MaxPool2D.cpp
index 759fc37ccd..8cbef97e2f 100644
--- a/src/armnnTfLiteParser/test/MaxPool2D.cpp
+++ b/src/armnnTfLiteParser/test/MaxPool2D.cpp
@@ -98,7 +98,7 @@ struct MaxPoolLiteFixtureUint2DOutput : MaxPool2DFixture
BOOST_FIXTURE_TEST_CASE(MaxPoolLiteUint1DOutput, MaxPoolLiteFixtureUint1DOutput)
{
- RunTest<4, armnn::DataType::QuantisedAsymm8>(0, { 2, 3, 5, 2 }, { 5 });
+ RunTest<4, armnn::DataType::QAsymmU8>(0, { 2, 3, 5, 2 }, { 5 });
}
BOOST_FIXTURE_TEST_CASE(MaxPoolLiteFloat1DOutput, MaxPoolLiteFixtureFloat1DOutput)
@@ -108,13 +108,13 @@ BOOST_FIXTURE_TEST_CASE(MaxPoolLiteFloat1DOutput, MaxPoolLiteFixtureFloat1DOutpu
BOOST_FIXTURE_TEST_CASE(MaxPoolLiteUint2DOutput, MaxPoolLiteFixtureUint2DOutput)
{
- RunTest<4, armnn::DataType::QuantisedAsymm8>(
+ RunTest<4, armnn::DataType::QAsymmU8>(
0, { 1, 2, 2, 3, 5, 6, 7, 8, 3, 2, 1, 0, 1, 2, 3, 4 }, { 6, 8, 3, 4 });
}
BOOST_FIXTURE_TEST_CASE(MaxPoolIncorrectDataTypeError, MaxPoolLiteFixtureFloat1DOutput)
{
- BOOST_CHECK_THROW((RunTest<4, armnn::DataType::QuantisedAsymm8>(0, { 2, 3, 5, 2 }, { 5 })), armnn::Exception);
+ BOOST_CHECK_THROW((RunTest<4, armnn::DataType::QAsymmU8>(0, { 2, 3, 5, 2 }, { 5 })), armnn::Exception);
}
BOOST_AUTO_TEST_SUITE_END()
diff --git a/src/armnnTfLiteParser/test/Reshape.cpp b/src/armnnTfLiteParser/test/Reshape.cpp
index 62fbad6953..6ed568ceff 100644
--- a/src/armnnTfLiteParser/test/Reshape.cpp
+++ b/src/armnnTfLiteParser/test/Reshape.cpp
@@ -86,7 +86,7 @@ struct ReshapeFixtureWithReshapeDims : ReshapeFixture
BOOST_FIXTURE_TEST_CASE(ParseReshapeWithReshapeDims, ReshapeFixtureWithReshapeDims)
{
SetupSingleInputSingleOutput("inputTensor", "outputTensor");
- RunTest<2, armnn::DataType::QuantisedAsymm8>(0,
+ RunTest<2, armnn::DataType::QAsymmU8>(0,
{ 1, 2, 3, 4, 5, 6, 7, 8, 9 },
{ 1, 2, 3, 4, 5, 6, 7, 8, 9 });
BOOST_TEST((m_Parser->GetNetworkOutputBindingInfo(0, "outputTensor").second.GetShape()
@@ -101,7 +101,7 @@ struct ReshapeFixtureWithReshapeDimsFlatten : ReshapeFixture
BOOST_FIXTURE_TEST_CASE(ParseReshapeWithReshapeDimsFlatten, ReshapeFixtureWithReshapeDimsFlatten)
{
SetupSingleInputSingleOutput("inputTensor", "outputTensor");
- RunTest<1, armnn::DataType::QuantisedAsymm8>(0,
+ RunTest<1, armnn::DataType::QAsymmU8>(0,
{ 1, 2, 3, 4, 5, 6, 7, 8, 9 },
{ 1, 2, 3, 4, 5, 6, 7, 8, 9 });
BOOST_TEST((m_Parser->GetNetworkOutputBindingInfo(0, "outputTensor").second.GetShape()
@@ -116,7 +116,7 @@ struct ReshapeFixtureWithReshapeDimsFlattenTwoDims : ReshapeFixture
BOOST_FIXTURE_TEST_CASE(ParseReshapeWithReshapeDimsFlattenTwoDims, ReshapeFixtureWithReshapeDimsFlattenTwoDims)
{
SetupSingleInputSingleOutput("inputTensor", "outputTensor");
- RunTest<2, armnn::DataType::QuantisedAsymm8>(0,
+ RunTest<2, armnn::DataType::QAsymmU8>(0,
{ 1, 1, 1, 2, 2, 2, 3, 3, 3, 4, 4, 4, 5, 5, 5, 6, 6, 6 },
{ 1, 1, 1, 2, 2, 2, 3, 3, 3, 4, 4, 4, 5, 5, 5, 6, 6, 6 });
BOOST_TEST((m_Parser->GetNetworkOutputBindingInfo(0, "outputTensor").second.GetShape()
@@ -131,7 +131,7 @@ struct ReshapeFixtureWithReshapeDimsFlattenOneDim : ReshapeFixture
BOOST_FIXTURE_TEST_CASE(ParseReshapeWithReshapeDimsFlattenOneDim, ReshapeFixtureWithReshapeDimsFlattenOneDim)
{
SetupSingleInputSingleOutput("inputTensor", "outputTensor");
- RunTest<3, armnn::DataType::QuantisedAsymm8>(0,
+ RunTest<3, armnn::DataType::QAsymmU8>(0,
{ 1, 1, 1, 2, 2, 2, 3, 3, 3, 4, 4, 4, 5, 5, 5, 6, 6, 6 },
{ 1, 1, 1, 2, 2, 2, 3, 3, 3, 4, 4, 4, 5, 5, 5, 6, 6, 6 });
BOOST_TEST((m_Parser->GetNetworkOutputBindingInfo(0, "outputTensor").second.GetShape()
diff --git a/src/armnnTfLiteParser/test/Softmax.cpp b/src/armnnTfLiteParser/test/Softmax.cpp
index dacd946352..c4d19f2ac0 100644
--- a/src/armnnTfLiteParser/test/Softmax.cpp
+++ b/src/armnnTfLiteParser/test/Softmax.cpp
@@ -71,7 +71,7 @@ struct SoftmaxFixture : public ParserFlatbuffersFixture
BOOST_FIXTURE_TEST_CASE(ParseSoftmaxLite, SoftmaxFixture)
{
- RunTest<2, armnn::DataType::QuantisedAsymm8>(0, { 0, 0, 100, 0, 0, 0, 0 }, { 0, 0, 255, 0, 0, 0, 0 });
+ RunTest<2, armnn::DataType::QAsymmU8>(0, { 0, 0, 100, 0, 0, 0, 0 }, { 0, 0, 255, 0, 0, 0, 0 });
}
BOOST_AUTO_TEST_SUITE_END()
diff --git a/src/armnnTfLiteParser/test/Split.cpp b/src/armnnTfLiteParser/test/Split.cpp
index 977bd7b6a4..5f23799fd6 100644
--- a/src/armnnTfLiteParser/test/Split.cpp
+++ b/src/armnnTfLiteParser/test/Split.cpp
@@ -179,7 +179,7 @@ struct SimpleSplitFixtureUint8 : SplitFixture
BOOST_FIXTURE_TEST_CASE(ParseAxisOneSplitTwoUint8, SimpleSplitFixtureUint8)
{
- RunTest<4, armnn::DataType::QuantisedAsymm8>(
+ RunTest<4, armnn::DataType::QAsymmU8>(
0,
{ {"inputTensor", { 1, 2, 3, 4, 5, 6, 7, 8,
9, 10, 11, 12, 13, 14, 15, 16 } } },
@@ -196,7 +196,7 @@ struct SimpleSplitAxisThreeFixtureUint8 : SplitFixture
BOOST_FIXTURE_TEST_CASE(ParseAxisThreeSplitTwoUint8, SimpleSplitAxisThreeFixtureUint8)
{
- RunTest<4, armnn::DataType::QuantisedAsymm8>(
+ RunTest<4, armnn::DataType::QAsymmU8>(
0,
{ {"inputTensor", { 1, 2, 3, 4, 5, 6, 7, 8,
9, 10, 11, 12, 13, 14, 15, 16 } } },
@@ -213,7 +213,7 @@ struct SimpleSplit2DFixtureUint8 : SplitFixture
BOOST_FIXTURE_TEST_CASE(SimpleSplit2DUint8, SimpleSplit2DFixtureUint8)
{
- RunTest<2, armnn::DataType::QuantisedAsymm8>(
+ RunTest<2, armnn::DataType::QAsymmU8>(
0,
{ {"inputTensor", { 1, 2, 3, 4, 5, 6, 7, 8 } } },
{ {"outputTensor1", { 1, 2, 3, 4 } },
@@ -229,7 +229,7 @@ struct SimpleSplit3DFixtureUint8 : SplitFixture
BOOST_FIXTURE_TEST_CASE(SimpleSplit3DUint8, SimpleSplit3DFixtureUint8)
{
- RunTest<3, armnn::DataType::QuantisedAsymm8>(
+ RunTest<3, armnn::DataType::QAsymmU8>(
0,
{ {"inputTensor", { 1, 2, 3, 4, 5, 6, 7, 8,
9, 10, 11, 12, 13, 14, 15, 16 } } },
diff --git a/src/armnnTfLiteParser/test/Squeeze.cpp b/src/armnnTfLiteParser/test/Squeeze.cpp
index 13261facf1..86a1966dd1 100644
--- a/src/armnnTfLiteParser/test/Squeeze.cpp
+++ b/src/armnnTfLiteParser/test/Squeeze.cpp
@@ -85,7 +85,7 @@ struct SqueezeFixtureWithSqueezeDims : SqueezeFixture
BOOST_FIXTURE_TEST_CASE(ParseSqueezeWithSqueezeDims, SqueezeFixtureWithSqueezeDims)
{
SetupSingleInputSingleOutput("inputTensor", "outputTensor");
- RunTest<3, armnn::DataType::QuantisedAsymm8>(0, { 1, 2, 3, 4 }, { 1, 2, 3, 4 });
+ RunTest<3, armnn::DataType::QAsymmU8>(0, { 1, 2, 3, 4 }, { 1, 2, 3, 4 });
BOOST_TEST((m_Parser->GetNetworkOutputBindingInfo(0, "outputTensor").second.GetShape()
== armnn::TensorShape({2,2,1})));
@@ -99,7 +99,7 @@ struct SqueezeFixtureWithoutSqueezeDims : SqueezeFixture
BOOST_FIXTURE_TEST_CASE(ParseSqueezeWithoutSqueezeDims, SqueezeFixtureWithoutSqueezeDims)
{
SetupSingleInputSingleOutput("inputTensor", "outputTensor");
- RunTest<2, armnn::DataType::QuantisedAsymm8>(0, { 1, 2, 3, 4 }, { 1, 2, 3, 4 });
+ RunTest<2, armnn::DataType::QAsymmU8>(0, { 1, 2, 3, 4 }, { 1, 2, 3, 4 });
BOOST_TEST((m_Parser->GetNetworkOutputBindingInfo(0, "outputTensor").second.GetShape()
== armnn::TensorShape({2,2})));
}
diff --git a/src/armnnTfLiteParser/test/Sub.cpp b/src/armnnTfLiteParser/test/Sub.cpp
index 0a3f58b519..6a251a5f74 100644
--- a/src/armnnTfLiteParser/test/Sub.cpp
+++ b/src/armnnTfLiteParser/test/Sub.cpp
@@ -97,7 +97,7 @@ struct SimpleSubFixture : SubFixture
BOOST_FIXTURE_TEST_CASE(SimpleSub, SimpleSubFixture)
{
- RunTest<2, armnn::DataType::QuantisedAsymm8>(
+ RunTest<2, armnn::DataType::QAsymmU8>(
0,
{{"inputTensor1", { 4, 5, 6, 7 }},
{"inputTensor2", { 3, 2, 1, 0 }}},
diff --git a/src/armnnTfLiteParser/test/TransposeConv.cpp b/src/armnnTfLiteParser/test/TransposeConv.cpp
index 46b02ac956..084a286dbd 100644
--- a/src/armnnTfLiteParser/test/TransposeConv.cpp
+++ b/src/armnnTfLiteParser/test/TransposeConv.cpp
@@ -118,7 +118,7 @@ struct SimpleTransposeConvFixture : TransposeConvFixture
BOOST_FIXTURE_TEST_CASE( ParseSimpleTransposeConv, SimpleTransposeConvFixture )
{
- RunTest<4, armnn::DataType::QuantisedAsymm8>(
+ RunTest<4, armnn::DataType::QAsymmU8>(
0,
{
1, 2,
diff --git a/src/armnnTfLiteParser/test/Unpack.cpp b/src/armnnTfLiteParser/test/Unpack.cpp
index 04fd50dc39..4fcd74f585 100644
--- a/src/armnnTfLiteParser/test/Unpack.cpp
+++ b/src/armnnTfLiteParser/test/Unpack.cpp
@@ -126,7 +126,7 @@ BOOST_FIXTURE_TEST_CASE(UnpackAxisZeroNumIsDefaultNotSpecified, DefaultUnpackAxi
BOOST_FIXTURE_TEST_CASE(UnpackAxisZeroNumIsDefaultNotSpecifiedUint8, DefaultUnpackAxisZeroUint8Fixture)
{
- RunTest<2, armnn::DataType::QuantisedAsymm8>(
+ RunTest<2, armnn::DataType::QAsymmU8>(
0,
{ {"inputTensor", { 1, 2, 3, 4, 5, 6,
7, 8, 9, 10, 11, 12,
@@ -165,7 +165,7 @@ BOOST_FIXTURE_TEST_CASE(UnpackLastAxisNumSix, DefaultUnpackLastAxisFixture)
}
BOOST_FIXTURE_TEST_CASE(UnpackLastAxisNumSixUint8, DefaultUnpackLastAxisUint8Fixture) {
- RunTest<2, armnn::DataType::QuantisedAsymm8>(
+ RunTest<2, armnn::DataType::QAsymmU8>(
0,
{{"inputTensor", { 1, 2, 3, 4, 5, 6,
7, 8, 9, 10, 11, 12,
diff --git a/src/armnnTfLiteParser/test/Unsupported.cpp b/src/armnnTfLiteParser/test/Unsupported.cpp
index 9a9cdc5156..39dee679fd 100644
--- a/src/armnnTfLiteParser/test/Unsupported.cpp
+++ b/src/armnnTfLiteParser/test/Unsupported.cpp
@@ -194,7 +194,7 @@ private:
switch (dataType)
{
case DataType::Float32: return "FLOAT32";
- case DataType::QuantisedAsymm8: return "UINT8";
+ case DataType::QAsymmU8: return "UINT8";
default: return "UNKNOWN";
}
}