diff options
author | Derek Lamberti <derek.lamberti@arm.com> | 2020-01-10 17:14:08 +0000 |
---|---|---|
committer | Kevin May <kevin.may@arm.com> | 2020-01-13 18:18:12 +0000 |
commit | f90c56d72de4848a2dc5844a97458aaf09df07c2 (patch) | |
tree | 71f1c6f16a4687286614f5526ed70938a611b27d /src/armnnTfLiteParser/test | |
parent | 842e0dbd40114e19bf26916fefe06c869dbe416d (diff) | |
download | armnn-f90c56d72de4848a2dc5844a97458aaf09df07c2.tar.gz |
Rename quantized data types to remove ambiguity for signed/unsigned payloads
!android-nn-driver:2572
Change-Id: I8fe52ceb09987b3d05c539409510f535165455cc
Signed-off-by: Derek Lamberti <derek.lamberti@arm.com>
Diffstat (limited to 'src/armnnTfLiteParser/test')
-rw-r--r-- | src/armnnTfLiteParser/test/Addition.cpp | 2 | ||||
-rw-r--r-- | src/armnnTfLiteParser/test/AvgPool2D.cpp | 6 | ||||
-rw-r--r-- | src/armnnTfLiteParser/test/Concatenation.cpp | 16 | ||||
-rw-r--r-- | src/armnnTfLiteParser/test/Constant.cpp | 2 | ||||
-rw-r--r-- | src/armnnTfLiteParser/test/Conv2D.cpp | 8 | ||||
-rw-r--r-- | src/armnnTfLiteParser/test/DepthwiseConvolution2D.cpp | 6 | ||||
-rw-r--r-- | src/armnnTfLiteParser/test/Dequantize.cpp | 4 | ||||
-rw-r--r-- | src/armnnTfLiteParser/test/DetectionPostProcess.cpp | 6 | ||||
-rw-r--r-- | src/armnnTfLiteParser/test/FullyConnected.cpp | 6 | ||||
-rw-r--r-- | src/armnnTfLiteParser/test/MaxPool2D.cpp | 6 | ||||
-rw-r--r-- | src/armnnTfLiteParser/test/Reshape.cpp | 8 | ||||
-rw-r--r-- | src/armnnTfLiteParser/test/Softmax.cpp | 2 | ||||
-rw-r--r-- | src/armnnTfLiteParser/test/Split.cpp | 8 | ||||
-rw-r--r-- | src/armnnTfLiteParser/test/Squeeze.cpp | 4 | ||||
-rw-r--r-- | src/armnnTfLiteParser/test/Sub.cpp | 2 | ||||
-rw-r--r-- | src/armnnTfLiteParser/test/TransposeConv.cpp | 2 | ||||
-rw-r--r-- | src/armnnTfLiteParser/test/Unpack.cpp | 4 | ||||
-rw-r--r-- | src/armnnTfLiteParser/test/Unsupported.cpp | 2 |
18 files changed, 47 insertions, 47 deletions
diff --git a/src/armnnTfLiteParser/test/Addition.cpp b/src/armnnTfLiteParser/test/Addition.cpp index 94389d3134..deeb707a2f 100644 --- a/src/armnnTfLiteParser/test/Addition.cpp +++ b/src/armnnTfLiteParser/test/Addition.cpp @@ -97,7 +97,7 @@ struct SimpleAddFixture : AddFixture BOOST_FIXTURE_TEST_CASE(SimpleAdd, SimpleAddFixture) { - RunTest<2, armnn::DataType::QuantisedAsymm8>( + RunTest<2, armnn::DataType::QAsymmU8>( 0, {{"inputTensor1", { 0, 1, 2, 3 }}, {"inputTensor2", { 4, 5, 6, 7 }}}, diff --git a/src/armnnTfLiteParser/test/AvgPool2D.cpp b/src/armnnTfLiteParser/test/AvgPool2D.cpp index a39c088d44..a56e7e7362 100644 --- a/src/armnnTfLiteParser/test/AvgPool2D.cpp +++ b/src/armnnTfLiteParser/test/AvgPool2D.cpp @@ -98,7 +98,7 @@ struct AvgPoolLiteFixture2DOutput : AvgPool2DFixture BOOST_FIXTURE_TEST_CASE(AvgPoolLite1DOutput, AvgPoolLiteFixtureUint1DOutput) { - RunTest<4, armnn::DataType::QuantisedAsymm8>(0, {2, 3, 5, 2 }, { 3 }); + RunTest<4, armnn::DataType::QAsymmU8>(0, {2, 3, 5, 2 }, { 3 }); } BOOST_FIXTURE_TEST_CASE(AvgPoolLiteFloat1DOutput, AvgPoolLiteFixtureFloat1DOutput) @@ -108,13 +108,13 @@ BOOST_FIXTURE_TEST_CASE(AvgPoolLiteFloat1DOutput, AvgPoolLiteFixtureFloat1DOutpu BOOST_FIXTURE_TEST_CASE(AvgPoolLite2DOutput, AvgPoolLiteFixture2DOutput) { - RunTest<4, armnn::DataType::QuantisedAsymm8>( + RunTest<4, armnn::DataType::QAsymmU8>( 0, { 1, 2, 2, 3, 5, 6, 7, 8, 3, 2, 1, 0, 1, 2, 3, 4 }, { 4, 5, 2, 2 }); } BOOST_FIXTURE_TEST_CASE(IncorrectDataTypeError, AvgPoolLiteFixtureFloat1DOutput) { - BOOST_CHECK_THROW((RunTest<4, armnn::DataType::QuantisedAsymm8>(0, {2, 3, 5, 2 }, { 3 })), armnn::Exception); + BOOST_CHECK_THROW((RunTest<4, armnn::DataType::QAsymmU8>(0, {2, 3, 5, 2 }, { 3 })), armnn::Exception); } BOOST_AUTO_TEST_SUITE_END() diff --git a/src/armnnTfLiteParser/test/Concatenation.cpp b/src/armnnTfLiteParser/test/Concatenation.cpp index d3d571f174..8e31a3edb0 100644 --- a/src/armnnTfLiteParser/test/Concatenation.cpp +++ b/src/armnnTfLiteParser/test/Concatenation.cpp @@ -100,7 +100,7 @@ struct ConcatenationFixtureNegativeDim : ConcatenationFixture BOOST_FIXTURE_TEST_CASE(ParseConcatenationNegativeDim, ConcatenationFixtureNegativeDim) { - RunTest<4, armnn::DataType::QuantisedAsymm8>( + RunTest<4, armnn::DataType::QAsymmU8>( 0, {{"inputTensor1", { 0, 1, 2, 3 }}, {"inputTensor2", { 4, 5, 6, 7 }}}, @@ -114,7 +114,7 @@ struct ConcatenationFixtureNCHW : ConcatenationFixture BOOST_FIXTURE_TEST_CASE(ParseConcatenationNCHW, ConcatenationFixtureNCHW) { - RunTest<4, armnn::DataType::QuantisedAsymm8>( + RunTest<4, armnn::DataType::QAsymmU8>( 0, {{"inputTensor1", { 0, 1, 2, 3 }}, {"inputTensor2", { 4, 5, 6, 7 }}}, @@ -128,7 +128,7 @@ struct ConcatenationFixtureNHWC : ConcatenationFixture BOOST_FIXTURE_TEST_CASE(ParseConcatenationNHWC, ConcatenationFixtureNHWC) { - RunTest<4, armnn::DataType::QuantisedAsymm8>( + RunTest<4, armnn::DataType::QAsymmU8>( 0, {{"inputTensor1", { 0, 1, 2, 3 }}, {"inputTensor2", { 4, 5, 6, 7 }}}, @@ -142,7 +142,7 @@ struct ConcatenationFixtureDim1 : ConcatenationFixture BOOST_FIXTURE_TEST_CASE(ParseConcatenationDim1, ConcatenationFixtureDim1) { - RunTest<4, armnn::DataType::QuantisedAsymm8>( + RunTest<4, armnn::DataType::QAsymmU8>( 0, { { "inputTensor1", { 0, 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16, 17, 18, 19, 20, 21, 22, 23 } }, @@ -161,7 +161,7 @@ struct ConcatenationFixtureDim3 : ConcatenationFixture BOOST_FIXTURE_TEST_CASE(ParseConcatenationDim3, ConcatenationFixtureDim3) { - RunTest<4, armnn::DataType::QuantisedAsymm8>( + RunTest<4, armnn::DataType::QAsymmU8>( 0, { { "inputTensor1", { 0, 1, 2, 3, 4, 5, 6, 7, @@ -196,7 +196,7 @@ struct ConcatenationFixture3DDim0 : ConcatenationFixture BOOST_FIXTURE_TEST_CASE(ParseConcatenation3DDim0, ConcatenationFixture3DDim0) { - RunTest<3, armnn::DataType::QuantisedAsymm8>( + RunTest<3, armnn::DataType::QAsymmU8>( 0, { { "inputTensor1", { 0, 1, 2, 3, 4, 5 } }, { "inputTensor2", { 6, 7, 8, 9, 10, 11, @@ -213,7 +213,7 @@ struct ConcatenationFixture3DDim1 : ConcatenationFixture BOOST_FIXTURE_TEST_CASE(ParseConcatenation3DDim1, ConcatenationFixture3DDim1) { - RunTest<3, armnn::DataType::QuantisedAsymm8>( + RunTest<3, armnn::DataType::QAsymmU8>( 0, { { "inputTensor1", { 0, 1, 2, 3, 4, 5 } }, { "inputTensor2", { 6, 7, 8, 9, 10, 11, @@ -230,7 +230,7 @@ struct ConcatenationFixture3DDim2 : ConcatenationFixture BOOST_FIXTURE_TEST_CASE(ParseConcatenation3DDim2, ConcatenationFixture3DDim2) { - RunTest<3, armnn::DataType::QuantisedAsymm8>( + RunTest<3, armnn::DataType::QAsymmU8>( 0, { { "inputTensor1", { 0, 1, 2, 3, 4, 5 } }, diff --git a/src/armnnTfLiteParser/test/Constant.cpp b/src/armnnTfLiteParser/test/Constant.cpp index 356e1b769d..cc89223469 100644 --- a/src/armnnTfLiteParser/test/Constant.cpp +++ b/src/armnnTfLiteParser/test/Constant.cpp @@ -103,7 +103,7 @@ struct SimpleConstantAddFixture : ConstantAddFixture BOOST_FIXTURE_TEST_CASE(SimpleConstantAdd, SimpleConstantAddFixture) { - RunTest<2, armnn::DataType::QuantisedAsymm8>( + RunTest<2, armnn::DataType::QAsymmU8>( 0, {{"InputTensor", { 0, 1, 2, 3 }}}, {{"OutputTensor", { 4, 6, 8, 10 }}} diff --git a/src/armnnTfLiteParser/test/Conv2D.cpp b/src/armnnTfLiteParser/test/Conv2D.cpp index 38c6675ddb..2eae5f5a1a 100644 --- a/src/armnnTfLiteParser/test/Conv2D.cpp +++ b/src/armnnTfLiteParser/test/Conv2D.cpp @@ -89,7 +89,7 @@ struct SimpleConv2DFixture : public ParserFlatbuffersFixture BOOST_FIXTURE_TEST_CASE( ParseSimpleConv2D, SimpleConv2DFixture ) { - RunTest<4, armnn::DataType::QuantisedAsymm8>( + RunTest<4, armnn::DataType::QAsymmU8>( 0, { 1, 2, 3, @@ -219,7 +219,7 @@ struct SimpleConv2DWithBiasesFixture : Conv2DWithBiasesFixture BOOST_FIXTURE_TEST_CASE( ParseConv2DWithBias, SimpleConv2DWithBiasesFixture ) { - RunTest<4, armnn::DataType::QuantisedAsymm8>( + RunTest<4, armnn::DataType::QAsymmU8>( 0, { 1, 2, @@ -290,7 +290,7 @@ BOOST_FIXTURE_TEST_CASE( ParseConv2DAndReluWithBias, ReluConv2DWithBiasesFixture uint8_t outZero = 20; uint8_t fz = 4; // filter zero point - RunTest<4, armnn::DataType::QuantisedAsymm8>( + RunTest<4, armnn::DataType::QAsymmU8>( 0, { 1, 2, @@ -331,7 +331,7 @@ BOOST_FIXTURE_TEST_CASE( ParseConv2DAndRelu6WithBias, Relu6Conv2DWithBiasesFixtu { uint8_t relu6Min = 6 / 2; // divide by output scale - RunTest<4, armnn::DataType::QuantisedAsymm8>( + RunTest<4, armnn::DataType::QAsymmU8>( 0, { 1, 2, diff --git a/src/armnnTfLiteParser/test/DepthwiseConvolution2D.cpp b/src/armnnTfLiteParser/test/DepthwiseConvolution2D.cpp index c0767801b3..2bf08fa79f 100644 --- a/src/armnnTfLiteParser/test/DepthwiseConvolution2D.cpp +++ b/src/armnnTfLiteParser/test/DepthwiseConvolution2D.cpp @@ -133,7 +133,7 @@ struct DepthwiseConvolution2dSameFixture : DepthwiseConvolution2dFixture BOOST_FIXTURE_TEST_CASE(ParseDepthwiseConv2DSame, DepthwiseConvolution2dSameFixture) { - RunTest<4, armnn::DataType::QuantisedAsymm8>( + RunTest<4, armnn::DataType::QAsymmU8>( 0, { 0, 1, 2, 3, 4, 5, @@ -160,7 +160,7 @@ struct DepthwiseConvolution2dValidFixture : DepthwiseConvolution2dFixture BOOST_FIXTURE_TEST_CASE(ParseDepthwiseConv2DValid, DepthwiseConvolution2dValidFixture) { - RunTest<4, armnn::DataType::QuantisedAsymm8>( + RunTest<4, armnn::DataType::QAsymmU8>( 0, { 0, 1, 2, 3, 4, 5, @@ -185,7 +185,7 @@ struct DepthwiseConvolution2dSameBiasFixture : DepthwiseConvolution2dFixture BOOST_FIXTURE_TEST_CASE(ParseDepthwiseConv2DSameBias, DepthwiseConvolution2dSameBiasFixture) { - RunTest<4, armnn::DataType::QuantisedAsymm8>( + RunTest<4, armnn::DataType::QAsymmU8>( 0, { 0, 1, 2, 3, 4, 5, diff --git a/src/armnnTfLiteParser/test/Dequantize.cpp b/src/armnnTfLiteParser/test/Dequantize.cpp index 2f98c07a66..79dfe2e26a 100644 --- a/src/armnnTfLiteParser/test/Dequantize.cpp +++ b/src/armnnTfLiteParser/test/Dequantize.cpp @@ -82,7 +82,7 @@ BOOST_AUTO_TEST_SUITE(TensorflowLiteParser) BOOST_FIXTURE_TEST_CASE(SimpleDequantizeQAsymm8, SimpleDequantizeFixtureQAsymm8) { - RunTest<2, armnn::DataType::QuantisedAsymm8 , armnn::DataType::Float32>( + RunTest<2, armnn::DataType::QAsymmU8 , armnn::DataType::Float32>( 0, {{"inputTensor", { 0u, 1u, 5u, 100u, 200u, 255u }}}, {{"outputTensor", { 0.0f, 1.5f, 7.5f, 150.0f, 300.0f, 382.5f }}}); @@ -97,7 +97,7 @@ BOOST_AUTO_TEST_SUITE(TensorflowLiteParser) BOOST_FIXTURE_TEST_CASE(SimpleDequantizeQsymm16, SimpleDequantizeFixtureQSymm16) { - RunTest<2, armnn::DataType::QuantisedSymm16 , armnn::DataType::Float32>( + RunTest<2, armnn::DataType::QSymmS16 , armnn::DataType::Float32>( 0, {{"inputTensor", { 0, 1, 5, 32767, -1, -32768 }}}, {{"outputTensor", { 0.0f, 1.5f, 7.5f, 49150.5f, -1.5f,-49152.0f }}}); diff --git a/src/armnnTfLiteParser/test/DetectionPostProcess.cpp b/src/armnnTfLiteParser/test/DetectionPostProcess.cpp index 1ec87f97d5..f12b2b94d6 100644 --- a/src/armnnTfLiteParser/test/DetectionPostProcess.cpp +++ b/src/armnnTfLiteParser/test/DetectionPostProcess.cpp @@ -220,7 +220,7 @@ BOOST_FIXTURE_TEST_CASE( ParseDetectionPostProcess, ParseDetectionPostProcessCus { "num_detections", numDetections} }; - RunTest<armnn::DataType::QuantisedAsymm8, armnn::DataType::Float32>(0, input, output); + RunTest<armnn::DataType::QAsymmU8, armnn::DataType::Float32>(0, input, output); } BOOST_FIXTURE_TEST_CASE(DetectionPostProcessGraphStructureTest, ParseDetectionPostProcessCustomOptions) @@ -288,8 +288,8 @@ BOOST_FIXTURE_TEST_CASE(DetectionPostProcessGraphStructureTest, ParseDetectionPo BOOST_TEST(CheckNumberOfOutputSlot(numDetectionsLayer, 0)); // Check the connections - armnn::TensorInfo boxEncodingTensor(armnn::TensorShape({ 1, 6, 4 }), armnn::DataType::QuantisedAsymm8, 1, 1); - armnn::TensorInfo scoresTensor(armnn::TensorShape({ 1, 6, 3 }), armnn::DataType::QuantisedAsymm8, + armnn::TensorInfo boxEncodingTensor(armnn::TensorShape({ 1, 6, 4 }), armnn::DataType::QAsymmU8, 1, 1); + armnn::TensorInfo scoresTensor(armnn::TensorShape({ 1, 6, 3 }), armnn::DataType::QAsymmU8, 0.00999999978f, 0); armnn::TensorInfo detectionBoxesTensor(armnn::TensorShape({ 1, 3, 4 }), armnn::DataType::Float32, 0, 0); diff --git a/src/armnnTfLiteParser/test/FullyConnected.cpp b/src/armnnTfLiteParser/test/FullyConnected.cpp index 54d7bcb1dc..d1223d5af2 100644 --- a/src/armnnTfLiteParser/test/FullyConnected.cpp +++ b/src/armnnTfLiteParser/test/FullyConnected.cpp @@ -125,7 +125,7 @@ struct FullyConnectedWithNoBiasFixture : FullyConnectedFixture BOOST_FIXTURE_TEST_CASE(FullyConnectedWithNoBias, FullyConnectedWithNoBiasFixture) { - RunTest<2, armnn::DataType::QuantisedAsymm8>( + RunTest<2, armnn::DataType::QAsymmU8>( 0, { 10, 20, 30, 40 }, { 400/2 }); @@ -145,7 +145,7 @@ struct FullyConnectedWithBiasFixture : FullyConnectedFixture BOOST_FIXTURE_TEST_CASE(ParseFullyConnectedWithBias, FullyConnectedWithBiasFixture) { - RunTest<2, armnn::DataType::QuantisedAsymm8>( + RunTest<2, armnn::DataType::QAsymmU8>( 0, { 10, 20, 30, 40 }, { (400+10)/2 }); @@ -165,7 +165,7 @@ struct FullyConnectedWithBiasMultipleOutputsFixture : FullyConnectedFixture BOOST_FIXTURE_TEST_CASE(FullyConnectedWithBiasMultipleOutputs, FullyConnectedWithBiasMultipleOutputsFixture) { - RunTest<2, armnn::DataType::QuantisedAsymm8>( + RunTest<2, armnn::DataType::QAsymmU8>( 0, { 1, 2, 3, 4, 10, 20, 30, 40 }, { (40+10)/2, (400+10)/2 }); diff --git a/src/armnnTfLiteParser/test/MaxPool2D.cpp b/src/armnnTfLiteParser/test/MaxPool2D.cpp index 759fc37ccd..8cbef97e2f 100644 --- a/src/armnnTfLiteParser/test/MaxPool2D.cpp +++ b/src/armnnTfLiteParser/test/MaxPool2D.cpp @@ -98,7 +98,7 @@ struct MaxPoolLiteFixtureUint2DOutput : MaxPool2DFixture BOOST_FIXTURE_TEST_CASE(MaxPoolLiteUint1DOutput, MaxPoolLiteFixtureUint1DOutput) { - RunTest<4, armnn::DataType::QuantisedAsymm8>(0, { 2, 3, 5, 2 }, { 5 }); + RunTest<4, armnn::DataType::QAsymmU8>(0, { 2, 3, 5, 2 }, { 5 }); } BOOST_FIXTURE_TEST_CASE(MaxPoolLiteFloat1DOutput, MaxPoolLiteFixtureFloat1DOutput) @@ -108,13 +108,13 @@ BOOST_FIXTURE_TEST_CASE(MaxPoolLiteFloat1DOutput, MaxPoolLiteFixtureFloat1DOutpu BOOST_FIXTURE_TEST_CASE(MaxPoolLiteUint2DOutput, MaxPoolLiteFixtureUint2DOutput) { - RunTest<4, armnn::DataType::QuantisedAsymm8>( + RunTest<4, armnn::DataType::QAsymmU8>( 0, { 1, 2, 2, 3, 5, 6, 7, 8, 3, 2, 1, 0, 1, 2, 3, 4 }, { 6, 8, 3, 4 }); } BOOST_FIXTURE_TEST_CASE(MaxPoolIncorrectDataTypeError, MaxPoolLiteFixtureFloat1DOutput) { - BOOST_CHECK_THROW((RunTest<4, armnn::DataType::QuantisedAsymm8>(0, { 2, 3, 5, 2 }, { 5 })), armnn::Exception); + BOOST_CHECK_THROW((RunTest<4, armnn::DataType::QAsymmU8>(0, { 2, 3, 5, 2 }, { 5 })), armnn::Exception); } BOOST_AUTO_TEST_SUITE_END() diff --git a/src/armnnTfLiteParser/test/Reshape.cpp b/src/armnnTfLiteParser/test/Reshape.cpp index 62fbad6953..6ed568ceff 100644 --- a/src/armnnTfLiteParser/test/Reshape.cpp +++ b/src/armnnTfLiteParser/test/Reshape.cpp @@ -86,7 +86,7 @@ struct ReshapeFixtureWithReshapeDims : ReshapeFixture BOOST_FIXTURE_TEST_CASE(ParseReshapeWithReshapeDims, ReshapeFixtureWithReshapeDims) { SetupSingleInputSingleOutput("inputTensor", "outputTensor"); - RunTest<2, armnn::DataType::QuantisedAsymm8>(0, + RunTest<2, armnn::DataType::QAsymmU8>(0, { 1, 2, 3, 4, 5, 6, 7, 8, 9 }, { 1, 2, 3, 4, 5, 6, 7, 8, 9 }); BOOST_TEST((m_Parser->GetNetworkOutputBindingInfo(0, "outputTensor").second.GetShape() @@ -101,7 +101,7 @@ struct ReshapeFixtureWithReshapeDimsFlatten : ReshapeFixture BOOST_FIXTURE_TEST_CASE(ParseReshapeWithReshapeDimsFlatten, ReshapeFixtureWithReshapeDimsFlatten) { SetupSingleInputSingleOutput("inputTensor", "outputTensor"); - RunTest<1, armnn::DataType::QuantisedAsymm8>(0, + RunTest<1, armnn::DataType::QAsymmU8>(0, { 1, 2, 3, 4, 5, 6, 7, 8, 9 }, { 1, 2, 3, 4, 5, 6, 7, 8, 9 }); BOOST_TEST((m_Parser->GetNetworkOutputBindingInfo(0, "outputTensor").second.GetShape() @@ -116,7 +116,7 @@ struct ReshapeFixtureWithReshapeDimsFlattenTwoDims : ReshapeFixture BOOST_FIXTURE_TEST_CASE(ParseReshapeWithReshapeDimsFlattenTwoDims, ReshapeFixtureWithReshapeDimsFlattenTwoDims) { SetupSingleInputSingleOutput("inputTensor", "outputTensor"); - RunTest<2, armnn::DataType::QuantisedAsymm8>(0, + RunTest<2, armnn::DataType::QAsymmU8>(0, { 1, 1, 1, 2, 2, 2, 3, 3, 3, 4, 4, 4, 5, 5, 5, 6, 6, 6 }, { 1, 1, 1, 2, 2, 2, 3, 3, 3, 4, 4, 4, 5, 5, 5, 6, 6, 6 }); BOOST_TEST((m_Parser->GetNetworkOutputBindingInfo(0, "outputTensor").second.GetShape() @@ -131,7 +131,7 @@ struct ReshapeFixtureWithReshapeDimsFlattenOneDim : ReshapeFixture BOOST_FIXTURE_TEST_CASE(ParseReshapeWithReshapeDimsFlattenOneDim, ReshapeFixtureWithReshapeDimsFlattenOneDim) { SetupSingleInputSingleOutput("inputTensor", "outputTensor"); - RunTest<3, armnn::DataType::QuantisedAsymm8>(0, + RunTest<3, armnn::DataType::QAsymmU8>(0, { 1, 1, 1, 2, 2, 2, 3, 3, 3, 4, 4, 4, 5, 5, 5, 6, 6, 6 }, { 1, 1, 1, 2, 2, 2, 3, 3, 3, 4, 4, 4, 5, 5, 5, 6, 6, 6 }); BOOST_TEST((m_Parser->GetNetworkOutputBindingInfo(0, "outputTensor").second.GetShape() diff --git a/src/armnnTfLiteParser/test/Softmax.cpp b/src/armnnTfLiteParser/test/Softmax.cpp index dacd946352..c4d19f2ac0 100644 --- a/src/armnnTfLiteParser/test/Softmax.cpp +++ b/src/armnnTfLiteParser/test/Softmax.cpp @@ -71,7 +71,7 @@ struct SoftmaxFixture : public ParserFlatbuffersFixture BOOST_FIXTURE_TEST_CASE(ParseSoftmaxLite, SoftmaxFixture) { - RunTest<2, armnn::DataType::QuantisedAsymm8>(0, { 0, 0, 100, 0, 0, 0, 0 }, { 0, 0, 255, 0, 0, 0, 0 }); + RunTest<2, armnn::DataType::QAsymmU8>(0, { 0, 0, 100, 0, 0, 0, 0 }, { 0, 0, 255, 0, 0, 0, 0 }); } BOOST_AUTO_TEST_SUITE_END() diff --git a/src/armnnTfLiteParser/test/Split.cpp b/src/armnnTfLiteParser/test/Split.cpp index 977bd7b6a4..5f23799fd6 100644 --- a/src/armnnTfLiteParser/test/Split.cpp +++ b/src/armnnTfLiteParser/test/Split.cpp @@ -179,7 +179,7 @@ struct SimpleSplitFixtureUint8 : SplitFixture BOOST_FIXTURE_TEST_CASE(ParseAxisOneSplitTwoUint8, SimpleSplitFixtureUint8) { - RunTest<4, armnn::DataType::QuantisedAsymm8>( + RunTest<4, armnn::DataType::QAsymmU8>( 0, { {"inputTensor", { 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16 } } }, @@ -196,7 +196,7 @@ struct SimpleSplitAxisThreeFixtureUint8 : SplitFixture BOOST_FIXTURE_TEST_CASE(ParseAxisThreeSplitTwoUint8, SimpleSplitAxisThreeFixtureUint8) { - RunTest<4, armnn::DataType::QuantisedAsymm8>( + RunTest<4, armnn::DataType::QAsymmU8>( 0, { {"inputTensor", { 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16 } } }, @@ -213,7 +213,7 @@ struct SimpleSplit2DFixtureUint8 : SplitFixture BOOST_FIXTURE_TEST_CASE(SimpleSplit2DUint8, SimpleSplit2DFixtureUint8) { - RunTest<2, armnn::DataType::QuantisedAsymm8>( + RunTest<2, armnn::DataType::QAsymmU8>( 0, { {"inputTensor", { 1, 2, 3, 4, 5, 6, 7, 8 } } }, { {"outputTensor1", { 1, 2, 3, 4 } }, @@ -229,7 +229,7 @@ struct SimpleSplit3DFixtureUint8 : SplitFixture BOOST_FIXTURE_TEST_CASE(SimpleSplit3DUint8, SimpleSplit3DFixtureUint8) { - RunTest<3, armnn::DataType::QuantisedAsymm8>( + RunTest<3, armnn::DataType::QAsymmU8>( 0, { {"inputTensor", { 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 14, 15, 16 } } }, diff --git a/src/armnnTfLiteParser/test/Squeeze.cpp b/src/armnnTfLiteParser/test/Squeeze.cpp index 13261facf1..86a1966dd1 100644 --- a/src/armnnTfLiteParser/test/Squeeze.cpp +++ b/src/armnnTfLiteParser/test/Squeeze.cpp @@ -85,7 +85,7 @@ struct SqueezeFixtureWithSqueezeDims : SqueezeFixture BOOST_FIXTURE_TEST_CASE(ParseSqueezeWithSqueezeDims, SqueezeFixtureWithSqueezeDims) { SetupSingleInputSingleOutput("inputTensor", "outputTensor"); - RunTest<3, armnn::DataType::QuantisedAsymm8>(0, { 1, 2, 3, 4 }, { 1, 2, 3, 4 }); + RunTest<3, armnn::DataType::QAsymmU8>(0, { 1, 2, 3, 4 }, { 1, 2, 3, 4 }); BOOST_TEST((m_Parser->GetNetworkOutputBindingInfo(0, "outputTensor").second.GetShape() == armnn::TensorShape({2,2,1}))); @@ -99,7 +99,7 @@ struct SqueezeFixtureWithoutSqueezeDims : SqueezeFixture BOOST_FIXTURE_TEST_CASE(ParseSqueezeWithoutSqueezeDims, SqueezeFixtureWithoutSqueezeDims) { SetupSingleInputSingleOutput("inputTensor", "outputTensor"); - RunTest<2, armnn::DataType::QuantisedAsymm8>(0, { 1, 2, 3, 4 }, { 1, 2, 3, 4 }); + RunTest<2, armnn::DataType::QAsymmU8>(0, { 1, 2, 3, 4 }, { 1, 2, 3, 4 }); BOOST_TEST((m_Parser->GetNetworkOutputBindingInfo(0, "outputTensor").second.GetShape() == armnn::TensorShape({2,2}))); } diff --git a/src/armnnTfLiteParser/test/Sub.cpp b/src/armnnTfLiteParser/test/Sub.cpp index 0a3f58b519..6a251a5f74 100644 --- a/src/armnnTfLiteParser/test/Sub.cpp +++ b/src/armnnTfLiteParser/test/Sub.cpp @@ -97,7 +97,7 @@ struct SimpleSubFixture : SubFixture BOOST_FIXTURE_TEST_CASE(SimpleSub, SimpleSubFixture) { - RunTest<2, armnn::DataType::QuantisedAsymm8>( + RunTest<2, armnn::DataType::QAsymmU8>( 0, {{"inputTensor1", { 4, 5, 6, 7 }}, {"inputTensor2", { 3, 2, 1, 0 }}}, diff --git a/src/armnnTfLiteParser/test/TransposeConv.cpp b/src/armnnTfLiteParser/test/TransposeConv.cpp index 46b02ac956..084a286dbd 100644 --- a/src/armnnTfLiteParser/test/TransposeConv.cpp +++ b/src/armnnTfLiteParser/test/TransposeConv.cpp @@ -118,7 +118,7 @@ struct SimpleTransposeConvFixture : TransposeConvFixture BOOST_FIXTURE_TEST_CASE( ParseSimpleTransposeConv, SimpleTransposeConvFixture ) { - RunTest<4, armnn::DataType::QuantisedAsymm8>( + RunTest<4, armnn::DataType::QAsymmU8>( 0, { 1, 2, diff --git a/src/armnnTfLiteParser/test/Unpack.cpp b/src/armnnTfLiteParser/test/Unpack.cpp index 04fd50dc39..4fcd74f585 100644 --- a/src/armnnTfLiteParser/test/Unpack.cpp +++ b/src/armnnTfLiteParser/test/Unpack.cpp @@ -126,7 +126,7 @@ BOOST_FIXTURE_TEST_CASE(UnpackAxisZeroNumIsDefaultNotSpecified, DefaultUnpackAxi BOOST_FIXTURE_TEST_CASE(UnpackAxisZeroNumIsDefaultNotSpecifiedUint8, DefaultUnpackAxisZeroUint8Fixture) { - RunTest<2, armnn::DataType::QuantisedAsymm8>( + RunTest<2, armnn::DataType::QAsymmU8>( 0, { {"inputTensor", { 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, @@ -165,7 +165,7 @@ BOOST_FIXTURE_TEST_CASE(UnpackLastAxisNumSix, DefaultUnpackLastAxisFixture) } BOOST_FIXTURE_TEST_CASE(UnpackLastAxisNumSixUint8, DefaultUnpackLastAxisUint8Fixture) { - RunTest<2, armnn::DataType::QuantisedAsymm8>( + RunTest<2, armnn::DataType::QAsymmU8>( 0, {{"inputTensor", { 1, 2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, diff --git a/src/armnnTfLiteParser/test/Unsupported.cpp b/src/armnnTfLiteParser/test/Unsupported.cpp index 9a9cdc5156..39dee679fd 100644 --- a/src/armnnTfLiteParser/test/Unsupported.cpp +++ b/src/armnnTfLiteParser/test/Unsupported.cpp @@ -194,7 +194,7 @@ private: switch (dataType) { case DataType::Float32: return "FLOAT32"; - case DataType::QuantisedAsymm8: return "UINT8"; + case DataType::QAsymmU8: return "UINT8"; default: return "UNKNOWN"; } } |