diff options
Diffstat (limited to 'src/backends/reference')
15 files changed, 434 insertions, 434 deletions
diff --git a/src/backends/reference/RefLayerSupport.cpp b/src/backends/reference/RefLayerSupport.cpp index ebcd1f633e..4767aa0b3b 100644 --- a/src/backends/reference/RefLayerSupport.cpp +++ b/src/backends/reference/RefLayerSupport.cpp @@ -75,8 +75,8 @@ bool RefLayerSupport::IsAbsSupported(const TensorInfo& input, const TensorInfo& { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported, @@ -105,8 +105,8 @@ bool RefLayerSupport::IsActivationSupported(const TensorInfo& input, std::array<DataType,4> supportedTypes = { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported, @@ -168,8 +168,8 @@ bool RefLayerSupport::IsAdditionSupported(const TensorInfo& input0, std::array<DataType,4> supportedTypes = { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input0, supportedTypes), reasonIfUnsupported, @@ -202,8 +202,8 @@ bool RefLayerSupport::IsArgMinMaxSupported(const armnn::TensorInfo &input, const std::array<DataType, 4> supportedTypes = { DataType::Float32, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16, + DataType::QAsymmU8, + DataType::QSymmS16, DataType::Signed32 }; @@ -232,8 +232,8 @@ bool RefLayerSupport::IsBatchNormalizationSupported(const TensorInfo& input, { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; bool supported = true; @@ -280,8 +280,8 @@ bool RefLayerSupport::IsBatchToSpaceNdSupported(const TensorInfo& input, { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported, @@ -322,8 +322,8 @@ bool RefLayerSupport::IsComparisonSupported(const TensorInfo& input0, { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; bool supported = true; @@ -351,8 +351,8 @@ bool RefLayerSupport::IsConcatSupported(const std::vector<const TensorInfo*> inp { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported, @@ -377,8 +377,8 @@ bool RefLayerSupport::IsConstantSupported(const TensorInfo& output, { DataType::Float32, DataType::Signed32, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; return CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported, @@ -439,8 +439,8 @@ bool RefLayerSupport::IsConvolution2dSupported(const TensorInfo& input, { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported, @@ -453,11 +453,11 @@ bool RefLayerSupport::IsConvolution2dSupported(const TensorInfo& input, "Reference convolution2d: input and output types mismatched."); const DataType inputType = input.GetDataType(); - if (inputType == DataType::QuantisedAsymm8) + if (inputType == DataType::QAsymmU8) { std::array<DataType, 2> supportedWeightTypes = { - DataType::QuantisedAsymm8, + DataType::QAsymmU8, DataType::QuantizedSymm8PerAxis }; @@ -500,8 +500,8 @@ bool RefLayerSupport::IsDebugSupported(const TensorInfo& input, { DataType::Float16, DataType::Float32, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported, @@ -528,8 +528,8 @@ bool RefLayerSupport::IsDepthToSpaceSupported(const TensorInfo& input, { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported, @@ -558,8 +558,8 @@ bool RefLayerSupport::IsDepthwiseConvolutionSupported(const TensorInfo& input, { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported, @@ -572,11 +572,11 @@ bool RefLayerSupport::IsDepthwiseConvolutionSupported(const TensorInfo& input, "Reference DepthwiseConvolution2d: input and output types mismatched."); const DataType inputType = input.GetDataType(); - if (inputType == DataType::QuantisedAsymm8) + if (inputType == DataType::QAsymmU8) { std::array<DataType, 2> supportedWeightTypes = { - DataType::QuantisedAsymm8, + DataType::QAsymmU8, DataType::QuantizedSymm8PerAxis }; @@ -616,9 +616,9 @@ bool RefLayerSupport::IsDequantizeSupported(const TensorInfo& input, bool supported = true; std::array<DataType,3> supportedInputTypes = { - DataType::QuantisedAsymm8, + DataType::QAsymmU8, DataType::QSymmS8, - DataType::QuantisedSymm16 + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input, supportedInputTypes), reasonIfUnsupported, @@ -655,8 +655,8 @@ bool RefLayerSupport::IsDetectionPostProcessSupported(const TensorInfo& boxEncod std::array<DataType,3> supportedInputTypes = { DataType::Float32, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(boxEncodings, supportedInputTypes), reasonIfUnsupported, @@ -688,8 +688,8 @@ bool RefLayerSupport::IsDivisionSupported(const TensorInfo& input0, std::array<DataType,4> supportedTypes = { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input0, supportedTypes), reasonIfUnsupported, @@ -754,7 +754,7 @@ bool RefLayerSupport::IsFloorSupported(const TensorInfo& input, { DataType::Float32, DataType::Float16, - DataType::QuantisedSymm16 + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported, @@ -780,8 +780,8 @@ bool RefLayerSupport::IsFullyConnectedSupported(const TensorInfo& input, { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported, @@ -834,8 +834,8 @@ bool RefLayerSupport::IsGatherSupported(const armnn::TensorInfo& input0, { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input0, supportedTypes), reasonIfUnsupported, @@ -913,8 +913,8 @@ bool RefLayerSupport::IsL2NormalizationSupported(const TensorInfo& input, { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; bool supported = true; @@ -979,7 +979,7 @@ bool RefLayerSupport::IsLstmSupported(const TensorInfo& input, std::array<DataType,2> supportedTypes = { DataType::Float32, - DataType::QuantisedSymm16 + DataType::QSymmS16 }; // check inputs and outputs @@ -1081,8 +1081,8 @@ bool RefLayerSupport::IsMaximumSupported(const TensorInfo& input0, std::array<DataType,4> supportedTypes = { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input0, supportedTypes), reasonIfUnsupported, @@ -1119,8 +1119,8 @@ bool RefLayerSupport::IsMeanSupported(const TensorInfo& input, { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported, @@ -1185,8 +1185,8 @@ bool RefLayerSupport::IsMemCopySupported(const TensorInfo &input, { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16, + DataType::QAsymmU8, + DataType::QSymmS16, DataType::Boolean }; @@ -1212,8 +1212,8 @@ bool RefLayerSupport::IsMinimumSupported(const TensorInfo& input0, std::array<DataType,4> supportedTypes = { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input0, supportedTypes), reasonIfUnsupported, @@ -1247,8 +1247,8 @@ bool RefLayerSupport::IsMultiplicationSupported(const TensorInfo& input0, std::array<DataType,4> supportedTypes = { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input0, supportedTypes), reasonIfUnsupported, @@ -1284,8 +1284,8 @@ bool RefLayerSupport::IsNormalizationSupported(const TensorInfo& input, { DataType::Float16, DataType::Float32, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; bool supported = true; @@ -1322,8 +1322,8 @@ bool RefLayerSupport::IsPadSupported(const TensorInfo& input, { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported, @@ -1350,8 +1350,8 @@ bool RefLayerSupport::IsPermuteSupported(const TensorInfo& input, std::array<DataType,3> supportedTypes = { DataType::Float32, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported, @@ -1379,8 +1379,8 @@ bool RefLayerSupport::IsPooling2dSupported(const TensorInfo& input, { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported, @@ -1411,9 +1411,9 @@ bool RefLayerSupport::IsQuantizeSupported(const TensorInfo& input, // Define supported output types. std::array<DataType,3> supportedOutputTypes = { - DataType::QuantisedAsymm8, + DataType::QAsymmU8, DataType::QSymmS8, - DataType::QuantisedSymm16 + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(output, supportedOutputTypes), reasonIfUnsupported, "Reference quantize: output type not supported."); @@ -1435,8 +1435,8 @@ bool RefLayerSupport::IsReshapeSupported(const TensorInfo& input, DataType::Float32, DataType::Float16, DataType::Signed32, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; return CheckSupportRule(TypeAnyOf(input, supportedOutputTypes), reasonIfUnsupported, "Reference reshape: input type not supported."); @@ -1451,8 +1451,8 @@ bool RefLayerSupport::IsResizeBilinearSupported(const TensorInfo& input, { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported, @@ -1478,8 +1478,8 @@ bool RefLayerSupport::IsResizeSupported(const TensorInfo& input, { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported, @@ -1503,8 +1503,8 @@ bool RefLayerSupport::IsRsqrtSupported(const TensorInfo& input, { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported, @@ -1533,8 +1533,8 @@ bool RefLayerSupport::IsSliceSupported(const TensorInfo& input, std::array<DataType, 3> supportedTypes = { DataType::Float32, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported, @@ -1560,8 +1560,8 @@ bool RefLayerSupport::IsSoftmaxSupported(const TensorInfo& input, { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported, @@ -1587,8 +1587,8 @@ bool RefLayerSupport::IsSpaceToBatchNdSupported(const TensorInfo& input, { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported, @@ -1616,8 +1616,8 @@ bool RefLayerSupport::IsSpaceToDepthSupported(const TensorInfo& input, { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported, @@ -1642,8 +1642,8 @@ bool RefLayerSupport::IsSplitterSupported(const TensorInfo& input, { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported, @@ -1663,8 +1663,8 @@ bool RefLayerSupport::IsSplitterSupported(const TensorInfo& input, { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported, @@ -1693,8 +1693,8 @@ bool RefLayerSupport::IsStackSupported(const std::vector<const TensorInfo*>& inp { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(output, supportedTypes), reasonIfUnsupported, @@ -1723,8 +1723,8 @@ bool RefLayerSupport::IsStridedSliceSupported(const TensorInfo& input, std::array<DataType,3> supportedTypes = { DataType::Float32, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported, @@ -1749,8 +1749,8 @@ bool RefLayerSupport::IsSubtractionSupported(const TensorInfo& input0, std::array<DataType,4> supportedTypes = { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input0, supportedTypes), reasonIfUnsupported, @@ -1785,8 +1785,8 @@ bool RefLayerSupport::IsPreluSupported(const TensorInfo& input, { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported, @@ -1821,8 +1821,8 @@ bool RefLayerSupport::IsTransposeConvolution2dSupported(const TensorInfo& input, { DataType::Float32, DataType::Float16, - DataType::QuantisedAsymm8, - DataType::QuantisedSymm16 + DataType::QAsymmU8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input, supportedTypes), reasonIfUnsupported, @@ -1836,11 +1836,11 @@ bool RefLayerSupport::IsTransposeConvolution2dSupported(const TensorInfo& input, const DataType inputType = input.GetDataType(); - if (inputType == DataType::QuantisedAsymm8) + if (inputType == DataType::QAsymmU8) { std::array<DataType, 2> supportedWeightTypes = { - DataType::QuantisedAsymm8, + DataType::QAsymmU8, DataType::QuantizedSymm8PerAxis }; diff --git a/src/backends/reference/RefWorkloadFactory.cpp b/src/backends/reference/RefWorkloadFactory.cpp index dffb13db2d..b3a0c859fb 100644 --- a/src/backends/reference/RefWorkloadFactory.cpp +++ b/src/backends/reference/RefWorkloadFactory.cpp @@ -51,7 +51,7 @@ bool IsFloat16(const WorkloadInfo& info) bool IsQSymm16(const WorkloadInfo& info) { - return IsDataType<DataType::QuantisedSymm16>(info); + return IsDataType<DataType::QSymmS16>(info); } RefWorkloadFactory::RefWorkloadFactory(const std::shared_ptr<RefMemoryManager>& memoryManager) diff --git a/src/backends/reference/test/RefCreateWorkloadTests.cpp b/src/backends/reference/test/RefCreateWorkloadTests.cpp index 580d8550f0..23a8e9b9e9 100644 --- a/src/backends/reference/test/RefCreateWorkloadTests.cpp +++ b/src/backends/reference/test/RefCreateWorkloadTests.cpp @@ -68,7 +68,7 @@ BOOST_AUTO_TEST_CASE(CreateActivationFloat32Workload) BOOST_AUTO_TEST_CASE(CreateActivationUint8Workload) { - RefCreateActivationWorkloadTest<RefActivationWorkload, armnn::DataType::QuantisedAsymm8>(); + RefCreateActivationWorkloadTest<RefActivationWorkload, armnn::DataType::QAsymmU8>(); } template <typename WorkloadType, @@ -101,7 +101,7 @@ BOOST_AUTO_TEST_CASE(CreateAdditionUint8Workload) RefCreateElementwiseWorkloadTest<RefAdditionWorkload, AdditionQueueDescriptor, AdditionLayer, - armnn::DataType::QuantisedAsymm8>(); + armnn::DataType::QAsymmU8>(); } BOOST_AUTO_TEST_CASE(CreateAdditionInt16Workload) @@ -109,7 +109,7 @@ BOOST_AUTO_TEST_CASE(CreateAdditionInt16Workload) RefCreateElementwiseWorkloadTest<RefAdditionWorkload, AdditionQueueDescriptor, AdditionLayer, - armnn::DataType::QuantisedSymm16>(); + armnn::DataType::QSymmS16>(); } BOOST_AUTO_TEST_CASE(CreateSubtractionFloat32Workload) @@ -133,7 +133,7 @@ BOOST_AUTO_TEST_CASE(CreateSubtractionUint8Workload) RefCreateElementwiseWorkloadTest<RefSubtractionWorkload, SubtractionQueueDescriptor, SubtractionLayer, - armnn::DataType::QuantisedAsymm8>(); + armnn::DataType::QAsymmU8>(); } BOOST_AUTO_TEST_CASE(CreateSubtractionInt16Workload) @@ -141,7 +141,7 @@ BOOST_AUTO_TEST_CASE(CreateSubtractionInt16Workload) RefCreateElementwiseWorkloadTest<RefSubtractionWorkload, SubtractionQueueDescriptor, SubtractionLayer, - armnn::DataType::QuantisedSymm16>(); + armnn::DataType::QSymmS16>(); } BOOST_AUTO_TEST_CASE(CreateMultiplicationFloatWorkload) @@ -157,7 +157,7 @@ BOOST_AUTO_TEST_CASE(CreateMultiplicationUint8Workload) RefCreateElementwiseWorkloadTest<RefMultiplicationWorkload, MultiplicationQueueDescriptor, MultiplicationLayer, - armnn::DataType::QuantisedAsymm8>(); + armnn::DataType::QAsymmU8>(); } BOOST_AUTO_TEST_CASE(CreateMultiplicationInt16Workload) @@ -165,7 +165,7 @@ BOOST_AUTO_TEST_CASE(CreateMultiplicationInt16Workload) RefCreateElementwiseWorkloadTest<RefMultiplicationWorkload, MultiplicationQueueDescriptor, MultiplicationLayer, - armnn::DataType::QuantisedSymm16>(); + armnn::DataType::QSymmS16>(); } BOOST_AUTO_TEST_CASE(CreateDivisionFloat32Workload) @@ -189,7 +189,7 @@ BOOST_AUTO_TEST_CASE(CreateDivisionUint8Workload) RefCreateElementwiseWorkloadTest<RefDivisionWorkload, DivisionQueueDescriptor, DivisionLayer, - armnn::DataType::QuantisedAsymm8>(); + armnn::DataType::QAsymmU8>(); } BOOST_AUTO_TEST_CASE(CreateDivisionInt16Workload) @@ -197,7 +197,7 @@ BOOST_AUTO_TEST_CASE(CreateDivisionInt16Workload) RefCreateElementwiseWorkloadTest<RefDivisionWorkload, DivisionQueueDescriptor, DivisionLayer, - armnn::DataType::QuantisedSymm16>(); + armnn::DataType::QSymmS16>(); } template <typename BatchNormalizationWorkloadType, armnn::DataType DataType> @@ -255,25 +255,25 @@ BOOST_AUTO_TEST_CASE(CreateBatchNormalizationFloat16WorkloadNhwc) BOOST_AUTO_TEST_CASE(CreateBatchNormalizationUint8Workload) { - RefCreateBatchNormalizationWorkloadTest<RefBatchNormalizationWorkload, armnn::DataType::QuantisedAsymm8> + RefCreateBatchNormalizationWorkloadTest<RefBatchNormalizationWorkload, armnn::DataType::QAsymmU8> (DataLayout::NCHW); } BOOST_AUTO_TEST_CASE(CreateBatchNormalizationUint8WorkloadNhwc) { - RefCreateBatchNormalizationWorkloadTest<RefBatchNormalizationWorkload, armnn::DataType::QuantisedAsymm8> + RefCreateBatchNormalizationWorkloadTest<RefBatchNormalizationWorkload, armnn::DataType::QAsymmU8> (DataLayout::NHWC); } BOOST_AUTO_TEST_CASE(CreateBatchNormalizationInt16Workload) { - RefCreateBatchNormalizationWorkloadTest<RefBatchNormalizationWorkload, armnn::DataType::QuantisedSymm16> + RefCreateBatchNormalizationWorkloadTest<RefBatchNormalizationWorkload, armnn::DataType::QSymmS16> (DataLayout::NCHW); } BOOST_AUTO_TEST_CASE(CreateBatchNormalizationInt16WorkloadNhwc) { - RefCreateBatchNormalizationWorkloadTest<RefBatchNormalizationWorkload, armnn::DataType::QuantisedSymm16> + RefCreateBatchNormalizationWorkloadTest<RefBatchNormalizationWorkload, armnn::DataType::QSymmS16> (DataLayout::NHWC); } @@ -358,8 +358,8 @@ static void RefCreateFullyConnectedWorkloadTest() auto workload = CreateFullyConnectedWorkloadTest<FullyConnectedWorkloadType, DataType>(factory, graph); // Checks that outputs and inputs are as we expect them (see definition of CreateFullyConnectedWorkloadTest). - float inputsQScale = DataType == armnn::DataType::QuantisedAsymm8 ? 1.0f : 0.0; - float outputQScale = DataType == armnn::DataType::QuantisedAsymm8 ? 2.0f : 0.0; + float inputsQScale = DataType == armnn::DataType::QAsymmU8 ? 1.0f : 0.0; + float outputQScale = DataType == armnn::DataType::QAsymmU8 ? 2.0f : 0.0; CheckInputOutput(std::move(workload), TensorInfo({ 3, 1, 4, 5 }, DataType, inputsQScale), TensorInfo({ 3, 7 }, DataType, outputQScale)); @@ -372,12 +372,12 @@ BOOST_AUTO_TEST_CASE(CreateFullyConnectedWorkloadFloat32) BOOST_AUTO_TEST_CASE(CreateFullyConnectedWorkloadQuantisedAsymm8) { - RefCreateFullyConnectedWorkloadTest<RefFullyConnectedWorkload, armnn::DataType::QuantisedAsymm8>(); + RefCreateFullyConnectedWorkloadTest<RefFullyConnectedWorkload, armnn::DataType::QAsymmU8>(); } BOOST_AUTO_TEST_CASE(CreateFullyConnectedWorkloadQuantisedSymm16) { - RefCreateFullyConnectedWorkloadTest<RefFullyConnectedWorkload, armnn::DataType::QuantisedSymm16>(); + RefCreateFullyConnectedWorkloadTest<RefFullyConnectedWorkload, armnn::DataType::QSymmS16>(); } template <typename NormalizationWorkloadType, armnn::DataType DataType> @@ -419,22 +419,22 @@ BOOST_AUTO_TEST_CASE(CreateRefNormalizationFloat32NhwcWorkload) BOOST_AUTO_TEST_CASE(CreateRefNormalizationUint8NchwWorkload) { - RefCreateNormalizationWorkloadTest<RefNormalizationWorkload, armnn::DataType::QuantisedAsymm8>(DataLayout::NCHW); + RefCreateNormalizationWorkloadTest<RefNormalizationWorkload, armnn::DataType::QAsymmU8>(DataLayout::NCHW); } BOOST_AUTO_TEST_CASE(CreateRefNormalizationUint8NhwcWorkload) { - RefCreateNormalizationWorkloadTest<RefNormalizationWorkload, armnn::DataType::QuantisedAsymm8>(DataLayout::NHWC); + RefCreateNormalizationWorkloadTest<RefNormalizationWorkload, armnn::DataType::QAsymmU8>(DataLayout::NHWC); } BOOST_AUTO_TEST_CASE(CreateRefNormalizationInt16NchwWorkload) { - RefCreateNormalizationWorkloadTest<RefNormalizationWorkload, armnn::DataType::QuantisedSymm16>(DataLayout::NCHW); + RefCreateNormalizationWorkloadTest<RefNormalizationWorkload, armnn::DataType::QSymmS16>(DataLayout::NCHW); } BOOST_AUTO_TEST_CASE(CreateRefNormalizationInt16NhwcWorkload) { - RefCreateNormalizationWorkloadTest<RefNormalizationWorkload, armnn::DataType::QuantisedSymm16>(DataLayout::NHWC); + RefCreateNormalizationWorkloadTest<RefNormalizationWorkload, armnn::DataType::QSymmS16>(DataLayout::NHWC); } template <typename Pooling2dWorkloadType, armnn::DataType DataType> @@ -477,22 +477,22 @@ BOOST_AUTO_TEST_CASE(CreatePooling2dFloat32NhwcWorkload) BOOST_AUTO_TEST_CASE(CreatePooling2dUint8Workload) { - RefCreatePooling2dWorkloadTest<RefPooling2dWorkload, armnn::DataType::QuantisedAsymm8>(DataLayout::NCHW); + RefCreatePooling2dWorkloadTest<RefPooling2dWorkload, armnn::DataType::QAsymmU8>(DataLayout::NCHW); } BOOST_AUTO_TEST_CASE(CreatePooling2dUint8NhwcWorkload) { - RefCreatePooling2dWorkloadTest<RefPooling2dWorkload, armnn::DataType::QuantisedAsymm8>(DataLayout::NHWC); + RefCreatePooling2dWorkloadTest<RefPooling2dWorkload, armnn::DataType::QAsymmU8>(DataLayout::NHWC); } BOOST_AUTO_TEST_CASE(CreatePooling2dInt16Workload) { - RefCreatePooling2dWorkloadTest<RefPooling2dWorkload, armnn::DataType::QuantisedSymm16>(DataLayout::NCHW); + RefCreatePooling2dWorkloadTest<RefPooling2dWorkload, armnn::DataType::QSymmS16>(DataLayout::NCHW); } BOOST_AUTO_TEST_CASE(CreatePooling2dInt16NhwcWorkload) { - RefCreatePooling2dWorkloadTest<RefPooling2dWorkload, armnn::DataType::QuantisedSymm16>(DataLayout::NHWC); + RefCreatePooling2dWorkloadTest<RefPooling2dWorkload, armnn::DataType::QSymmS16>(DataLayout::NHWC); } template <typename SoftmaxWorkloadType, armnn::DataType DataType> @@ -521,12 +521,12 @@ BOOST_AUTO_TEST_CASE(CreateSoftmaxFloat16Workload) BOOST_AUTO_TEST_CASE(CreateSoftmaxQuantisedAsymm8Workload) { - RefCreateSoftmaxWorkloadTest<RefSoftmaxWorkload, armnn::DataType::QuantisedAsymm8>(); + RefCreateSoftmaxWorkloadTest<RefSoftmaxWorkload, armnn::DataType::QAsymmU8>(); } BOOST_AUTO_TEST_CASE(CreateSoftmaxQuantisedSymm16Workload) { - RefCreateSoftmaxWorkloadTest<RefSoftmaxWorkload, armnn::DataType::QuantisedSymm16>(); + RefCreateSoftmaxWorkloadTest<RefSoftmaxWorkload, armnn::DataType::QSymmS16>(); } template <typename SplitterWorkloadType, armnn::DataType DataType> @@ -563,7 +563,7 @@ BOOST_AUTO_TEST_CASE(CreateSplitterFloat16Workload) BOOST_AUTO_TEST_CASE(CreateSplitterUint8Workload) { - RefCreateSplitterWorkloadTest<RefSplitterWorkload, armnn::DataType::QuantisedAsymm8>(); + RefCreateSplitterWorkloadTest<RefSplitterWorkload, armnn::DataType::QAsymmU8>(); } template <typename SplitterWorkloadType, typename ConcatWorkloadType, armnn::DataType DataType> @@ -611,7 +611,7 @@ BOOST_AUTO_TEST_CASE(CreateSplitterConcatFloat16) BOOST_AUTO_TEST_CASE(CreateSplitterConcatUint8) { - RefCreateSplitterConcatWorkloadTest<RefSplitterWorkload, RefConcatWorkload, DataType::QuantisedAsymm8>(); + RefCreateSplitterConcatWorkloadTest<RefSplitterWorkload, RefConcatWorkload, DataType::QAsymmU8>(); } template <typename SplitterWorkloadType, typename ActivationWorkloadType, armnn::DataType DataType> @@ -661,7 +661,7 @@ BOOST_AUTO_TEST_CASE(CreateSingleOutputMultipleInputsFloat32) BOOST_AUTO_TEST_CASE(CreateSingleOutputMultipleInputsUint8) { RefCreateSingleOutputMultipleInputsTest<RefSplitterWorkload, RefActivationWorkload, - armnn::DataType::QuantisedAsymm8>(); + armnn::DataType::QAsymmU8>(); } template <typename ResizeBilinearWorkloadType, armnn::DataType DataType> @@ -704,12 +704,12 @@ BOOST_AUTO_TEST_CASE(CreateResizeBilinearFloat16) BOOST_AUTO_TEST_CASE(CreateResizeBilinearUint8) { - RefCreateResizeBilinearTest<RefResizeWorkload, armnn::DataType::QuantisedAsymm8>(DataLayout::NCHW); + RefCreateResizeBilinearTest<RefResizeWorkload, armnn::DataType::QAsymmU8>(DataLayout::NCHW); } BOOST_AUTO_TEST_CASE(CreateResizeBilinearQuantisedAsymm16) { - RefCreateResizeBilinearTest<RefResizeWorkload, armnn::DataType::QuantisedSymm16>(DataLayout::NCHW); + RefCreateResizeBilinearTest<RefResizeWorkload, armnn::DataType::QSymmS16>(DataLayout::NCHW); } BOOST_AUTO_TEST_CASE(CreateResizeBilinearFloat32Nhwc) @@ -744,12 +744,12 @@ BOOST_AUTO_TEST_CASE(CreateRsqrtFloat16) BOOST_AUTO_TEST_CASE(CreateRsqrtUint8) { - RefCreateRsqrtTest<RefRsqrtWorkload, armnn::DataType::QuantisedAsymm8>(); + RefCreateRsqrtTest<RefRsqrtWorkload, armnn::DataType::QAsymmU8>(); } BOOST_AUTO_TEST_CASE(CreateRsqrtQsymm16) { - RefCreateRsqrtTest<RefRsqrtWorkload, armnn::DataType::QuantisedSymm16>(); + RefCreateRsqrtTest<RefRsqrtWorkload, armnn::DataType::QSymmS16>(); } template <typename BatchToSpaceNdWorkloadType, armnn::DataType DataType> @@ -777,12 +777,12 @@ BOOST_AUTO_TEST_CASE(CreateBatchToSpaceNdFloat16) BOOST_AUTO_TEST_CASE(CreateBatchToSpaceNdUint8) { - RefCreateBatchToSpaceNdTest<RefBatchToSpaceNdWorkload, armnn::DataType::QuantisedAsymm8>(); + RefCreateBatchToSpaceNdTest<RefBatchToSpaceNdWorkload, armnn::DataType::QAsymmU8>(); } BOOST_AUTO_TEST_CASE(CreateBatchToSpaceNdQSymm16) { - RefCreateBatchToSpaceNdTest<RefBatchToSpaceNdWorkload, armnn::DataType::QuantisedSymm16>(); + RefCreateBatchToSpaceNdTest<RefBatchToSpaceNdWorkload, armnn::DataType::QSymmS16>(); } template <typename L2NormalizationWorkloadType, armnn::DataType DataType> @@ -825,22 +825,22 @@ BOOST_AUTO_TEST_CASE(CreateL2NormalizationFloat32Nhwc) BOOST_AUTO_TEST_CASE(CreateL2NormalizationInt16) { - RefCreateL2NormalizationTest<RefL2NormalizationWorkload, armnn::DataType::QuantisedSymm16>(DataLayout::NCHW); + RefCreateL2NormalizationTest<RefL2NormalizationWorkload, armnn::DataType::QSymmS16>(DataLayout::NCHW); } BOOST_AUTO_TEST_CASE(CreateL2NormalizationInt16Nhwc) { - RefCreateL2NormalizationTest<RefL2NormalizationWorkload, armnn::DataType::QuantisedSymm16>(DataLayout::NHWC); + RefCreateL2NormalizationTest<RefL2NormalizationWorkload, armnn::DataType::QSymmS16>(DataLayout::NHWC); } BOOST_AUTO_TEST_CASE(CreateL2NormalizationUint8) { - RefCreateL2NormalizationTest<RefL2NormalizationWorkload, armnn::DataType::QuantisedAsymm8>(DataLayout::NCHW); + RefCreateL2NormalizationTest<RefL2NormalizationWorkload, armnn::DataType::QAsymmU8>(DataLayout::NCHW); } BOOST_AUTO_TEST_CASE(CreateL2NormalizationUint8Nhwc) { - RefCreateL2NormalizationTest<RefL2NormalizationWorkload, armnn::DataType::QuantisedAsymm8>(DataLayout::NHWC); + RefCreateL2NormalizationTest<RefL2NormalizationWorkload, armnn::DataType::QAsymmU8>(DataLayout::NHWC); } template <typename ReshapeWorkloadType, armnn::DataType DataType> @@ -864,12 +864,12 @@ BOOST_AUTO_TEST_CASE(CreateReshapeWorkloadFloat32) BOOST_AUTO_TEST_CASE(CreateReshapeWorkloadQuantisedAsymm8) { - RefCreateReshapeWorkloadTest<RefReshapeWorkload, armnn::DataType::QuantisedAsymm8>(); + RefCreateReshapeWorkloadTest<RefReshapeWorkload, armnn::DataType::QAsymmU8>(); } BOOST_AUTO_TEST_CASE(CreateReshapeWorkloadQuantisedSymm16) { - RefCreateReshapeWorkloadTest<RefReshapeWorkload, armnn::DataType::QuantisedSymm16>(); + RefCreateReshapeWorkloadTest<RefReshapeWorkload, armnn::DataType::QSymmS16>(); } template <typename ConcatWorkloadType, armnn::DataType DataType> @@ -898,12 +898,12 @@ BOOST_AUTO_TEST_CASE(CreateConcatDim0Float16Workload) BOOST_AUTO_TEST_CASE(CreateConcatDim0Uint8Workload) { - RefCreateConcatWorkloadTest<RefConcatWorkload, armnn::DataType::QuantisedAsymm8>({ 4, 3, 2, 5 }, 0); + RefCreateConcatWorkloadTest<RefConcatWorkload, armnn::DataType::QAsymmU8>({ 4, 3, 2, 5 }, 0); } BOOST_AUTO_TEST_CASE(CreateConcatDim0Uint16Workload) { - RefCreateConcatWorkloadTest<RefConcatWorkload, armnn::DataType::QuantisedSymm16>({ 4, 3, 2, 5 }, 0); + RefCreateConcatWorkloadTest<RefConcatWorkload, armnn::DataType::QSymmS16>({ 4, 3, 2, 5 }, 0); } BOOST_AUTO_TEST_CASE(CreateConcatDim1Float32Workload) @@ -913,7 +913,7 @@ BOOST_AUTO_TEST_CASE(CreateConcatDim1Float32Workload) BOOST_AUTO_TEST_CASE(CreateConcatDim1Uint8Workload) { - RefCreateConcatWorkloadTest<RefConcatWorkload, armnn::DataType::QuantisedAsymm8>({ 2, 6, 2, 5 }, 1); + RefCreateConcatWorkloadTest<RefConcatWorkload, armnn::DataType::QAsymmU8>({ 2, 6, 2, 5 }, 1); } BOOST_AUTO_TEST_CASE(CreateConcatDim2Float32Workload) @@ -923,7 +923,7 @@ BOOST_AUTO_TEST_CASE(CreateConcatDim2Float32Workload) BOOST_AUTO_TEST_CASE(CreateConcatDim2Uint8Workload) { - RefCreateConcatWorkloadTest<RefConcatWorkload, armnn::DataType::QuantisedAsymm8>({ 2, 3, 4, 5 }, 2); + RefCreateConcatWorkloadTest<RefConcatWorkload, armnn::DataType::QAsymmU8>({ 2, 3, 4, 5 }, 2); } BOOST_AUTO_TEST_CASE(CreateConcatDim3Float32Workload) @@ -933,7 +933,7 @@ BOOST_AUTO_TEST_CASE(CreateConcatDim3Float32Workload) BOOST_AUTO_TEST_CASE(CreateConcatDim3Uint8Workload) { - RefCreateConcatWorkloadTest<RefConcatWorkload, armnn::DataType::QuantisedAsymm8>({ 2, 3, 2, 10 }, 3); + RefCreateConcatWorkloadTest<RefConcatWorkload, armnn::DataType::QAsymmU8>({ 2, 3, 2, 10 }, 3); } template <typename ConstantWorkloadType, armnn::DataType DataType> @@ -951,12 +951,12 @@ static void RefCreateConstantWorkloadTest(const armnn::TensorShape& outputShape) BOOST_AUTO_TEST_CASE(CreateConstantUint8Workload) { - RefCreateConstantWorkloadTest<RefConstantWorkload, armnn::DataType::QuantisedAsymm8>({ 2, 3, 2, 10 }); + RefCreateConstantWorkloadTest<RefConstantWorkload, armnn::DataType::QAsymmU8>({ 2, 3, 2, 10 }); } BOOST_AUTO_TEST_CASE(CreateConstantInt16Workload) { - RefCreateConstantWorkloadTest<RefConstantWorkload, armnn::DataType::QuantisedSymm16>({ 2, 3, 2, 10 }); + RefCreateConstantWorkloadTest<RefConstantWorkload, armnn::DataType::QSymmS16>({ 2, 3, 2, 10 }); } BOOST_AUTO_TEST_CASE(CreateConstantFloat32Workload) @@ -1001,12 +1001,12 @@ BOOST_AUTO_TEST_CASE(CreatePreluFloat16Workload) BOOST_AUTO_TEST_CASE(CreatePreluUint8Workload) { - RefCreatePreluWorkloadTest({ 1, 4, 1, 2 }, { 5, 4, 3, 1 }, { 5, 4, 3, 2 }, armnn::DataType::QuantisedAsymm8); + RefCreatePreluWorkloadTest({ 1, 4, 1, 2 }, { 5, 4, 3, 1 }, { 5, 4, 3, 2 }, armnn::DataType::QAsymmU8); } BOOST_AUTO_TEST_CASE(CreatePreluInt16Workload) { - RefCreatePreluWorkloadTest({ 1, 4, 1, 2 }, { 5, 4, 3, 1 }, { 5, 4, 3, 2 }, armnn::DataType::QuantisedSymm16); + RefCreatePreluWorkloadTest({ 1, 4, 1, 2 }, { 5, 4, 3, 1 }, { 5, 4, 3, 2 }, armnn::DataType::QSymmS16); } BOOST_AUTO_TEST_CASE(CreatePreluFloat32NoBroadcastWorkload) @@ -1026,14 +1026,14 @@ BOOST_AUTO_TEST_CASE(CreatePreluFloat16NoBroadcastWorkload) BOOST_AUTO_TEST_CASE(CreatePreluUint8NoBroadcastWorkload) { BOOST_CHECK_THROW(RefCreatePreluWorkloadTest({ 1, 4, 7, 2 }, { 5, 4, 3, 1 }, { 5, 4, 3, 2 }, - armnn::DataType::QuantisedAsymm8), + armnn::DataType::QAsymmU8), armnn::InvalidArgumentException); } BOOST_AUTO_TEST_CASE(CreatePreluInt16NoBroadcastWorkload) { BOOST_CHECK_THROW(RefCreatePreluWorkloadTest({ 1, 4, 7, 2 }, { 5, 4, 3, 1 }, { 5, 4, 3, 2 }, - armnn::DataType::QuantisedSymm16), + armnn::DataType::QSymmS16), armnn::InvalidArgumentException); } @@ -1062,12 +1062,12 @@ BOOST_AUTO_TEST_CASE(CreateSpaceToDepthWorkloadFloat16) BOOST_AUTO_TEST_CASE(CreateSpaceToDepthWorkloadQASymm8) { - RefCreateSpaceToDepthWorkloadTest<RefSpaceToDepthWorkload, armnn::DataType::QuantisedAsymm8>(); + RefCreateSpaceToDepthWorkloadTest<RefSpaceToDepthWorkload, armnn::DataType::QAsymmU8>(); } BOOST_AUTO_TEST_CASE(CreateSpaceToDepthWorkloadQSymm16) { - RefCreateSpaceToDepthWorkloadTest<RefSpaceToDepthWorkload, armnn::DataType::QuantisedSymm16>(); + RefCreateSpaceToDepthWorkloadTest<RefSpaceToDepthWorkload, armnn::DataType::QSymmS16>(); } template <armnn::DataType DataType> @@ -1103,12 +1103,12 @@ BOOST_AUTO_TEST_CASE(CreateStackFloat32Workload) BOOST_AUTO_TEST_CASE(CreateStackUint8Workload) { - RefCreateStackWorkloadTest<armnn::DataType::QuantisedAsymm8>({ 3, 4, 5 }, { 3, 4, 2, 5 }, 2, 2); + RefCreateStackWorkloadTest<armnn::DataType::QAsymmU8>({ 3, 4, 5 }, { 3, 4, 2, 5 }, 2, 2); } BOOST_AUTO_TEST_CASE(CreateStackUint16Workload) { - RefCreateStackWorkloadTest<armnn::DataType::QuantisedSymm16>({ 3, 4, 5 }, { 3, 4, 2, 5 }, 2, 2); + RefCreateStackWorkloadTest<armnn::DataType::QSymmS16>({ 3, 4, 5 }, { 3, 4, 2, 5 }, 2, 2); } BOOST_AUTO_TEST_SUITE_END() diff --git a/src/backends/reference/test/RefEndToEndTests.cpp b/src/backends/reference/test/RefEndToEndTests.cpp index 4d8c82d900..75eccdee88 100644 --- a/src/backends/reference/test/RefEndToEndTests.cpp +++ b/src/backends/reference/test/RefEndToEndTests.cpp @@ -37,12 +37,12 @@ BOOST_AUTO_TEST_CASE(RefAbsEndToEndTestFloat32) BOOST_AUTO_TEST_CASE(RefAbsEndToEndTestUint8) { - AbsEndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends); + AbsEndToEnd<armnn::DataType::QAsymmU8>(defaultBackends); } BOOST_AUTO_TEST_CASE(RefAbsEndToEndTestInt16) { - AbsEndToEnd<armnn::DataType::QuantisedSymm16>(defaultBackends); + AbsEndToEnd<armnn::DataType::QSymmS16>(defaultBackends); } // Constant @@ -75,12 +75,12 @@ BOOST_AUTO_TEST_CASE(Unsigned8) softmax->GetOutputSlot(0).Connect(output->GetInputSlot(0)); // Sets the tensors in the network. - TensorInfo inputTensorInfo(TensorShape({1, 5}), DataType::QuantisedAsymm8); + TensorInfo inputTensorInfo(TensorShape({1, 5}), DataType::QAsymmU8); inputTensorInfo.SetQuantizationOffset(100); inputTensorInfo.SetQuantizationScale(10000.0f); input->GetOutputSlot(0).SetTensorInfo(inputTensorInfo); - TensorInfo outputTensorInfo(TensorShape({1, 5}), DataType::QuantisedAsymm8); + TensorInfo outputTensorInfo(TensorShape({1, 5}), DataType::QAsymmU8); outputTensorInfo.SetQuantizationOffset(0); outputTensorInfo.SetQuantizationScale(1.0f/255.0f); softmax->GetOutputSlot(0).SetTensorInfo(outputTensorInfo); @@ -369,7 +369,7 @@ BOOST_AUTO_TEST_CASE(RefEqualSimpleEndToEndUint8Test) const std::vector<uint8_t> expectedOutput({ 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 1, 1, 1, 1 }); - ComparisonSimpleEndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends, + ComparisonSimpleEndToEnd<armnn::DataType::QAsymmU8>(defaultBackends, ComparisonOperation::Equal, expectedOutput); } @@ -379,7 +379,7 @@ BOOST_AUTO_TEST_CASE(RefGreaterSimpleEndToEndUint8Test) const std::vector<uint8_t> expectedOutput({ 0, 0, 0, 0, 1, 1, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0 }); - ComparisonSimpleEndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends, + ComparisonSimpleEndToEnd<armnn::DataType::QAsymmU8>(defaultBackends, ComparisonOperation::Greater, expectedOutput); } @@ -409,7 +409,7 @@ BOOST_AUTO_TEST_CASE(RefEqualBroadcastEndToEndUint8Test) const std::vector<uint8_t > expectedOutput({ 1, 0, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0 }); - ComparisonBroadcastEndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends, + ComparisonBroadcastEndToEnd<armnn::DataType::QAsymmU8>(defaultBackends, ComparisonOperation::Equal, expectedOutput); } @@ -419,7 +419,7 @@ BOOST_AUTO_TEST_CASE(RefGreaterBroadcastEndToEndUint8Test) const std::vector<uint8_t> expectedOutput({ 0, 1, 0, 0, 0, 1, 1, 1, 1, 1, 1, 1 }); - ComparisonBroadcastEndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends, + ComparisonBroadcastEndToEnd<armnn::DataType::QAsymmU8>(defaultBackends, ComparisonOperation::Greater, expectedOutput); } @@ -431,12 +431,12 @@ BOOST_AUTO_TEST_CASE(RefBatchToSpaceNdEndToEndFloat32NHWCTest) BOOST_AUTO_TEST_CASE(RefBatchToSpaceNdEndToEndUint8NHWCTest) { - BatchToSpaceNdEndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends, armnn::DataLayout::NHWC); + BatchToSpaceNdEndToEnd<armnn::DataType::QAsymmU8>(defaultBackends, armnn::DataLayout::NHWC); } BOOST_AUTO_TEST_CASE(RefBatchToSpaceNdEndToEndQSymm16NHWCTest) { - BatchToSpaceNdEndToEnd<armnn::DataType::QuantisedSymm16>(defaultBackends, armnn::DataLayout::NHWC); + BatchToSpaceNdEndToEnd<armnn::DataType::QSymmS16>(defaultBackends, armnn::DataLayout::NHWC); } BOOST_AUTO_TEST_CASE(RefBatchToSpaceNdEndToEndFloat32NCHWTest) @@ -446,12 +446,12 @@ BOOST_AUTO_TEST_CASE(RefBatchToSpaceNdEndToEndFloat32NCHWTest) BOOST_AUTO_TEST_CASE(RefBatchToSpaceNdEndToEndUint8NCHWTest) { - BatchToSpaceNdEndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends, armnn::DataLayout::NCHW); + BatchToSpaceNdEndToEnd<armnn::DataType::QAsymmU8>(defaultBackends, armnn::DataLayout::NCHW); } BOOST_AUTO_TEST_CASE(RefBatchToSpaceNdEndToEndQSymm16NCHWTest) { - BatchToSpaceNdEndToEnd<armnn::DataType::QuantisedSymm16>(defaultBackends, armnn::DataLayout::NCHW); + BatchToSpaceNdEndToEnd<armnn::DataType::QSymmS16>(defaultBackends, armnn::DataLayout::NCHW); } BOOST_AUTO_TEST_CASE(RefBatchToSpaceNdEndToEndComplexFloat32NHWCTest) @@ -461,12 +461,12 @@ BOOST_AUTO_TEST_CASE(RefBatchToSpaceNdEndToEndComplexFloat32NHWCTest) BOOST_AUTO_TEST_CASE(RefBatchToSpaceNdEndToEndComplexUint8NHWCTest) { - BatchToSpaceNdComplexEndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends, armnn::DataLayout::NHWC); + BatchToSpaceNdComplexEndToEnd<armnn::DataType::QAsymmU8>(defaultBackends, armnn::DataLayout::NHWC); } BOOST_AUTO_TEST_CASE(RefBatchToSpaceNdEndToEndComplexQSymm16NHWCTest) { - BatchToSpaceNdComplexEndToEnd<armnn::DataType::QuantisedSymm16>(defaultBackends, armnn::DataLayout::NHWC); + BatchToSpaceNdComplexEndToEnd<armnn::DataType::QSymmS16>(defaultBackends, armnn::DataLayout::NHWC); } BOOST_AUTO_TEST_CASE(RefBatchToSpaceNdEndToEndComplexFloat32NCHWTest) @@ -476,12 +476,12 @@ BOOST_AUTO_TEST_CASE(RefBatchToSpaceNdEndToEndComplexFloat32NCHWTest) BOOST_AUTO_TEST_CASE(RefBatchToSpaceNdEndToEndComplexUint8NCHWTest) { - BatchToSpaceNdComplexEndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends, armnn::DataLayout::NCHW); + BatchToSpaceNdComplexEndToEnd<armnn::DataType::QAsymmU8>(defaultBackends, armnn::DataLayout::NCHW); } BOOST_AUTO_TEST_CASE(RefBatchToSpaceNdEndToEndComplexQSymm16NCHWTest) { - BatchToSpaceNdComplexEndToEnd<armnn::DataType::QuantisedSymm16>(defaultBackends, armnn::DataLayout::NCHW); + BatchToSpaceNdComplexEndToEnd<armnn::DataType::QSymmS16>(defaultBackends, armnn::DataLayout::NCHW); } BOOST_AUTO_TEST_CASE(RefConcatEndToEndDim0Test) @@ -491,7 +491,7 @@ BOOST_AUTO_TEST_CASE(RefConcatEndToEndDim0Test) BOOST_AUTO_TEST_CASE(RefConcatEndToEndDim0Uint8Test) { - ConcatDim0EndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends); + ConcatDim0EndToEnd<armnn::DataType::QAsymmU8>(defaultBackends); } BOOST_AUTO_TEST_CASE(RefConcatEndToEndDim1Test) @@ -501,7 +501,7 @@ BOOST_AUTO_TEST_CASE(RefConcatEndToEndDim1Test) BOOST_AUTO_TEST_CASE(RefConcatEndToEndDim1Uint8Test) { - ConcatDim1EndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends); + ConcatDim1EndToEnd<armnn::DataType::QAsymmU8>(defaultBackends); } BOOST_AUTO_TEST_CASE(RefConcatEndToEndDim2Test) @@ -511,7 +511,7 @@ BOOST_AUTO_TEST_CASE(RefConcatEndToEndDim2Test) BOOST_AUTO_TEST_CASE(RefConcatEndToEndDim2Uint8Test) { - ConcatDim2EndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends); + ConcatDim2EndToEnd<armnn::DataType::QAsymmU8>(defaultBackends); } BOOST_AUTO_TEST_CASE(RefConcatEndToEndDim3Test) @@ -521,7 +521,7 @@ BOOST_AUTO_TEST_CASE(RefConcatEndToEndDim3Test) BOOST_AUTO_TEST_CASE(RefConcatEndToEndDim3Uint8Test) { - ConcatDim3EndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends); + ConcatDim3EndToEnd<armnn::DataType::QAsymmU8>(defaultBackends); } BOOST_AUTO_TEST_CASE(RefGatherFloatTest) @@ -531,12 +531,12 @@ BOOST_AUTO_TEST_CASE(RefGatherFloatTest) BOOST_AUTO_TEST_CASE(RefGatherUint8Test) { - GatherEndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends); + GatherEndToEnd<armnn::DataType::QAsymmU8>(defaultBackends); } BOOST_AUTO_TEST_CASE(RefGatherInt16Test) { - GatherEndToEnd<armnn::DataType::QuantisedSymm16>(defaultBackends); + GatherEndToEnd<armnn::DataType::QSymmS16>(defaultBackends); } BOOST_AUTO_TEST_CASE(RefGatherMultiDimFloatTest) @@ -546,12 +546,12 @@ BOOST_AUTO_TEST_CASE(RefGatherMultiDimFloatTest) BOOST_AUTO_TEST_CASE(RefGatherMultiDimUint8Test) { - GatherMultiDimEndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends); + GatherMultiDimEndToEnd<armnn::DataType::QAsymmU8>(defaultBackends); } BOOST_AUTO_TEST_CASE(RefGatherMultiDimInt16Test) { - GatherMultiDimEndToEnd<armnn::DataType::QuantisedSymm16>(defaultBackends); + GatherMultiDimEndToEnd<armnn::DataType::QSymmS16>(defaultBackends); } // DepthToSpace @@ -567,12 +567,12 @@ BOOST_AUTO_TEST_CASE(DephtToSpaceEndToEndNchwFloat16) BOOST_AUTO_TEST_CASE(DephtToSpaceEndToEndNchwUint8) { - DepthToSpaceEndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends, armnn::DataLayout::NCHW); + DepthToSpaceEndToEnd<armnn::DataType::QAsymmU8>(defaultBackends, armnn::DataLayout::NCHW); } BOOST_AUTO_TEST_CASE(DephtToSpaceEndToEndNchwInt16) { - DepthToSpaceEndToEnd<armnn::DataType::QuantisedSymm16>(defaultBackends, armnn::DataLayout::NCHW); + DepthToSpaceEndToEnd<armnn::DataType::QSymmS16>(defaultBackends, armnn::DataLayout::NCHW); } BOOST_AUTO_TEST_CASE(DephtToSpaceEndToEndNhwcFloat32) @@ -587,33 +587,33 @@ BOOST_AUTO_TEST_CASE(DephtToSpaceEndToEndNhwcFloat16) BOOST_AUTO_TEST_CASE(DephtToSpaceEndToEndNhwcUint8) { - DepthToSpaceEndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends, armnn::DataLayout::NHWC); + DepthToSpaceEndToEnd<armnn::DataType::QAsymmU8>(defaultBackends, armnn::DataLayout::NHWC); } BOOST_AUTO_TEST_CASE(DephtToSpaceEndToEndNhwcInt16) { - DepthToSpaceEndToEnd<armnn::DataType::QuantisedSymm16>(defaultBackends, armnn::DataLayout::NHWC); + DepthToSpaceEndToEnd<armnn::DataType::QSymmS16>(defaultBackends, armnn::DataLayout::NHWC); } // Dequantize BOOST_AUTO_TEST_CASE(DequantizeEndToEndSimpleTest) { - DequantizeEndToEndSimple<armnn::DataType::QuantisedAsymm8>(defaultBackends); + DequantizeEndToEndSimple<armnn::DataType::QAsymmU8>(defaultBackends); } BOOST_AUTO_TEST_CASE(DequantizeEndToEndOffsetTest) { - DequantizeEndToEndOffset<armnn::DataType::QuantisedAsymm8>(defaultBackends); + DequantizeEndToEndOffset<armnn::DataType::QAsymmU8>(defaultBackends); } BOOST_AUTO_TEST_CASE(DequantizeEndToEndSimpleInt16Test) { - DequantizeEndToEndSimple<armnn::DataType::QuantisedSymm16>(defaultBackends); + DequantizeEndToEndSimple<armnn::DataType::QSymmS16>(defaultBackends); } BOOST_AUTO_TEST_CASE(DequantizeEndToEndOffsetInt16Test) { - DequantizeEndToEndOffset<armnn::DataType::QuantisedSymm16>(defaultBackends); + DequantizeEndToEndOffset<armnn::DataType::QSymmS16>(defaultBackends); } BOOST_AUTO_TEST_CASE(RefDetectionPostProcessRegularNmsTest) @@ -697,7 +697,7 @@ BOOST_AUTO_TEST_CASE(RefDetectionPostProcessRegularNmsUint8Test) QuantizeData(qBoxEncodings.data(), boxEncodings.data(), boxEncodingsInfo); QuantizeData(qScores.data(), scores.data(), scoresInfo); QuantizeData(qAnchors.data(), anchors.data(), anchorsInfo); - DetectionPostProcessRegularNmsEndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends, qBoxEncodings, + DetectionPostProcessRegularNmsEndToEnd<armnn::DataType::QAsymmU8>(defaultBackends, qBoxEncodings, qScores, qAnchors, 1.0f, 1, 0.01f, 0, 0.5f, 0); } @@ -775,7 +775,7 @@ BOOST_AUTO_TEST_CASE(RefDetectionPostProcessFastNmsUint8Test) QuantizeData(qBoxEncodings.data(), boxEncodings.data(), boxEncodingsInfo); QuantizeData(qScores.data(), scores.data(), scoresInfo); QuantizeData(qAnchors.data(), anchors.data(), anchorsInfo); - DetectionPostProcessFastNmsEndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends, qBoxEncodings, + DetectionPostProcessFastNmsEndToEnd<armnn::DataType::QAsymmU8>(defaultBackends, qBoxEncodings, qScores, qAnchors, 1.0f, 1, 0.01f, 0, 0.5f, 0); } @@ -793,12 +793,12 @@ BOOST_AUTO_TEST_CASE(RefPreluEndToEndTestFloat32) BOOST_AUTO_TEST_CASE(RefPreluEndToEndTestUint8) { - PreluEndToEndPositiveTest<armnn::DataType::QuantisedAsymm8>(defaultBackends); + PreluEndToEndPositiveTest<armnn::DataType::QAsymmU8>(defaultBackends); } BOOST_AUTO_TEST_CASE(RefPreluEndToEndTestQSymm16) { - PreluEndToEndPositiveTest<armnn::DataType::QuantisedSymm16>(defaultBackends); + PreluEndToEndPositiveTest<armnn::DataType::QSymmS16>(defaultBackends); } BOOST_AUTO_TEST_CASE(RefSpaceToDepthNhwcEndToEndTest1) @@ -829,7 +829,7 @@ BOOST_AUTO_TEST_CASE(RefSplitter1dEndToEndTest) BOOST_AUTO_TEST_CASE(RefSplitter1dEndToEndUint8Test) { - Splitter1dEndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends); + Splitter1dEndToEnd<armnn::DataType::QAsymmU8>(defaultBackends); } BOOST_AUTO_TEST_CASE(RefSplitter2dDim0EndToEndTest) @@ -844,12 +844,12 @@ BOOST_AUTO_TEST_CASE(RefSplitter2dDim1EndToEndTest) BOOST_AUTO_TEST_CASE(RefSplitter2dDim0EndToEndUint8Test) { - Splitter2dDim0EndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends); + Splitter2dDim0EndToEnd<armnn::DataType::QAsymmU8>(defaultBackends); } BOOST_AUTO_TEST_CASE(RefSplitter2dDim1EndToEndUint8Test) { - Splitter2dDim1EndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends); + Splitter2dDim1EndToEnd<armnn::DataType::QAsymmU8>(defaultBackends); } BOOST_AUTO_TEST_CASE(RefSplitter3dDim0EndToEndTest) @@ -869,17 +869,17 @@ BOOST_AUTO_TEST_CASE(RefSplitter3dDim2EndToEndTest) BOOST_AUTO_TEST_CASE(RefSplitter3dDim0EndToEndUint8Test) { - Splitter3dDim0EndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends); + Splitter3dDim0EndToEnd<armnn::DataType::QAsymmU8>(defaultBackends); } BOOST_AUTO_TEST_CASE(RefSplitter3dDim1EndToEndUint8Test) { - Splitter3dDim1EndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends); + Splitter3dDim1EndToEnd<armnn::DataType::QAsymmU8>(defaultBackends); } BOOST_AUTO_TEST_CASE(RefSplitter3dDim2EndToEndUint8Test) { - Splitter3dDim2EndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends); + Splitter3dDim2EndToEnd<armnn::DataType::QAsymmU8>(defaultBackends); } BOOST_AUTO_TEST_CASE(RefSplitter4dDim0EndToEndTest) @@ -904,22 +904,22 @@ BOOST_AUTO_TEST_CASE(RefSplitter4dDim3EndToEndTest) BOOST_AUTO_TEST_CASE(RefSplitter4dDim0EndToEndUint8Test) { - Splitter4dDim0EndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends); + Splitter4dDim0EndToEnd<armnn::DataType::QAsymmU8>(defaultBackends); } BOOST_AUTO_TEST_CASE(RefSplitter4dDim1EndToEndUint8Test) { - Splitter4dDim1EndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends); + Splitter4dDim1EndToEnd<armnn::DataType::QAsymmU8>(defaultBackends); } BOOST_AUTO_TEST_CASE(RefSplitter4dDim2EndToEndUint8Test) { - Splitter4dDim2EndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends); + Splitter4dDim2EndToEnd<armnn::DataType::QAsymmU8>(defaultBackends); } BOOST_AUTO_TEST_CASE(RefSplitter4dDim3EndToEndUint8Test) { - Splitter4dDim3EndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends); + Splitter4dDim3EndToEnd<armnn::DataType::QAsymmU8>(defaultBackends); } // TransposeConvolution2d @@ -931,13 +931,13 @@ BOOST_AUTO_TEST_CASE(RefTransposeConvolution2dEndToEndFloatNchwTest) BOOST_AUTO_TEST_CASE(RefTransposeConvolution2dEndToEndUint8NchwTest) { - TransposeConvolution2dEndToEnd<armnn::DataType::QuantisedAsymm8, armnn::DataType::Signed32>( + TransposeConvolution2dEndToEnd<armnn::DataType::QAsymmU8, armnn::DataType::Signed32>( defaultBackends, armnn::DataLayout::NCHW); } BOOST_AUTO_TEST_CASE(RefTransposeConvolution2dEndToEndInt16NchwTest) { - TransposeConvolution2dEndToEnd<armnn::DataType::QuantisedSymm16, armnn::DataType::Signed32>( + TransposeConvolution2dEndToEnd<armnn::DataType::QSymmS16, armnn::DataType::Signed32>( defaultBackends, armnn::DataLayout::NCHW); } @@ -949,13 +949,13 @@ BOOST_AUTO_TEST_CASE(RefTransposeConvolution2dEndToEndFloatNhwcTest) BOOST_AUTO_TEST_CASE(RefTransposeConvolution2dEndToEndUint8NhwcTest) { - TransposeConvolution2dEndToEnd<armnn::DataType::QuantisedAsymm8, armnn::DataType::Signed32>( + TransposeConvolution2dEndToEnd<armnn::DataType::QAsymmU8, armnn::DataType::Signed32>( defaultBackends, armnn::DataLayout::NHWC); } BOOST_AUTO_TEST_CASE(RefTransposeConvolution2dEndToEndInt16NhwcTest) { - TransposeConvolution2dEndToEnd<armnn::DataType::QuantisedSymm16, armnn::DataType::Signed32>( + TransposeConvolution2dEndToEnd<armnn::DataType::QSymmS16, armnn::DataType::Signed32>( defaultBackends, armnn::DataLayout::NHWC); } @@ -967,12 +967,12 @@ BOOST_AUTO_TEST_CASE(RefResizeBilinearEndToEndFloatNchwTest) BOOST_AUTO_TEST_CASE(RefResizeBilinearEndToEndUint8NchwTest) { - ResizeBilinearEndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends, armnn::DataLayout::NCHW); + ResizeBilinearEndToEnd<armnn::DataType::QAsymmU8>(defaultBackends, armnn::DataLayout::NCHW); } BOOST_AUTO_TEST_CASE(RefResizeBilinearEndToEndInt16NchwTest) { - ResizeBilinearEndToEnd<armnn::DataType::QuantisedSymm16>(defaultBackends, armnn::DataLayout::NCHW); + ResizeBilinearEndToEnd<armnn::DataType::QSymmS16>(defaultBackends, armnn::DataLayout::NCHW); } BOOST_AUTO_TEST_CASE(RefResizeBilinearEndToEndFloatNhwcTest) @@ -982,12 +982,12 @@ BOOST_AUTO_TEST_CASE(RefResizeBilinearEndToEndFloatNhwcTest) BOOST_AUTO_TEST_CASE(RefResizeBilinearEndToEndUint8NhwcTest) { - ResizeBilinearEndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends, armnn::DataLayout::NHWC); + ResizeBilinearEndToEnd<armnn::DataType::QAsymmU8>(defaultBackends, armnn::DataLayout::NHWC); } BOOST_AUTO_TEST_CASE(RefResizeBilinearEndToEndInt16NhwcTest) { - ResizeBilinearEndToEnd<armnn::DataType::QuantisedSymm16>(defaultBackends, armnn::DataLayout::NHWC); + ResizeBilinearEndToEnd<armnn::DataType::QSymmS16>(defaultBackends, armnn::DataLayout::NHWC); } // Resize NearestNeighbor @@ -998,12 +998,12 @@ BOOST_AUTO_TEST_CASE(RefResizeNearestNeighborEndToEndFloatNchwTest) BOOST_AUTO_TEST_CASE(RefResizeNearestNeighborEndToEndUint8NchwTest) { - ResizeNearestNeighborEndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends, armnn::DataLayout::NCHW); + ResizeNearestNeighborEndToEnd<armnn::DataType::QAsymmU8>(defaultBackends, armnn::DataLayout::NCHW); } BOOST_AUTO_TEST_CASE(RefResizeNearestNeighborEndToEndInt16NchwTest) { - ResizeNearestNeighborEndToEnd<armnn::DataType::QuantisedSymm16>(defaultBackends, armnn::DataLayout::NCHW); + ResizeNearestNeighborEndToEnd<armnn::DataType::QSymmS16>(defaultBackends, armnn::DataLayout::NCHW); } BOOST_AUTO_TEST_CASE(RefResizeNearestNeighborEndToEndFloatNhwcTest) @@ -1013,12 +1013,12 @@ BOOST_AUTO_TEST_CASE(RefResizeNearestNeighborEndToEndFloatNhwcTest) BOOST_AUTO_TEST_CASE(RefResizeNearestNeighborEndToEndUint8NhwcTest) { - ResizeNearestNeighborEndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends, armnn::DataLayout::NHWC); + ResizeNearestNeighborEndToEnd<armnn::DataType::QAsymmU8>(defaultBackends, armnn::DataLayout::NHWC); } BOOST_AUTO_TEST_CASE(RefResizeNearestNeighborEndToEndInt16NhwcTest) { - ResizeNearestNeighborEndToEnd<armnn::DataType::QuantisedSymm16>(defaultBackends, armnn::DataLayout::NHWC); + ResizeNearestNeighborEndToEnd<armnn::DataType::QSymmS16>(defaultBackends, armnn::DataLayout::NHWC); } // InstanceNormalization @@ -1050,7 +1050,7 @@ BOOST_AUTO_TEST_CASE(RefArgMaxSimpleTest) BOOST_AUTO_TEST_CASE(RefArgMaxSimpleUint8Test) { - ArgMaxEndToEndSimple<armnn::DataType::QuantisedAsymm8>(defaultBackends); + ArgMaxEndToEndSimple<armnn::DataType::QAsymmU8>(defaultBackends); } BOOST_AUTO_TEST_CASE(RefArgMinSimpleTest) @@ -1060,7 +1060,7 @@ BOOST_AUTO_TEST_CASE(RefArgMinSimpleTest) BOOST_AUTO_TEST_CASE(RefArgMinSimpleUint8Test) { - ArgMinEndToEndSimple<armnn::DataType::QuantisedAsymm8>(defaultBackends); + ArgMinEndToEndSimple<armnn::DataType::QAsymmU8>(defaultBackends); } BOOST_AUTO_TEST_CASE(RefArgMaxAxis0Test) @@ -1070,7 +1070,7 @@ BOOST_AUTO_TEST_CASE(RefArgMaxAxis0Test) BOOST_AUTO_TEST_CASE(RefArgMaxAxis0Uint8Test) { - ArgMaxAxis0EndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends); + ArgMaxAxis0EndToEnd<armnn::DataType::QAsymmU8>(defaultBackends); } BOOST_AUTO_TEST_CASE(RefArgMinAxis0Test) @@ -1081,7 +1081,7 @@ BOOST_AUTO_TEST_CASE(RefArgMinAxis0Test) BOOST_AUTO_TEST_CASE(RefArgMinAxis0Uint8Test) { - ArgMinAxis0EndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends); + ArgMinAxis0EndToEnd<armnn::DataType::QAsymmU8>(defaultBackends); } BOOST_AUTO_TEST_CASE(RefArgMaxAxis1Test) @@ -1091,7 +1091,7 @@ BOOST_AUTO_TEST_CASE(RefArgMaxAxis1Test) BOOST_AUTO_TEST_CASE(RefArgMaxAxis1Uint8Test) { - ArgMaxAxis1EndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends); + ArgMaxAxis1EndToEnd<armnn::DataType::QAsymmU8>(defaultBackends); } BOOST_AUTO_TEST_CASE(RefArgMinAxis1Test) @@ -1102,7 +1102,7 @@ BOOST_AUTO_TEST_CASE(RefArgMinAxis1Test) BOOST_AUTO_TEST_CASE(RefArgMinAxis1Uint8Test) { - ArgMinAxis1EndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends); + ArgMinAxis1EndToEnd<armnn::DataType::QAsymmU8>(defaultBackends); } BOOST_AUTO_TEST_CASE(RefArgMaxAxis2Test) @@ -1112,7 +1112,7 @@ BOOST_AUTO_TEST_CASE(RefArgMaxAxis2Test) BOOST_AUTO_TEST_CASE(RefArgMaxAxis2Uint8Test) { - ArgMaxAxis2EndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends); + ArgMaxAxis2EndToEnd<armnn::DataType::QAsymmU8>(defaultBackends); } BOOST_AUTO_TEST_CASE(RefArgMinAxis2Test) @@ -1123,7 +1123,7 @@ BOOST_AUTO_TEST_CASE(RefArgMinAxis2Test) BOOST_AUTO_TEST_CASE(RefArgMinAxis2Uint8Test) { - ArgMinAxis2EndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends); + ArgMinAxis2EndToEnd<armnn::DataType::QAsymmU8>(defaultBackends); } BOOST_AUTO_TEST_CASE(RefArgMaxAxis3Test) @@ -1133,7 +1133,7 @@ BOOST_AUTO_TEST_CASE(RefArgMaxAxis3Test) BOOST_AUTO_TEST_CASE(RefArgMaxAxis3Uint8Test) { - ArgMaxAxis3EndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends); + ArgMaxAxis3EndToEnd<armnn::DataType::QAsymmU8>(defaultBackends); } BOOST_AUTO_TEST_CASE(RefArgMinAxis3Test) @@ -1144,7 +1144,7 @@ BOOST_AUTO_TEST_CASE(RefArgMinAxis3Test) BOOST_AUTO_TEST_CASE(RefArgMinAxis3Uint8Test) { - ArgMinAxis3EndToEnd<armnn::DataType::QuantisedAsymm8>(defaultBackends); + ArgMinAxis3EndToEnd<armnn::DataType::QAsymmU8>(defaultBackends); } #if !defined(__ANDROID__) diff --git a/src/backends/reference/test/RefLayerSupportTests.cpp b/src/backends/reference/test/RefLayerSupportTests.cpp index 106a7f6157..ab0809d90b 100644 --- a/src/backends/reference/test/RefLayerSupportTests.cpp +++ b/src/backends/reference/test/RefLayerSupportTests.cpp @@ -63,13 +63,13 @@ BOOST_AUTO_TEST_CASE(IsLayerSupportedFloat32Reference) BOOST_AUTO_TEST_CASE(IsLayerSupportedUint8Reference) { armnn::RefWorkloadFactory factory; - IsLayerSupportedTests<armnn::RefWorkloadFactory, armnn::DataType::QuantisedAsymm8>(&factory); + IsLayerSupportedTests<armnn::RefWorkloadFactory, armnn::DataType::QAsymmU8>(&factory); } BOOST_AUTO_TEST_CASE(IsLayerSupportedInt16Reference) { armnn::RefWorkloadFactory factory; - IsLayerSupportedTests<armnn::RefWorkloadFactory, armnn::DataType::QuantisedSymm16>(&factory); + IsLayerSupportedTests<armnn::RefWorkloadFactory, armnn::DataType::QSymmS16>(&factory); } BOOST_AUTO_TEST_CASE(IsConvertFp16ToFp32SupportedReference) diff --git a/src/backends/reference/test/RefLayerTests.cpp b/src/backends/reference/test/RefLayerTests.cpp index b88f432acf..b0d8db802e 100644 --- a/src/backends/reference/test/RefLayerTests.cpp +++ b/src/backends/reference/test/RefLayerTests.cpp @@ -79,19 +79,19 @@ ARMNN_AUTO_TEST_CASE(Convolution2d3x3Dilation3x3Nhwc, false, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(Convolution2d3x3Dilation3x3Uint8, - Convolution2d3x3Dilation3x3Test<DataType::QuantisedAsymm8, DataType::Signed32>, + Convolution2d3x3Dilation3x3Test<DataType::QAsymmU8, DataType::Signed32>, false, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(Convolution2d3x3Dilation3x3NhwcUint8, - Convolution2d3x3Dilation3x3Test<DataType::QuantisedAsymm8, DataType::Signed32>, + Convolution2d3x3Dilation3x3Test<DataType::QAsymmU8, DataType::Signed32>, false, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(Convolution2d3x3Dilation3x3Int16, - Convolution2d3x3Dilation3x3Test<DataType::QuantisedSymm16, DataType::Signed32>, + Convolution2d3x3Dilation3x3Test<DataType::QSymmS16, DataType::Signed32>, false, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(Convolution2d3x3Dilation3x3NhwcInt16, - Convolution2d3x3Dilation3x3Test<DataType::QuantisedSymm16, DataType::Signed32>, + Convolution2d3x3Dilation3x3Test<DataType::QSymmS16, DataType::Signed32>, false, DataLayout::NHWC) @@ -104,19 +104,19 @@ ARMNN_AUTO_TEST_CASE(Convolution2d2x3x3Dilation3x3Nhwc, false, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(Convolution2d2x3x3Dilation3x3Uint8, - Convolution2d2x3x3Dilation3x3Test<DataType::QuantisedAsymm8, DataType::Signed32>, + Convolution2d2x3x3Dilation3x3Test<DataType::QAsymmU8, DataType::Signed32>, false, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(Convolution2d2x3x3Dilation3x3NhwcUint8, - Convolution2d2x3x3Dilation3x3Test<DataType::QuantisedAsymm8, DataType::Signed32>, + Convolution2d2x3x3Dilation3x3Test<DataType::QAsymmU8, DataType::Signed32>, false, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(Convolution2d2x3x3Dilation3x3Int16, - Convolution2d2x3x3Dilation3x3Test<DataType::QuantisedSymm16, DataType::Signed32>, + Convolution2d2x3x3Dilation3x3Test<DataType::QSymmS16, DataType::Signed32>, false, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(Convolution2d2x3x3Dilation3x3NhwcInt16, - Convolution2d2x3x3Dilation3x3Test<DataType::QuantisedSymm16, DataType::Signed32>, + Convolution2d2x3x3Dilation3x3Test<DataType::QSymmS16, DataType::Signed32>, false, DataLayout::NHWC) @@ -129,19 +129,19 @@ ARMNN_AUTO_TEST_CASE(Convolution2d2x2Dilation2x2Padding2x2Stride3x3Nhwc, false, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(Convolution2d2x2Dilation2x2Padding2x2Stride3x3Uint8, - Convolution2d2x2Dilation2x2Padding2x2Stride3x3Test<DataType::QuantisedAsymm8, DataType::Signed32>, + Convolution2d2x2Dilation2x2Padding2x2Stride3x3Test<DataType::QAsymmU8, DataType::Signed32>, false, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(Convolution2d2x2Dilation2x2Padding2x2Stride3x3NhwcUint8, - Convolution2d2x2Dilation2x2Padding2x2Stride3x3Test<DataType::QuantisedAsymm8, DataType::Signed32>, + Convolution2d2x2Dilation2x2Padding2x2Stride3x3Test<DataType::QAsymmU8, DataType::Signed32>, false, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(Convolution2d2x2Dilation2x2Padding2x2Stride3x3Int16, - Convolution2d2x2Dilation2x2Padding2x2Stride3x3Test<DataType::QuantisedSymm16, DataType::Signed32>, + Convolution2d2x2Dilation2x2Padding2x2Stride3x3Test<DataType::QSymmS16, DataType::Signed32>, false, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(Convolution2d2x2Dilation2x2Padding2x2Stride3x3NhwcInt16, - Convolution2d2x2Dilation2x2Padding2x2Stride3x3Test<DataType::QuantisedSymm16, DataType::Signed32>, + Convolution2d2x2Dilation2x2Padding2x2Stride3x3Test<DataType::QSymmS16, DataType::Signed32>, false, DataLayout::NHWC) @@ -180,19 +180,19 @@ ARMNN_AUTO_TEST_CASE(DepthwiseConvolution2d3x3Dilation3x3Nhwc, false, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(DepthwiseConvolution2d3x3Dilation3x3Uint8, - DepthwiseConvolution2d3x3Dilation3x3Test<DataType::QuantisedAsymm8, DataType::Signed32>, + DepthwiseConvolution2d3x3Dilation3x3Test<DataType::QAsymmU8, DataType::Signed32>, false, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(DepthwiseConvolution2d3x3Dilation3x3NhwcUint8, - DepthwiseConvolution2d3x3Dilation3x3Test<DataType::QuantisedAsymm8, DataType::Signed32>, + DepthwiseConvolution2d3x3Dilation3x3Test<DataType::QAsymmU8, DataType::Signed32>, false, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(DepthwiseConvolution2d3x3Dilation3x3Int16, - DepthwiseConvolution2d3x3Dilation3x3Test<DataType::QuantisedSymm16, DataType::Signed32>, + DepthwiseConvolution2d3x3Dilation3x3Test<DataType::QSymmS16, DataType::Signed32>, false, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(DepthwiseConvolution2d3x3Dilation3x3NhwcInt16, - DepthwiseConvolution2d3x3Dilation3x3Test<DataType::QuantisedSymm16, DataType::Signed32>, + DepthwiseConvolution2d3x3Dilation3x3Test<DataType::QSymmS16, DataType::Signed32>, false, DataLayout::NHWC) @@ -205,19 +205,19 @@ ARMNN_AUTO_TEST_CASE(DepthwiseConvolution2d2x3x3Dilation3x3Nhwc, false, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(DepthwiseConvolution2d2x3x3Dilation3x3Uint8, - DepthwiseConvolution2d2x3x3Dilation3x3Test<DataType::QuantisedAsymm8, DataType::Signed32>, + DepthwiseConvolution2d2x3x3Dilation3x3Test<DataType::QAsymmU8, DataType::Signed32>, false, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(DepthwiseConvolution2d2x3x3Dilation3x3NhwcUint8, - DepthwiseConvolution2d2x3x3Dilation3x3Test<DataType::QuantisedAsymm8, DataType::Signed32>, + DepthwiseConvolution2d2x3x3Dilation3x3Test<DataType::QAsymmU8, DataType::Signed32>, false, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(DepthwiseConvolution2d2x3x3Dilation3x3Int16, - DepthwiseConvolution2d2x3x3Dilation3x3Test<DataType::QuantisedSymm16, DataType::Signed32>, + DepthwiseConvolution2d2x3x3Dilation3x3Test<DataType::QSymmS16, DataType::Signed32>, false, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(DepthwiseConvolution2d2x3x3Dilation3x3NhwcInt16, - DepthwiseConvolution2d2x3x3Dilation3x3Test<DataType::QuantisedSymm16, DataType::Signed32>, + DepthwiseConvolution2d2x3x3Dilation3x3Test<DataType::QSymmS16, DataType::Signed32>, false, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(DepthwiseConvolution2dMult4, @@ -465,11 +465,11 @@ ARMNN_AUTO_TEST_CASE(TanhInt16, TanhInt16Test) // Fully Connected ARMNN_AUTO_TEST_CASE(SimpleFullyConnected, FullyConnectedFloat32Test, false, false) -ARMNN_AUTO_TEST_CASE(FullyConnectedUint8, FullyConnectedTest<DataType::QuantisedAsymm8>, false) -ARMNN_AUTO_TEST_CASE(FullyConnectedQSymm16, FullyConnectedTest<DataType::QuantisedSymm16>, false) +ARMNN_AUTO_TEST_CASE(FullyConnectedUint8, FullyConnectedTest<DataType::QAsymmU8>, false) +ARMNN_AUTO_TEST_CASE(FullyConnectedQSymm16, FullyConnectedTest<DataType::QSymmS16>, false) ARMNN_AUTO_TEST_CASE(SimpleFullyConnectedWithBias, FullyConnectedFloat32Test, true, false) -ARMNN_AUTO_TEST_CASE(FullyConnectedBiasedUint8, FullyConnectedTest<DataType::QuantisedAsymm8>, true) -ARMNN_AUTO_TEST_CASE(FullyConnectedBiasedQSymm16, FullyConnectedTest<DataType::QuantisedSymm16>, true) +ARMNN_AUTO_TEST_CASE(FullyConnectedBiasedUint8, FullyConnectedTest<DataType::QAsymmU8>, true) +ARMNN_AUTO_TEST_CASE(FullyConnectedBiasedQSymm16, FullyConnectedTest<DataType::QSymmS16>, true) ARMNN_AUTO_TEST_CASE(SimpleFullyConnectedWithTranspose, FullyConnectedFloat32Test, false, true) ARMNN_AUTO_TEST_CASE(FullyConnectedLarge, FullyConnectedLargeTest, false) @@ -493,9 +493,9 @@ ARMNN_AUTO_TEST_CASE(ConcatUint8, ConcatUint8Test) ARMNN_AUTO_TEST_CASE(ConcatUint8DifferentQParams, ConcatUint8DifferentQParamsTest) ARMNN_AUTO_TEST_CASE(ConcatUint16, ConcatUint16Test) ARMNN_AUTO_TEST_CASE(ConcatUint8DifferentInputOutputQParam, - ConcatDifferentInputOutputQParamTest<DataType::QuantisedAsymm8>, true) + ConcatDifferentInputOutputQParamTest<DataType::QAsymmU8>, true) ARMNN_AUTO_TEST_CASE(ConcatInt16DifferentInputOutputQParam, - ConcatDifferentInputOutputQParamTest<DataType::QuantisedSymm16>, true) + ConcatDifferentInputOutputQParamTest<DataType::QSymmS16>, true) // Add ARMNN_AUTO_TEST_CASE(SimpleAdd, AdditionTest) @@ -681,10 +681,10 @@ ARMNN_AUTO_TEST_CASE(SimpleResizeBilinearFloat16, SimpleResizeBilinearTest<DataType::Float16>, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(SimpleResizeBilinearUint8, - SimpleResizeBilinearTest<DataType::QuantisedAsymm8>, + SimpleResizeBilinearTest<DataType::QAsymmU8>, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(SimpleResizeBilinearUint16, - SimpleResizeBilinearTest<DataType::QuantisedSymm16>, + SimpleResizeBilinearTest<DataType::QSymmS16>, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(ResizeBilinearNop, ResizeBilinearNopTest<DataType::Float32>, @@ -693,10 +693,10 @@ ARMNN_AUTO_TEST_CASE(ResizeBilinearNopFloat16, ResizeBilinearNopTest<DataType::Float16>, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(ResizeBilinearNopUint8, - ResizeBilinearNopTest<DataType::QuantisedAsymm8>, + ResizeBilinearNopTest<DataType::QAsymmU8>, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(esizeBilinearNopUint16, - SimpleResizeBilinearTest<DataType::QuantisedSymm16>, + SimpleResizeBilinearTest<DataType::QSymmS16>, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(ResizeBilinearSqMin, ResizeBilinearSqMinTest<DataType::Float32>, @@ -705,10 +705,10 @@ ARMNN_AUTO_TEST_CASE(ResizeBilinearSqMinFloat16, ResizeBilinearSqMinTest<DataType::Float16>, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(ResizeBilinearSqMinUint8, - ResizeBilinearSqMinTest<DataType::QuantisedAsymm8>, + ResizeBilinearSqMinTest<DataType::QAsymmU8>, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(ResizeBilinearSqMinUint16, - SimpleResizeBilinearTest<DataType::QuantisedSymm16>, + SimpleResizeBilinearTest<DataType::QSymmS16>, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(ResizeBilinearMin, ResizeBilinearMinTest<DataType::Float32>, @@ -717,10 +717,10 @@ ARMNN_AUTO_TEST_CASE(ResizeBilinearMinFloat16, ResizeBilinearMinTest<DataType::Float16>, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(ResizeBilinearMinUint8, - ResizeBilinearMinTest<DataType::QuantisedAsymm8>, + ResizeBilinearMinTest<DataType::QAsymmU8>, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(ResizeBilinearMinUint16, - SimpleResizeBilinearTest<DataType::QuantisedSymm16>, + SimpleResizeBilinearTest<DataType::QSymmS16>, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(ResizeBilinearMag, ResizeBilinearMagTest<DataType::Float32>, @@ -729,10 +729,10 @@ ARMNN_AUTO_TEST_CASE(ResizeBilinearMagFloat16, ResizeBilinearMagTest<DataType::Float16>, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(ResizeBilinearMagUint8, - ResizeBilinearMagTest<DataType::QuantisedAsymm8>, + ResizeBilinearMagTest<DataType::QAsymmU8>, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(ResizeBilinearMagUint16, - SimpleResizeBilinearTest<DataType::QuantisedSymm16>, + SimpleResizeBilinearTest<DataType::QSymmS16>, DataLayout::NCHW) // Resize Bilinear - NHWC @@ -743,10 +743,10 @@ ARMNN_AUTO_TEST_CASE(ResizeBilinearNopNhwcFloat16, ResizeBilinearNopTest<DataType::Float16>, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(ResizeBilinearNopUint8Nhwc, - ResizeBilinearNopTest<DataType::QuantisedAsymm8>, + ResizeBilinearNopTest<DataType::QAsymmU8>, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(ResizeBilinearNopUint16Nhwc, - ResizeBilinearNopTest<DataType::QuantisedSymm16>, + ResizeBilinearNopTest<DataType::QSymmS16>, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(SimpleResizeBilinearNhwc, SimpleResizeBilinearTest<DataType::Float32>, @@ -755,10 +755,10 @@ ARMNN_AUTO_TEST_CASE(SimpleResizeBilinearNhwcFloat16, SimpleResizeBilinearTest<DataType::Float16>, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(SimpleResizeBilinearUint8Nhwc, - SimpleResizeBilinearTest<DataType::QuantisedAsymm8>, + SimpleResizeBilinearTest<DataType::QAsymmU8>, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(SimpleResizeBilinearUint16Nhwc, - ResizeBilinearNopTest<DataType::QuantisedSymm16>, + ResizeBilinearNopTest<DataType::QSymmS16>, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(ResizeBilinearSqMinNhwc, ResizeBilinearSqMinTest<DataType::Float32>, @@ -767,10 +767,10 @@ ARMNN_AUTO_TEST_CASE(ResizeBilinearSqMinNhwcFloat16, ResizeBilinearSqMinTest<DataType::Float16>, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(ResizeBilinearSqMinUint8Nhwc, - ResizeBilinearSqMinTest<DataType::QuantisedAsymm8>, + ResizeBilinearSqMinTest<DataType::QAsymmU8>, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(ResizeBilinearSqMinUint16Nhwc, - ResizeBilinearNopTest<DataType::QuantisedSymm16>, + ResizeBilinearNopTest<DataType::QSymmS16>, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(ResizeBilinearMinNhwc, ResizeBilinearMinTest<DataType::Float32>, @@ -779,10 +779,10 @@ ARMNN_AUTO_TEST_CASE(ResizeBilinearMinNhwcFloat16, ResizeBilinearMinTest<DataType::Float16>, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(ResizeBilinearMinUint8Nhwc, - ResizeBilinearMinTest<DataType::QuantisedAsymm8>, + ResizeBilinearMinTest<DataType::QAsymmU8>, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(ResizeBilinearMinUint16Nhwc, - ResizeBilinearNopTest<DataType::QuantisedSymm16>, + ResizeBilinearNopTest<DataType::QSymmS16>, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(ResizeBilinearMagNhwc, ResizeBilinearMagTest<DataType::Float32>, @@ -791,10 +791,10 @@ ARMNN_AUTO_TEST_CASE(ResizeBilinearMagNhwcFloat16, ResizeBilinearMagTest<DataType::Float16>, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(ResizeBilinearMagUint8Nhwc, - ResizeBilinearMagTest<DataType::QuantisedAsymm8>, + ResizeBilinearMagTest<DataType::QAsymmU8>, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(ResizeBilinearMagUint16Nhwc, - ResizeBilinearNopTest<DataType::QuantisedSymm16>, + ResizeBilinearNopTest<DataType::QSymmS16>, DataLayout::NHWC) // Resize NearestNeighbor - NCHW @@ -802,46 +802,46 @@ ARMNN_AUTO_TEST_CASE(SimpleResizeNearestNeighbor, SimpleResizeNearestNeighborTest<DataType::Float32>, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(SimpleResizeNearestNeighborUint8, - SimpleResizeNearestNeighborTest<DataType::QuantisedAsymm8>, + SimpleResizeNearestNeighborTest<DataType::QAsymmU8>, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(SimpleResizeNearestNeighborUint16, - SimpleResizeNearestNeighborTest<DataType::QuantisedSymm16>, + SimpleResizeNearestNeighborTest<DataType::QSymmS16>, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(ResizeNearestNeighborNop, ResizeNearestNeighborNopTest<DataType::Float32>, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(ResizeNearestNeighborNopUint8, - ResizeNearestNeighborNopTest<DataType::QuantisedAsymm8>, + ResizeNearestNeighborNopTest<DataType::QAsymmU8>, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(esizeNearestNeighborNopUint16, - SimpleResizeNearestNeighborTest<DataType::QuantisedSymm16>, + SimpleResizeNearestNeighborTest<DataType::QSymmS16>, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(ResizeNearestNeighborSqMin, ResizeNearestNeighborSqMinTest<DataType::Float32>, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(ResizeNearestNeighborSqMinUint8, - ResizeNearestNeighborSqMinTest<DataType::QuantisedAsymm8>, + ResizeNearestNeighborSqMinTest<DataType::QAsymmU8>, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(ResizeNearestNeighborSqMinUint16, - SimpleResizeNearestNeighborTest<DataType::QuantisedSymm16>, + SimpleResizeNearestNeighborTest<DataType::QSymmS16>, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(ResizeNearestNeighborMin, ResizeNearestNeighborMinTest<DataType::Float32>, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(ResizeNearestNeighborMinUint8, - ResizeNearestNeighborMinTest<DataType::QuantisedAsymm8>, + ResizeNearestNeighborMinTest<DataType::QAsymmU8>, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(ResizeNearestNeighborMinUint16, - SimpleResizeNearestNeighborTest<DataType::QuantisedSymm16>, + SimpleResizeNearestNeighborTest<DataType::QSymmS16>, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(ResizeNearestNeighborMag, ResizeNearestNeighborMagTest<DataType::Float32>, DataLayout::NCHW, 0.10f, 50, 0.11f, 20) ARMNN_AUTO_TEST_CASE(ResizeNearestNeighborMagUint8, - ResizeNearestNeighborMagTest<DataType::QuantisedAsymm8>, + ResizeNearestNeighborMagTest<DataType::QAsymmU8>, DataLayout::NCHW, 0.10f, 50, 0.11f, 20) ARMNN_AUTO_TEST_CASE(ResizeNearestNeighborMagUint16, - SimpleResizeNearestNeighborTest<DataType::QuantisedSymm16>, + SimpleResizeNearestNeighborTest<DataType::QSymmS16>, DataLayout::NCHW) // Resize NearestNeighbor - NHWC @@ -849,46 +849,46 @@ ARMNN_AUTO_TEST_CASE(ResizeNearestNeighborNopNhwc, ResizeNearestNeighborNopTest<DataType::Float32>, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(ResizeNearestNeighborNopUint8Nhwc, - ResizeNearestNeighborNopTest<DataType::QuantisedAsymm8>, + ResizeNearestNeighborNopTest<DataType::QAsymmU8>, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(ResizeNearestNeighborNopUint16Nhwc, - ResizeNearestNeighborNopTest<DataType::QuantisedSymm16>, + ResizeNearestNeighborNopTest<DataType::QSymmS16>, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(SimpleResizeNearestNeighborNhwc, SimpleResizeNearestNeighborTest<DataType::Float32>, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(SimpleResizeNearestNeighborUint8Nhwc, - SimpleResizeNearestNeighborTest<DataType::QuantisedAsymm8>, + SimpleResizeNearestNeighborTest<DataType::QAsymmU8>, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(SimpleResizeNearestNeighborUint16Nhwc, - ResizeNearestNeighborNopTest<DataType::QuantisedSymm16>, + ResizeNearestNeighborNopTest<DataType::QSymmS16>, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(ResizeNearestNeighborSqMinNhwc, ResizeNearestNeighborSqMinTest<DataType::Float32>, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(ResizeNearestNeighborSqMinUint8Nhwc, - ResizeNearestNeighborSqMinTest<DataType::QuantisedAsymm8>, + ResizeNearestNeighborSqMinTest<DataType::QAsymmU8>, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(ResizeNearestNeighborSqMinUint16Nhwc, - ResizeNearestNeighborNopTest<DataType::QuantisedSymm16>, + ResizeNearestNeighborNopTest<DataType::QSymmS16>, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(ResizeNearestNeighborMinNhwc, ResizeNearestNeighborMinTest<DataType::Float32>, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(ResizeNearestNeighborMinUint8Nhwc, - ResizeNearestNeighborMinTest<DataType::QuantisedAsymm8>, + ResizeNearestNeighborMinTest<DataType::QAsymmU8>, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(ResizeNearestNeighborMinUint16Nhwc, - ResizeNearestNeighborNopTest<DataType::QuantisedSymm16>, + ResizeNearestNeighborNopTest<DataType::QSymmS16>, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(ResizeNearestNeighborMagNhwc, ResizeNearestNeighborMagTest<DataType::Float32>, DataLayout::NHWC, 0.10f, 50, 0.11f, 20) ARMNN_AUTO_TEST_CASE(ResizeNearestNeighborMagUint8Nhwc, - ResizeNearestNeighborMagTest<DataType::QuantisedAsymm8>, + ResizeNearestNeighborMagTest<DataType::QAsymmU8>, DataLayout::NHWC, 0.10f, 50, 0.11f, 20) ARMNN_AUTO_TEST_CASE(ResizeNearestNeighborMagUint16Nhwc, - ResizeNearestNeighborNopTest<DataType::QuantisedSymm16>, + ResizeNearestNeighborNopTest<DataType::QSymmS16>, DataLayout::NHWC) // Fake Quantization @@ -952,10 +952,10 @@ ARMNN_AUTO_TEST_CASE(PadUint82dCustomPadding, PadUint82dCustomPaddingTest) ARMNN_AUTO_TEST_CASE(PadUint83d, PadUint83dTest) ARMNN_AUTO_TEST_CASE(PadUint84d, PadUint84dTest) -ARMNN_AUTO_TEST_CASE(Pad2dQSymm16, Pad2dTestCommon<DataType::QuantisedSymm16>, 2.0f, 0, 0.0f) -ARMNN_AUTO_TEST_CASE(Pad2dQSymm16CustomPadding, Pad2dTestCommon<DataType::QuantisedSymm16>, 2.0f, 0, 1.0f) -ARMNN_AUTO_TEST_CASE(Pad3dQSymm16, Pad3dTestCommon<DataType::QuantisedSymm16>, 2.0f, 0) -ARMNN_AUTO_TEST_CASE(Pad4dQSymm16, Pad4dTestCommon<DataType::QuantisedSymm16>, 2.0f, 0) +ARMNN_AUTO_TEST_CASE(Pad2dQSymm16, Pad2dTestCommon<DataType::QSymmS16>, 2.0f, 0, 0.0f) +ARMNN_AUTO_TEST_CASE(Pad2dQSymm16CustomPadding, Pad2dTestCommon<DataType::QSymmS16>, 2.0f, 0, 1.0f) +ARMNN_AUTO_TEST_CASE(Pad3dQSymm16, Pad3dTestCommon<DataType::QSymmS16>, 2.0f, 0) +ARMNN_AUTO_TEST_CASE(Pad4dQSymm16, Pad4dTestCommon<DataType::QSymmS16>, 2.0f, 0) // Constant ARMNN_AUTO_TEST_CASE(Constant, ConstantTest) @@ -1011,12 +1011,12 @@ ARMNN_AUTO_TEST_CASE(Concat4dDiffShapeDim3Uint8, Concat4dDiffShapeDim3Uint8Test, // Floor ARMNN_AUTO_TEST_CASE(SimpleFloor, SimpleFloorTest<DataType::Float32>) ARMNN_AUTO_TEST_CASE(SimpleFloorFloat16, SimpleFloorTest<DataType::Float16>) -ARMNN_AUTO_TEST_CASE(SimpleFloorQuantisedSymm16, SimpleFloorTest<DataType::QuantisedSymm16>) +ARMNN_AUTO_TEST_CASE(SimpleFloorQuantisedSymm16, SimpleFloorTest<DataType::QSymmS16>) // Reshape ARMNN_AUTO_TEST_CASE(SimpleReshapeFloat32, SimpleReshapeTest<DataType::Float32>) -ARMNN_AUTO_TEST_CASE(SimpleReshapeQuantisedAsymm8, SimpleReshapeTest<DataType::QuantisedAsymm8>) -ARMNN_AUTO_TEST_CASE(SimpleReshapeQuantisedSymm16, SimpleReshapeTest<DataType::QuantisedSymm16>) +ARMNN_AUTO_TEST_CASE(SimpleReshapeQuantisedAsymm8, SimpleReshapeTest<DataType::QAsymmU8>) +ARMNN_AUTO_TEST_CASE(SimpleReshapeQuantisedSymm16, SimpleReshapeTest<DataType::QSymmS16>) ARMNN_AUTO_TEST_CASE(Reshape5d, Reshape5dTest<DataType::Float32>) // Rsqrt @@ -1026,24 +1026,24 @@ ARMNN_AUTO_TEST_CASE(RsqrtZero, RsqrtZeroTest<DataType::Float32>) ARMNN_AUTO_TEST_CASE(RsqrtNegative, RsqrtNegativeTest<DataType::Float32>) ARMNN_AUTO_TEST_CASE(Rsqrt2dFloat16, Rsqrt2dTest<DataType::Float16>) ARMNN_AUTO_TEST_CASE(Rsqrt3dFloat16, Rsqrt3dTest<DataType::Float16>) -ARMNN_AUTO_TEST_CASE(Rsqrt2dQuantisedAsymm8, Rsqrt2dTest<DataType::QuantisedAsymm8>) -ARMNN_AUTO_TEST_CASE(Rsqrt3dQuantisedAsymm8, Rsqrt3dTest<DataType::QuantisedAsymm8>) -ARMNN_AUTO_TEST_CASE(Rsqrt2dQuantisedSymm16, Rsqrt2dTest<DataType::QuantisedSymm16>) -ARMNN_AUTO_TEST_CASE(Rsqrt3dQuantisedSymm16, Rsqrt3dTest<DataType::QuantisedSymm16>) +ARMNN_AUTO_TEST_CASE(Rsqrt2dQuantisedAsymm8, Rsqrt2dTest<DataType::QAsymmU8>) +ARMNN_AUTO_TEST_CASE(Rsqrt3dQuantisedAsymm8, Rsqrt3dTest<DataType::QAsymmU8>) +ARMNN_AUTO_TEST_CASE(Rsqrt2dQuantisedSymm16, Rsqrt2dTest<DataType::QSymmS16>) +ARMNN_AUTO_TEST_CASE(Rsqrt3dQuantisedSymm16, Rsqrt3dTest<DataType::QSymmS16>) // Permute ARMNN_AUTO_TEST_CASE(SimplePermuteFloat32, SimplePermuteTest<DataType::Float32>) ARMNN_AUTO_TEST_CASE(PermuteFloat32ValueSet1Test, PermuteValueSet1Test<DataType::Float32>) ARMNN_AUTO_TEST_CASE(PermuteFloat32ValueSet2Test, PermuteValueSet2Test<DataType::Float32>) ARMNN_AUTO_TEST_CASE(PermuteFloat32ValueSet3Test, PermuteValueSet3Test<DataType::Float32>) -ARMNN_AUTO_TEST_CASE(SimplePermuteQASymm8, SimplePermuteTest<DataType::QuantisedAsymm8>) -ARMNN_AUTO_TEST_CASE(PermuteQASymm8ValueSet1Test, PermuteValueSet1Test<DataType::QuantisedAsymm8>) -ARMNN_AUTO_TEST_CASE(PermuteQASymm8ValueSet2Test, PermuteValueSet2Test<DataType::QuantisedAsymm8>) -ARMNN_AUTO_TEST_CASE(PermuteQASymm8ValueSet3Test, PermuteValueSet3Test<DataType::QuantisedAsymm8>) -ARMNN_AUTO_TEST_CASE(SimplePermuteQSymm16, SimplePermuteTest<DataType::QuantisedSymm16>) -ARMNN_AUTO_TEST_CASE(PermuteQSymm16ValueSet1Test, PermuteValueSet1Test<DataType::QuantisedSymm16>) -ARMNN_AUTO_TEST_CASE(PermuteQSymm16ValueSet2Test, PermuteValueSet2Test<DataType::QuantisedSymm16>) -ARMNN_AUTO_TEST_CASE(PermuteQSymm16ValueSet3Test, PermuteValueSet3Test<DataType::QuantisedSymm16>) +ARMNN_AUTO_TEST_CASE(SimplePermuteQASymm8, SimplePermuteTest<DataType::QAsymmU8>) +ARMNN_AUTO_TEST_CASE(PermuteQASymm8ValueSet1Test, PermuteValueSet1Test<DataType::QAsymmU8>) +ARMNN_AUTO_TEST_CASE(PermuteQASymm8ValueSet2Test, PermuteValueSet2Test<DataType::QAsymmU8>) +ARMNN_AUTO_TEST_CASE(PermuteQASymm8ValueSet3Test, PermuteValueSet3Test<DataType::QAsymmU8>) +ARMNN_AUTO_TEST_CASE(SimplePermuteQSymm16, SimplePermuteTest<DataType::QSymmS16>) +ARMNN_AUTO_TEST_CASE(PermuteQSymm16ValueSet1Test, PermuteValueSet1Test<DataType::QSymmS16>) +ARMNN_AUTO_TEST_CASE(PermuteQSymm16ValueSet2Test, PermuteValueSet2Test<DataType::QSymmS16>) +ARMNN_AUTO_TEST_CASE(PermuteQSymm16ValueSet3Test, PermuteValueSet3Test<DataType::QSymmS16>) // Lstm BOOST_AUTO_TEST_CASE(LstmUtilsZeroVector) { @@ -1090,21 +1090,21 @@ ARMNN_AUTO_TEST_CASE(MeanVts1Float32, MeanVts1Test<DataType::Float32>) ARMNN_AUTO_TEST_CASE(MeanVts2Float32, MeanVts2Test<DataType::Float32>) ARMNN_AUTO_TEST_CASE(MeanVts3Float32, MeanVts3Test<DataType::Float32>) -ARMNN_AUTO_TEST_CASE(MeanSimpleQuantisedAsymm8, MeanSimpleTest<DataType::QuantisedAsymm8>) -ARMNN_AUTO_TEST_CASE(MeanSimpleAxisQuantisedAsymm8, MeanSimpleAxisTest<DataType::QuantisedAsymm8>) -ARMNN_AUTO_TEST_CASE(MeanKeepDimsQuantisedAsymm8, MeanKeepDimsTest<DataType::QuantisedAsymm8>) -ARMNN_AUTO_TEST_CASE(MeanMultipleDimsQuantisedAsymm8, MeanMultipleDimsTest<DataType::QuantisedAsymm8>) -ARMNN_AUTO_TEST_CASE(MeanVts1QuantisedAsymm8, MeanVts1Test<DataType::QuantisedAsymm8>) -ARMNN_AUTO_TEST_CASE(MeanVts2QuantisedAsymm8, MeanVts2Test<DataType::QuantisedAsymm8>) -ARMNN_AUTO_TEST_CASE(MeanVts3QuantisedAsymm8, MeanVts3Test<DataType::QuantisedAsymm8>) - -ARMNN_AUTO_TEST_CASE(MeanSimpleQuantisedSymm16, MeanSimpleTest<DataType::QuantisedSymm16>) -ARMNN_AUTO_TEST_CASE(MeanSimpleAxisQuantisedSymm16, MeanSimpleAxisTest<DataType::QuantisedSymm16>) -ARMNN_AUTO_TEST_CASE(MeanKeepDimsQuantisedSymm16, MeanKeepDimsTest<DataType::QuantisedSymm16>) -ARMNN_AUTO_TEST_CASE(MeanMultipleDimsQuantisedSymm16, MeanMultipleDimsTest<DataType::QuantisedSymm16>) -ARMNN_AUTO_TEST_CASE(MeanVts1QuantisedSymm16, MeanVts1Test<DataType::QuantisedSymm16>) -ARMNN_AUTO_TEST_CASE(MeanVts2QuantisedSymm16, MeanVts2Test<DataType::QuantisedSymm16>) -ARMNN_AUTO_TEST_CASE(MeanVts3QuantisedSymm16, MeanVts3Test<DataType::QuantisedSymm16>) +ARMNN_AUTO_TEST_CASE(MeanSimpleQuantisedAsymm8, MeanSimpleTest<DataType::QAsymmU8>) +ARMNN_AUTO_TEST_CASE(MeanSimpleAxisQuantisedAsymm8, MeanSimpleAxisTest<DataType::QAsymmU8>) +ARMNN_AUTO_TEST_CASE(MeanKeepDimsQuantisedAsymm8, MeanKeepDimsTest<DataType::QAsymmU8>) +ARMNN_AUTO_TEST_CASE(MeanMultipleDimsQuantisedAsymm8, MeanMultipleDimsTest<DataType::QAsymmU8>) +ARMNN_AUTO_TEST_CASE(MeanVts1QuantisedAsymm8, MeanVts1Test<DataType::QAsymmU8>) +ARMNN_AUTO_TEST_CASE(MeanVts2QuantisedAsymm8, MeanVts2Test<DataType::QAsymmU8>) +ARMNN_AUTO_TEST_CASE(MeanVts3QuantisedAsymm8, MeanVts3Test<DataType::QAsymmU8>) + +ARMNN_AUTO_TEST_CASE(MeanSimpleQuantisedSymm16, MeanSimpleTest<DataType::QSymmS16>) +ARMNN_AUTO_TEST_CASE(MeanSimpleAxisQuantisedSymm16, MeanSimpleAxisTest<DataType::QSymmS16>) +ARMNN_AUTO_TEST_CASE(MeanKeepDimsQuantisedSymm16, MeanKeepDimsTest<DataType::QSymmS16>) +ARMNN_AUTO_TEST_CASE(MeanMultipleDimsQuantisedSymm16, MeanMultipleDimsTest<DataType::QSymmS16>) +ARMNN_AUTO_TEST_CASE(MeanVts1QuantisedSymm16, MeanVts1Test<DataType::QSymmS16>) +ARMNN_AUTO_TEST_CASE(MeanVts2QuantisedSymm16, MeanVts2Test<DataType::QSymmS16>) +ARMNN_AUTO_TEST_CASE(MeanVts3QuantisedSymm16, MeanVts3Test<DataType::QSymmS16>) ARMNN_AUTO_TEST_CASE(AdditionAfterMaxPool, AdditionAfterMaxPoolTest) @@ -1123,15 +1123,15 @@ ARMNN_AUTO_TEST_CASE(ArgMaxChannelSigned32, ArgMaxChannelTest<DataType::Signed32 ARMNN_AUTO_TEST_CASE(ArgMaxHeightSigned32, ArgMaxHeightTest<DataType::Signed32>) ARMNN_AUTO_TEST_CASE(ArgMinWidthSigned32, ArgMinWidthTest<DataType::Signed32>) -ARMNN_AUTO_TEST_CASE(ArgMaxSimpleQuantisedAsymm8, ArgMaxSimpleTest<DataType::QuantisedAsymm8>) -ARMNN_AUTO_TEST_CASE(ArgMinSimpleQuantisedAsymm8, ArgMinSimpleTest<DataType::QuantisedAsymm8>) -ARMNN_AUTO_TEST_CASE(ArgMinChannelQuantisedAsymm8, ArgMinChannelTest<DataType::QuantisedAsymm8>) -ARMNN_AUTO_TEST_CASE(ArgMaxChannelQuantisedAsymm8, ArgMaxChannelTest<DataType::QuantisedAsymm8>) +ARMNN_AUTO_TEST_CASE(ArgMaxSimpleQuantisedAsymm8, ArgMaxSimpleTest<DataType::QAsymmU8>) +ARMNN_AUTO_TEST_CASE(ArgMinSimpleQuantisedAsymm8, ArgMinSimpleTest<DataType::QAsymmU8>) +ARMNN_AUTO_TEST_CASE(ArgMinChannelQuantisedAsymm8, ArgMinChannelTest<DataType::QAsymmU8>) +ARMNN_AUTO_TEST_CASE(ArgMaxChannelQuantisedAsymm8, ArgMaxChannelTest<DataType::QAsymmU8>) -ARMNN_AUTO_TEST_CASE(ArgMaxSimpleQuantisedSymm16, ArgMaxSimpleTest<DataType::QuantisedSymm16>) -ARMNN_AUTO_TEST_CASE(ArgMinSimpleQuantisedSymm16, ArgMinSimpleTest<DataType::QuantisedSymm16>) -ARMNN_AUTO_TEST_CASE(ArgMinChannelQuantisedSymm16, ArgMinChannelTest<DataType::QuantisedSymm16>) -ARMNN_AUTO_TEST_CASE(ArgMaxChannelQuantisedSymm16, ArgMaxChannelTest<DataType::QuantisedSymm16>) +ARMNN_AUTO_TEST_CASE(ArgMaxSimpleQuantisedSymm16, ArgMaxSimpleTest<DataType::QSymmS16>) +ARMNN_AUTO_TEST_CASE(ArgMinSimpleQuantisedSymm16, ArgMinSimpleTest<DataType::QSymmS16>) +ARMNN_AUTO_TEST_CASE(ArgMinChannelQuantisedSymm16, ArgMinChannelTest<DataType::QSymmS16>) +ARMNN_AUTO_TEST_CASE(ArgMaxChannelQuantisedSymm16, ArgMaxChannelTest<DataType::QSymmS16>) // Space To Batch Nd ARMNN_AUTO_TEST_CASE(SpaceToBatchNdSimpleFloat32, SpaceToBatchNdSimpleFloat32Test) @@ -1191,21 +1191,21 @@ ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNhwcFloat16_5, BatchToSpaceNdNhwcTest5<DataTy ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNhwcFloat16_6, BatchToSpaceNdNhwcTest6<DataType::Float16>) ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNhwcFloat16_7, BatchToSpaceNdNhwcTest7<DataType::Float16>) -ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNhwcUint1, BatchToSpaceNdNhwcTest1<DataType::QuantisedAsymm8>) -ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNhwcUint2, BatchToSpaceNdNhwcTest2<DataType::QuantisedAsymm8>) -ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNhwcUint3, BatchToSpaceNdNhwcTest3<DataType::QuantisedAsymm8>) -ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNhwcUint4, BatchToSpaceNdNhwcTest4<DataType::QuantisedAsymm8>) -ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNhwcUint5, BatchToSpaceNdNhwcTest5<DataType::QuantisedAsymm8>) -ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNhwcUint6, BatchToSpaceNdNhwcTest6<DataType::QuantisedAsymm8>) -ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNhwcUint7, BatchToSpaceNdNhwcTest7<DataType::QuantisedAsymm8>) - -ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNhwcQsymm16_1, BatchToSpaceNdNhwcTest1<DataType::QuantisedSymm16>) -ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNhwcQsymm16_2, BatchToSpaceNdNhwcTest2<DataType::QuantisedSymm16>) -ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNhwcQsymm16_3, BatchToSpaceNdNhwcTest3<DataType::QuantisedSymm16>) -ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNhwcQsymm16_4, BatchToSpaceNdNhwcTest4<DataType::QuantisedSymm16>) -ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNhwcQsymm16_5, BatchToSpaceNdNhwcTest5<DataType::QuantisedSymm16>) -ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNhwcQsymm16_6, BatchToSpaceNdNhwcTest6<DataType::QuantisedSymm16>) -ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNhwcQsymm16_7, BatchToSpaceNdNhwcTest7<DataType::QuantisedSymm16>) +ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNhwcUint1, BatchToSpaceNdNhwcTest1<DataType::QAsymmU8>) +ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNhwcUint2, BatchToSpaceNdNhwcTest2<DataType::QAsymmU8>) +ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNhwcUint3, BatchToSpaceNdNhwcTest3<DataType::QAsymmU8>) +ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNhwcUint4, BatchToSpaceNdNhwcTest4<DataType::QAsymmU8>) +ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNhwcUint5, BatchToSpaceNdNhwcTest5<DataType::QAsymmU8>) +ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNhwcUint6, BatchToSpaceNdNhwcTest6<DataType::QAsymmU8>) +ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNhwcUint7, BatchToSpaceNdNhwcTest7<DataType::QAsymmU8>) + +ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNhwcQsymm16_1, BatchToSpaceNdNhwcTest1<DataType::QSymmS16>) +ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNhwcQsymm16_2, BatchToSpaceNdNhwcTest2<DataType::QSymmS16>) +ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNhwcQsymm16_3, BatchToSpaceNdNhwcTest3<DataType::QSymmS16>) +ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNhwcQsymm16_4, BatchToSpaceNdNhwcTest4<DataType::QSymmS16>) +ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNhwcQsymm16_5, BatchToSpaceNdNhwcTest5<DataType::QSymmS16>) +ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNhwcQsymm16_6, BatchToSpaceNdNhwcTest6<DataType::QSymmS16>) +ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNhwcQsymm16_7, BatchToSpaceNdNhwcTest7<DataType::QSymmS16>) ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwFloat16_1, BatchToSpaceNdNchwTest1<DataType::Float16>) ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwFloat16_2, BatchToSpaceNdNchwTest2<DataType::Float16>) @@ -1215,21 +1215,21 @@ ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwFloat16_5, BatchToSpaceNdNchwTest5<DataTy ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwFloat16_6, BatchToSpaceNdNchwTest6<DataType::Float16>) ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwFloat16_7, BatchToSpaceNdNchwTest7<DataType::Float16>) -ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwUint1, BatchToSpaceNdNchwTest1<DataType::QuantisedAsymm8>) -ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwUint2, BatchToSpaceNdNchwTest2<DataType::QuantisedAsymm8>) -ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwUint3, BatchToSpaceNdNchwTest3<DataType::QuantisedAsymm8>) -ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwUint4, BatchToSpaceNdNchwTest4<DataType::QuantisedAsymm8>) -ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwUint5, BatchToSpaceNdNchwTest5<DataType::QuantisedAsymm8>) -ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwUint6, BatchToSpaceNdNchwTest6<DataType::QuantisedAsymm8>) -ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwUint7, BatchToSpaceNdNchwTest7<DataType::QuantisedAsymm8>) - -ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwQsymm16_1, BatchToSpaceNdNchwTest1<DataType::QuantisedSymm16>) -ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwQsymm16_2, BatchToSpaceNdNchwTest2<DataType::QuantisedSymm16>) -ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwQsymm16_3, BatchToSpaceNdNchwTest3<DataType::QuantisedSymm16>) -ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwQsymm16_4, BatchToSpaceNdNchwTest4<DataType::QuantisedSymm16>) -ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwQsymm16_5, BatchToSpaceNdNchwTest5<DataType::QuantisedSymm16>) -ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwQsymm16_6, BatchToSpaceNdNchwTest6<DataType::QuantisedSymm16>) -ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwQsymm16_7, BatchToSpaceNdNchwTest7<DataType::QuantisedSymm16>) +ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwUint1, BatchToSpaceNdNchwTest1<DataType::QAsymmU8>) +ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwUint2, BatchToSpaceNdNchwTest2<DataType::QAsymmU8>) +ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwUint3, BatchToSpaceNdNchwTest3<DataType::QAsymmU8>) +ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwUint4, BatchToSpaceNdNchwTest4<DataType::QAsymmU8>) +ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwUint5, BatchToSpaceNdNchwTest5<DataType::QAsymmU8>) +ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwUint6, BatchToSpaceNdNchwTest6<DataType::QAsymmU8>) +ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwUint7, BatchToSpaceNdNchwTest7<DataType::QAsymmU8>) + +ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwQsymm16_1, BatchToSpaceNdNchwTest1<DataType::QSymmS16>) +ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwQsymm16_2, BatchToSpaceNdNchwTest2<DataType::QSymmS16>) +ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwQsymm16_3, BatchToSpaceNdNchwTest3<DataType::QSymmS16>) +ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwQsymm16_4, BatchToSpaceNdNchwTest4<DataType::QSymmS16>) +ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwQsymm16_5, BatchToSpaceNdNchwTest5<DataType::QSymmS16>) +ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwQsymm16_6, BatchToSpaceNdNchwTest6<DataType::QSymmS16>) +ARMNN_AUTO_TEST_CASE(BatchToSpaceNdNchwQsymm16_7, BatchToSpaceNdNchwTest7<DataType::QSymmS16>) // DepthToSpace ARMNN_AUTO_TEST_CASE(DepthToSpaceNchwFloat32_1, DepthToSpaceTest1<DataType::Float32>, DataLayout::NCHW); @@ -1242,15 +1242,15 @@ ARMNN_AUTO_TEST_CASE(DepthToSpaceNchwFloat16_2, DepthToSpaceTest2<DataType::Floa ARMNN_AUTO_TEST_CASE(DepthToSpaceNchwFloat16_3, DepthToSpaceTest3<DataType::Float16>, DataLayout::NCHW); ARMNN_AUTO_TEST_CASE(DepthToSpaceNchwFloat16_4, DepthToSpaceTest4<DataType::Float16>, DataLayout::NCHW); -ARMNN_AUTO_TEST_CASE(DepthToSpaceNchwUint8_1, DepthToSpaceTest1<DataType::QuantisedAsymm8>, DataLayout::NCHW); -ARMNN_AUTO_TEST_CASE(DepthToSpaceNchwUint8_2, DepthToSpaceTest2<DataType::QuantisedAsymm8>, DataLayout::NCHW); -ARMNN_AUTO_TEST_CASE(DepthToSpaceNchwUint8_3, DepthToSpaceTest3<DataType::QuantisedAsymm8>, DataLayout::NCHW); -ARMNN_AUTO_TEST_CASE(DepthToSpaceNchwUint8_4, DepthToSpaceTest4<DataType::QuantisedAsymm8>, DataLayout::NCHW); +ARMNN_AUTO_TEST_CASE(DepthToSpaceNchwUint8_1, DepthToSpaceTest1<DataType::QAsymmU8>, DataLayout::NCHW); +ARMNN_AUTO_TEST_CASE(DepthToSpaceNchwUint8_2, DepthToSpaceTest2<DataType::QAsymmU8>, DataLayout::NCHW); +ARMNN_AUTO_TEST_CASE(DepthToSpaceNchwUint8_3, DepthToSpaceTest3<DataType::QAsymmU8>, DataLayout::NCHW); +ARMNN_AUTO_TEST_CASE(DepthToSpaceNchwUint8_4, DepthToSpaceTest4<DataType::QAsymmU8>, DataLayout::NCHW); -ARMNN_AUTO_TEST_CASE(DepthToSpaceNchwInt16_1, DepthToSpaceTest1<DataType::QuantisedSymm16>, DataLayout::NCHW); -ARMNN_AUTO_TEST_CASE(DepthToSpaceNchwInt16_2, DepthToSpaceTest2<DataType::QuantisedSymm16>, DataLayout::NCHW); -ARMNN_AUTO_TEST_CASE(DepthToSpaceNchwInt16_3, DepthToSpaceTest3<DataType::QuantisedSymm16>, DataLayout::NCHW); -ARMNN_AUTO_TEST_CASE(DepthToSpaceNchwInt16_4, DepthToSpaceTest4<DataType::QuantisedSymm16>, DataLayout::NCHW); +ARMNN_AUTO_TEST_CASE(DepthToSpaceNchwInt16_1, DepthToSpaceTest1<DataType::QSymmS16>, DataLayout::NCHW); +ARMNN_AUTO_TEST_CASE(DepthToSpaceNchwInt16_2, DepthToSpaceTest2<DataType::QSymmS16>, DataLayout::NCHW); +ARMNN_AUTO_TEST_CASE(DepthToSpaceNchwInt16_3, DepthToSpaceTest3<DataType::QSymmS16>, DataLayout::NCHW); +ARMNN_AUTO_TEST_CASE(DepthToSpaceNchwInt16_4, DepthToSpaceTest4<DataType::QSymmS16>, DataLayout::NCHW); ARMNN_AUTO_TEST_CASE(DepthToSpaceNhwcFloat32_1, DepthToSpaceTest1<DataType::Float32>, DataLayout::NHWC); ARMNN_AUTO_TEST_CASE(DepthToSpaceNhwcFloat32_2, DepthToSpaceTest2<DataType::Float32>, DataLayout::NHWC); @@ -1262,15 +1262,15 @@ ARMNN_AUTO_TEST_CASE(DepthToSpaceNhwcFloat16_2, DepthToSpaceTest2<DataType::Floa ARMNN_AUTO_TEST_CASE(DepthToSpaceNhwcFloat16_3, DepthToSpaceTest3<DataType::Float16>, DataLayout::NHWC); ARMNN_AUTO_TEST_CASE(DepthToSpaceNhwcFloat16_4, DepthToSpaceTest4<DataType::Float16>, DataLayout::NHWC); -ARMNN_AUTO_TEST_CASE(DepthToSpaceNhwcUint8_1, DepthToSpaceTest1<DataType::QuantisedAsymm8>, DataLayout::NHWC); -ARMNN_AUTO_TEST_CASE(DepthToSpaceNhwcUint8_2, DepthToSpaceTest2<DataType::QuantisedAsymm8>, DataLayout::NHWC); -ARMNN_AUTO_TEST_CASE(DepthToSpaceNhwcUint8_3, DepthToSpaceTest3<DataType::QuantisedAsymm8>, DataLayout::NHWC); -ARMNN_AUTO_TEST_CASE(DepthToSpaceNhwcUint8_4, DepthToSpaceTest4<DataType::QuantisedAsymm8>, DataLayout::NHWC); +ARMNN_AUTO_TEST_CASE(DepthToSpaceNhwcUint8_1, DepthToSpaceTest1<DataType::QAsymmU8>, DataLayout::NHWC); +ARMNN_AUTO_TEST_CASE(DepthToSpaceNhwcUint8_2, DepthToSpaceTest2<DataType::QAsymmU8>, DataLayout::NHWC); +ARMNN_AUTO_TEST_CASE(DepthToSpaceNhwcUint8_3, DepthToSpaceTest3<DataType::QAsymmU8>, DataLayout::NHWC); +ARMNN_AUTO_TEST_CASE(DepthToSpaceNhwcUint8_4, DepthToSpaceTest4<DataType::QAsymmU8>, DataLayout::NHWC); -ARMNN_AUTO_TEST_CASE(DepthToSpaceNhwcInt16_1, DepthToSpaceTest1<DataType::QuantisedSymm16>, DataLayout::NHWC); -ARMNN_AUTO_TEST_CASE(DepthToSpaceNhwcInt16_2, DepthToSpaceTest2<DataType::QuantisedSymm16>, DataLayout::NHWC); -ARMNN_AUTO_TEST_CASE(DepthToSpaceNhwcInt16_3, DepthToSpaceTest3<DataType::QuantisedSymm16>, DataLayout::NHWC); -ARMNN_AUTO_TEST_CASE(DepthToSpaceNhwcInt16_4, DepthToSpaceTest4<DataType::QuantisedSymm16>, DataLayout::NHWC); +ARMNN_AUTO_TEST_CASE(DepthToSpaceNhwcInt16_1, DepthToSpaceTest1<DataType::QSymmS16>, DataLayout::NHWC); +ARMNN_AUTO_TEST_CASE(DepthToSpaceNhwcInt16_2, DepthToSpaceTest2<DataType::QSymmS16>, DataLayout::NHWC); +ARMNN_AUTO_TEST_CASE(DepthToSpaceNhwcInt16_3, DepthToSpaceTest3<DataType::QSymmS16>, DataLayout::NHWC); +ARMNN_AUTO_TEST_CASE(DepthToSpaceNhwcInt16_4, DepthToSpaceTest4<DataType::QSymmS16>, DataLayout::NHWC); // SpaceToDepth ARMNN_AUTO_TEST_CASE(SpaceToDepthNchwAsymmQ8, SpaceToDepthNchwAsymmQ8Test) @@ -1381,10 +1381,10 @@ ARMNN_AUTO_TEST_CASE(Abs3d, Abs3dTest<DataType::Float32>) ARMNN_AUTO_TEST_CASE(AbsZero, AbsZeroTest<DataType::Float32>) ARMNN_AUTO_TEST_CASE(Abs2dFloat16, Abs2dTest<DataType::Float16>) ARMNN_AUTO_TEST_CASE(Abs3dFloat16, Abs3dTest<DataType::Float16>) -ARMNN_AUTO_TEST_CASE(Abs2dQuantisedAsymm8, Abs2dTest<DataType::QuantisedAsymm8>) -ARMNN_AUTO_TEST_CASE(Abs3dQuantisedAsymm8, Abs3dTest<DataType::QuantisedAsymm8>) -ARMNN_AUTO_TEST_CASE(Abs2dQuantisedSymm16, Abs2dTest<DataType::QuantisedSymm16>) -ARMNN_AUTO_TEST_CASE(Abs3dQuantisedSymm16, Abs3dTest<DataType::QuantisedSymm16>) +ARMNN_AUTO_TEST_CASE(Abs2dQuantisedAsymm8, Abs2dTest<DataType::QAsymmU8>) +ARMNN_AUTO_TEST_CASE(Abs3dQuantisedAsymm8, Abs3dTest<DataType::QAsymmU8>) +ARMNN_AUTO_TEST_CASE(Abs2dQuantisedSymm16, Abs2dTest<DataType::QSymmS16>) +ARMNN_AUTO_TEST_CASE(Abs3dQuantisedSymm16, Abs3dTest<DataType::QSymmS16>) // Detection PostProcess BOOST_AUTO_TEST_CASE(DetectionPostProcessRegularNmsFloat) @@ -1398,22 +1398,22 @@ BOOST_AUTO_TEST_CASE(DetectionPostProcessFastNmsFloat) BOOST_AUTO_TEST_CASE(DetectionPostProcessRegularNmsUint8) { DetectionPostProcessRegularNmsQuantizedTest< - RefWorkloadFactory, DataType::QuantisedAsymm8>(); + RefWorkloadFactory, DataType::QAsymmU8>(); } BOOST_AUTO_TEST_CASE(DetectionPostProcessFastNmsUint8) { DetectionPostProcessRegularNmsQuantizedTest< - RefWorkloadFactory, DataType::QuantisedAsymm8>(); + RefWorkloadFactory, DataType::QAsymmU8>(); } BOOST_AUTO_TEST_CASE(DetectionPostProcessRegularNmsInt16) { DetectionPostProcessRegularNmsQuantizedTest< - RefWorkloadFactory, DataType::QuantisedSymm16>(); + RefWorkloadFactory, DataType::QSymmS16>(); } BOOST_AUTO_TEST_CASE(DetectionPostProcessFastNmsInt16) { DetectionPostProcessFastNmsQuantizedTest< - RefWorkloadFactory, DataType::QuantisedSymm16>(); + RefWorkloadFactory, DataType::QSymmS16>(); } // Dequantize @@ -1434,8 +1434,8 @@ ARMNN_AUTO_TEST_CASE(QuantizeClampInt16, QuantizeClampInt16Test) // PReLU ARMNN_AUTO_TEST_CASE(PreluFloat32, PreluTest<DataType::Float32>) ARMNN_AUTO_TEST_CASE(PreluFloat16, PreluTest<DataType::Float16>) -ARMNN_AUTO_TEST_CASE(PreluUint8, PreluTest<DataType::QuantisedAsymm8>) -ARMNN_AUTO_TEST_CASE(PreluInt16, PreluTest<DataType::QuantisedSymm16>) +ARMNN_AUTO_TEST_CASE(PreluUint8, PreluTest<DataType::QAsymmU8>) +ARMNN_AUTO_TEST_CASE(PreluInt16, PreluTest<DataType::QSymmS16>) // Slice ARMNN_AUTO_TEST_CASE(Slice4dFloat32, Slice4dFloat32Test) @@ -1463,19 +1463,19 @@ ARMNN_AUTO_TEST_CASE(SimpleTransposeConvolution2dFloatNhwc, true, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(SimpleTransposeConvolution2dUint8Nchw, - SimpleTransposeConvolution2dTest<DataType::QuantisedAsymm8, DataType::Signed32>, + SimpleTransposeConvolution2dTest<DataType::QAsymmU8, DataType::Signed32>, true, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(SimpleTransposeConvolution2dUint8Nhwc, - SimpleTransposeConvolution2dTest<DataType::QuantisedAsymm8, DataType::Signed32>, + SimpleTransposeConvolution2dTest<DataType::QAsymmU8, DataType::Signed32>, true, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(SimpleTransposeConvolution2dInt16Nchw, - SimpleTransposeConvolution2dTest<DataType::QuantisedSymm16, DataType::Signed32>, + SimpleTransposeConvolution2dTest<DataType::QSymmS16, DataType::Signed32>, true, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(SimpleTransposeConvolution2dInt16Nhwc, - SimpleTransposeConvolution2dTest<DataType::QuantisedSymm16, DataType::Signed32>, + SimpleTransposeConvolution2dTest<DataType::QSymmS16, DataType::Signed32>, true, DataLayout::NCHW) @@ -1488,19 +1488,19 @@ ARMNN_AUTO_TEST_CASE(UnbiasedSimpleTransposeConvolution2dFloatNhwc, true, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(UnbiasedSimpleTransposeConvolution2dUint8Nchw, - SimpleTransposeConvolution2dTest<DataType::QuantisedAsymm8, DataType::Signed32>, + SimpleTransposeConvolution2dTest<DataType::QAsymmU8, DataType::Signed32>, true, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(UnbiasedSimpleTransposeConvolution2dUint8Nhwc, - SimpleTransposeConvolution2dTest<DataType::QuantisedAsymm8, DataType::Signed32>, + SimpleTransposeConvolution2dTest<DataType::QAsymmU8, DataType::Signed32>, true, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(UnbiasedSimpleTransposeConvolution2dInt16Nchw, - SimpleTransposeConvolution2dTest<DataType::QuantisedSymm16, DataType::Signed32>, + SimpleTransposeConvolution2dTest<DataType::QSymmS16, DataType::Signed32>, true, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(UnbiasedSimpleTransposeConvolution2dInt16Nhwc, - SimpleTransposeConvolution2dTest<DataType::QuantisedSymm16, DataType::Signed32>, + SimpleTransposeConvolution2dTest<DataType::QSymmS16, DataType::Signed32>, true, DataLayout::NCHW) @@ -1513,19 +1513,19 @@ ARMNN_AUTO_TEST_CASE(PaddedTransposeConvolution2dFloatNhwc, true, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(PaddedTransposeConvolution2dUint8Nchw, - PaddedTransposeConvolution2dTest<DataType::QuantisedAsymm8, DataType::Signed32>, + PaddedTransposeConvolution2dTest<DataType::QAsymmU8, DataType::Signed32>, true, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(PaddedTransposeConvolution2dUint8Nhwc, - PaddedTransposeConvolution2dTest<DataType::QuantisedAsymm8, DataType::Signed32>, + PaddedTransposeConvolution2dTest<DataType::QAsymmU8, DataType::Signed32>, true, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(PaddedTransposeConvolution2dInt16Nchw, - PaddedTransposeConvolution2dTest<DataType::QuantisedSymm16, DataType::Signed32>, + PaddedTransposeConvolution2dTest<DataType::QSymmS16, DataType::Signed32>, true, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(PaddedTransposeConvolution2dInt16Nhwc, - PaddedTransposeConvolution2dTest<DataType::QuantisedSymm16, DataType::Signed32>, + PaddedTransposeConvolution2dTest<DataType::QSymmS16, DataType::Signed32>, true, DataLayout::NCHW) @@ -1538,19 +1538,19 @@ ARMNN_AUTO_TEST_CASE(UnbiasedPaddedTransposeConvolution2dFloatNhwc, true, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(UnbiasedPaddedTransposeConvolution2dUint8Nchw, - PaddedTransposeConvolution2dTest<DataType::QuantisedAsymm8, DataType::Signed32>, + PaddedTransposeConvolution2dTest<DataType::QAsymmU8, DataType::Signed32>, true, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(UnbiasedPaddedTransposeConvolution2dUint8Nhwc, - PaddedTransposeConvolution2dTest<DataType::QuantisedAsymm8, DataType::Signed32>, + PaddedTransposeConvolution2dTest<DataType::QAsymmU8, DataType::Signed32>, true, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(UnbiasedPaddedTransposeConvolution2dInt16Nchw, - PaddedTransposeConvolution2dTest<DataType::QuantisedSymm16, DataType::Signed32>, + PaddedTransposeConvolution2dTest<DataType::QSymmS16, DataType::Signed32>, true, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(UnbiasedPaddedTransposeConvolution2dInt16Nhwc, - PaddedTransposeConvolution2dTest<DataType::QuantisedSymm16, DataType::Signed32>, + PaddedTransposeConvolution2dTest<DataType::QSymmS16, DataType::Signed32>, true, DataLayout::NCHW) @@ -1563,19 +1563,19 @@ ARMNN_AUTO_TEST_CASE(StridedTransposeConvolution2dFloatNhwc, true, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(StridedTransposeConvolution2dUint8Nchw, - StridedTransposeConvolution2dTest<DataType::QuantisedAsymm8, DataType::Signed32>, + StridedTransposeConvolution2dTest<DataType::QAsymmU8, DataType::Signed32>, true, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(StridedTransposeConvolution2dUint8Nhwc, - StridedTransposeConvolution2dTest<DataType::QuantisedAsymm8, DataType::Signed32>, + StridedTransposeConvolution2dTest<DataType::QAsymmU8, DataType::Signed32>, true, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(StridedTransposeConvolution2dInt16Nchw, - StridedTransposeConvolution2dTest<DataType::QuantisedSymm16, DataType::Signed32>, + StridedTransposeConvolution2dTest<DataType::QSymmS16, DataType::Signed32>, true, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(StridedTransposeConvolution2dInt16Nhwc, - StridedTransposeConvolution2dTest<DataType::QuantisedSymm16, DataType::Signed32>, + StridedTransposeConvolution2dTest<DataType::QSymmS16, DataType::Signed32>, true, DataLayout::NCHW) @@ -1588,19 +1588,19 @@ ARMNN_AUTO_TEST_CASE(UnbiasedStridedTransposeConvolution2dFloatNhwc, true, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(UnbiasedStridedTransposeConvolution2dUint8Nchw, - StridedTransposeConvolution2dTest<DataType::QuantisedAsymm8, DataType::Signed32>, + StridedTransposeConvolution2dTest<DataType::QAsymmU8, DataType::Signed32>, true, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(UnbiasedStridedTransposeConvolution2dUint8Nhwc, - StridedTransposeConvolution2dTest<DataType::QuantisedAsymm8, DataType::Signed32>, + StridedTransposeConvolution2dTest<DataType::QAsymmU8, DataType::Signed32>, true, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(UnbiasedStridedTransposeConvolution2dInt16Nchw, - StridedTransposeConvolution2dTest<DataType::QuantisedSymm16, DataType::Signed32>, + StridedTransposeConvolution2dTest<DataType::QSymmS16, DataType::Signed32>, true, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(UnbiasedStridedTransposeConvolution2dInt16Nhwc, - StridedTransposeConvolution2dTest<DataType::QuantisedSymm16, DataType::Signed32>, + StridedTransposeConvolution2dTest<DataType::QSymmS16, DataType::Signed32>, true, DataLayout::NCHW) @@ -1611,16 +1611,16 @@ ARMNN_AUTO_TEST_CASE(MultiChannelTransposeConvolution2dFloatNhwc, MultiChannelTransposeConvolution2dTest<DataType::Float32, DataType::Float32>, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(MultiChannelTransposeConvolution2dUint8Nchw, - MultiChannelTransposeConvolution2dTest<DataType::QuantisedAsymm8, DataType::Signed32>, + MultiChannelTransposeConvolution2dTest<DataType::QAsymmU8, DataType::Signed32>, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(MultiChannelTransposeConvolution2dUint8Nhwc, - MultiChannelTransposeConvolution2dTest<DataType::QuantisedAsymm8, DataType::Signed32>, + MultiChannelTransposeConvolution2dTest<DataType::QAsymmU8, DataType::Signed32>, DataLayout::NHWC) ARMNN_AUTO_TEST_CASE(MultiChannelTransposeConvolution2dInt16Nchw, - MultiChannelTransposeConvolution2dTest<DataType::QuantisedSymm16, DataType::Signed32>, + MultiChannelTransposeConvolution2dTest<DataType::QSymmS16, DataType::Signed32>, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(MultiChannelTransposeConvolution2dInt16Nhwc, - MultiChannelTransposeConvolution2dTest<DataType::QuantisedSymm16, DataType::Signed32>, + MultiChannelTransposeConvolution2dTest<DataType::QSymmS16, DataType::Signed32>, DataLayout::NCHW) ARMNN_AUTO_TEST_CASE(TransposeConvolution2dPerAxisQuantTestNchw, diff --git a/src/backends/reference/workloads/Decoders.hpp b/src/backends/reference/workloads/Decoders.hpp index 9d41c9e9e7..faabdcdb3f 100644 --- a/src/backends/reference/workloads/Decoders.hpp +++ b/src/backends/reference/workloads/Decoders.hpp @@ -79,14 +79,14 @@ inline std::unique_ptr<Decoder<float>> MakeDecoder(const TensorInfo& info, const params.second, params.first); } - case DataType::QuantisedAsymm8: + case DataType::QAsymmU8: { return std::make_unique<QASymm8Decoder>( static_cast<const uint8_t*>(data), info.GetQuantizationScale(), info.GetQuantizationOffset()); } - case DataType::QuantisedSymm16: + case DataType::QSymmS16: { return std::make_unique<QSymm16Decoder>( static_cast<const int16_t*>(data), diff --git a/src/backends/reference/workloads/Encoders.hpp b/src/backends/reference/workloads/Encoders.hpp index 92493ed641..4fe202f0bf 100644 --- a/src/backends/reference/workloads/Encoders.hpp +++ b/src/backends/reference/workloads/Encoders.hpp @@ -30,7 +30,7 @@ inline std::unique_ptr<Encoder<float>> MakeEncoder(const TensorInfo& info, void* params.second, params.first); } - case armnn::DataType::QuantisedAsymm8: + case armnn::DataType::QAsymmU8: { return std::make_unique<QASymm8Encoder>( static_cast<uint8_t*>(data), @@ -44,7 +44,7 @@ inline std::unique_ptr<Encoder<float>> MakeEncoder(const TensorInfo& info, void* info.GetQuantizationScale(), info.GetQuantizationOffset()); } - case armnn::DataType::QuantisedSymm16: + case armnn::DataType::QSymmS16: { return std::make_unique<QSymm16Encoder>( static_cast<int16_t*>(data), diff --git a/src/backends/reference/workloads/RefDebugWorkload.cpp b/src/backends/reference/workloads/RefDebugWorkload.cpp index 2a3883f8f7..9572f9a2a5 100644 --- a/src/backends/reference/workloads/RefDebugWorkload.cpp +++ b/src/backends/reference/workloads/RefDebugWorkload.cpp @@ -46,7 +46,7 @@ void RefDebugWorkload<DataType>::RegisterDebugCallback(const DebugCallbackFuncti template class RefDebugWorkload<DataType::Float16>; template class RefDebugWorkload<DataType::Float32>; -template class RefDebugWorkload<DataType::QuantisedAsymm8>; -template class RefDebugWorkload<DataType::QuantisedSymm16>; +template class RefDebugWorkload<DataType::QAsymmU8>; +template class RefDebugWorkload<DataType::QSymmS16>; } // namespace armnn diff --git a/src/backends/reference/workloads/RefDebugWorkload.hpp b/src/backends/reference/workloads/RefDebugWorkload.hpp index 0964515b2c..fc154e9457 100644 --- a/src/backends/reference/workloads/RefDebugWorkload.hpp +++ b/src/backends/reference/workloads/RefDebugWorkload.hpp @@ -39,7 +39,7 @@ private: using RefDebugFloat16Workload = RefDebugWorkload<DataType::Float16>; using RefDebugFloat32Workload = RefDebugWorkload<DataType::Float32>; -using RefDebugQAsymm8Workload = RefDebugWorkload<DataType::QuantisedAsymm8>; -using RefDebugQSymm16Workload = RefDebugWorkload<DataType::QuantisedSymm16>; +using RefDebugQAsymm8Workload = RefDebugWorkload<DataType::QAsymmU8>; +using RefDebugQSymm16Workload = RefDebugWorkload<DataType::QSymmS16>; } // namespace armnn diff --git a/src/backends/reference/workloads/RefPadWorkload.cpp b/src/backends/reference/workloads/RefPadWorkload.cpp index c4b9daeb4c..356f6b1172 100644 --- a/src/backends/reference/workloads/RefPadWorkload.cpp +++ b/src/backends/reference/workloads/RefPadWorkload.cpp @@ -35,7 +35,7 @@ void RefPadWorkload<DataType>::Execute() const template class RefPadWorkload<DataType::Float32>; template class RefPadWorkload<DataType::Float16>; -template class RefPadWorkload<DataType::QuantisedAsymm8>; -template class RefPadWorkload<DataType::QuantisedSymm16>; +template class RefPadWorkload<DataType::QAsymmU8>; +template class RefPadWorkload<DataType::QSymmS16>; } //namespace armnn
\ No newline at end of file diff --git a/src/backends/reference/workloads/RefPadWorkload.hpp b/src/backends/reference/workloads/RefPadWorkload.hpp index d1521f4f8d..28fb55386e 100644 --- a/src/backends/reference/workloads/RefPadWorkload.hpp +++ b/src/backends/reference/workloads/RefPadWorkload.hpp @@ -32,7 +32,7 @@ public: using RefPadFloat32Workload = RefPadWorkload<DataType::Float32>; using RefPadFloat16Workload = RefPadWorkload<DataType::Float16>; -using RefPadQAsymm8Workload = RefPadWorkload<DataType::QuantisedAsymm8>; -using RefPadQSymm16Workload = RefPadWorkload<DataType::QuantisedSymm16>; +using RefPadQAsymm8Workload = RefPadWorkload<DataType::QAsymmU8>; +using RefPadQSymm16Workload = RefPadWorkload<DataType::QSymmS16>; } //namespace armnn diff --git a/src/backends/reference/workloads/RefPermuteWorkload.cpp b/src/backends/reference/workloads/RefPermuteWorkload.cpp index 4e7b76bf0a..d0e1431ffd 100644 --- a/src/backends/reference/workloads/RefPermuteWorkload.cpp +++ b/src/backends/reference/workloads/RefPermuteWorkload.cpp @@ -30,7 +30,7 @@ void RefPermuteWorkload<DataType>::Execute() const template class RefPermuteWorkload<DataType::Float16>; template class RefPermuteWorkload<DataType::Float32>; -template class RefPermuteWorkload<DataType::QuantisedAsymm8>; -template class RefPermuteWorkload<DataType::QuantisedSymm16>; +template class RefPermuteWorkload<DataType::QAsymmU8>; +template class RefPermuteWorkload<DataType::QSymmS16>; } //namespace armnn diff --git a/src/backends/reference/workloads/RefPermuteWorkload.hpp b/src/backends/reference/workloads/RefPermuteWorkload.hpp index 1e69afb0ed..00a33850aa 100644 --- a/src/backends/reference/workloads/RefPermuteWorkload.hpp +++ b/src/backends/reference/workloads/RefPermuteWorkload.hpp @@ -29,7 +29,7 @@ public: using RefPermuteFloat16Workload = RefPermuteWorkload<DataType::Float16>; using RefPermuteFloat32Workload = RefPermuteWorkload<DataType::Float32>; -using RefPermuteQAsymm8Workload = RefPermuteWorkload<DataType::QuantisedAsymm8>; -using RefPermuteQSymm16Workload = RefPermuteWorkload<DataType::QuantisedSymm16>; +using RefPermuteQAsymm8Workload = RefPermuteWorkload<DataType::QAsymmU8>; +using RefPermuteQSymm16Workload = RefPermuteWorkload<DataType::QSymmS16>; } //namespace armnn
\ No newline at end of file diff --git a/src/backends/reference/workloads/RefQuantizeWorkload.cpp b/src/backends/reference/workloads/RefQuantizeWorkload.cpp index a78804b709..31534abe3e 100644 --- a/src/backends/reference/workloads/RefQuantizeWorkload.cpp +++ b/src/backends/reference/workloads/RefQuantizeWorkload.cpp @@ -43,7 +43,7 @@ void RefQuantizeWorkload::Execute() const switch(m_TargetType) { - case DataType::QuantisedAsymm8: + case DataType::QAsymmU8: { QuantizeImpl<uint8_t>(input, output, m_NumElements, m_Scale, m_Offset); break; @@ -53,7 +53,7 @@ void RefQuantizeWorkload::Execute() const QuantizeImpl<int8_t>(input, output, m_NumElements, m_Scale, m_Offset); break; } - case DataType::QuantisedSymm16: + case DataType::QSymmS16: { QuantizeImpl<int16_t>(input, output, m_NumElements, m_Scale, 0); break; |