From fd2710651ada27fc82f28c07fb1e09effc3bda2d Mon Sep 17 00:00:00 2001 From: Finn Williams Date: Wed, 4 Dec 2019 14:27:27 +0000 Subject: IVGCVSW-4211 Add Signed 8 bit Quantisation support into the Reference backend !android-nn-driver:2435 Signed-off-by: Finn Williams Change-Id: I10ecd4a8937725953396805f33a3562a5384c4d4 --- src/backends/reference/RefLayerSupport.cpp | 8 ++-- src/backends/reference/test/RefLayerTests.cpp | 3 ++ src/backends/reference/workloads/BaseIterator.hpp | 43 ++++++++++++++++++++++ src/backends/reference/workloads/Decoders.hpp | 7 ++++ src/backends/reference/workloads/Encoders.hpp | 7 ++++ .../reference/workloads/RefQuantizeWorkload.cpp | 5 +++ 6 files changed, 70 insertions(+), 3 deletions(-) (limited to 'src/backends/reference') diff --git a/src/backends/reference/RefLayerSupport.cpp b/src/backends/reference/RefLayerSupport.cpp index 299503ddc6..19b76152f3 100644 --- a/src/backends/reference/RefLayerSupport.cpp +++ b/src/backends/reference/RefLayerSupport.cpp @@ -615,8 +615,9 @@ bool RefLayerSupport::IsDequantizeSupported(const TensorInfo& input, { bool supported = true; - std::array supportedInputTypes = { + std::array supportedInputTypes = { DataType::QuantisedAsymm8, + DataType::QSymmS8, DataType::QuantisedSymm16 }; @@ -1398,7 +1399,7 @@ bool RefLayerSupport::IsQuantizeSupported(const TensorInfo& input, { bool supported = true; - // Define supported output types. + // Define supported input types. std::array supportedInputTypes = { DataType::Float32, }; @@ -1407,8 +1408,9 @@ bool RefLayerSupport::IsQuantizeSupported(const TensorInfo& input, "Reference quantize: input type not supported."); // Define supported output types. - std::array supportedOutputTypes = { + std::array supportedOutputTypes = { DataType::QuantisedAsymm8, + DataType::QSymmS8, DataType::QuantisedSymm16 }; supported &= CheckSupportRule(TypeAnyOf(output, supportedOutputTypes), reasonIfUnsupported, diff --git a/src/backends/reference/test/RefLayerTests.cpp b/src/backends/reference/test/RefLayerTests.cpp index a397e935c1..b88f432acf 100644 --- a/src/backends/reference/test/RefLayerTests.cpp +++ b/src/backends/reference/test/RefLayerTests.cpp @@ -1419,13 +1419,16 @@ BOOST_AUTO_TEST_CASE(DetectionPostProcessFastNmsInt16) // Dequantize ARMNN_AUTO_TEST_CASE(DequantizeSimpleUint8, DequantizeSimpleUint8Test) ARMNN_AUTO_TEST_CASE(DequantizeOffsetUint8, DequantizeOffsetUint8Test) +ARMNN_AUTO_TEST_CASE(DequantizeSimpleInt8, DequantizeSimpleInt8Test) ARMNN_AUTO_TEST_CASE(DequantizeSimpleInt16, DequantizeSimpleInt16Test) ARMNN_AUTO_TEST_CASE(DequantizeSimpleUint8ToFp16, DequantizeSimpleUint8ToFp16Test) +ARMNN_AUTO_TEST_CASE(DequantizeSimpleInt8ToFp16, DequantizeSimpleInt8ToFp16Test) ARMNN_AUTO_TEST_CASE(DequantizeSimpleInt16ToFp16, DequantizeSimpleInt16ToFp16Test) // Quantize ARMNN_AUTO_TEST_CASE(QuantizeSimpleUint8, QuantizeSimpleUint8Test) ARMNN_AUTO_TEST_CASE(QuantizeClampUint8, QuantizeClampUint8Test) +ARMNN_AUTO_TEST_CASE(QuantizeClampInt8, QuantizeClampInt8Test) ARMNN_AUTO_TEST_CASE(QuantizeClampInt16, QuantizeClampInt16Test) // PReLU diff --git a/src/backends/reference/workloads/BaseIterator.hpp b/src/backends/reference/workloads/BaseIterator.hpp index ca5110c2fd..ca6d3cbc60 100644 --- a/src/backends/reference/workloads/BaseIterator.hpp +++ b/src/backends/reference/workloads/BaseIterator.hpp @@ -137,6 +137,25 @@ private: const int32_t m_Offset; }; +class QSymmS8Decoder : public TypedIterator> +{ +public: + QSymmS8Decoder(const int8_t* data, const float scale, const int32_t offset) + : TypedIterator(data), m_Scale(scale), m_Offset(offset) {} + + QSymmS8Decoder(const float scale, const int32_t offset) + : QSymmS8Decoder(nullptr, scale, offset) {} + + float Get() const override + { + return armnn::Dequantize(*m_Iterator, m_Scale, m_Offset); + } + +private: + const float m_Scale; + const int32_t m_Offset; +}; + class QSymm16Decoder : public TypedIterator> { public: @@ -245,6 +264,30 @@ private: const int32_t m_Offset; }; +class QSymmS8Encoder : public TypedIterator> +{ +public: + QSymmS8Encoder(int8_t* data, const float scale, const int32_t offset) + : TypedIterator(data), m_Scale(scale), m_Offset(offset) {} + + QSymmS8Encoder(const float scale, const int32_t offset) + : QSymmS8Encoder(nullptr, scale, offset) {} + + void Set(float right) override + { + *m_Iterator = armnn::Quantize(right, m_Scale, m_Offset); + } + + float Get() const override + { + return armnn::Dequantize(*m_Iterator, m_Scale, m_Offset); + } + +private: + const float m_Scale; + const int32_t m_Offset; +}; + class QSymm16Encoder : public TypedIterator> { public: diff --git a/src/backends/reference/workloads/Decoders.hpp b/src/backends/reference/workloads/Decoders.hpp index b9cd7f9573..9d41c9e9e7 100644 --- a/src/backends/reference/workloads/Decoders.hpp +++ b/src/backends/reference/workloads/Decoders.hpp @@ -105,6 +105,13 @@ inline std::unique_ptr> MakeDecoder(const TensorInfo& info, const { return MakeSigned32Decoder(info, data); } + case DataType::QSymmS8: + { + return std::make_unique( + static_cast(data), + info.GetQuantizationScale(), + info.GetQuantizationOffset()); + } default: { BOOST_ASSERT_MSG(false, "Unsupported Data Type!"); diff --git a/src/backends/reference/workloads/Encoders.hpp b/src/backends/reference/workloads/Encoders.hpp index 0d578d68de..92493ed641 100644 --- a/src/backends/reference/workloads/Encoders.hpp +++ b/src/backends/reference/workloads/Encoders.hpp @@ -37,6 +37,13 @@ inline std::unique_ptr> MakeEncoder(const TensorInfo& info, void* info.GetQuantizationScale(), info.GetQuantizationOffset()); } + case DataType::QSymmS8: + { + return std::make_unique( + static_cast(data), + info.GetQuantizationScale(), + info.GetQuantizationOffset()); + } case armnn::DataType::QuantisedSymm16: { return std::make_unique( diff --git a/src/backends/reference/workloads/RefQuantizeWorkload.cpp b/src/backends/reference/workloads/RefQuantizeWorkload.cpp index b7ace32e14..a78804b709 100644 --- a/src/backends/reference/workloads/RefQuantizeWorkload.cpp +++ b/src/backends/reference/workloads/RefQuantizeWorkload.cpp @@ -48,6 +48,11 @@ void RefQuantizeWorkload::Execute() const QuantizeImpl(input, output, m_NumElements, m_Scale, m_Offset); break; } + case DataType::QSymmS8: + { + QuantizeImpl(input, output, m_NumElements, m_Scale, m_Offset); + break; + } case DataType::QuantisedSymm16: { QuantizeImpl(input, output, m_NumElements, m_Scale, 0); -- cgit v1.2.1