diff options
author | Keith Davis <keith.davis@arm.com> | 2020-01-29 16:52:59 +0000 |
---|---|---|
committer | Keith Davis <keith.davis@arm.com> | 2020-01-31 16:43:52 +0000 |
commit | 5e51cd803c7d926b2a9bd7c680c8d59fe100c2f2 (patch) | |
tree | 963982112d739a85d25fcc08bed8b16be50f6a94 /src/backends/reference | |
parent | c809a29e5d5bd62fac05c7d6059e1fa409ab4dc1 (diff) | |
download | armnn-5e51cd803c7d926b2a9bd7c680c8d59fe100c2f2.tar.gz |
IVGCVSW-4390 Refactor QUANTIZE to make use of Decoder/Encoder types
* Add no-ops for CL/NEON Uint8
* Refactor Quantize workload to Decoder/Encoder types
Signed-off-by: Keith Davis <keith.davis@arm.com>
Change-Id: I80b09de528299b925e2ac38acd9a5019b8d3e4ac
Diffstat (limited to 'src/backends/reference')
-rw-r--r-- | src/backends/reference/RefLayerSupport.cpp | 7 | ||||
-rw-r--r-- | src/backends/reference/workloads/RefQuantizeWorkload.cpp | 55 | ||||
-rw-r--r-- | src/backends/reference/workloads/RefQuantizeWorkload.hpp | 10 |
3 files changed, 31 insertions, 41 deletions
diff --git a/src/backends/reference/RefLayerSupport.cpp b/src/backends/reference/RefLayerSupport.cpp index 8410c303ae..c6a3af46bf 100644 --- a/src/backends/reference/RefLayerSupport.cpp +++ b/src/backends/reference/RefLayerSupport.cpp @@ -1439,9 +1439,12 @@ bool RefLayerSupport::IsQuantizeSupported(const TensorInfo& input, bool supported = true; // Define supported input types. - std::array<DataType,2> supportedInputTypes = { + std::array<DataType,5> supportedInputTypes = { DataType::QSymmS8, - DataType::Float32 + DataType::Float32, + DataType::QAsymmU8, + DataType::QSymmS8, + DataType::QSymmS16 }; supported &= CheckSupportRule(TypeAnyOf(input, supportedInputTypes), reasonIfUnsupported, diff --git a/src/backends/reference/workloads/RefQuantizeWorkload.cpp b/src/backends/reference/workloads/RefQuantizeWorkload.cpp index ab2ee7fc4e..2eef5f33db 100644 --- a/src/backends/reference/workloads/RefQuantizeWorkload.cpp +++ b/src/backends/reference/workloads/RefQuantizeWorkload.cpp @@ -5,6 +5,8 @@ #include "RefQuantizeWorkload.hpp" +#include "RefWorkloadUtils.hpp" + #include <armnn/TypesUtils.hpp> @@ -14,14 +16,13 @@ namespace armnn namespace { -template<typename T> -void QuantizeImpl(const void *input, void *output, size_t numValues, float scale, int offset) +void QuantizeImpl(Decoder<float>& in, Encoder<float>& out, size_t numValues) { - auto in = static_cast<const float *>(input); - auto out = static_cast<T *>(output); - for (size_t i = 0; i < numValues; i++, in++, out++) + for (unsigned int i = 0; i < numValues; i++) { - *out = armnn::Quantize<T>(*in, scale, offset); + in[i]; + out[i]; + out.Set(in.Get()); } } @@ -30,42 +31,24 @@ void QuantizeImpl(const void *input, void *output, size_t numValues, float scale RefQuantizeWorkload::RefQuantizeWorkload(const QuantizeQueueDescriptor& descriptor, const WorkloadInfo &info) : BaseWorkload(descriptor, info) , m_NumElements(info.m_InputTensorInfos[0].GetNumElements()) - , m_TargetType(info.m_OutputTensorInfos[0].GetDataType()) - , m_Scale(info.m_OutputTensorInfos[0].GetQuantizationScale()) - , m_Offset(info.m_OutputTensorInfos[0].GetQuantizationOffset()) { } -void RefQuantizeWorkload::Execute() const +void RefQuantizeWorkload::PostAllocationConfigure() { - const void* input = m_Data.m_Inputs[0]->Map(true); - void* output = m_Data.m_Outputs[0]->Map(true); + const TensorInfo& inputInfo = armnn::GetTensorInfo(m_Data.m_Inputs[0]); + m_InputDecoder = MakeDecoder<float>(inputInfo); - switch(m_TargetType) - { - case DataType::QAsymmU8: - { - QuantizeImpl<uint8_t>(input, output, m_NumElements, m_Scale, m_Offset); - break; - } - case DataType::QSymmS8: - { - QuantizeImpl<int8_t>(input, output, m_NumElements, m_Scale, 0); - break; - } - case DataType::QSymmS16: - { - QuantizeImpl<int16_t>(input, output, m_NumElements, m_Scale, 0); - break; - } - default: - { - BOOST_ASSERT_MSG(false, "RefQuantizeWorkload: Non quantized output type encountered"); - } - } + const TensorInfo& outputInfo = armnn::GetTensorInfo(m_Data.m_Outputs[0]); + m_OutputEncoder = MakeEncoder<float>(outputInfo); +} + +void RefQuantizeWorkload::Execute() const +{ + m_InputDecoder->Reset(m_Data.m_Inputs[0]->Map()); + m_OutputEncoder->Reset(m_Data.m_Outputs[0]->Map()); - m_Data.m_Inputs[0]->Unmap(); - m_Data.m_Outputs[0]->Unmap(); + QuantizeImpl(*m_InputDecoder, *m_OutputEncoder, m_NumElements); } } //namespace armnn
\ No newline at end of file diff --git a/src/backends/reference/workloads/RefQuantizeWorkload.hpp b/src/backends/reference/workloads/RefQuantizeWorkload.hpp index 6a43b8471d..9ae107607b 100644 --- a/src/backends/reference/workloads/RefQuantizeWorkload.hpp +++ b/src/backends/reference/workloads/RefQuantizeWorkload.hpp @@ -7,6 +7,8 @@ #include <backendsCommon/Workload.hpp> #include <backendsCommon/WorkloadData.hpp> +#include "Decoders.hpp" +#include "Encoders.hpp" namespace armnn { @@ -14,13 +16,15 @@ class RefQuantizeWorkload : public BaseWorkload<QuantizeQueueDescriptor> { public: RefQuantizeWorkload(const QuantizeQueueDescriptor& descriptor, const WorkloadInfo &info); + void PostAllocationConfigure() override; void Execute() const override; private: + + std::unique_ptr<Decoder<float>> m_InputDecoder; + std::unique_ptr<Encoder<float>> m_OutputEncoder; + size_t m_NumElements; - armnn::DataType m_TargetType; - float m_Scale; - int m_Offset; }; } //namespace armnn
\ No newline at end of file |