From 8a54ac09dc4b9be1c1826bc63d4baf5101dafd53 Mon Sep 17 00:00:00 2001 From: Nattapat Chaimanowong Date: Fri, 29 Mar 2019 15:25:04 +0000 Subject: IVGCVSW-2875 Reference implementation and unit tests for Dequantize Change-Id: Ie4ade0519cb0bbe35dc36be6c9cd749b9171c74b Signed-off-by: Nattapat Chaimanowong --- src/backends/backendsCommon/MakeWorkloadHelper.hpp | 5 -- src/backends/backendsCommon/test/CMakeLists.txt | 1 + .../backendsCommon/test/DequantizeTestImpl.hpp | 95 ++++++++++++++++++++++ src/backends/backendsCommon/test/LayerTests.cpp | 8 ++ src/backends/backendsCommon/test/LayerTests.hpp | 4 + src/backends/reference/RefLayerSupport.cpp | 10 +++ src/backends/reference/RefLayerSupport.hpp | 4 + src/backends/reference/RefWorkloadFactory.cpp | 6 ++ src/backends/reference/RefWorkloadFactory.hpp | 3 + src/backends/reference/backend.mk | 1 + src/backends/reference/test/RefLayerTests.cpp | 2 + src/backends/reference/workloads/CMakeLists.txt | 2 + .../reference/workloads/RefDequantizeWorkload.cpp | 34 ++++++++ .../reference/workloads/RefDequantizeWorkload.hpp | 22 +++++ .../reference/workloads/RefWorkloadUtils.hpp | 9 ++ src/backends/reference/workloads/RefWorkloads.hpp | 1 + 16 files changed, 202 insertions(+), 5 deletions(-) create mode 100644 src/backends/backendsCommon/test/DequantizeTestImpl.hpp create mode 100644 src/backends/reference/workloads/RefDequantizeWorkload.cpp create mode 100644 src/backends/reference/workloads/RefDequantizeWorkload.hpp (limited to 'src/backends') diff --git a/src/backends/backendsCommon/MakeWorkloadHelper.hpp b/src/backends/backendsCommon/MakeWorkloadHelper.hpp index 2d54335355..9cd8774984 100644 --- a/src/backends/backendsCommon/MakeWorkloadHelper.hpp +++ b/src/backends/backendsCommon/MakeWorkloadHelper.hpp @@ -47,11 +47,6 @@ std::unique_ptr MakeWorkloadHelper(const QueueDescriptorType& descrip info.m_InputTensorInfos[0].GetDataType() : info.m_OutputTensorInfos[0].GetDataType(); - BOOST_ASSERT(info.m_InputTensorInfos.empty() || - info.m_OutputTensorInfos.empty() || - ((info.m_InputTensorInfos[0].GetDataType() == info.m_OutputTensorInfos[0].GetDataType()) || - info.m_OutputTensorInfos[0].GetDataType() == armnn::DataType::Boolean)); - switch (dataType) { case DataType::Float16: diff --git a/src/backends/backendsCommon/test/CMakeLists.txt b/src/backends/backendsCommon/test/CMakeLists.txt index 124d8f2744..47d52312b6 100644 --- a/src/backends/backendsCommon/test/CMakeLists.txt +++ b/src/backends/backendsCommon/test/CMakeLists.txt @@ -14,6 +14,7 @@ list(APPEND armnnBackendsCommonUnitTests_sources ConvertFp16ToFp32TestImpl.hpp ConvertFp32ToFp16TestImpl.hpp DebugTestImpl.hpp + DequantizeTestImpl.hpp DetectionPostProcessLayerTestImpl.hpp DetectionPostProcessTestImpl.hpp EndToEndTestImpl.hpp diff --git a/src/backends/backendsCommon/test/DequantizeTestImpl.hpp b/src/backends/backendsCommon/test/DequantizeTestImpl.hpp new file mode 100644 index 0000000000..5c10c751e6 --- /dev/null +++ b/src/backends/backendsCommon/test/DequantizeTestImpl.hpp @@ -0,0 +1,95 @@ +// +// Copyright © 2017 Arm Ltd. All rights reserved. +// SPDX-License-Identifier: MIT +// +#pragma once + +#include "WorkloadTestUtils.hpp" + +#include +#include +#include + +#include +#include +#include + +#include + +namespace +{ + +template +LayerTestResult DequantizeTestImpl( + armnn::IWorkloadFactory& workloadFactory, + const armnn::IBackendInternal::IMemoryManagerSharedPtr& memoryManager, + const armnn::TensorInfo& inputTensorInfo, + const armnn::TensorInfo& outputTensorInfo, + const std::vector& inputData, + const std::vector& expectedOutputData, + armnn::DequantizeQueueDescriptor descriptor) +{ + boost::multi_array input = MakeTensor(inputTensorInfo, inputData); + + LayerTestResult ret(outputTensorInfo); + ret.outputExpected = MakeTensor(outputTensorInfo, expectedOutputData); + + std::unique_ptr inputHandle = workloadFactory.CreateTensorHandle(inputTensorInfo); + std::unique_ptr outputHandle = workloadFactory.CreateTensorHandle(outputTensorInfo); + + armnn::WorkloadInfo info; + AddInputToWorkload(descriptor, info, inputTensorInfo, inputHandle.get()); + AddOutputToWorkload(descriptor, info, outputTensorInfo, outputHandle.get()); + + std::unique_ptr workload = workloadFactory.CreateDequantize(descriptor, info); + + inputHandle->Allocate(); + outputHandle->Allocate(); + + CopyDataToITensorHandle(inputHandle.get(), input.data()); + + ExecuteWorkload(*workload, memoryManager); + + CopyDataFromITensorHandle(ret.output.data(), outputHandle.get()); + + return ret; +} + +template +LayerTestResult DequantizeSimpleTest( + armnn::IWorkloadFactory& workloadFactory, + const armnn::IBackendInternal::IMemoryManagerSharedPtr& memoryManager) +{ + using T = armnn::ResolveType; + + armnn::DequantizeQueueDescriptor desc; + + const armnn::TensorInfo inputTensorInfo({1, 2, 2, 3}, ArmnnInputType, 0.5f, 1); + const armnn::TensorInfo outputTensorInfo({1, 2, 2, 3}, armnn::DataType::Float32); + + std::vector inputData = std::vector( + { + 3, 5, 7, + 9, 11, 13, + 15, 17, 19, + 21, 23, 25, + }); + + std::vector expectedOutputData = std::vector( + { + 1.0f, 2.0f, 3.0f, + 4.0f, 5.0f, 6.0f, + 7.0f, 8.0f, 9.0f, + 10.0f, 11.0f, 12.0f, + }); + + return DequantizeTestImpl(workloadFactory, + memoryManager, + inputTensorInfo, + outputTensorInfo, + inputData, + expectedOutputData, + desc); +} + +} // anonymous namespace diff --git a/src/backends/backendsCommon/test/LayerTests.cpp b/src/backends/backendsCommon/test/LayerTests.cpp index a088aaa947..be1d43541c 100644 --- a/src/backends/backendsCommon/test/LayerTests.cpp +++ b/src/backends/backendsCommon/test/LayerTests.cpp @@ -43,6 +43,7 @@ #include "ConvertFp16ToFp32TestImpl.hpp" #include "ConvertFp32ToFp16TestImpl.hpp" #include "DebugTestImpl.hpp" +#include "DequantizeTestImpl.hpp" // 3-channel 16x8 image used as common input data for a number of Conv2d tests. static std::vector ConvInput3x8x16({ @@ -8969,3 +8970,10 @@ LayerTestResult GatherMultiDimParamsMultiDimIndicesUint8Test( return GatherMultiDimParamsMultiDimIndicesTestImpl( workloadFactory, memoryManager); } + +LayerTestResult DequantizeUint8Test( + armnn::IWorkloadFactory& workloadFactory, + const armnn::IBackendInternal::IMemoryManagerSharedPtr& memoryManager) +{ + return DequantizeSimpleTest(workloadFactory, memoryManager); +} diff --git a/src/backends/backendsCommon/test/LayerTests.hpp b/src/backends/backendsCommon/test/LayerTests.hpp index a8cb553ce9..c14ba7b35f 100644 --- a/src/backends/backendsCommon/test/LayerTests.hpp +++ b/src/backends/backendsCommon/test/LayerTests.hpp @@ -1439,3 +1439,7 @@ LayerTestResult GatherMultiDimParamsMultiDimIndicesFloatTest( LayerTestResult GatherMultiDimParamsMultiDimIndicesUint8Test( armnn::IWorkloadFactory& workloadFactory, const armnn::IBackendInternal::IMemoryManagerSharedPtr& memoryManager); + +LayerTestResult DequantizeUint8Test( + armnn::IWorkloadFactory& workloadFactory, + const armnn::IBackendInternal::IMemoryManagerSharedPtr& memoryManager); diff --git a/src/backends/reference/RefLayerSupport.cpp b/src/backends/reference/RefLayerSupport.cpp index 820f36b7ac..532c8eaf98 100644 --- a/src/backends/reference/RefLayerSupport.cpp +++ b/src/backends/reference/RefLayerSupport.cpp @@ -380,6 +380,16 @@ bool RefLayerSupport::IsDepthwiseConvolutionSupported(const TensorInfo& input, &TrueFunc<>); } +bool RefLayerSupport::IsDequantizeSupported(const TensorInfo& input, + const TensorInfo& output, + Optional reasonIfUnsupported) const +{ + return IsSupportedForDataTypeRef(reasonIfUnsupported, + input.GetDataType(), + &FalseFunc<>, + &TrueFunc<>); +} + bool RefLayerSupport::IsDetectionPostProcessSupported(const armnn::TensorInfo& input0, const armnn::TensorInfo& input1, const armnn::DetectionPostProcessDescriptor& descriptor, diff --git a/src/backends/reference/RefLayerSupport.hpp b/src/backends/reference/RefLayerSupport.hpp index c0d7fcf4aa..42a5a4455e 100644 --- a/src/backends/reference/RefLayerSupport.hpp +++ b/src/backends/reference/RefLayerSupport.hpp @@ -65,6 +65,10 @@ public: const Optional& biases, Optional reasonIfUnsupported = EmptyOptional()) const override; + bool IsDequantizeSupported(const TensorInfo& input, + const TensorInfo& output, + Optional reasonIfUnsupported = EmptyOptional()) const override; + bool IsDetectionPostProcessSupported(const TensorInfo& input0, const TensorInfo& input1, const DetectionPostProcessDescriptor& descriptor, diff --git a/src/backends/reference/RefWorkloadFactory.cpp b/src/backends/reference/RefWorkloadFactory.cpp index 3bf83bd9be..dda1819adf 100644 --- a/src/backends/reference/RefWorkloadFactory.cpp +++ b/src/backends/reference/RefWorkloadFactory.cpp @@ -348,4 +348,10 @@ std::unique_ptr RefWorkloadFactory::CreatePreCompiled(const PreCompil return nullptr; } +std::unique_ptr RefWorkloadFactory::CreateDequantize(const DequantizeQueueDescriptor& descriptor, + const WorkloadInfo& info) const +{ + return MakeWorkload(descriptor, info); +} + } // namespace armnn diff --git a/src/backends/reference/RefWorkloadFactory.hpp b/src/backends/reference/RefWorkloadFactory.hpp index 211f93b2d5..14d3178e74 100644 --- a/src/backends/reference/RefWorkloadFactory.hpp +++ b/src/backends/reference/RefWorkloadFactory.hpp @@ -177,6 +177,9 @@ public: std::unique_ptr CreateGather(const GatherQueueDescriptor& descriptor, const WorkloadInfo& info) const override; + std::unique_ptr CreateDequantize(const DequantizeQueueDescriptor& descriptor, + const WorkloadInfo& info) const override; + private: template diff --git a/src/backends/reference/backend.mk b/src/backends/reference/backend.mk index a75146b946..90aa63a62c 100644 --- a/src/backends/reference/backend.mk +++ b/src/backends/reference/backend.mk @@ -39,6 +39,7 @@ BACKEND_SOURCES := \ workloads/RefDebugWorkload.cpp \ workloads/RefDepthwiseConvolution2dFloat32Workload.cpp \ workloads/RefDepthwiseConvolution2dUint8Workload.cpp \ + workloads/RefDequantizeWorkload.cpp \ workloads/RefDetectionPostProcessFloat32Workload.cpp \ workloads/RefDetectionPostProcessUint8Workload.cpp \ workloads/RefElementwiseWorkload.cpp \ diff --git a/src/backends/reference/test/RefLayerTests.cpp b/src/backends/reference/test/RefLayerTests.cpp index 13ea82de3c..c0e2e785da 100644 --- a/src/backends/reference/test/RefLayerTests.cpp +++ b/src/backends/reference/test/RefLayerTests.cpp @@ -530,5 +530,7 @@ BOOST_AUTO_TEST_CASE(DetectionPostProcessFastNmsUint8) DetectionPostProcessFastNmsUint8Test(); } +// Dequantize +ARMNN_AUTO_TEST_CASE(DequantizeUint8, DequantizeUint8Test) BOOST_AUTO_TEST_SUITE_END() diff --git a/src/backends/reference/workloads/CMakeLists.txt b/src/backends/reference/workloads/CMakeLists.txt index 89aed9194b..c4fc202250 100644 --- a/src/backends/reference/workloads/CMakeLists.txt +++ b/src/backends/reference/workloads/CMakeLists.txt @@ -63,6 +63,8 @@ list(APPEND armnnRefBackendWorkloads_sources RefDepthwiseConvolution2dFloat32Workload.hpp RefDepthwiseConvolution2dUint8Workload.cpp RefDepthwiseConvolution2dUint8Workload.hpp + RefDequantizeWorkload.cpp + RefDequantizeWorkload.hpp RefDetectionPostProcessUint8Workload.cpp RefDetectionPostProcessUint8Workload.hpp RefDetectionPostProcessFloat32Workload.cpp diff --git a/src/backends/reference/workloads/RefDequantizeWorkload.cpp b/src/backends/reference/workloads/RefDequantizeWorkload.cpp new file mode 100644 index 0000000000..d861c50730 --- /dev/null +++ b/src/backends/reference/workloads/RefDequantizeWorkload.cpp @@ -0,0 +1,34 @@ +// +// Copyright © 2017 Arm Ltd. All rights reserved. +// SPDX-License-Identifier: MIT +// + +#include "RefDequantizeWorkload.hpp" +#include "RefWorkloadUtils.hpp" + +namespace armnn +{ + +void RefDequantizeWorkload::Execute() const +{ + ARMNN_SCOPED_PROFILING_EVENT(Compute::CpuRef, "RefDequantizeWorkload_Execute"); + + const TensorInfo& inputInfo = GetTensorInfo(m_Data.m_Inputs[0]); + const DataType& inputDataType = inputInfo.GetDataType(); + + float* outputData = GetOutputTensorData(0, m_Data); + + switch (inputDataType) + { + case DataType::QuantisedAsymm8: + Dequantize(GetInputTensorData(0, m_Data), outputData, inputInfo); + break; + case DataType::QuantisedSymm16: + Dequantize(GetInputTensorData(0, m_Data), outputData, inputInfo); + break; + default: + throw InvalidArgumentException("RefDequantizeWorkload: Unsupported input data type"); + } +} + +} // namespace armnn diff --git a/src/backends/reference/workloads/RefDequantizeWorkload.hpp b/src/backends/reference/workloads/RefDequantizeWorkload.hpp new file mode 100644 index 0000000000..8d019e3220 --- /dev/null +++ b/src/backends/reference/workloads/RefDequantizeWorkload.hpp @@ -0,0 +1,22 @@ +// +// Copyright © 2017 Arm Ltd. All rights reserved. +// SPDX-License-Identifier: MIT +// + +#pragma once + +#include + +namespace armnn +{ + +class RefDequantizeWorkload : public BaseWorkload +{ +public: + using BaseWorkload::m_Data; + using BaseWorkload::BaseWorkload; + + void Execute() const override; +}; + +} // namespace armnn diff --git a/src/backends/reference/workloads/RefWorkloadUtils.hpp b/src/backends/reference/workloads/RefWorkloadUtils.hpp index feb43290b4..ce796160f2 100644 --- a/src/backends/reference/workloads/RefWorkloadUtils.hpp +++ b/src/backends/reference/workloads/RefWorkloadUtils.hpp @@ -127,6 +127,15 @@ std::vector Dequantize(const T* quant, const TensorInfo& info) return ret; } +template +inline void Dequantize(const T* inputData, float* outputData, const TensorInfo& info) +{ + for (unsigned int i = 0; i < info.GetNumElements(); i++) + { + outputData[i] = Dequantize(inputData[i], info.GetQuantizationScale(), info.GetQuantizationOffset()); + } +} + inline void Quantize(uint8_t* quant, const float* dequant, const TensorInfo& info) { for (size_t i = 0; i < info.GetNumElements(); i++) diff --git a/src/backends/reference/workloads/RefWorkloads.hpp b/src/backends/reference/workloads/RefWorkloads.hpp index 2156388ba2..7d2e813f6b 100644 --- a/src/backends/reference/workloads/RefWorkloads.hpp +++ b/src/backends/reference/workloads/RefWorkloads.hpp @@ -63,3 +63,4 @@ #include "RefDebugWorkload.hpp" #include "RefRsqrtFloat32Workload.hpp" #include "RefComparisonWorkload.hpp" +#include "RefDequantizeWorkload.hpp" -- cgit v1.2.1