From f71079328ae72a65c91e410b2bd35eabb67cb6d1 Mon Sep 17 00:00:00 2001 From: Jan Eilers Date: Fri, 1 Nov 2019 11:09:36 +0000 Subject: Add fp16 support for dequantize * Changed RefDequantizeWorkload to use Encoder/Decoder * Added related unit tests for Cl, Neon and Ref Signed-off-by: Jan Eilers Change-Id: Ic2fd4103090dd2127c6859b49305736f7b2dfb05 --- .../reference/workloads/RefDequantizeWorkload.cpp | 20 +++++++------------- 1 file changed, 7 insertions(+), 13 deletions(-) (limited to 'src/backends/reference/workloads/RefDequantizeWorkload.cpp') diff --git a/src/backends/reference/workloads/RefDequantizeWorkload.cpp b/src/backends/reference/workloads/RefDequantizeWorkload.cpp index d861c50730..e6f5c6b359 100644 --- a/src/backends/reference/workloads/RefDequantizeWorkload.cpp +++ b/src/backends/reference/workloads/RefDequantizeWorkload.cpp @@ -5,6 +5,9 @@ #include "RefDequantizeWorkload.hpp" #include "RefWorkloadUtils.hpp" +#include "Encoders.hpp" +#include "Decoders.hpp" +#include "Dequantize.hpp" namespace armnn { @@ -14,21 +17,12 @@ void RefDequantizeWorkload::Execute() const ARMNN_SCOPED_PROFILING_EVENT(Compute::CpuRef, "RefDequantizeWorkload_Execute"); const TensorInfo& inputInfo = GetTensorInfo(m_Data.m_Inputs[0]); - const DataType& inputDataType = inputInfo.GetDataType(); + const TensorInfo& outputInfo = GetTensorInfo(m_Data.m_Outputs[0]); - float* outputData = GetOutputTensorData(0, m_Data); + auto inputDecoder = MakeDecoder(inputInfo, m_Data.m_Inputs[0]->Map()); + auto outputEncoder = MakeEncoder(outputInfo, m_Data.m_Outputs[0]->Map()); - switch (inputDataType) - { - case DataType::QuantisedAsymm8: - Dequantize(GetInputTensorData(0, m_Data), outputData, inputInfo); - break; - case DataType::QuantisedSymm16: - Dequantize(GetInputTensorData(0, m_Data), outputData, inputInfo); - break; - default: - throw InvalidArgumentException("RefDequantizeWorkload: Unsupported input data type"); - } + Dequantize(*inputDecoder, *outputEncoder, inputInfo, outputInfo); } } // namespace armnn -- cgit v1.2.1