aboutsummaryrefslogtreecommitdiff
path: root/src/backends/reference/workloads/RefDequantizeWorkload.cpp
diff options
context:
space:
mode:
authorJan Eilers <jan.eilers@arm.com>2019-11-01 11:09:36 +0000
committerJan Eilers <jan.eilers@arm.com>2019-11-04 12:09:08 +0000
commitf71079328ae72a65c91e410b2bd35eabb67cb6d1 (patch)
treee5460c94ea84f0ffb6ec09df820912cd9bd750ec /src/backends/reference/workloads/RefDequantizeWorkload.cpp
parent7ff9a6096e3c1facbd6786993a6437b9f72069d2 (diff)
downloadarmnn-f71079328ae72a65c91e410b2bd35eabb67cb6d1.tar.gz
Add fp16 support for dequantize
* Changed RefDequantizeWorkload to use Encoder/Decoder * Added related unit tests for Cl, Neon and Ref Signed-off-by: Jan Eilers <jan.eilers@arm.com> Change-Id: Ic2fd4103090dd2127c6859b49305736f7b2dfb05
Diffstat (limited to 'src/backends/reference/workloads/RefDequantizeWorkload.cpp')
-rw-r--r--src/backends/reference/workloads/RefDequantizeWorkload.cpp20
1 files changed, 7 insertions, 13 deletions
diff --git a/src/backends/reference/workloads/RefDequantizeWorkload.cpp b/src/backends/reference/workloads/RefDequantizeWorkload.cpp
index d861c50730..e6f5c6b359 100644
--- a/src/backends/reference/workloads/RefDequantizeWorkload.cpp
+++ b/src/backends/reference/workloads/RefDequantizeWorkload.cpp
@@ -5,6 +5,9 @@
#include "RefDequantizeWorkload.hpp"
#include "RefWorkloadUtils.hpp"
+#include "Encoders.hpp"
+#include "Decoders.hpp"
+#include "Dequantize.hpp"
namespace armnn
{
@@ -14,21 +17,12 @@ void RefDequantizeWorkload::Execute() const
ARMNN_SCOPED_PROFILING_EVENT(Compute::CpuRef, "RefDequantizeWorkload_Execute");
const TensorInfo& inputInfo = GetTensorInfo(m_Data.m_Inputs[0]);
- const DataType& inputDataType = inputInfo.GetDataType();
+ const TensorInfo& outputInfo = GetTensorInfo(m_Data.m_Outputs[0]);
- float* outputData = GetOutputTensorData<float>(0, m_Data);
+ auto inputDecoder = MakeDecoder<float>(inputInfo, m_Data.m_Inputs[0]->Map());
+ auto outputEncoder = MakeEncoder<float>(outputInfo, m_Data.m_Outputs[0]->Map());
- switch (inputDataType)
- {
- case DataType::QuantisedAsymm8:
- Dequantize<uint8_t>(GetInputTensorData<uint8_t>(0, m_Data), outputData, inputInfo);
- break;
- case DataType::QuantisedSymm16:
- Dequantize<int16_t>(GetInputTensorData<int16_t>(0, m_Data), outputData, inputInfo);
- break;
- default:
- throw InvalidArgumentException("RefDequantizeWorkload: Unsupported input data type");
- }
+ Dequantize(*inputDecoder, *outputEncoder, inputInfo, outputInfo);
}
} // namespace armnn