From dd6aceaa884815e68ed69fca71de81babd3204da Mon Sep 17 00:00:00 2001 From: Mohamed Nour Abouelseoud Date: Thu, 18 Oct 2018 12:26:19 +0100 Subject: IVGCVSW-2013 Add a UInt8 Reference Implementation for the PAD Operator Change-Id: I41f3606198db1fda8d72aaf5169594ba9156eb38 --- src/backends/reference/workloads/RefPadWorkload.cpp | 17 ++++++++++------- 1 file changed, 10 insertions(+), 7 deletions(-) (limited to 'src/backends/reference/workloads/RefPadWorkload.cpp') diff --git a/src/backends/reference/workloads/RefPadWorkload.cpp b/src/backends/reference/workloads/RefPadWorkload.cpp index 233fbe4f34..b41c2de9af 100644 --- a/src/backends/reference/workloads/RefPadWorkload.cpp +++ b/src/backends/reference/workloads/RefPadWorkload.cpp @@ -10,28 +10,31 @@ #include "Profiling.hpp" +#include "TypeUtils.hpp" + #include namespace armnn { -RefPadWorkload::RefPadWorkload(const PadQueueDescriptor& descriptor, const WorkloadInfo& info) - :BaseWorkload(descriptor, info) {} - - -void RefPadWorkload::Execute() const +template +void RefPadWorkload::Execute() const { + using T = ResolveType; ARMNN_SCOPED_PROFILING_EVENT(Compute::CpuRef, "RefPadWorkload_Execute"); const TensorInfo& inputInfo = GetTensorInfo(m_Data.m_Inputs[0]); const TensorInfo& outputInfo = GetTensorInfo(m_Data.m_Outputs[0]); - const float* inputData = GetInputTensorDataFloat(0, m_Data); - float* outputData = GetOutputTensorDataFloat(0, m_Data); + const T* inputData = GetInputTensorData(0, m_Data); + T* outputData = GetOutputTensorData(0, m_Data); Pad(inputInfo, outputInfo, m_Data.m_Parameters.m_PadList, inputData, outputData); } +template class RefPadWorkload; +template class RefPadWorkload; + } //namespace armnn \ No newline at end of file -- cgit v1.2.1