// // Copyright © 2017 Arm Ltd. All rights reserved. // SPDX-License-Identifier: MIT // #include "RefFakeQuantizationFloat32Workload.hpp" #include "RefWorkloadUtils.hpp" #include "Profiling.hpp" #include namespace armnn { void FakeQuantization(const float* inputData, float* outputData, uint32_t numElements, float min, float max) { float scale = (max - min) / 255.f; int32_t offset = boost::numeric_cast((-min * 255.f) / (max - min)); for (uint32_t i = 0; i < numElements; i++) { outputData[i] = static_cast(armnn::Quantize(inputData[i], scale, offset)); } } void RefFakeQuantizationFloat32Workload::Execute() const { ARMNN_SCOPED_PROFILING_EVENT(Compute::CpuRef, "RefFakeQuantizationFloat32Workload_Execute"); const TensorInfo& inputInfo = GetTensorInfo(m_Data.m_Inputs[0]); const float* inputData = GetInputTensorDataFloat(0, m_Data); float* outputData = GetOutputTensorDataFloat(0, m_Data); FakeQuantization(inputData, outputData, inputInfo.GetNumElements(), m_Data.m_Parameters.m_Min, m_Data.m_Parameters.m_Max); } } //namespace armnn