12 #include <boost/numeric/conversion/cast.hpp> 17 void FakeQuantization(
const float* inputData,
float* outputData, uint32_t numElements,
float min,
float max)
19 float scale = (max - min) / 255.f;
22 for (uint32_t i = 0; i < numElements; i++)
24 outputData[i] =
static_cast<float>(armnn::Quantize<uint8_t>(inputData[i], scale, offset));
39 m_Data.m_Parameters.m_Max);
const float * GetInputTensorDataFloat(unsigned int idx, const PayloadType &data)
CPU Execution: Reference C++ kernels.
virtual void Execute() const override
const QueueDescriptor m_Data
const TensorInfo & GetTensorInfo(const ITensorHandle *tensorHandle)
float32 helpers
Copyright (c) 2020 ARM Limited.
void FakeQuantization(const float *inputData, float *outputData, uint32_t numElements, float min, float max)
#define ARMNN_SCOPED_PROFILING_EVENT(backendId, name)
std::enable_if_t< std::is_unsigned< Source >::value &&std::is_unsigned< Dest >::value, Dest > numeric_cast(Source source)
std::vector< ITensorHandle * > m_Inputs
float * GetOutputTensorDataFloat(unsigned int idx, const PayloadType &data)
unsigned int GetNumElements() const