19 void QuantizeImpl(Decoder<float>& in, Encoder<float>& out,
size_t numValues)
21 for (
unsigned int i = 0; i < numValues; i++)
33 , m_NumElements(info.m_InputTensorInfos[0].GetNumElements())
40 m_InputDecoder = MakeDecoder<float>(inputInfo);
43 m_OutputEncoder = MakeEncoder<float>(outputInfo);
51 QuantizeImpl(*m_InputDecoder, *m_OutputEncoder, m_NumElements);
const QuantizeQueueDescriptor m_Data
const TensorInfo & GetTensorInfo(const ITensorHandle *tensorHandle)
float32 helpers
Copyright (c) 2020 ARM Limited.
void PostAllocationConfigure() override
void Execute() const override
std::vector< ITensorHandle * > m_Outputs
Contains information about inputs and outputs to a layer.
RefQuantizeWorkload(const QuantizeQueueDescriptor &descriptor, const WorkloadInfo &info)
std::vector< ITensorHandle * > m_Inputs