19 void QuantizeImpl(Decoder<float>& in, Encoder<float>& out,
size_t numValues)
21 for (
unsigned int i = 0; i < numValues; i++)
33 , m_NumElements(info.m_InputTensorInfos[0].GetNumElements())
40 m_InputDecoder = MakeDecoder<float>(inputInfo);
43 m_OutputEncoder = MakeEncoder<float>(outputInfo);
51 QuantizeImpl(*m_InputDecoder, *m_OutputEncoder, m_NumElements);
void PostAllocationConfigure() override
void Execute() const override
RefQuantizeWorkload(const QuantizeQueueDescriptor &descriptor, const WorkloadInfo &info)
const QuantizeQueueDescriptor m_Data
std::vector< ITensorHandle * > m_Outputs
std::vector< ITensorHandle * > m_Inputs
const TensorInfo & GetTensorInfo(const ITensorHandle *tensorHandle)
float32 helpers