16 #include <boost/polymorphic_cast.hpp> 29 boost::polymorphic_downcast<const RefTensorHandle*>(tensorHandle);
33 template <
typename DataType,
typename PayloadType>
37 return reinterpret_cast<const DataType*
>(tensorHandle->
Map());
40 template <
typename DataType,
typename PayloadType>
44 return reinterpret_cast<DataType*
>(tensorHandle->
Map());
47 template <
typename PayloadType>
50 return GetInputTensorData<float>(idx, data);
53 template <
typename PayloadType>
56 return GetOutputTensorData<float>(idx, data);
59 template <
typename PayloadType>
62 return GetInputTensorData<Half>(idx, data);
65 template <
typename PayloadType>
68 return GetOutputTensorData<Half>(idx, data);
const DataType * GetInputTensorData(unsigned int idx, const PayloadType &data)
int32_t GetQuantizationOffset() const
DataType * GetOutputTensorData(unsigned int idx, const PayloadType &data)
QuantizedType Quantize(float value, float scale, int32_t offset)
Explicit specialization of Quantize for int8_t.
float * GetOutputTensorDataFloat(unsigned int idx, const PayloadType &data)
const float * GetInputTensorDataFloat(unsigned int idx, const PayloadType &data)
Half * GetOutputTensorDataHalf(unsigned int idx, const PayloadType &data)
const TensorInfo & GetTensorInfo() const
virtual const void * Map(bool blocking=true) const =0
float Dequantize(QuantizedType value, float scale, int32_t offset)
const Half * GetInputTensorDataHalf(unsigned int idx, const PayloadType &data)
unsigned int GetNumElements() const
const TensorInfo & GetTensorInfo(const ITensorHandle *tensorHandle)
float32 helpers
float GetQuantizationScale() const