14 namespace optimizations
17 template <
typename LayerT>
20 LayerT* layer = PolymorphicDowncast<LayerT*>(l);
31 layer->m_Weight->template GetConstTensor<float>(),
58 ConvertWeight<Convolution2dLayer>(&layer);
66 ConvertWeight<FullyConnectedLayer>(&layer);
~ConvertFp32NetworkToBf16Impl()=default
Copyright (c) 2021 ARM Limited and Contributors.
LayerT * ConvertWeight(Layer *l)
DataType GetDataType() const
A tensor defined by a TensorInfo (shape and data type) and an immutable backing store.
LayerType GetType() const override
Returns the armnn::LayerType of this layer.
static void ConvertFloat32ToBFloat16(const float *srcFloat32Buffer, size_t numElements, void *dstBFloat16Buffer)
std::vector< ConvertFp32ToBf16Layer * > InsertConvertFp32ToBf16LayersBefore(Graph &graph, Layer &layer, bool expectCorrectInputType)
ConvertFp32NetworkToBf16Impl()=default
DataType GetDataType() const
void Run(Graph &graph, Layer &layer) const
unsigned int GetNumElements() const