// // Copyright © 2020 Arm Ltd. All rights reserved. // SPDX-License-Identifier: MIT // #pragma once #include namespace armnn { /// This layer converts data type Float32 to BFloat16. class ConvertFp32ToBf16Layer : public Layer { public: /// Makes a workload for the ConvertFp32ToBf16Layer type. /// @param [in] factory The workload factory which will create the workload. /// @return A pointer to the created workload, or nullptr if not created. virtual std::unique_ptr CreateWorkload(const IWorkloadFactory& factory) const override; /// Creates a dynamically-allocated copy of this layer. /// @param [in] graph The graph into which this layer is being cloned. ConvertFp32ToBf16Layer* Clone(Graph& graph) const override; /// Check if the input tensor shape(s) /// will lead to a valid configuration of @ref ConvertFp32ToBf16Layer. /// @param [in] shapeInferenceMethod Indicates if output shape shall be overwritten or just validated. void ValidateTensorShapesFromInputs( ShapeInferenceMethod shapeInferenceMethod = ShapeInferenceMethod::ValidateOnly) override; void Accept(ILayerVisitor& visitor) const override; protected: /// Constructor to create a ConvertFp32ToBf16Layer. /// @param [in] name Optional name for the layer. ConvertFp32ToBf16Layer(const char* name); /// Default destructor ~ConvertFp32ToBf16Layer() = default; }; } // namespace