From 2f1366a931a8641d0f8c4ce28dc85dfa818236ed Mon Sep 17 00:00:00 2001 From: Georgios Pinitas Date: Tue, 31 Jul 2018 16:33:06 +0100 Subject: COMPMID-1188: Add quantization info support in graph FC layer. Change-Id: Ie9a6a896da142198243139fb9f8be0f83b87ccce Reviewed-on: https://eu-gerrit-1.euhpc.arm.com/142130 Tested-by: Jenkins Reviewed-by: Vidhya Sudhan Loganathan Reviewed-by: Anthony Barbier --- arm_compute/graph/GraphBuilder.h | 18 ++++++++----- arm_compute/graph/backends/FunctionHelpers.h | 2 ++ arm_compute/graph/frontend/Layers.h | 33 +++++++++++++++-------- arm_compute/graph/nodes/FullyConnectedLayerNode.h | 24 ++++++++++++----- src/graph/GraphBuilder.cpp | 7 ++--- src/graph/nodes/FullyConnectedLayer.cpp | 25 +++++++++++++---- 6 files changed, 76 insertions(+), 33 deletions(-) diff --git a/arm_compute/graph/GraphBuilder.h b/arm_compute/graph/GraphBuilder.h index 191848c15f..5bb1df4a11 100644 --- a/arm_compute/graph/GraphBuilder.h +++ b/arm_compute/graph/GraphBuilder.h @@ -212,17 +212,21 @@ public: static NodeID add_flatten_node(Graph &g, NodeParams params, NodeIdxPair input); /** Adds a fully connected layer node to the graph * - * @param[in] g Graph to add the layer to - * @param[in] params Common node parameters - * @param[in] input Input to the fully connected layer node as a NodeID-Index pair - * @param[in] num_outputs Number of output neurons - * @param[in] weights_accessor (Optional) Accessor of the weights node data - * @param[in] bias_accessor (Optional) Accessor of the bias node data + * @param[in] g Graph to add the layer to + * @param[in] params Common node parameters + * @param[in] input Input to the fully connected layer node as a NodeID-Index pair + * @param[in] num_outputs Number of output neurons + * @param[in] weights_accessor (Optional) Accessor of the weights node data + * @param[in] bias_accessor (Optional) Accessor of the bias node data + * @param[in] weights_quant_info (Optional) Weights quantization info + * @param[in] out_quant_info (Optional) Output quantization info * * @return Node ID of the created node, EmptyNodeID in case of error */ static NodeID add_fully_connected_layer(Graph &g, NodeParams params, NodeIdxPair input, unsigned int num_outputs, - ITensorAccessorUPtr weights_accessor = nullptr, ITensorAccessorUPtr bias_accessor = nullptr); + ITensorAccessorUPtr weights_accessor = nullptr, ITensorAccessorUPtr bias_accessor = nullptr, + const QuantizationInfo weights_quant_info = QuantizationInfo(), + const QuantizationInfo out_quant_info = QuantizationInfo()); /** Adds a normalization layer node to the graph * * @param[in] g Graph to add the node to diff --git a/arm_compute/graph/backends/FunctionHelpers.h b/arm_compute/graph/backends/FunctionHelpers.h index 172f00277e..b7c9c579de 100644 --- a/arm_compute/graph/backends/FunctionHelpers.h +++ b/arm_compute/graph/backends/FunctionHelpers.h @@ -541,6 +541,8 @@ std::unique_ptr create_fully_connected_layer(FullyConnectedLayerNode ARM_COMPUTE_LOG_GRAPH_INFO("Instantiated " << node.type() << " Target " << TargetInfo::TargetType << " Data Type: " << input->info()->data_type() + << " Input QuantInfo: " << input->info()->quantization_info() + << " Weights QuantInfo: " << weights->info()->quantization_info() << " Input shape: " << input->info()->tensor_shape() << " Weights shape: " << weights->info()->tensor_shape() << " Output shape: " << output->info()->tensor_shape() diff --git a/arm_compute/graph/frontend/Layers.h b/arm_compute/graph/frontend/Layers.h index 02ef56952d..a222c8546e 100644 --- a/arm_compute/graph/frontend/Layers.h +++ b/arm_compute/graph/frontend/Layers.h @@ -381,14 +381,22 @@ class FullyConnectedLayer final : public ILayer public: /** Construct a fully connected layer. * - * @param[in] num_outputs Number of outputs. - * @param[in] weights Accessor to get weights from. - * @param[in] bias Accessor to get bias from. + * @param[in] num_outputs Number of outputs. + * @param[in] weights Accessor to get weights from. + * @param[in] bias Accessor to get bias from. + * @param[in] weights_quant_info (Optional) Weights quantization information + * @param[in] out_quant_info (Optional) Output quantization info */ - FullyConnectedLayer(unsigned int num_outputs, - ITensorAccessorUPtr weights, - ITensorAccessorUPtr bias) - : _num_outputs(num_outputs), _weights(std::move(weights)), _bias(std::move(bias)) + FullyConnectedLayer(unsigned int num_outputs, + ITensorAccessorUPtr weights, + ITensorAccessorUPtr bias, + const QuantizationInfo weights_quant_info = QuantizationInfo(), + const QuantizationInfo out_quant_info = QuantizationInfo()) + : _num_outputs(num_outputs), + _weights(std::move(weights)), + _bias(std::move(bias)), + _weights_quant_info(std::move(weights_quant_info)), + _out_quant_info(std::move(out_quant_info)) { } @@ -397,13 +405,16 @@ public: NodeParams common_params = { name(), s.hints().target_hint }; NodeIdxPair input = { s.tail_node(), 0 }; return GraphBuilder::add_fully_connected_layer(s.graph(), common_params, input, _num_outputs, - std::move(_weights), std::move(_bias)); + std::move(_weights), std::move(_bias), + std::move(_weights_quant_info), std::move(_out_quant_info)); } private: - unsigned int _num_outputs; - ITensorAccessorUPtr _weights; - ITensorAccessorUPtr _bias; + unsigned int _num_outputs; + ITensorAccessorUPtr _weights; + ITensorAccessorUPtr _bias; + const QuantizationInfo _weights_quant_info; + const QuantizationInfo _out_quant_info; }; /** Normalization Layer */ diff --git a/arm_compute/graph/nodes/FullyConnectedLayerNode.h b/arm_compute/graph/nodes/FullyConnectedLayerNode.h index 79201c8bdc..1bff6006c8 100644 --- a/arm_compute/graph/nodes/FullyConnectedLayerNode.h +++ b/arm_compute/graph/nodes/FullyConnectedLayerNode.h @@ -36,30 +36,39 @@ class FullyConnectedLayerNode final : public INode public: /** Constructor * - * @param[in] num_outputs Number of neurons in the layer - * @param[in] fc_info (Optional) Additional information about the fully connected layer + * @param[in] num_outputs Number of neurons in the layer + * @param[in] out_quant_info (Optional) Output quantization info + * @param[in] fc_info (Optional) Additional information about the fully connected layer */ - FullyConnectedLayerNode(unsigned int num_outputs, FullyConnectedLayerInfo fc_info = FullyConnectedLayerInfo()); + FullyConnectedLayerNode(unsigned int num_outputs, + QuantizationInfo out_quant_info = QuantizationInfo(), + FullyConnectedLayerInfo fc_info = FullyConnectedLayerInfo()); /** Computes weights descriptor * * @warning Works for inputs with 1D batch space * - * @param[in] input_descriptor Input descriptor - * @param[in] num_outputs Number of output neurons + * @param[in] input_descriptor Input descriptor + * @param[in] num_outputs Number of output neurons + * @param[in] weights_quant_info (Optional) Weights quantization info * * @return Weights descriptor */ - static TensorDescriptor compute_weights_descriptor(const TensorDescriptor &input_descriptor, unsigned int num_outputs); + static TensorDescriptor compute_weights_descriptor(const TensorDescriptor &input_descriptor, + unsigned int num_outputs, + QuantizationInfo weights_quant_info = QuantizationInfo()); /** Computes fully connected layer output descriptor * * @warning Works for inputs with 1D batch space * * @param[in] input_descriptor Input descriptor * @param[in] num_outputs Number of output neurons + * @param[in] out_quant_info (Optional) Weights quantization info * * @return Output descriptor */ - static TensorDescriptor compute_output_descriptor(const TensorDescriptor &input_descriptor, unsigned int num_outputs); + static TensorDescriptor compute_output_descriptor(const TensorDescriptor &input_descriptor, + unsigned int num_outputs, + QuantizationInfo out_quant_info = QuantizationInfo()); /** Fully connected layer addition information * * @return Additional information about the fully connected layer @@ -74,6 +83,7 @@ public: private: unsigned int _num_outputs; + QuantizationInfo _out_quant_info; FullyConnectedLayerInfo _info; }; } // namespace graph diff --git a/src/graph/GraphBuilder.cpp b/src/graph/GraphBuilder.cpp index b3721719d9..7f567fd559 100644 --- a/src/graph/GraphBuilder.cpp +++ b/src/graph/GraphBuilder.cpp @@ -423,7 +423,8 @@ NodeID GraphBuilder::add_flatten_node(Graph &g, NodeParams params, NodeIdxPair i } NodeID GraphBuilder::add_fully_connected_layer(Graph &g, NodeParams params, NodeIdxPair input, unsigned int num_outputs, - ITensorAccessorUPtr weights_accessor, ITensorAccessorUPtr bias_accessor) + ITensorAccessorUPtr weights_accessor, ITensorAccessorUPtr bias_accessor, + const QuantizationInfo weights_quant_info, const QuantizationInfo out_quant_info) { CHECK_NODEIDX_PAIR(input, g); ARM_COMPUTE_ERROR_ON(num_outputs == 0); @@ -434,7 +435,7 @@ NodeID GraphBuilder::add_fully_connected_layer(Graph &g, NodeParams params, Node const TensorDescriptor input_tensor_desc = get_tensor_descriptor(g, g.node(input.node_id)->outputs()[0]); // Create weights node - TensorDescriptor w_desc = FullyConnectedLayerNode::compute_weights_descriptor(input_tensor_desc, num_outputs); + TensorDescriptor w_desc = FullyConnectedLayerNode::compute_weights_descriptor(input_tensor_desc, num_outputs, weights_quant_info); NodeID w_nid = add_const_node_with_name(g, params, "Weights", w_desc, std::move(weights_accessor)); // Create bias nodes @@ -456,7 +457,7 @@ NodeID GraphBuilder::add_fully_connected_layer(Graph &g, NodeParams params, Node fc_info.weights_trained_layout = DataLayout::NCHW; // Create fully connected node and connect - NodeID fc_nid = g.add_node(num_outputs, fc_info); + NodeID fc_nid = g.add_node(num_outputs, out_quant_info, fc_info); g.add_connection(input.node_id, input.index, fc_nid, 0); g.add_connection(w_nid, 0, fc_nid, 1); if(has_bias) diff --git a/src/graph/nodes/FullyConnectedLayer.cpp b/src/graph/nodes/FullyConnectedLayer.cpp index 6ba4eefa88..f41e1a3c22 100644 --- a/src/graph/nodes/FullyConnectedLayer.cpp +++ b/src/graph/nodes/FullyConnectedLayer.cpp @@ -31,15 +31,16 @@ namespace arm_compute { namespace graph { -FullyConnectedLayerNode::FullyConnectedLayerNode(unsigned int num_outputs, FullyConnectedLayerInfo fc_info) - : _num_outputs(num_outputs), _info(fc_info) +FullyConnectedLayerNode::FullyConnectedLayerNode(unsigned int num_outputs, QuantizationInfo out_quant_info, FullyConnectedLayerInfo fc_info) + : _num_outputs(num_outputs), _out_quant_info(out_quant_info), _info(fc_info) { _input_edges.resize(3, EmptyEdgeID); _outputs.resize(1, NullTensorID); } TensorDescriptor FullyConnectedLayerNode::compute_weights_descriptor(const TensorDescriptor &input_descriptor, - unsigned int num_outputs) + unsigned int num_outputs, + QuantizationInfo weights_quant_info) { unsigned int num_weights = 1; unsigned int num_dimensions = input_descriptor.shape.num_dimensions(); @@ -56,11 +57,18 @@ TensorDescriptor FullyConnectedLayerNode::compute_weights_descriptor(const Tenso TensorDescriptor weights_descriptor = input_descriptor; weights_descriptor.shape = TensorShape(num_weights, num_outputs); + // Set quantization info if present + if(!weights_quant_info.empty()) + { + weights_descriptor.quant_info = weights_quant_info; + } + return weights_descriptor; } TensorDescriptor FullyConnectedLayerNode::compute_output_descriptor(const TensorDescriptor &input_descriptor, - unsigned int num_outputs) + unsigned int num_outputs, + QuantizationInfo out_quant_info) { // Note: Only 1D batch space is supported at the moment unsigned int batches = input_descriptor.shape[1]; @@ -69,9 +77,16 @@ TensorDescriptor FullyConnectedLayerNode::compute_output_descriptor(const Tensor batches = input_descriptor.shape[3]; } + // Set descriptor shape TensorDescriptor output_descriptor = input_descriptor; output_descriptor.shape = TensorShape(num_outputs, batches); + // Set quantization info if present + if(!out_quant_info.empty()) + { + output_descriptor.quant_info = out_quant_info; + } + return output_descriptor; } @@ -98,7 +113,7 @@ TensorDescriptor FullyConnectedLayerNode::configure_output(size_t idx) const const Tensor *src = input(0); ARM_COMPUTE_ERROR_ON(src == nullptr); - return compute_output_descriptor(src->desc(), _num_outputs); + return compute_output_descriptor(src->desc(), _num_outputs, _out_quant_info); } NodeType FullyConnectedLayerNode::type() const -- cgit v1.2.1