From e2c82fee3b6d38f6e79412c78176792b817defd0 Mon Sep 17 00:00:00 2001 From: Georgios Pinitas Date: Mon, 2 Oct 2017 18:51:47 +0100 Subject: COMPMID-550: Adds support for branches. Change-Id: I778007c9221ce3156400284c4039b90245eb2b7f Reviewed-on: http://mpd-gerrit.cambridge.arm.com/90043 Tested-by: Kaizen Reviewed-by: Anthony Barbier --- src/graph/nodes/FullyConnectedLayer.cpp | 46 ++++++++++++++++++--------------- 1 file changed, 25 insertions(+), 21 deletions(-) (limited to 'src/graph/nodes/FullyConnectedLayer.cpp') diff --git a/src/graph/nodes/FullyConnectedLayer.cpp b/src/graph/nodes/FullyConnectedLayer.cpp index 6b21810a36..fa5ead8bdd 100644 --- a/src/graph/nodes/FullyConnectedLayer.cpp +++ b/src/graph/nodes/FullyConnectedLayer.cpp @@ -45,7 +45,7 @@ TensorShape calculate_fullyconnected_layer_output_shape(const TensorShape &input return TensorShape(output_neurons, batches); } template -std::unique_ptr instantiate_function(ITensor *input, Tensor &weights, Tensor &biases, ITensor *output) +std::unique_ptr instantiate_function(arm_compute::ITensor *input, Tensor &weights, Tensor &biases, arm_compute::ITensor *output) { bool weights_are_loaded = weights.tensor() != nullptr; bool biases_are_loaded = biases.tensor() != nullptr; @@ -69,27 +69,33 @@ std::unique_ptr instantiate_function(ITensor *input, Ten } template -std::unique_ptr instantiate(ITensor *input, Tensor &weights, Tensor &biases, ITensor *output); +std::unique_ptr instantiate(arm_compute::ITensor *input, Tensor &weights, Tensor &biases, arm_compute::ITensor *output); template <> -std::unique_ptr instantiate(ITensor *input, Tensor &weights, Tensor &biases, ITensor *output) +std::unique_ptr instantiate(arm_compute::ITensor *input, Tensor &weights, Tensor &biases, arm_compute::ITensor *output) { - return instantiate_function(input, weights, biases, output); + return instantiate_function(input, weights, biases, output); } template <> -std::unique_ptr instantiate(ITensor *input, Tensor &weights, Tensor &biases, ITensor *output) +std::unique_ptr instantiate(arm_compute::ITensor *input, Tensor &weights, Tensor &biases, arm_compute::ITensor *output) { - return instantiate_function(input, weights, biases, output); + return instantiate_function(input, weights, biases, output); } } // namespace -std::unique_ptr FullyConnectedLayer::instantiate_node(GraphContext &ctx, ITensor *input, ITensor *output) +std::unique_ptr FullyConnectedLayer::instantiate_node(GraphContext &ctx, ITensorObject *input, ITensorObject *output) { + ARM_COMPUTE_ERROR_ON(input == nullptr || input->tensor() == nullptr); + ARM_COMPUTE_ERROR_ON(output == nullptr || output->tensor() == nullptr); + + arm_compute::ITensor *in = input->tensor(); + arm_compute::ITensor *out = output->tensor(); + if(_weights.tensor() == nullptr) { unsigned int num_weights = 1; - unsigned int num_dimensions = input->info()->num_dimensions(); + unsigned int num_dimensions = in->info()->num_dimensions(); // Ignore the batch dimension if there is one: if(num_dimensions == 2 || num_dimensions == 4) { @@ -97,39 +103,37 @@ std::unique_ptr FullyConnectedLayer::instantiate_node(Gr } for(unsigned int i = 0; i < num_dimensions; i++) { - num_weights *= input->info()->dimension(i); + num_weights *= in->info()->dimension(i); } - _weights.set_info(TensorInfo(TensorShape(num_weights, _num_neurons), input->info()->num_channels(), input->info()->data_type(), input->info()->fixed_point_position())); + _weights.set_info(TensorInfo(TensorShape(num_weights, _num_neurons), in->info()->num_channels(), in->info()->data_type(), in->info()->fixed_point_position())); } if(_biases.tensor() == nullptr) { - _biases.set_info(TensorInfo(TensorShape(_num_neurons), input->info()->num_channels(), input->info()->data_type(), input->info()->fixed_point_position())); + _biases.set_info(TensorInfo(TensorShape(_num_neurons), in->info()->num_channels(), in->info()->data_type(), in->info()->fixed_point_position())); } // Auto configure output - arm_compute::auto_init_if_empty(*output->info(), - calculate_fullyconnected_layer_output_shape(input->info()->tensor_shape(), _num_neurons), - input->info()->num_channels(), input->info()->data_type(), input->info()->fixed_point_position()); + arm_compute::auto_init_if_empty(*out->info(), + calculate_fullyconnected_layer_output_shape(in->info()->tensor_shape(), _num_neurons), + in->info()->num_channels(), in->info()->data_type(), in->info()->fixed_point_position()); std::unique_ptr func; _target_hint = ctx.hints().target_hint(); if(_target_hint == TargetHint::OPENCL) { - func = instantiate(input, _weights, _biases, output); - ARM_COMPUTE_LOG("Instantiating CLFullyConnectedLayer"); + func = instantiate(in, _weights, _biases, out); } else { - func = instantiate(input, _weights, _biases, output); - ARM_COMPUTE_LOG("Instantiating NEFullyConnectedLayer"); + func = instantiate(in, _weights, _biases, out); } - ARM_COMPUTE_LOG(" Type: " << input->info()->data_type() - << " Input Shape: " << input->info()->tensor_shape() + ARM_COMPUTE_LOG(" Type: " << in->info()->data_type() + << " Input Shape: " << in->info()->tensor_shape() << " Weights shape: " << _weights.info().tensor_shape() << " Biases Shape: " << _biases.info().tensor_shape() - << " Output Shape: " << output->info()->tensor_shape() + << " Output Shape: " << out->info()->tensor_shape() << std::endl); return func; -- cgit v1.2.1