From ff421f2100e0e9e532f5fe78585300546af61690 Mon Sep 17 00:00:00 2001 From: Georgios Pinitas Date: Wed, 4 Oct 2017 16:53:58 +0100 Subject: COMPMID-601: Add GraphContext GraphContext hold all the information about the hints that need to be passed in the nodes. As these might expand, it serves as a centralized class for such information. Change-Id: I0b5527630fb97cc5fa500db0bac8307ff2ea36e6 Reviewed-on: http://mpd-gerrit.cambridge.arm.com/90300 Tested-by: Kaizen Reviewed-by: Anthony Barbier --- src/graph/nodes/ActivationLayer.cpp | 28 ++++++++-------- src/graph/nodes/ConvolutionLayer.cpp | 58 +++++++++++++++++---------------- src/graph/nodes/FullyConnectedLayer.cpp | 32 +++++++++--------- src/graph/nodes/NormalizationLayer.cpp | 28 ++++++++-------- src/graph/nodes/PoolingLayer.cpp | 28 ++++++++-------- src/graph/nodes/SoftmaxLayer.cpp | 28 ++++++++-------- 6 files changed, 102 insertions(+), 100 deletions(-) (limited to 'src/graph/nodes') diff --git a/src/graph/nodes/ActivationLayer.cpp b/src/graph/nodes/ActivationLayer.cpp index b71e22c601..da2dac04e2 100644 --- a/src/graph/nodes/ActivationLayer.cpp +++ b/src/graph/nodes/ActivationLayer.cpp @@ -34,7 +34,7 @@ using namespace arm_compute::graph; namespace { -template +template std::unique_ptr instantiate_function(ITensor *input, ITensor *output, const ActivationLayerInfo &activation_info) { auto activation = arm_compute::support::cpp14::make_unique(); @@ -46,19 +46,19 @@ std::unique_ptr instantiate_function(ITensor *input, ITe return std::move(activation); } -template +template std::unique_ptr instantiate(ITensor *input, ITensor *output, const ActivationLayerInfo &activation_info); template <> -std::unique_ptr instantiate(ITensor *input, ITensor *output, const ActivationLayerInfo &activation_info) +std::unique_ptr instantiate(ITensor *input, ITensor *output, const ActivationLayerInfo &activation_info) { - return instantiate_function(input, output, activation_info); + return instantiate_function(input, output, activation_info); } template <> -std::unique_ptr instantiate(ITensor *input, ITensor *output, const ActivationLayerInfo &activation_info) +std::unique_ptr instantiate(ITensor *input, ITensor *output, const ActivationLayerInfo &activation_info) { - return instantiate_function(input, output, activation_info); + return instantiate_function(input, output, activation_info); } } // namespace @@ -67,27 +67,27 @@ ActivationLayer::ActivationLayer(const ActivationLayerInfo activation_info) { } -std::unique_ptr ActivationLayer::instantiate_node(Hint hint, ITensor *input, ITensor *output) +std::unique_ptr ActivationLayer::instantiate_node(GraphContext &ctx, ITensor *input, ITensor *output) { std::unique_ptr func; - _hint = hint; - _input = input; - _output = output; + _target_hint = ctx.hints().target_hint(); + _input = input; + _output = output; - if(_hint == Hint::OPENCL) + if(_target_hint == TargetHint::OPENCL) { - func = instantiate(input, output, _activation_info); + func = instantiate(input, output, _activation_info); } else { - func = instantiate(input, output, _activation_info); + func = instantiate(input, output, _activation_info); } return func; } void ActivationLayer::print_info() { - if(_hint == Hint::OPENCL) + if(_target_hint == TargetHint::OPENCL) { std::cout << "Instantiating CLActivationLayer"; } diff --git a/src/graph/nodes/ConvolutionLayer.cpp b/src/graph/nodes/ConvolutionLayer.cpp index ce9f096719..a992095786 100644 --- a/src/graph/nodes/ConvolutionLayer.cpp +++ b/src/graph/nodes/ConvolutionLayer.cpp @@ -65,7 +65,7 @@ TensorShape calculate_convolution_layer_output_shape(const TensorShape &input_sh } // Instantiate GEMM based convolution layer -template +template std::unique_ptr instantiate_function(ITensor *input, ITensor *weights, ITensor *biases, ITensor *output, const PadStrideInfo &conv_info, const WeightsInfo &weights_info) { auto conv = arm_compute::support::cpp14::make_unique(); @@ -79,7 +79,7 @@ std::unique_ptr instantiate_function(ITensor *input, ITe } // Instantiate direct convolution layer -template +template std::unique_ptr instantiate_direct_function(ITensor *input, ITensor *weights, ITensor *biases, ITensor *output, const PadStrideInfo &conv_info) { auto conv = arm_compute::support::cpp14::make_unique(); @@ -92,35 +92,37 @@ std::unique_ptr instantiate_direct_function(ITensor *inp return std::move(conv); } -template +template std::unique_ptr instantiate(ITensor *input, ITensor *weights, ITensor *biases, ITensor *output, const PadStrideInfo &conv_info, const WeightsInfo &weights_info, ConvolutionMethodHint conv_method); template <> -std::unique_ptr instantiate(ITensor *input, ITensor *weights, ITensor *biases, ITensor *output, const PadStrideInfo &conv_info, const WeightsInfo &weights_info, - ConvolutionMethodHint conv_method) +std::unique_ptr instantiate(ITensor *input, ITensor *weights, ITensor *biases, ITensor *output, const PadStrideInfo &conv_info, + const WeightsInfo &weights_info, + ConvolutionMethodHint conv_method) { if(conv_method == ConvolutionMethodHint::GEMM) { - return instantiate_function(input, weights, biases, output, conv_info, weights_info); + return instantiate_function(input, weights, biases, output, conv_info, weights_info); } else { - return instantiate_direct_function(input, weights, biases, output, conv_info); + return instantiate_direct_function(input, weights, biases, output, conv_info); } } template <> -std::unique_ptr instantiate(ITensor *input, ITensor *weights, ITensor *biases, ITensor *output, const PadStrideInfo &conv_info, const WeightsInfo &weights_info, - ConvolutionMethodHint conv_method) +std::unique_ptr instantiate(ITensor *input, ITensor *weights, ITensor *biases, ITensor *output, const PadStrideInfo &conv_info, + const WeightsInfo &weights_info, + ConvolutionMethodHint conv_method) { if(conv_method == ConvolutionMethodHint::GEMM) { - return instantiate_function(input, weights, biases, output, conv_info, weights_info); + return instantiate_function(input, weights, biases, output, conv_info, weights_info); } else { - return instantiate_direct_function(input, weights, biases, output, conv_info); + return instantiate_direct_function(input, weights, biases, output, conv_info); } } } // namespace @@ -166,7 +168,7 @@ private: std::vector> _convolutions; }; -std::unique_ptr ConvolutionLayer::instantiate_node(Hint hint, ITensor *input, ITensor *output) +std::unique_ptr ConvolutionLayer::instantiate_node(GraphContext &ctx, ITensor *input, ITensor *output) { // Set weights and biases info if(_weights.tensor() == nullptr) @@ -181,17 +183,18 @@ std::unique_ptr ConvolutionLayer::instantiate_node(Hint } std::unique_ptr func; - _hint = hint; - _input = input; - _output = output; + _target_hint = ctx.hints().target_hint(); + _input = input; + _output = output; + const ConvolutionMethodHint conv_method_hint = ctx.hints().convolution_method_hint(); // Check if the weights and biases are loaded bool weights_are_loaded = _weights.tensor() != nullptr; bool biases_are_loaded = _weights.tensor() != nullptr; // Set bias and weights target - _weights.set_target(_hint); - _biases.set_target(_hint); + _weights.set_target(_target_hint); + _biases.set_target(_target_hint); // Calculate output shape TensorShape output_shape = calculate_convolution_layer_output_shape(_input->info()->tensor_shape(), _weights.info().tensor_shape(), _conv_info); @@ -200,14 +203,13 @@ std::unique_ptr ConvolutionLayer::instantiate_node(Hint arm_compute::auto_init_if_empty(*_output->info(), output_shape, 1, _input->info()->data_type(), _input->info()->fixed_point_position()); // Create appropriate convolution function - // TODO(geopin01): Fix convolution layer hints once the GraphContext has been added if(_num_groups == 1) { - func = instantiate_convolution(ConvolutionMethodHint::GEMM); + func = instantiate_convolution(conv_method_hint); } else { - func = instantiate_grouped_convolution(ConvolutionMethodHint::GEMM); + func = instantiate_grouped_convolution(conv_method_hint); } // Fill weights @@ -226,7 +228,7 @@ std::unique_ptr ConvolutionLayer::instantiate_node(Hint void ConvolutionLayer::print_info() { - if(_hint == Hint::OPENCL) + if(_target_hint == TargetHint::OPENCL) { std::cout << "Instantiating CLConvolutionLayer"; } @@ -248,13 +250,13 @@ void ConvolutionLayer::print_info() std::unique_ptr ConvolutionLayer::instantiate_convolution(ConvolutionMethodHint conv_method_hint) { std::unique_ptr func; - if(_hint == Hint::OPENCL) + if(_target_hint == TargetHint::OPENCL) { - func = instantiate(_input, _weights.tensor(), _biases.tensor(), _output, _conv_info, _weights_info, conv_method_hint); + func = instantiate(_input, _weights.tensor(), _biases.tensor(), _output, _conv_info, _weights_info, conv_method_hint); } else { - func = instantiate(_input, _weights.tensor(), _biases.tensor(), _output, _conv_info, _weights_info, conv_method_hint); + func = instantiate(_input, _weights.tensor(), _biases.tensor(), _output, _conv_info, _weights_info, conv_method_hint); } return func; } @@ -306,20 +308,20 @@ std::unique_ptr ConvolutionLayer::instantiate_grouped_co Coordinates biases_coord(biases_split * i); // Create sub-tensors for input, output, weights and bias - auto hint_to_use = (_hint == Hint::OPENCL) ? Hint::OPENCL : Hint::NEON; + auto hint_to_use = (_target_hint == TargetHint::OPENCL) ? TargetHint::OPENCL : TargetHint::NEON; _is[i] = SubTensor(_input, input_shape, input_coord, hint_to_use); _os[i] = SubTensor(_output, output_shape, output_coord, hint_to_use); _ws[i] = SubTensor(_weights.tensor(), weights_shape, weights_coord, hint_to_use); _bs[i] = SubTensor(_biases.tensor(), biases_shape, biases_coord, hint_to_use); // Instantiate convolution function - if(_hint == Hint::OPENCL) + if(_target_hint == TargetHint::OPENCL) { - func = instantiate(_is[i].tensor(), _ws[i].tensor(), _bs[i].tensor(), _os[i].tensor(), _conv_info, _weights_info, conv_method_hint); + func = instantiate(_is[i].tensor(), _ws[i].tensor(), _bs[i].tensor(), _os[i].tensor(), _conv_info, _weights_info, conv_method_hint); } else { - func = instantiate(_is[i].tensor(), _ws[i].tensor(), _bs[i].tensor(), _os[i].tensor(), _conv_info, _weights_info, conv_method_hint); + func = instantiate(_is[i].tensor(), _ws[i].tensor(), _bs[i].tensor(), _os[i].tensor(), _conv_info, _weights_info, conv_method_hint); } // Add convolution function to the list of convolutions for the grouped convolution diff --git a/src/graph/nodes/FullyConnectedLayer.cpp b/src/graph/nodes/FullyConnectedLayer.cpp index fcc86be8fa..c317660b20 100644 --- a/src/graph/nodes/FullyConnectedLayer.cpp +++ b/src/graph/nodes/FullyConnectedLayer.cpp @@ -43,7 +43,7 @@ TensorShape calculate_fullyconnected_layer_output_shape(const TensorShape &input } return TensorShape(output_neurons, batches); } -template +template std::unique_ptr instantiate_function(ITensor *input, Tensor &weights, Tensor &biases, ITensor *output) { bool weights_are_loaded = weights.tensor() != nullptr; @@ -52,8 +52,8 @@ std::unique_ptr instantiate_function(ITensor *input, Ten auto conv = arm_compute::support::cpp14::make_unique(); conv->configure( dynamic_cast(input), - dynamic_cast(weights.set_target(hint)), - dynamic_cast(biases.set_target(hint)), + dynamic_cast(weights.set_target(target_hint)), + dynamic_cast(biases.set_target(target_hint)), dynamic_cast(output)); if(!weights_are_loaded) { @@ -67,23 +67,23 @@ std::unique_ptr instantiate_function(ITensor *input, Ten return std::move(conv); } -template +template std::unique_ptr instantiate(ITensor *input, Tensor &weights, Tensor &biases, ITensor *output); template <> -std::unique_ptr instantiate(ITensor *input, Tensor &weights, Tensor &biases, ITensor *output) +std::unique_ptr instantiate(ITensor *input, Tensor &weights, Tensor &biases, ITensor *output) { - return instantiate_function(input, weights, biases, output); + return instantiate_function(input, weights, biases, output); } template <> -std::unique_ptr instantiate(ITensor *input, Tensor &weights, Tensor &biases, ITensor *output) +std::unique_ptr instantiate(ITensor *input, Tensor &weights, Tensor &biases, ITensor *output) { - return instantiate_function(input, weights, biases, output); + return instantiate_function(input, weights, biases, output); } } // namespace -std::unique_ptr FullyConnectedLayer::instantiate_node(Hint hint, ITensor *input, ITensor *output) +std::unique_ptr FullyConnectedLayer::instantiate_node(GraphContext &ctx, ITensor *input, ITensor *output) { if(_weights.tensor() == nullptr) { @@ -111,17 +111,17 @@ std::unique_ptr FullyConnectedLayer::instantiate_node(Hi input->info()->num_channels(), input->info()->data_type(), input->info()->fixed_point_position()); std::unique_ptr func; - _hint = hint; - _input = input; - _output = output; + _target_hint = ctx.hints().target_hint(); + _input = input; + _output = output; - if(_hint == Hint::OPENCL) + if(_target_hint == TargetHint::OPENCL) { - func = instantiate(input, _weights, _biases, output); + func = instantiate(input, _weights, _biases, output); } else { - func = instantiate(input, _weights, _biases, output); + func = instantiate(input, _weights, _biases, output); } return func; @@ -129,7 +129,7 @@ std::unique_ptr FullyConnectedLayer::instantiate_node(Hi void FullyConnectedLayer::print_info() { - if(_hint == Hint::OPENCL) + if(_target_hint == TargetHint::OPENCL) { std::cout << "Instantiating CLFullyConnectedLayer"; } diff --git a/src/graph/nodes/NormalizationLayer.cpp b/src/graph/nodes/NormalizationLayer.cpp index 55ef9bf243..99d07dc8da 100644 --- a/src/graph/nodes/NormalizationLayer.cpp +++ b/src/graph/nodes/NormalizationLayer.cpp @@ -34,7 +34,7 @@ using namespace arm_compute::graph; namespace { -template +template std::unique_ptr instantiate_function(ITensor *input, ITensor *output, const NormalizationLayerInfo &norm_info) { auto norm = arm_compute::support::cpp14::make_unique(); @@ -46,19 +46,19 @@ std::unique_ptr instantiate_function(ITensor *input, ITe return std::move(norm); } -template +template std::unique_ptr instantiate(ITensor *input, ITensor *output, const NormalizationLayerInfo &norm_info); template <> -std::unique_ptr instantiate(ITensor *input, ITensor *output, const NormalizationLayerInfo &norm_info) +std::unique_ptr instantiate(ITensor *input, ITensor *output, const NormalizationLayerInfo &norm_info) { - return instantiate_function(input, output, norm_info); + return instantiate_function(input, output, norm_info); } template <> -std::unique_ptr instantiate(ITensor *input, ITensor *output, const NormalizationLayerInfo &norm_info) +std::unique_ptr instantiate(ITensor *input, ITensor *output, const NormalizationLayerInfo &norm_info) { - return instantiate_function(input, output, norm_info); + return instantiate_function(input, output, norm_info); } } // namespace @@ -67,20 +67,20 @@ NormalizationLayer::NormalizationLayer(const NormalizationLayerInfo norm_info) { } -std::unique_ptr NormalizationLayer::instantiate_node(Hint hint, ITensor *input, ITensor *output) +std::unique_ptr NormalizationLayer::instantiate_node(GraphContext &ctx, ITensor *input, ITensor *output) { std::unique_ptr func; - _hint = hint; - _input = input; - _output = output; + _target_hint = ctx.hints().target_hint(); + _input = input; + _output = output; - if(_hint == Hint::OPENCL) + if(_target_hint == TargetHint::OPENCL) { - func = instantiate(input, output, _norm_info); + func = instantiate(input, output, _norm_info); } else { - func = instantiate(input, output, _norm_info); + func = instantiate(input, output, _norm_info); } return func; @@ -88,7 +88,7 @@ std::unique_ptr NormalizationLayer::instantiate_node(Hin void NormalizationLayer::print_info() { - if(_hint == Hint::OPENCL) + if(_target_hint == TargetHint::OPENCL) { std::cout << "Instantiating CLNormalizationLayer"; } diff --git a/src/graph/nodes/PoolingLayer.cpp b/src/graph/nodes/PoolingLayer.cpp index f29332f65b..2a5e4cb3d8 100644 --- a/src/graph/nodes/PoolingLayer.cpp +++ b/src/graph/nodes/PoolingLayer.cpp @@ -34,7 +34,7 @@ using namespace arm_compute::graph; namespace { -template +template std::unique_ptr instantiate_function(ITensor *input, ITensor *output, const PoolingLayerInfo &pool_info) { auto pool = arm_compute::support::cpp14::make_unique(); @@ -46,19 +46,19 @@ std::unique_ptr instantiate_function(ITensor *input, ITe return std::move(pool); } -template +template std::unique_ptr instantiate(ITensor *input, ITensor *output, const PoolingLayerInfo &pool_info); template <> -std::unique_ptr instantiate(ITensor *input, ITensor *output, const PoolingLayerInfo &pool_info) +std::unique_ptr instantiate(ITensor *input, ITensor *output, const PoolingLayerInfo &pool_info) { - return instantiate_function(input, output, pool_info); + return instantiate_function(input, output, pool_info); } template <> -std::unique_ptr instantiate(ITensor *input, ITensor *output, const PoolingLayerInfo &pool_info) +std::unique_ptr instantiate(ITensor *input, ITensor *output, const PoolingLayerInfo &pool_info) { - return instantiate_function(input, output, pool_info); + return instantiate_function(input, output, pool_info); } } // namespace @@ -67,20 +67,20 @@ PoolingLayer::PoolingLayer(const PoolingLayerInfo pool_info) { } -std::unique_ptr PoolingLayer::instantiate_node(Hint hint, ITensor *input, ITensor *output) +std::unique_ptr PoolingLayer::instantiate_node(GraphContext &ctx, ITensor *input, ITensor *output) { std::unique_ptr func; - _hint = hint; - _input = input; - _output = output; + _target_hint = ctx.hints().target_hint(); + _input = input; + _output = output; - if(_hint == Hint::OPENCL) + if(_target_hint == TargetHint::OPENCL) { - func = instantiate(input, output, _pool_info); + func = instantiate(input, output, _pool_info); } else { - func = instantiate(input, output, _pool_info); + func = instantiate(input, output, _pool_info); } return func; @@ -88,7 +88,7 @@ std::unique_ptr PoolingLayer::instantiate_node(Hint hint void PoolingLayer::print_info() { - if(_hint == Hint::OPENCL) + if(_target_hint == TargetHint::OPENCL) { std::cout << "Instantiating CLPoolingLayer"; } diff --git a/src/graph/nodes/SoftmaxLayer.cpp b/src/graph/nodes/SoftmaxLayer.cpp index fee88970fc..9e798ef7cc 100644 --- a/src/graph/nodes/SoftmaxLayer.cpp +++ b/src/graph/nodes/SoftmaxLayer.cpp @@ -34,7 +34,7 @@ using namespace arm_compute::graph; namespace { -template +template std::unique_ptr instantiate_function(ITensor *input, ITensor *output) { auto softmax = arm_compute::support::cpp14::make_unique(); @@ -45,36 +45,36 @@ std::unique_ptr instantiate_function(ITensor *input, ITe return std::move(softmax); } -template +template std::unique_ptr instantiate(ITensor *input, ITensor *output); template <> -std::unique_ptr instantiate(ITensor *input, ITensor *output) +std::unique_ptr instantiate(ITensor *input, ITensor *output) { - return instantiate_function(input, output); + return instantiate_function(input, output); } template <> -std::unique_ptr instantiate(ITensor *input, ITensor *output) +std::unique_ptr instantiate(ITensor *input, ITensor *output) { - return instantiate_function(input, output); + return instantiate_function(input, output); } } // namespace -std::unique_ptr SoftmaxLayer::instantiate_node(Hint hint, ITensor *input, ITensor *output) +std::unique_ptr SoftmaxLayer::instantiate_node(GraphContext &ctx, ITensor *input, ITensor *output) { std::unique_ptr func; - _hint = hint; - _input = input; - _output = output; + _target_hint = ctx.hints().target_hint(); + _input = input; + _output = output; - if(_hint == Hint::OPENCL) + if(_target_hint == TargetHint::OPENCL) { - func = instantiate(input, output); + func = instantiate(input, output); } else { - func = instantiate(input, output); + func = instantiate(input, output); } return func; @@ -82,7 +82,7 @@ std::unique_ptr SoftmaxLayer::instantiate_node(Hint hint void SoftmaxLayer::print_info() { - if(_hint == Hint::OPENCL) + if(_target_hint == TargetHint::OPENCL) { std::cout << "Instantiating CLSoftmaxLayer"; } -- cgit v1.2.1