From 9204646e091ffc25eda61768537357916a4f7df4 Mon Sep 17 00:00:00 2001 From: Matthew Bentham Date: Sat, 7 Mar 2020 22:15:55 +0000 Subject: COMPMID-3069: Remove unnecessary uses of ToolchainSupport.h Change-Id: I9c09d1002043fd2f927493a85924298d54b1ad9c Signed-off-by: Georgios Pinitas Reviewed-on: https://review.mlplatform.org/c/ml/ComputeLibrary/+/2854 Tested-by: Arm Jenkins Reviewed-by: Michele Di Giorgio --- .../NEON/functions/NEWinogradConvolutionLayer.cpp | 41 ++++++++++------------ 1 file changed, 19 insertions(+), 22 deletions(-) (limited to 'src/runtime/NEON/functions/NEWinogradConvolutionLayer.cpp') diff --git a/src/runtime/NEON/functions/NEWinogradConvolutionLayer.cpp b/src/runtime/NEON/functions/NEWinogradConvolutionLayer.cpp index 6983c1c01b..81190fbf0e 100644 --- a/src/runtime/NEON/functions/NEWinogradConvolutionLayer.cpp +++ b/src/runtime/NEON/functions/NEWinogradConvolutionLayer.cpp @@ -1,5 +1,5 @@ /* - * Copyright (c) 2017-2019 ARM Limited. + * Copyright (c) 2017-2020 ARM Limited. * * SPDX-License-Identifier: MIT * @@ -31,7 +31,7 @@ #include "arm_compute/core/utils/misc/ShapeCalculator.h" #include "arm_compute/runtime/NEON/NEScheduler.h" #include "arm_compute/runtime/NEON/functions/NEGEMMAssemblyDispatch.h" -#include "support/ToolchainSupport.h" +#include "support/MemorySupport.h" #include "arm_compute/core/NEON/kernels/convolution/common/utils.hpp" #include "arm_compute/core/NEON/kernels/convolution/winograd/winograd.hpp" @@ -235,27 +235,26 @@ bool check_support_fast_math(const Size2D &output_tile, const Size2D &kernel_siz inline bool fuse_function_supported(const ActivationLayerInfo &act_info) { - return act_info.activation() == ActivationLayerInfo::ActivationFunction::RELU || - act_info.activation() == ActivationLayerInfo::ActivationFunction::BOUNDED_RELU; + return act_info.activation() == ActivationLayerInfo::ActivationFunction::RELU || act_info.activation() == ActivationLayerInfo::ActivationFunction::BOUNDED_RELU; } arm_gemm::Activation arm_gemm_activation_from_acl_activation(const ActivationLayerInfo &act_info) { - switch(act_info.activation()) + switch(act_info.activation()) + { + case ActivationLayerInfo::ActivationFunction::RELU: + { + return arm_gemm::Activation(arm_gemm::Activation::Type::ReLU, act_info.a(), act_info.b()); + } + case ActivationLayerInfo::ActivationFunction::BOUNDED_RELU: + { + return arm_gemm::Activation(arm_gemm::Activation::Type::BoundedReLU, act_info.a(), act_info.b()); + } + default: { - case ActivationLayerInfo::ActivationFunction::RELU: - { - return arm_gemm::Activation(arm_gemm::Activation::Type::ReLU, act_info.a(), act_info.b()); - } - case ActivationLayerInfo::ActivationFunction::BOUNDED_RELU: - { - return arm_gemm::Activation(arm_gemm::Activation::Type::BoundedReLU, act_info.a(), act_info.b()); - } - default: - { - return arm_gemm::Activation(arm_gemm::Activation::Type::None); - } + return arm_gemm::Activation(arm_gemm::Activation::Type::None); } + } } } //namespace @@ -283,8 +282,6 @@ void NEWinogradConvolutionLayer::configure(const ITensor *input, const ITensor * const Size2D kernel_size = Size2D(weights->info()->dimension(width_idx), weights->info()->dimension(height_idx)); const Size2D output_tile = winograd_output_tile(input_dims, kernel_size); - - // Check if the Winograd configuration requires fast math if(!enable_fast_math) { @@ -511,7 +508,7 @@ void NEWinogradConvolutionLayer::configure(const ITensor *input, const ITensor * _memory_group.manage(&_output_nhwc); output_to_use = &_output_nhwc; } - const arm_gemm::Activation activation = arm_gemm_activation_from_acl_activation(act_info); + const arm_gemm::Activation activation = arm_gemm_activation_from_acl_activation(act_info); transform_output_kernel->configure(biases, &_output_transformed, @@ -542,7 +539,7 @@ void NEWinogradConvolutionLayer::configure(const ITensor *input, const ITensor * _transform_output_kernel = std::move(transform_output_kernel); //Configure Activation Layer - _is_activationlayer_enabled = act_info.enabled() && ! fuse_function_supported(act_info); + _is_activationlayer_enabled = act_info.enabled() && !fuse_function_supported(act_info); if(_is_activationlayer_enabled) { _activationlayer_function.configure(_output, nullptr, act_info); @@ -578,7 +575,7 @@ void NEWinogradConvolutionLayer::run() _permute_output.run(); } - if(_is_activationlayer_enabled ) + if(_is_activationlayer_enabled) { _activationlayer_function.run(); } -- cgit v1.2.1