From 33f41fabd30fb444aaa0cf3e65b61794d498d151 Mon Sep 17 00:00:00 2001 From: Michele Di Giorgio Date: Tue, 9 Mar 2021 14:09:08 +0000 Subject: Fix trademarks throughout the codebase Resolves: COMPMID-4299 Change-Id: Ie6a52c1371b9a2a7b5bb4f019ecd5e70a2008567 Signed-off-by: Michele Di Giorgio Reviewed-on: https://review.mlplatform.org/c/ml/ComputeLibrary/+/5338 Reviewed-by: Georgios Pinitas Comments-Addressed: Arm Jenkins Tested-by: Arm Jenkins --- arm_compute/runtime/NEON/functions/NEArgMinMaxLayer.h | 2 +- .../runtime/NEON/functions/NEBoundingBoxTransform.h | 8 ++------ .../runtime/NEON/functions/NEConvolutionLayer.h | 2 +- .../runtime/NEON/functions/NEDeconvolutionLayer.h | 2 +- .../NEON/functions/NEDepthwiseConvolutionLayer.h | 4 ++-- .../runtime/NEON/functions/NEDirectConvolutionLayer.h | 2 +- arm_compute/runtime/NEON/functions/NEFFT1D.h | 2 +- arm_compute/runtime/NEON/functions/NEFFT2D.h | 2 +- .../runtime/NEON/functions/NEFFTConvolutionLayer.h | 6 +++--- .../runtime/NEON/functions/NEFullyConnectedLayer.h | 4 ++-- arm_compute/runtime/NEON/functions/NEGEMM.h | 2 +- arm_compute/runtime/NEON/functions/NEGEMMConv2d.h | 2 +- .../runtime/NEON/functions/NEGEMMConvolutionLayer.h | 2 +- .../NEON/functions/NEGEMMLowpMatrixMultiplyCore.h | 3 +-- .../runtime/NEON/functions/NEGEMMLowpOutputStage.h | 18 +++++++++--------- .../runtime/NEON/functions/NELSTMLayerQuantized.h | 2 +- .../runtime/NEON/functions/NEMaxUnpoolingLayer.h | 2 +- .../runtime/NEON/functions/NENormalizationLayer.h | 2 +- arm_compute/runtime/NEON/functions/NEPadLayer.h | 2 +- arm_compute/runtime/NEON/functions/NEPermute.h | 2 +- arm_compute/runtime/NEON/functions/NEPoolingLayer.h | 2 +- arm_compute/runtime/NEON/functions/NEQLSTMLayer.h | 2 +- .../runtime/NEON/functions/NEQuantizationLayer.h | 7 +------ arm_compute/runtime/NEON/functions/NEROIAlignLayer.h | 7 +------ arm_compute/runtime/NEON/functions/NEROIPoolingLayer.h | 7 +------ .../runtime/NEON/functions/NEReductionOperation.h | 2 +- arm_compute/runtime/NEON/functions/NERemap.h | 5 +---- .../runtime/NEON/functions/NESpaceToBatchLayer.h | 2 +- .../runtime/NEON/functions/NESpaceToDepthLayer.h | 5 +---- .../NEON/functions/NEWinogradConvolutionLayer.h | 3 ++- 30 files changed, 44 insertions(+), 69 deletions(-) (limited to 'arm_compute/runtime/NEON/functions') diff --git a/arm_compute/runtime/NEON/functions/NEArgMinMaxLayer.h b/arm_compute/runtime/NEON/functions/NEArgMinMaxLayer.h index 8235185a8e..cbf1d5b444 100644 --- a/arm_compute/runtime/NEON/functions/NEArgMinMaxLayer.h +++ b/arm_compute/runtime/NEON/functions/NEArgMinMaxLayer.h @@ -37,7 +37,7 @@ class ITensor; /** Function to calculate the index of the minimum or maximum values in a * tensor based on an axis. * - * This function calls the following Neon kernels: + * This function calls the following kernels: * * -# @ref NEReductionOperationKernel * -# @ref NEFillBorderKernel diff --git a/arm_compute/runtime/NEON/functions/NEBoundingBoxTransform.h b/arm_compute/runtime/NEON/functions/NEBoundingBoxTransform.h index de8dfef4ed..c377520a12 100644 --- a/arm_compute/runtime/NEON/functions/NEBoundingBoxTransform.h +++ b/arm_compute/runtime/NEON/functions/NEBoundingBoxTransform.h @@ -1,5 +1,5 @@ /* - * Copyright (c) 2019-2020 Arm Limited. + * Copyright (c) 2019-2021 Arm Limited. * * SPDX-License-Identifier: MIT * @@ -32,11 +32,7 @@ namespace arm_compute class ITensor; class ITensorInfo; -/** Basic function to run @ref NEBoundingBoxTransformKernel. - * - * This function calls the following Neon kernels: - * -# @ref NEBoundingBoxTransformKernel - */ +/** Basic function to run @ref NEBoundingBoxTransformKernel. */ class NEBoundingBoxTransform : public INESimpleFunctionNoBorder { public: diff --git a/arm_compute/runtime/NEON/functions/NEConvolutionLayer.h b/arm_compute/runtime/NEON/functions/NEConvolutionLayer.h index a387255b6c..d2d41c1e8a 100644 --- a/arm_compute/runtime/NEON/functions/NEConvolutionLayer.h +++ b/arm_compute/runtime/NEON/functions/NEConvolutionLayer.h @@ -37,7 +37,7 @@ namespace arm_compute // Forward declarations class ITensor; -/** Basic function to simulate a convolution layer. This function calls one of the following Neon functions: +/** Basic function to simulate a convolution layer. This function calls one of the following functions: * -# @ref NEGEMMConvolutionLayer (executed only in case GEMM is required for the operation) * -# @ref NEWinogradConvolutionLayer (executed only in case Winograd is required for the operation) * -# @ref NEDirectConvolutionLayer (executed only in case Direct Convolution is required for the operation) diff --git a/arm_compute/runtime/NEON/functions/NEDeconvolutionLayer.h b/arm_compute/runtime/NEON/functions/NEDeconvolutionLayer.h index 02a0f784ec..3864a663c2 100644 --- a/arm_compute/runtime/NEON/functions/NEDeconvolutionLayer.h +++ b/arm_compute/runtime/NEON/functions/NEDeconvolutionLayer.h @@ -64,7 +64,7 @@ namespace arm_compute * The weights used by Deconvolution are supposed to be the same as the ones used for Convolution. Therefore, it will be necessary to use the weights in the * reverse order to perform an actual convolution. This is achieved by using @ref NEReverse. * - * This function calls the following Neon kernels/functions: + * This function calls the following kernels/functions: * * -# @ref CPPUpsample * -# @ref NEConvolutionLayer diff --git a/arm_compute/runtime/NEON/functions/NEDepthwiseConvolutionLayer.h b/arm_compute/runtime/NEON/functions/NEDepthwiseConvolutionLayer.h index 98fffe0b33..9aa8f04eb8 100644 --- a/arm_compute/runtime/NEON/functions/NEDepthwiseConvolutionLayer.h +++ b/arm_compute/runtime/NEON/functions/NEDepthwiseConvolutionLayer.h @@ -110,7 +110,7 @@ private: const PadStrideInfo &conv_info, unsigned int depth_multiplier = 1, ActivationLayerInfo act_info = ActivationLayerInfo(), const Size2D &dilation = Size2D(1U, 1U)); - /** Basic function to execute optimized depthwise convolution routines. This function calls the following Neon kernels: + /** Basic function to execute optimized depthwise convolution routines. This function calls the following kernels: * * @note At the moment 3x3 and 5x5 convolution of stride 1, 2 are supported * @@ -192,7 +192,7 @@ private: bool _is_prepared; }; - /** Basic function to execute a generic depthwise convolution. This function calls the following Neon kernel: + /** Basic function to execute a generic depthwise convolution. This function calls the following kernel: * * -# @ref NEDepthwiseConvolutionLayerNativeKernel * diff --git a/arm_compute/runtime/NEON/functions/NEDirectConvolutionLayer.h b/arm_compute/runtime/NEON/functions/NEDirectConvolutionLayer.h index ff0c3054d8..86914fa0bc 100644 --- a/arm_compute/runtime/NEON/functions/NEDirectConvolutionLayer.h +++ b/arm_compute/runtime/NEON/functions/NEDirectConvolutionLayer.h @@ -41,7 +41,7 @@ class NEFillBorderKernel; /** Function to run the direct convolution. * - * This function calls the following Neon kernels: + * This function calls the following kernels: * * -# @ref NEFillBorderKernel for the input * -# @ref NEDirectConvolutionLayerOutputStageKernel diff --git a/arm_compute/runtime/NEON/functions/NEFFT1D.h b/arm_compute/runtime/NEON/functions/NEFFT1D.h index 04e8f81b69..a533aa7f48 100644 --- a/arm_compute/runtime/NEON/functions/NEFFT1D.h +++ b/arm_compute/runtime/NEON/functions/NEFFT1D.h @@ -40,7 +40,7 @@ class NEFFTDigitReverseKernel; class NEFFTRadixStageKernel; class NEFFTScaleKernel; -/** Basic function to execute one dimensional FFT. This function calls the following Neon kernels: +/** Basic function to execute one dimensional FFT. This function calls the following kernels: * * -# @ref NEFFTDigitReverseKernel Performs digit reverse * -# @ref NEFFTRadixStageKernel A list of FFT kernels depending on the radix decomposition diff --git a/arm_compute/runtime/NEON/functions/NEFFT2D.h b/arm_compute/runtime/NEON/functions/NEFFT2D.h index 218401b429..ce84a85105 100644 --- a/arm_compute/runtime/NEON/functions/NEFFT2D.h +++ b/arm_compute/runtime/NEON/functions/NEFFT2D.h @@ -36,7 +36,7 @@ namespace arm_compute // Forward declaration class ITensor; -/** Basic function to execute two dimensional FFT. This function calls the following Neon kernels: +/** Basic function to execute two dimensional FFT. This function calls the following kernels: * * -# @ref NEFFT1D 1D FFT is performed on the first given axis * -# @ref NEFFT1D 1D FFT is performed on the second given axis diff --git a/arm_compute/runtime/NEON/functions/NEFFTConvolutionLayer.h b/arm_compute/runtime/NEON/functions/NEFFTConvolutionLayer.h index 8967363e7f..213fa6093b 100644 --- a/arm_compute/runtime/NEON/functions/NEFFTConvolutionLayer.h +++ b/arm_compute/runtime/NEON/functions/NEFFTConvolutionLayer.h @@ -43,7 +43,7 @@ namespace arm_compute // Forward declarations class ITensor; -/** Basic function to execute FFT-based convolution on Neon. This function calls the following Neon functions/kernels: +/** Basic function to execute FFT-based convolution on CPU. This function calls the following functions/kernels: * * -# @ref NEPermute Permute input if NHWC(only NCHW is supported). * -# @ref NEPadLayer Pad input. @@ -84,7 +84,7 @@ public: * Data types supported: Same as @p input. * @param[in] conv_info Contains padding and stride information described in @ref PadStrideInfo. * @param[in] act_info (Optional) Activation layer information in case of a fused activation. - * @param[in] enable_fast_math (Optional) Enable fast math computation. Unused for Neon backend. + * @param[in] enable_fast_math (Optional) Enable fast math computation. Unused for CPU backend. */ void configure(ITensor *input, const ITensor *weights, const ITensor *biases, ITensor *output, const PadStrideInfo &conv_info, const ActivationLayerInfo &act_info = ActivationLayerInfo(), bool enable_fast_math = false); @@ -101,7 +101,7 @@ public: * Data types supported: Same as @p input. * @param[in] conv_info Contains padding and stride information described in @ref PadStrideInfo. * @param[in] act_info (Optional) Activation layer information in case of a fused activation. - * @param[in] enable_fast_math (Optional) Enable fast math computation. Unused for Neon backend. + * @param[in] enable_fast_math (Optional) Enable fast math computation. Unused for CPU backend. * * @return a status */ diff --git a/arm_compute/runtime/NEON/functions/NEFullyConnectedLayer.h b/arm_compute/runtime/NEON/functions/NEFullyConnectedLayer.h index 1b3f36d866..58b11744a8 100644 --- a/arm_compute/runtime/NEON/functions/NEFullyConnectedLayer.h +++ b/arm_compute/runtime/NEON/functions/NEFullyConnectedLayer.h @@ -35,7 +35,7 @@ namespace arm_compute { -/** Basic function to reshape the weights of Fully Connected layer with Neon. This function calls the following kernels: +/** Basic function to reshape the weights of Fully Connected layer. This function calls the following kernels: * * @note The fully connected layer accepts "weights" tensors only with 2 dimensions. */ @@ -117,7 +117,7 @@ private: }; } // namespace weights_transformations -/** Basic function to compute a Fully Connected layer on Neon. This function calls the following Neon kernels: +/** Basic function to compute a Fully Connected layer. This function calls the following kernels: * -# @ref NEIm2ColKernel (called when the input comes from a convolutional layer) * -# @ref NEFullyConnectedLayerReshapeWeights (if @p are_weights_reshaped is set to false and transpose_weights is set to true ) (called once) * -# @ref NEGEMMMatrixMultiplyKernel or @ref NEGEMMLowpMatrixMultiplyCore (if quantized asymmetric) diff --git a/arm_compute/runtime/NEON/functions/NEGEMM.h b/arm_compute/runtime/NEON/functions/NEGEMM.h index 6f7951eece..a6c3436656 100644 --- a/arm_compute/runtime/NEON/functions/NEGEMM.h +++ b/arm_compute/runtime/NEON/functions/NEGEMM.h @@ -43,7 +43,7 @@ class NEGEMMMatrixMultiplyKernel; class NEGEMMTranspose1xWKernel; class NEGEMMAssemblyDispatch; -/** Basic function to execute GEMM on Neon. This function calls the following Neon kernels: +/** Basic function to execute GEMM. This function calls the following kernels: * * If optimized assembly is available: * -# @ref NEGEMMAssemblyDispatch diff --git a/arm_compute/runtime/NEON/functions/NEGEMMConv2d.h b/arm_compute/runtime/NEON/functions/NEGEMMConv2d.h index 2bd233f520..8c3ba4f0c8 100644 --- a/arm_compute/runtime/NEON/functions/NEGEMMConv2d.h +++ b/arm_compute/runtime/NEON/functions/NEGEMMConv2d.h @@ -38,7 +38,7 @@ namespace arm_compute class ITensor; class NEGEMMAssemblyDispatch; -/** Basic function to compute the convolution layer. This function calls the following Neon kernels/functions: +/** Basic function to compute the convolution layer. This function calls the following kernels/functions: * * Supports only NHWC data layout * diff --git a/arm_compute/runtime/NEON/functions/NEGEMMConvolutionLayer.h b/arm_compute/runtime/NEON/functions/NEGEMMConvolutionLayer.h index 33f00c087c..9897bf1d4d 100644 --- a/arm_compute/runtime/NEON/functions/NEGEMMConvolutionLayer.h +++ b/arm_compute/runtime/NEON/functions/NEGEMMConvolutionLayer.h @@ -150,7 +150,7 @@ private: }; } // namespace weights_transformations -/** Basic function to compute the convolution layer. This function calls the following Neon kernels/functions: +/** Basic function to compute the convolution layer. This function calls the following kernels/functions: * * -# @ref NEIm2ColKernel * -# @ref NEGEMM (if the data type is BFLOAT16/FP16/FP32) diff --git a/arm_compute/runtime/NEON/functions/NEGEMMLowpMatrixMultiplyCore.h b/arm_compute/runtime/NEON/functions/NEGEMMLowpMatrixMultiplyCore.h index 821b498dad..b2b77bd406 100644 --- a/arm_compute/runtime/NEON/functions/NEGEMMLowpMatrixMultiplyCore.h +++ b/arm_compute/runtime/NEON/functions/NEGEMMLowpMatrixMultiplyCore.h @@ -37,7 +37,6 @@ namespace arm_compute { class ITensor; class NEConvertQuantizedSignednessKernel; -class NEConvertQuantizedSignednessKernel; class NEGEMMInterleave4x4Kernel; class NEGEMMLowpMatrixMultiplyKernel; class NEGEMMLowpOffsetContributionKernel; @@ -47,7 +46,7 @@ class NEGEMMLowpMatrixBReductionKernel; class NEGEMMTranspose1xWKernel; class NEGEMMAssemblyDispatch; -/** Basic function to execute GEMMLowpMatrixMultiplyCore on Neon. This function calls the following Neon kernels if the DOT product instruction is not available: +/** Basic function to execute GEMMLowpMatrixMultiplyCore. This function calls the following kernels if the DOT product instruction is not available: * * -# @ref NEGEMMInterleave4x4Kernel * -# @ref NEGEMMTranspose1xWKernel diff --git a/arm_compute/runtime/NEON/functions/NEGEMMLowpOutputStage.h b/arm_compute/runtime/NEON/functions/NEGEMMLowpOutputStage.h index 79b427ea6f..c22ed1b5c4 100644 --- a/arm_compute/runtime/NEON/functions/NEGEMMLowpOutputStage.h +++ b/arm_compute/runtime/NEON/functions/NEGEMMLowpOutputStage.h @@ -27,7 +27,7 @@ #include "arm_compute/core/Types.h" #include "arm_compute/runtime/NEON/INESimpleFunctionNoBorder.h" -/** This file contains all available output stages for GEMMLowp on Neon. +/** This file contains all available output stages for GEMMLowp. * * In gemmlowp, the "output stage" is the process that takes a final int32 accumulator value (the output of @ref NEGEMMLowpMatrixMultiplyCore), * and processes it to obtain the final ASYMM8 value. @@ -40,7 +40,7 @@ namespace arm_compute class ITensor; class ITensorInfo; -/** Basic function to execute NEGEMMLowpQuantizeDownInt32ToUint8ScaleByFixedPoint on Neon. +/** Basic function to execute NEGEMMLowpQuantizeDownInt32ToUint8ScaleByFixedPoint. * * NEGEMMLowpQuantizeDownInt32ToUint8ScaleByFixedPoint depends on 3 parameters: * @@ -61,7 +61,7 @@ class ITensorInfo; * * ((FixedPointMul(input[i][k] + bias[k], result_fixedpoint_multiplier)) >> result_shift) + result_offset_after_shift * - * This function calls the following Neon kernels: + * This function calls the following kernels: * * -# @ref NEGEMMLowpQuantizeDownInt32ToUint8ScaleByFixedPointKernel * @@ -112,7 +112,7 @@ public: */ static Status validate(const ITensorInfo *input, const ITensorInfo *bias, const ITensorInfo *output, int min = std::numeric_limits::lowest(), int max = std::numeric_limits::max()); }; -/** Basic function to execute NEGEMMLowpQuantizeDownInt32ToInt8ScaleByFixedPoint on Neon. +/** Basic function to execute NEGEMMLowpQuantizeDownInt32ToInt8ScaleByFixedPoint. * * NEGEMMLowpQuantizeDownInt32ToInt8ScaleByFixedPoint depends on 3 parameters: * @@ -133,7 +133,7 @@ public: * * ((FixedPointMul(input[i][k] + bias[k], result_fixedpoint_multiplier)) >> result_shift) + result_offset_after_shift * - * This function calls the following Neon kernels: + * This function calls the following kernels: * * -# @ref NEGEMMLowpQuantizeDownInt32ToInt8ScaleByFixedPointKernel * @@ -184,7 +184,7 @@ public: */ static Status validate(const ITensorInfo *input, const ITensorInfo *bias, const ITensorInfo *output, int min = std::numeric_limits::lowest(), int max = std::numeric_limits::max()); }; -/** Basic function to execute NEGEMMLowpQuantizeDownInt32ToInt16ScaleByFixedPoint on Neon. +/** Basic function to execute NEGEMMLowpQuantizeDownInt32ToInt16ScaleByFixedPoint. * * NEGEMMLowpQuantizeDownInt32ToInt16ScaleByFixedPoint depends on 2 parameters: * @@ -205,7 +205,7 @@ public: * * ((FixedPointMul(input[i][k] + bias[k], result_fixedpoint_multiplier)) >> result_shift) + result_offset_after_shift * - * This function calls the following Neon kernels: + * This function calls the following kernels: * * -# @ref NEGEMMLowpQuantizeDownInt32ToInt16ScaleByFixedPointKernel * @@ -256,9 +256,9 @@ public: static Status validate(const ITensorInfo *input, const ITensorInfo *bias, const ITensorInfo *output, int min = std::numeric_limits::lowest(), int max = std::numeric_limits::max()); }; -/** Basic function to execute GEMMLowpQuantizeDown kernels on Neon. +/** Basic function to execute GEMMLowpQuantizeDown kernels. * - * This function calls the following Neon kernels: + * This function calls the following kernels: * * -# @ref NEGEMMLowpQuantizeDownInt32ScaleKernel * -# @ref NEGEMMLowpQuantizeDownInt32ToUint8ScaleByFixedPointKernel diff --git a/arm_compute/runtime/NEON/functions/NELSTMLayerQuantized.h b/arm_compute/runtime/NEON/functions/NELSTMLayerQuantized.h index a59dcf88cc..53a024ae04 100644 --- a/arm_compute/runtime/NEON/functions/NELSTMLayerQuantized.h +++ b/arm_compute/runtime/NEON/functions/NELSTMLayerQuantized.h @@ -47,7 +47,7 @@ class ITensor; /** Basic function to run @ref NELSTMLayerQuantized * - * This function calls the following Neon functions/kernels: + * This function calls the following functions/kernels: * * -# @ref NEGEMMLowpMatrixMultiplyCore Quantized matrix multiplication core. Accumulators are 32-bit integers * -# @ref NEGEMMLowpQuantizeDownInt32ToInt16ScaleByFixedPoint Convert 32-bit integers into QSYMM16 diff --git a/arm_compute/runtime/NEON/functions/NEMaxUnpoolingLayer.h b/arm_compute/runtime/NEON/functions/NEMaxUnpoolingLayer.h index 7b1f7e9ca1..fae26b3c93 100644 --- a/arm_compute/runtime/NEON/functions/NEMaxUnpoolingLayer.h +++ b/arm_compute/runtime/NEON/functions/NEMaxUnpoolingLayer.h @@ -35,7 +35,7 @@ class ITensorInfo; class NEFill; class NEMaxUnpoolingLayerKernel; -/** Function to perform MaxUnpooling. This function calls the following Neon kernels: +/** Function to perform MaxUnpooling. This function calls the following kernels: * * -# @ref NEFill * -# @ref NEMaxUnpoolingLayerKernel diff --git a/arm_compute/runtime/NEON/functions/NENormalizationLayer.h b/arm_compute/runtime/NEON/functions/NENormalizationLayer.h index fbc2f6f95b..8c4ad1516e 100644 --- a/arm_compute/runtime/NEON/functions/NENormalizationLayer.h +++ b/arm_compute/runtime/NEON/functions/NENormalizationLayer.h @@ -39,7 +39,7 @@ namespace arm_compute class ITensor; class NENormalizationLayerKernel; -/** Basic function to compute a normalization layer. This function calls the following Neon kernels: +/** Basic function to compute a normalization layer. This function calls the following kernels: * * -# @ref NEPixelWiseMultiplication * -# @ref NEFillBorderKernel diff --git a/arm_compute/runtime/NEON/functions/NEPadLayer.h b/arm_compute/runtime/NEON/functions/NEPadLayer.h index 242625604f..76ff0643a0 100644 --- a/arm_compute/runtime/NEON/functions/NEPadLayer.h +++ b/arm_compute/runtime/NEON/functions/NEPadLayer.h @@ -38,7 +38,7 @@ namespace arm_compute { class NEPadLayerKernel; -/** Basic function to pad a tensor. This function calls the following Neon functions/kernels: +/** Basic function to pad a tensor. This function calls the following functions/kernels: * * - For padding mode = PaddingMode::CONSTANT: * -# @ref NEPadLayerKernel diff --git a/arm_compute/runtime/NEON/functions/NEPermute.h b/arm_compute/runtime/NEON/functions/NEPermute.h index fb95e45bdb..2508458a3d 100644 --- a/arm_compute/runtime/NEON/functions/NEPermute.h +++ b/arm_compute/runtime/NEON/functions/NEPermute.h @@ -52,7 +52,7 @@ public: NEPermute &operator=(const NEPermute &) = delete; /** Default move assignment operator */ NEPermute &operator=(NEPermute &&) = default; - /** Configure the permute Neon kernel + /** Configure the permute function * * @note Arbitrary permutation vectors are supported with rank not greater than 4 * diff --git a/arm_compute/runtime/NEON/functions/NEPoolingLayer.h b/arm_compute/runtime/NEON/functions/NEPoolingLayer.h index e374348f98..cb136ebca9 100644 --- a/arm_compute/runtime/NEON/functions/NEPoolingLayer.h +++ b/arm_compute/runtime/NEON/functions/NEPoolingLayer.h @@ -36,7 +36,7 @@ namespace arm_compute class ITensor; class ITensorInfo; -/** Basic function to simulate a pooling layer with the specified pooling operation. This function calls the following Neon kernels: +/** Basic function to simulate a pooling layer with the specified pooling operation. This function calls the following kernels: * * -# @ref NEFillBorderKernel (executed if padding size is different from zero) * -# @ref cpu::kernels::CpuPoolingKernel diff --git a/arm_compute/runtime/NEON/functions/NEQLSTMLayer.h b/arm_compute/runtime/NEON/functions/NEQLSTMLayer.h index 954aceba1a..e706179415 100644 --- a/arm_compute/runtime/NEON/functions/NEQLSTMLayer.h +++ b/arm_compute/runtime/NEON/functions/NEQLSTMLayer.h @@ -47,7 +47,7 @@ class NEGEMMLowpMatrixAReductionKernel; /** Basic function to run @ref NEQLSTMLayer * - * This function calls the following Neon functions/kernels: + * This function calls the following kernels: * * -# @ref NEActivationLayer Activation functions (tanh and logistic) * -# @ref NEArithmeticAddition Elementwise addition diff --git a/arm_compute/runtime/NEON/functions/NEQuantizationLayer.h b/arm_compute/runtime/NEON/functions/NEQuantizationLayer.h index 54ec76b177..9e2d9ecf24 100644 --- a/arm_compute/runtime/NEON/functions/NEQuantizationLayer.h +++ b/arm_compute/runtime/NEON/functions/NEQuantizationLayer.h @@ -35,12 +35,7 @@ namespace arm_compute class ITensor; class ITensorInfo; -/** Basic function to simulate a quantization layer. This function calls the following Arm(R) Neon(TM) implementation layers: - * - * - * -# @ref cpu::CpuQuantization - * - */ +/** Basic function to run a quantization layer using @ref cpu::CpuQuantization */ class NEQuantizationLayer : public IFunction { public: diff --git a/arm_compute/runtime/NEON/functions/NEROIAlignLayer.h b/arm_compute/runtime/NEON/functions/NEROIAlignLayer.h index 9d934588fb..c72cd494d2 100644 --- a/arm_compute/runtime/NEON/functions/NEROIAlignLayer.h +++ b/arm_compute/runtime/NEON/functions/NEROIAlignLayer.h @@ -32,12 +32,7 @@ namespace arm_compute class ITensor; class ITensorInfo; -/** Basic function to run @ref NEROIAlignLayerKernel. - * - * This function calls the following Neon kernels: - * -# @ref NEROIAlignLayerKernel - * - */ +/** Basic function to run @ref NEROIAlignLayerKernel. */ class NEROIAlignLayer : public INESimpleFunctionNoBorder { public: diff --git a/arm_compute/runtime/NEON/functions/NEROIPoolingLayer.h b/arm_compute/runtime/NEON/functions/NEROIPoolingLayer.h index 510c89caf2..214dd43402 100644 --- a/arm_compute/runtime/NEON/functions/NEROIPoolingLayer.h +++ b/arm_compute/runtime/NEON/functions/NEROIPoolingLayer.h @@ -35,12 +35,7 @@ class ITensorInfo; class NEROIPoolingLayerKernel; class ROIPoolingLayerInfo; -/** Basic function to run @ref NEROIPoolingLayerKernel. - * - * This function calls the following Neon kernels: - * -# @ref NEROIPoolingLayerKernel - * - */ +/** Basic function to run @ref NEROIPoolingLayerKernel. */ class NEROIPoolingLayer : public IFunction { public: diff --git a/arm_compute/runtime/NEON/functions/NEReductionOperation.h b/arm_compute/runtime/NEON/functions/NEReductionOperation.h index f30cc810f1..b96b70926c 100644 --- a/arm_compute/runtime/NEON/functions/NEReductionOperation.h +++ b/arm_compute/runtime/NEON/functions/NEReductionOperation.h @@ -35,7 +35,7 @@ namespace arm_compute class ITensor; class NEReductionOperationKernel; -/** Basic function to simulate a reduction operation. This function calls the following Neon kernels: +/** Basic function to simulate a reduction operation. This function calls the following kernels: * * -# @ref NEReshapeLayer * -# @ref NEReductionOperationKernel diff --git a/arm_compute/runtime/NEON/functions/NERemap.h b/arm_compute/runtime/NEON/functions/NERemap.h index 84d0f2ee92..835ebfab7e 100644 --- a/arm_compute/runtime/NEON/functions/NERemap.h +++ b/arm_compute/runtime/NEON/functions/NERemap.h @@ -34,13 +34,10 @@ namespace arm_compute { class ITensor; -/** Basic function to execute remap. This function calls the following Neon kernels: +/** Basic function to execute remap. This function calls the following kernels: * * -# @ref NEFillBorderKernel (executed if border_mode == CONSTANT or border_mode == REPLICATE) * -# @ref NERemapKernel - * - * @deprecated This function is deprecated and is intended to be removed in 21.05 release - * */ class NERemap : public INESimpleFunction { diff --git a/arm_compute/runtime/NEON/functions/NESpaceToBatchLayer.h b/arm_compute/runtime/NEON/functions/NESpaceToBatchLayer.h index aeeaefcc38..27c1ddf8e3 100644 --- a/arm_compute/runtime/NEON/functions/NESpaceToBatchLayer.h +++ b/arm_compute/runtime/NEON/functions/NESpaceToBatchLayer.h @@ -36,7 +36,7 @@ class ITensorInfo; class NESpaceToBatchLayerKernel; class NEFill; -/** Basic function to spatial divide a tensor. This function calls the following Neon kernels/functions: +/** Basic function to spatial divide a tensor. This function calls the following kernels/functions: * * -# @ref NEFill * -# @ref NESpaceToBatchLayerKernel diff --git a/arm_compute/runtime/NEON/functions/NESpaceToDepthLayer.h b/arm_compute/runtime/NEON/functions/NESpaceToDepthLayer.h index d76fc48204..73c228d8ee 100644 --- a/arm_compute/runtime/NEON/functions/NESpaceToDepthLayer.h +++ b/arm_compute/runtime/NEON/functions/NESpaceToDepthLayer.h @@ -35,10 +35,7 @@ class ITensor; class ITensorInfo; class NESpaceToDepthLayerKernel; -/** This function calls the following Neon kernels/functions: - * - * -# @ref NESpaceToDepthLayerKernel - */ +/** Basic function to run @ref NESpaceToDepthLayerKernel. */ class NESpaceToDepthLayer : public IFunction { public: diff --git a/arm_compute/runtime/NEON/functions/NEWinogradConvolutionLayer.h b/arm_compute/runtime/NEON/functions/NEWinogradConvolutionLayer.h index e41cdbd0ac..befc373646 100644 --- a/arm_compute/runtime/NEON/functions/NEWinogradConvolutionLayer.h +++ b/arm_compute/runtime/NEON/functions/NEWinogradConvolutionLayer.h @@ -42,7 +42,8 @@ namespace arm_compute class ITensor; class ICPPKernel; -/** Basic function to simulate a convolution layer. This function calls the following Neon kernels: +/** Basic function to simulate a convolution layer. This function calls the following kernels: + * * -# @ref NEWinogradLayerTransformWeightsKernel (executed only once in the first call to the run() method ) * -# @ref NEWinogradLayerTransformInputKernel * -# @ref NEWinogradLayerTransformOutputKernel -- cgit v1.2.1