diff options
Diffstat (limited to 'arm_compute/runtime/NEON')
34 files changed, 48 insertions, 73 deletions
diff --git a/arm_compute/runtime/NEON/INEOperator.h b/arm_compute/runtime/NEON/INEOperator.h index 184a5959b4..5637d831a3 100644 --- a/arm_compute/runtime/NEON/INEOperator.h +++ b/arm_compute/runtime/NEON/INEOperator.h @@ -39,7 +39,7 @@ class Window; using INEKernel = ICPPKernel; namespace experimental { -/** Basic interface for functions which have a single async Neon kernel */ +/** Basic interface for functions which have a single async CPU kernel */ class INEOperator : public IOperator { public: diff --git a/arm_compute/runtime/NEON/INESimpleFunction.h b/arm_compute/runtime/NEON/INESimpleFunction.h index 8c7cf6512c..7512759bd0 100644 --- a/arm_compute/runtime/NEON/INESimpleFunction.h +++ b/arm_compute/runtime/NEON/INESimpleFunction.h @@ -33,7 +33,7 @@ namespace arm_compute class ICPPKernel; class NEFillBorderKernel; using INEKernel = ICPPKernel; -/** Basic interface for functions which have a single Neon kernel */ +/** Basic interface for functions which have a single CPU kernel */ class INESimpleFunction : public IFunction { public: diff --git a/arm_compute/runtime/NEON/INESimpleFunctionNoBorder.h b/arm_compute/runtime/NEON/INESimpleFunctionNoBorder.h index 52bd5f333b..dc4bac17e4 100644 --- a/arm_compute/runtime/NEON/INESimpleFunctionNoBorder.h +++ b/arm_compute/runtime/NEON/INESimpleFunctionNoBorder.h @@ -33,7 +33,7 @@ namespace arm_compute { class ICPPKernel; using INEKernel = ICPPKernel; -/** Basic interface for functions which have a single Neon kernel and no border */ +/** Basic interface for functions which have a single CPU kernel and no border */ class INESimpleFunctionNoBorder : public IFunction { public: diff --git a/arm_compute/runtime/NEON/NEScheduler.h b/arm_compute/runtime/NEON/NEScheduler.h index 542142a30a..a3082d00f6 100644 --- a/arm_compute/runtime/NEON/NEScheduler.h +++ b/arm_compute/runtime/NEON/NEScheduler.h @@ -28,7 +28,7 @@ namespace arm_compute { -/** Neon Scheduler */ +/** CPU Scheduler */ using NEScheduler = Scheduler; } #endif /*ARM_COMPUTE_NESCHEDULER_H */ diff --git a/arm_compute/runtime/NEON/functions/NEArgMinMaxLayer.h b/arm_compute/runtime/NEON/functions/NEArgMinMaxLayer.h index 8235185a8e..cbf1d5b444 100644 --- a/arm_compute/runtime/NEON/functions/NEArgMinMaxLayer.h +++ b/arm_compute/runtime/NEON/functions/NEArgMinMaxLayer.h @@ -37,7 +37,7 @@ class ITensor; /** Function to calculate the index of the minimum or maximum values in a * tensor based on an axis. * - * This function calls the following Neon kernels: + * This function calls the following kernels: * * -# @ref NEReductionOperationKernel * -# @ref NEFillBorderKernel diff --git a/arm_compute/runtime/NEON/functions/NEBoundingBoxTransform.h b/arm_compute/runtime/NEON/functions/NEBoundingBoxTransform.h index de8dfef4ed..c377520a12 100644 --- a/arm_compute/runtime/NEON/functions/NEBoundingBoxTransform.h +++ b/arm_compute/runtime/NEON/functions/NEBoundingBoxTransform.h @@ -1,5 +1,5 @@ /* - * Copyright (c) 2019-2020 Arm Limited. + * Copyright (c) 2019-2021 Arm Limited. * * SPDX-License-Identifier: MIT * @@ -32,11 +32,7 @@ namespace arm_compute class ITensor; class ITensorInfo; -/** Basic function to run @ref NEBoundingBoxTransformKernel. - * - * This function calls the following Neon kernels: - * -# @ref NEBoundingBoxTransformKernel - */ +/** Basic function to run @ref NEBoundingBoxTransformKernel. */ class NEBoundingBoxTransform : public INESimpleFunctionNoBorder { public: diff --git a/arm_compute/runtime/NEON/functions/NEConvolutionLayer.h b/arm_compute/runtime/NEON/functions/NEConvolutionLayer.h index a387255b6c..d2d41c1e8a 100644 --- a/arm_compute/runtime/NEON/functions/NEConvolutionLayer.h +++ b/arm_compute/runtime/NEON/functions/NEConvolutionLayer.h @@ -37,7 +37,7 @@ namespace arm_compute // Forward declarations class ITensor; -/** Basic function to simulate a convolution layer. This function calls one of the following Neon functions: +/** Basic function to simulate a convolution layer. This function calls one of the following functions: * -# @ref NEGEMMConvolutionLayer (executed only in case GEMM is required for the operation) * -# @ref NEWinogradConvolutionLayer (executed only in case Winograd is required for the operation) * -# @ref NEDirectConvolutionLayer (executed only in case Direct Convolution is required for the operation) diff --git a/arm_compute/runtime/NEON/functions/NEDeconvolutionLayer.h b/arm_compute/runtime/NEON/functions/NEDeconvolutionLayer.h index 02a0f784ec..3864a663c2 100644 --- a/arm_compute/runtime/NEON/functions/NEDeconvolutionLayer.h +++ b/arm_compute/runtime/NEON/functions/NEDeconvolutionLayer.h @@ -64,7 +64,7 @@ namespace arm_compute * The weights used by Deconvolution are supposed to be the same as the ones used for Convolution. Therefore, it will be necessary to use the weights in the * reverse order to perform an actual convolution. This is achieved by using @ref NEReverse. * - * This function calls the following Neon kernels/functions: + * This function calls the following kernels/functions: * * -# @ref CPPUpsample * -# @ref NEConvolutionLayer diff --git a/arm_compute/runtime/NEON/functions/NEDepthwiseConvolutionLayer.h b/arm_compute/runtime/NEON/functions/NEDepthwiseConvolutionLayer.h index 98fffe0b33..9aa8f04eb8 100644 --- a/arm_compute/runtime/NEON/functions/NEDepthwiseConvolutionLayer.h +++ b/arm_compute/runtime/NEON/functions/NEDepthwiseConvolutionLayer.h @@ -110,7 +110,7 @@ private: const PadStrideInfo &conv_info, unsigned int depth_multiplier = 1, ActivationLayerInfo act_info = ActivationLayerInfo(), const Size2D &dilation = Size2D(1U, 1U)); - /** Basic function to execute optimized depthwise convolution routines. This function calls the following Neon kernels: + /** Basic function to execute optimized depthwise convolution routines. This function calls the following kernels: * * @note At the moment 3x3 and 5x5 convolution of stride 1, 2 are supported * @@ -192,7 +192,7 @@ private: bool _is_prepared; }; - /** Basic function to execute a generic depthwise convolution. This function calls the following Neon kernel: + /** Basic function to execute a generic depthwise convolution. This function calls the following kernel: * * -# @ref NEDepthwiseConvolutionLayerNativeKernel * diff --git a/arm_compute/runtime/NEON/functions/NEDirectConvolutionLayer.h b/arm_compute/runtime/NEON/functions/NEDirectConvolutionLayer.h index ff0c3054d8..86914fa0bc 100644 --- a/arm_compute/runtime/NEON/functions/NEDirectConvolutionLayer.h +++ b/arm_compute/runtime/NEON/functions/NEDirectConvolutionLayer.h @@ -41,7 +41,7 @@ class NEFillBorderKernel; /** Function to run the direct convolution. * - * This function calls the following Neon kernels: + * This function calls the following kernels: * * -# @ref NEFillBorderKernel for the input * -# @ref NEDirectConvolutionLayerOutputStageKernel diff --git a/arm_compute/runtime/NEON/functions/NEFFT1D.h b/arm_compute/runtime/NEON/functions/NEFFT1D.h index 04e8f81b69..a533aa7f48 100644 --- a/arm_compute/runtime/NEON/functions/NEFFT1D.h +++ b/arm_compute/runtime/NEON/functions/NEFFT1D.h @@ -40,7 +40,7 @@ class NEFFTDigitReverseKernel; class NEFFTRadixStageKernel; class NEFFTScaleKernel; -/** Basic function to execute one dimensional FFT. This function calls the following Neon kernels: +/** Basic function to execute one dimensional FFT. This function calls the following kernels: * * -# @ref NEFFTDigitReverseKernel Performs digit reverse * -# @ref NEFFTRadixStageKernel A list of FFT kernels depending on the radix decomposition diff --git a/arm_compute/runtime/NEON/functions/NEFFT2D.h b/arm_compute/runtime/NEON/functions/NEFFT2D.h index 218401b429..ce84a85105 100644 --- a/arm_compute/runtime/NEON/functions/NEFFT2D.h +++ b/arm_compute/runtime/NEON/functions/NEFFT2D.h @@ -36,7 +36,7 @@ namespace arm_compute // Forward declaration class ITensor; -/** Basic function to execute two dimensional FFT. This function calls the following Neon kernels: +/** Basic function to execute two dimensional FFT. This function calls the following kernels: * * -# @ref NEFFT1D 1D FFT is performed on the first given axis * -# @ref NEFFT1D 1D FFT is performed on the second given axis diff --git a/arm_compute/runtime/NEON/functions/NEFFTConvolutionLayer.h b/arm_compute/runtime/NEON/functions/NEFFTConvolutionLayer.h index 8967363e7f..213fa6093b 100644 --- a/arm_compute/runtime/NEON/functions/NEFFTConvolutionLayer.h +++ b/arm_compute/runtime/NEON/functions/NEFFTConvolutionLayer.h @@ -43,7 +43,7 @@ namespace arm_compute // Forward declarations class ITensor; -/** Basic function to execute FFT-based convolution on Neon. This function calls the following Neon functions/kernels: +/** Basic function to execute FFT-based convolution on CPU. This function calls the following functions/kernels: * * -# @ref NEPermute Permute input if NHWC(only NCHW is supported). * -# @ref NEPadLayer Pad input. @@ -84,7 +84,7 @@ public: * Data types supported: Same as @p input. * @param[in] conv_info Contains padding and stride information described in @ref PadStrideInfo. * @param[in] act_info (Optional) Activation layer information in case of a fused activation. - * @param[in] enable_fast_math (Optional) Enable fast math computation. Unused for Neon backend. + * @param[in] enable_fast_math (Optional) Enable fast math computation. Unused for CPU backend. */ void configure(ITensor *input, const ITensor *weights, const ITensor *biases, ITensor *output, const PadStrideInfo &conv_info, const ActivationLayerInfo &act_info = ActivationLayerInfo(), bool enable_fast_math = false); @@ -101,7 +101,7 @@ public: * Data types supported: Same as @p input. * @param[in] conv_info Contains padding and stride information described in @ref PadStrideInfo. * @param[in] act_info (Optional) Activation layer information in case of a fused activation. - * @param[in] enable_fast_math (Optional) Enable fast math computation. Unused for Neon backend. + * @param[in] enable_fast_math (Optional) Enable fast math computation. Unused for CPU backend. * * @return a status */ diff --git a/arm_compute/runtime/NEON/functions/NEFullyConnectedLayer.h b/arm_compute/runtime/NEON/functions/NEFullyConnectedLayer.h index 1b3f36d866..58b11744a8 100644 --- a/arm_compute/runtime/NEON/functions/NEFullyConnectedLayer.h +++ b/arm_compute/runtime/NEON/functions/NEFullyConnectedLayer.h @@ -35,7 +35,7 @@ namespace arm_compute { -/** Basic function to reshape the weights of Fully Connected layer with Neon. This function calls the following kernels: +/** Basic function to reshape the weights of Fully Connected layer. This function calls the following kernels: * * @note The fully connected layer accepts "weights" tensors only with 2 dimensions. */ @@ -117,7 +117,7 @@ private: }; } // namespace weights_transformations -/** Basic function to compute a Fully Connected layer on Neon. This function calls the following Neon kernels: +/** Basic function to compute a Fully Connected layer. This function calls the following kernels: * -# @ref NEIm2ColKernel (called when the input comes from a convolutional layer) * -# @ref NEFullyConnectedLayerReshapeWeights (if @p are_weights_reshaped is set to false and transpose_weights is set to true ) (called once) * -# @ref NEGEMMMatrixMultiplyKernel or @ref NEGEMMLowpMatrixMultiplyCore (if quantized asymmetric) diff --git a/arm_compute/runtime/NEON/functions/NEGEMM.h b/arm_compute/runtime/NEON/functions/NEGEMM.h index 6f7951eece..a6c3436656 100644 --- a/arm_compute/runtime/NEON/functions/NEGEMM.h +++ b/arm_compute/runtime/NEON/functions/NEGEMM.h @@ -43,7 +43,7 @@ class NEGEMMMatrixMultiplyKernel; class NEGEMMTranspose1xWKernel; class NEGEMMAssemblyDispatch; -/** Basic function to execute GEMM on Neon. This function calls the following Neon kernels: +/** Basic function to execute GEMM. This function calls the following kernels: * * If optimized assembly is available: * -# @ref NEGEMMAssemblyDispatch diff --git a/arm_compute/runtime/NEON/functions/NEGEMMConv2d.h b/arm_compute/runtime/NEON/functions/NEGEMMConv2d.h index 2bd233f520..8c3ba4f0c8 100644 --- a/arm_compute/runtime/NEON/functions/NEGEMMConv2d.h +++ b/arm_compute/runtime/NEON/functions/NEGEMMConv2d.h @@ -38,7 +38,7 @@ namespace arm_compute class ITensor; class NEGEMMAssemblyDispatch; -/** Basic function to compute the convolution layer. This function calls the following Neon kernels/functions: +/** Basic function to compute the convolution layer. This function calls the following kernels/functions: * * Supports only NHWC data layout * diff --git a/arm_compute/runtime/NEON/functions/NEGEMMConvolutionLayer.h b/arm_compute/runtime/NEON/functions/NEGEMMConvolutionLayer.h index 33f00c087c..9897bf1d4d 100644 --- a/arm_compute/runtime/NEON/functions/NEGEMMConvolutionLayer.h +++ b/arm_compute/runtime/NEON/functions/NEGEMMConvolutionLayer.h @@ -150,7 +150,7 @@ private: }; } // namespace weights_transformations -/** Basic function to compute the convolution layer. This function calls the following Neon kernels/functions: +/** Basic function to compute the convolution layer. This function calls the following kernels/functions: * * -# @ref NEIm2ColKernel * -# @ref NEGEMM (if the data type is BFLOAT16/FP16/FP32) diff --git a/arm_compute/runtime/NEON/functions/NEGEMMLowpMatrixMultiplyCore.h b/arm_compute/runtime/NEON/functions/NEGEMMLowpMatrixMultiplyCore.h index 821b498dad..b2b77bd406 100644 --- a/arm_compute/runtime/NEON/functions/NEGEMMLowpMatrixMultiplyCore.h +++ b/arm_compute/runtime/NEON/functions/NEGEMMLowpMatrixMultiplyCore.h @@ -37,7 +37,6 @@ namespace arm_compute { class ITensor; class NEConvertQuantizedSignednessKernel; -class NEConvertQuantizedSignednessKernel; class NEGEMMInterleave4x4Kernel; class NEGEMMLowpMatrixMultiplyKernel; class NEGEMMLowpOffsetContributionKernel; @@ -47,7 +46,7 @@ class NEGEMMLowpMatrixBReductionKernel; class NEGEMMTranspose1xWKernel; class NEGEMMAssemblyDispatch; -/** Basic function to execute GEMMLowpMatrixMultiplyCore on Neon. This function calls the following Neon kernels if the DOT product instruction is not available: +/** Basic function to execute GEMMLowpMatrixMultiplyCore. This function calls the following kernels if the DOT product instruction is not available: * * -# @ref NEGEMMInterleave4x4Kernel * -# @ref NEGEMMTranspose1xWKernel diff --git a/arm_compute/runtime/NEON/functions/NEGEMMLowpOutputStage.h b/arm_compute/runtime/NEON/functions/NEGEMMLowpOutputStage.h index 79b427ea6f..c22ed1b5c4 100644 --- a/arm_compute/runtime/NEON/functions/NEGEMMLowpOutputStage.h +++ b/arm_compute/runtime/NEON/functions/NEGEMMLowpOutputStage.h @@ -27,7 +27,7 @@ #include "arm_compute/core/Types.h" #include "arm_compute/runtime/NEON/INESimpleFunctionNoBorder.h" -/** This file contains all available output stages for GEMMLowp on Neon. +/** This file contains all available output stages for GEMMLowp. * * In gemmlowp, the "output stage" is the process that takes a final int32 accumulator value (the output of @ref NEGEMMLowpMatrixMultiplyCore), * and processes it to obtain the final ASYMM8 value. @@ -40,7 +40,7 @@ namespace arm_compute class ITensor; class ITensorInfo; -/** Basic function to execute NEGEMMLowpQuantizeDownInt32ToUint8ScaleByFixedPoint on Neon. +/** Basic function to execute NEGEMMLowpQuantizeDownInt32ToUint8ScaleByFixedPoint. * * NEGEMMLowpQuantizeDownInt32ToUint8ScaleByFixedPoint depends on 3 parameters: * @@ -61,7 +61,7 @@ class ITensorInfo; * * ((FixedPointMul(input[i][k] + bias[k], result_fixedpoint_multiplier)) >> result_shift) + result_offset_after_shift * - * This function calls the following Neon kernels: + * This function calls the following kernels: * * -# @ref NEGEMMLowpQuantizeDownInt32ToUint8ScaleByFixedPointKernel * @@ -112,7 +112,7 @@ public: */ static Status validate(const ITensorInfo *input, const ITensorInfo *bias, const ITensorInfo *output, int min = std::numeric_limits<int32_t>::lowest(), int max = std::numeric_limits<int32_t>::max()); }; -/** Basic function to execute NEGEMMLowpQuantizeDownInt32ToInt8ScaleByFixedPoint on Neon. +/** Basic function to execute NEGEMMLowpQuantizeDownInt32ToInt8ScaleByFixedPoint. * * NEGEMMLowpQuantizeDownInt32ToInt8ScaleByFixedPoint depends on 3 parameters: * @@ -133,7 +133,7 @@ public: * * ((FixedPointMul(input[i][k] + bias[k], result_fixedpoint_multiplier)) >> result_shift) + result_offset_after_shift * - * This function calls the following Neon kernels: + * This function calls the following kernels: * * -# @ref NEGEMMLowpQuantizeDownInt32ToInt8ScaleByFixedPointKernel * @@ -184,7 +184,7 @@ public: */ static Status validate(const ITensorInfo *input, const ITensorInfo *bias, const ITensorInfo *output, int min = std::numeric_limits<int32_t>::lowest(), int max = std::numeric_limits<int32_t>::max()); }; -/** Basic function to execute NEGEMMLowpQuantizeDownInt32ToInt16ScaleByFixedPoint on Neon. +/** Basic function to execute NEGEMMLowpQuantizeDownInt32ToInt16ScaleByFixedPoint. * * NEGEMMLowpQuantizeDownInt32ToInt16ScaleByFixedPoint depends on 2 parameters: * @@ -205,7 +205,7 @@ public: * * ((FixedPointMul(input[i][k] + bias[k], result_fixedpoint_multiplier)) >> result_shift) + result_offset_after_shift * - * This function calls the following Neon kernels: + * This function calls the following kernels: * * -# @ref NEGEMMLowpQuantizeDownInt32ToInt16ScaleByFixedPointKernel * @@ -256,9 +256,9 @@ public: static Status validate(const ITensorInfo *input, const ITensorInfo *bias, const ITensorInfo *output, int min = std::numeric_limits<int32_t>::lowest(), int max = std::numeric_limits<int32_t>::max()); }; -/** Basic function to execute GEMMLowpQuantizeDown kernels on Neon. +/** Basic function to execute GEMMLowpQuantizeDown kernels. * - * This function calls the following Neon kernels: + * This function calls the following kernels: * * -# @ref NEGEMMLowpQuantizeDownInt32ScaleKernel * -# @ref NEGEMMLowpQuantizeDownInt32ToUint8ScaleByFixedPointKernel diff --git a/arm_compute/runtime/NEON/functions/NELSTMLayerQuantized.h b/arm_compute/runtime/NEON/functions/NELSTMLayerQuantized.h index a59dcf88cc..53a024ae04 100644 --- a/arm_compute/runtime/NEON/functions/NELSTMLayerQuantized.h +++ b/arm_compute/runtime/NEON/functions/NELSTMLayerQuantized.h @@ -47,7 +47,7 @@ class ITensor; /** Basic function to run @ref NELSTMLayerQuantized * - * This function calls the following Neon functions/kernels: + * This function calls the following functions/kernels: * * -# @ref NEGEMMLowpMatrixMultiplyCore Quantized matrix multiplication core. Accumulators are 32-bit integers * -# @ref NEGEMMLowpQuantizeDownInt32ToInt16ScaleByFixedPoint Convert 32-bit integers into QSYMM16 diff --git a/arm_compute/runtime/NEON/functions/NEMaxUnpoolingLayer.h b/arm_compute/runtime/NEON/functions/NEMaxUnpoolingLayer.h index 7b1f7e9ca1..fae26b3c93 100644 --- a/arm_compute/runtime/NEON/functions/NEMaxUnpoolingLayer.h +++ b/arm_compute/runtime/NEON/functions/NEMaxUnpoolingLayer.h @@ -35,7 +35,7 @@ class ITensorInfo; class NEFill; class NEMaxUnpoolingLayerKernel; -/** Function to perform MaxUnpooling. This function calls the following Neon kernels: +/** Function to perform MaxUnpooling. This function calls the following kernels: * * -# @ref NEFill * -# @ref NEMaxUnpoolingLayerKernel diff --git a/arm_compute/runtime/NEON/functions/NENormalizationLayer.h b/arm_compute/runtime/NEON/functions/NENormalizationLayer.h index fbc2f6f95b..8c4ad1516e 100644 --- a/arm_compute/runtime/NEON/functions/NENormalizationLayer.h +++ b/arm_compute/runtime/NEON/functions/NENormalizationLayer.h @@ -39,7 +39,7 @@ namespace arm_compute class ITensor; class NENormalizationLayerKernel; -/** Basic function to compute a normalization layer. This function calls the following Neon kernels: +/** Basic function to compute a normalization layer. This function calls the following kernels: * * -# @ref NEPixelWiseMultiplication * -# @ref NEFillBorderKernel diff --git a/arm_compute/runtime/NEON/functions/NEPadLayer.h b/arm_compute/runtime/NEON/functions/NEPadLayer.h index 242625604f..76ff0643a0 100644 --- a/arm_compute/runtime/NEON/functions/NEPadLayer.h +++ b/arm_compute/runtime/NEON/functions/NEPadLayer.h @@ -38,7 +38,7 @@ namespace arm_compute { class NEPadLayerKernel; -/** Basic function to pad a tensor. This function calls the following Neon functions/kernels: +/** Basic function to pad a tensor. This function calls the following functions/kernels: * * - For padding mode = PaddingMode::CONSTANT: * -# @ref NEPadLayerKernel diff --git a/arm_compute/runtime/NEON/functions/NEPermute.h b/arm_compute/runtime/NEON/functions/NEPermute.h index fb95e45bdb..2508458a3d 100644 --- a/arm_compute/runtime/NEON/functions/NEPermute.h +++ b/arm_compute/runtime/NEON/functions/NEPermute.h @@ -52,7 +52,7 @@ public: NEPermute &operator=(const NEPermute &) = delete; /** Default move assignment operator */ NEPermute &operator=(NEPermute &&) = default; - /** Configure the permute Neon kernel + /** Configure the permute function * * @note Arbitrary permutation vectors are supported with rank not greater than 4 * diff --git a/arm_compute/runtime/NEON/functions/NEPoolingLayer.h b/arm_compute/runtime/NEON/functions/NEPoolingLayer.h index e374348f98..cb136ebca9 100644 --- a/arm_compute/runtime/NEON/functions/NEPoolingLayer.h +++ b/arm_compute/runtime/NEON/functions/NEPoolingLayer.h @@ -36,7 +36,7 @@ namespace arm_compute class ITensor; class ITensorInfo; -/** Basic function to simulate a pooling layer with the specified pooling operation. This function calls the following Neon kernels: +/** Basic function to simulate a pooling layer with the specified pooling operation. This function calls the following kernels: * * -# @ref NEFillBorderKernel (executed if padding size is different from zero) * -# @ref cpu::kernels::CpuPoolingKernel diff --git a/arm_compute/runtime/NEON/functions/NEQLSTMLayer.h b/arm_compute/runtime/NEON/functions/NEQLSTMLayer.h index 954aceba1a..e706179415 100644 --- a/arm_compute/runtime/NEON/functions/NEQLSTMLayer.h +++ b/arm_compute/runtime/NEON/functions/NEQLSTMLayer.h @@ -47,7 +47,7 @@ class NEGEMMLowpMatrixAReductionKernel; /** Basic function to run @ref NEQLSTMLayer * - * This function calls the following Neon functions/kernels: + * This function calls the following kernels: * * -# @ref NEActivationLayer Activation functions (tanh and logistic) * -# @ref NEArithmeticAddition Elementwise addition diff --git a/arm_compute/runtime/NEON/functions/NEQuantizationLayer.h b/arm_compute/runtime/NEON/functions/NEQuantizationLayer.h index 54ec76b177..9e2d9ecf24 100644 --- a/arm_compute/runtime/NEON/functions/NEQuantizationLayer.h +++ b/arm_compute/runtime/NEON/functions/NEQuantizationLayer.h @@ -35,12 +35,7 @@ namespace arm_compute class ITensor; class ITensorInfo; -/** Basic function to simulate a quantization layer. This function calls the following Arm(R) Neon(TM) implementation layers: - * - * - * -# @ref cpu::CpuQuantization - * - */ +/** Basic function to run a quantization layer using @ref cpu::CpuQuantization */ class NEQuantizationLayer : public IFunction { public: diff --git a/arm_compute/runtime/NEON/functions/NEROIAlignLayer.h b/arm_compute/runtime/NEON/functions/NEROIAlignLayer.h index 9d934588fb..c72cd494d2 100644 --- a/arm_compute/runtime/NEON/functions/NEROIAlignLayer.h +++ b/arm_compute/runtime/NEON/functions/NEROIAlignLayer.h @@ -32,12 +32,7 @@ namespace arm_compute class ITensor; class ITensorInfo; -/** Basic function to run @ref NEROIAlignLayerKernel. - * - * This function calls the following Neon kernels: - * -# @ref NEROIAlignLayerKernel - * - */ +/** Basic function to run @ref NEROIAlignLayerKernel. */ class NEROIAlignLayer : public INESimpleFunctionNoBorder { public: diff --git a/arm_compute/runtime/NEON/functions/NEROIPoolingLayer.h b/arm_compute/runtime/NEON/functions/NEROIPoolingLayer.h index 510c89caf2..214dd43402 100644 --- a/arm_compute/runtime/NEON/functions/NEROIPoolingLayer.h +++ b/arm_compute/runtime/NEON/functions/NEROIPoolingLayer.h @@ -35,12 +35,7 @@ class ITensorInfo; class NEROIPoolingLayerKernel; class ROIPoolingLayerInfo; -/** Basic function to run @ref NEROIPoolingLayerKernel. - * - * This function calls the following Neon kernels: - * -# @ref NEROIPoolingLayerKernel - * - */ +/** Basic function to run @ref NEROIPoolingLayerKernel. */ class NEROIPoolingLayer : public IFunction { public: diff --git a/arm_compute/runtime/NEON/functions/NEReductionOperation.h b/arm_compute/runtime/NEON/functions/NEReductionOperation.h index f30cc810f1..b96b70926c 100644 --- a/arm_compute/runtime/NEON/functions/NEReductionOperation.h +++ b/arm_compute/runtime/NEON/functions/NEReductionOperation.h @@ -35,7 +35,7 @@ namespace arm_compute class ITensor; class NEReductionOperationKernel; -/** Basic function to simulate a reduction operation. This function calls the following Neon kernels: +/** Basic function to simulate a reduction operation. This function calls the following kernels: * * -# @ref NEReshapeLayer * -# @ref NEReductionOperationKernel diff --git a/arm_compute/runtime/NEON/functions/NERemap.h b/arm_compute/runtime/NEON/functions/NERemap.h index 84d0f2ee92..835ebfab7e 100644 --- a/arm_compute/runtime/NEON/functions/NERemap.h +++ b/arm_compute/runtime/NEON/functions/NERemap.h @@ -34,13 +34,10 @@ namespace arm_compute { class ITensor; -/** Basic function to execute remap. This function calls the following Neon kernels: +/** Basic function to execute remap. This function calls the following kernels: * * -# @ref NEFillBorderKernel (executed if border_mode == CONSTANT or border_mode == REPLICATE) * -# @ref NERemapKernel - * - * @deprecated This function is deprecated and is intended to be removed in 21.05 release - * */ class NERemap : public INESimpleFunction { diff --git a/arm_compute/runtime/NEON/functions/NESpaceToBatchLayer.h b/arm_compute/runtime/NEON/functions/NESpaceToBatchLayer.h index aeeaefcc38..27c1ddf8e3 100644 --- a/arm_compute/runtime/NEON/functions/NESpaceToBatchLayer.h +++ b/arm_compute/runtime/NEON/functions/NESpaceToBatchLayer.h @@ -36,7 +36,7 @@ class ITensorInfo; class NESpaceToBatchLayerKernel; class NEFill; -/** Basic function to spatial divide a tensor. This function calls the following Neon kernels/functions: +/** Basic function to spatial divide a tensor. This function calls the following kernels/functions: * * -# @ref NEFill * -# @ref NESpaceToBatchLayerKernel diff --git a/arm_compute/runtime/NEON/functions/NESpaceToDepthLayer.h b/arm_compute/runtime/NEON/functions/NESpaceToDepthLayer.h index d76fc48204..73c228d8ee 100644 --- a/arm_compute/runtime/NEON/functions/NESpaceToDepthLayer.h +++ b/arm_compute/runtime/NEON/functions/NESpaceToDepthLayer.h @@ -35,10 +35,7 @@ class ITensor; class ITensorInfo; class NESpaceToDepthLayerKernel; -/** This function calls the following Neon kernels/functions: - * - * -# @ref NESpaceToDepthLayerKernel - */ +/** Basic function to run @ref NESpaceToDepthLayerKernel. */ class NESpaceToDepthLayer : public IFunction { public: diff --git a/arm_compute/runtime/NEON/functions/NEWinogradConvolutionLayer.h b/arm_compute/runtime/NEON/functions/NEWinogradConvolutionLayer.h index e41cdbd0ac..befc373646 100644 --- a/arm_compute/runtime/NEON/functions/NEWinogradConvolutionLayer.h +++ b/arm_compute/runtime/NEON/functions/NEWinogradConvolutionLayer.h @@ -42,7 +42,8 @@ namespace arm_compute class ITensor; class ICPPKernel; -/** Basic function to simulate a convolution layer. This function calls the following Neon kernels: +/** Basic function to simulate a convolution layer. This function calls the following kernels: + * * -# @ref NEWinogradLayerTransformWeightsKernel (executed only once in the first call to the run() method ) * -# @ref NEWinogradLayerTransformInputKernel * -# @ref NEWinogradLayerTransformOutputKernel |