From f22f67298d1e7bbc349d5179bceb70971b4bc226 Mon Sep 17 00:00:00 2001 From: Michele Di Giorgio Date: Fri, 3 Jul 2020 16:29:24 +0100 Subject: COMPMID-3532: Align data type support between doxygen and implementation - CPP The patch also removes some unused NEON kernels. Change-Id: I4a7622f31c88ee038b21874614a981764a03122a Signed-off-by: Michele Di Giorgio Reviewed-on: https://review.mlplatform.org/c/ml/ComputeLibrary/+/3509 Tested-by: Arm Jenkins Reviewed-by: Sheri Zhang Reviewed-by: Georgios Pinitas Comments-Addressed: Arm Jenkins --- .../runtime/CPP/functions/CPPBoxWithNonMaximaSuppressionLimit.h | 4 ++-- arm_compute/runtime/CPP/functions/CPPSplit.h | 4 ++-- arm_compute/runtime/CPP/functions/CPPTopKV.h | 4 ++-- arm_compute/runtime/CPP/functions/CPPUpsample.h | 4 ++-- 4 files changed, 8 insertions(+), 8 deletions(-) (limited to 'arm_compute/runtime') diff --git a/arm_compute/runtime/CPP/functions/CPPBoxWithNonMaximaSuppressionLimit.h b/arm_compute/runtime/CPP/functions/CPPBoxWithNonMaximaSuppressionLimit.h index b6f55b5bf8..b1ea84dcda 100644 --- a/arm_compute/runtime/CPP/functions/CPPBoxWithNonMaximaSuppressionLimit.h +++ b/arm_compute/runtime/CPP/functions/CPPBoxWithNonMaximaSuppressionLimit.h @@ -58,7 +58,7 @@ public: * @param[out] classes The classes output tensor of size [N]. Data types supported: Same as @p scores_in * @param[out] batch_splits_out (Optional) The batch splits output tensor. Data types supported: Same as @p scores_in * @param[out] keeps (Optional) The keeps output tensor of size [N]. Data types supported: Same as @p scores_in - * @param[in] keeps_size (Optional) Number of filtered indices per class tensor of size [num_classes]. Data types supported: Same as @p scores_in + * @param[in] keeps_size (Optional) Number of filtered indices per class tensor of size [num_classes]. Data types supported: U32. * @param[in] info (Optional) BoxNMSLimitInfo information. */ void configure(const ITensor *scores_in, const ITensor *boxes_in, const ITensor *batch_splits_in, ITensor *scores_out, ITensor *boxes_out, ITensor *classes, @@ -76,7 +76,7 @@ public: * @param[in] classes The classes output tensor of size [N]. Data types supported: Same as @p scores_in * @param[in] batch_splits_out (Optional) The batch splits output tensor. Data types supported: Same as @p scores_in * @param[in] keeps (Optional) The keeps output tensor of size [N]. Data types supported: Same as @p scores_in - * @param[in] keeps_size (Optional) Number of filtered indices per class tensor of size [num_classes]. Data types supported: Same as @p scores_in + * @param[in] keeps_size (Optional) Number of filtered indices per class tensor of size [num_classes]. Data types supported: U32. * @param[in] info (Optional) BoxNMSLimitInfo information. * * @return a status diff --git a/arm_compute/runtime/CPP/functions/CPPSplit.h b/arm_compute/runtime/CPP/functions/CPPSplit.h index 6adcbc3323..611cc41157 100644 --- a/arm_compute/runtime/CPP/functions/CPPSplit.h +++ b/arm_compute/runtime/CPP/functions/CPPSplit.h @@ -47,8 +47,8 @@ public: } /** Static function to check if given info will lead to a valid configuration of @ref CPPSplit * - * @param[in] input The input tensor info. Data types supported: U8/S8/QASYMM8/U16/S16/U32/S32/F16/F32. - * @param[in] outputs A vector containing the output tensors' info. Data types supported: Same as @p input. + * @param[in] input The input tensor info. Data types supported: All. + * @param[in] outputs A vector containing the output tensors' info. Data types supported: same as @p input. * The output tensors should match the input tensor dimensions for all shape dimensions apart * from the split dimension * @param[in] axis Axis on which to split the input. diff --git a/arm_compute/runtime/CPP/functions/CPPTopKV.h b/arm_compute/runtime/CPP/functions/CPPTopKV.h index c94e277312..fb93bdf764 100644 --- a/arm_compute/runtime/CPP/functions/CPPTopKV.h +++ b/arm_compute/runtime/CPP/functions/CPPTopKV.h @@ -39,7 +39,7 @@ public: /** Set the input and output of the kernel. * * @param[in] predictions A batch_size x classes tensor. Data types supported: F16/S32/F32/QASYMM8/QASYMM8_SIGNED - * @param[in] targets A batch_size 1D tensor of class ids. Data types supported: S32 + * @param[in] targets A batch_size 1D tensor of class ids. Data types supported: U32 * @param[out] output Computed precision at @p k as a bool 1D tensor. Data types supported: U8 * @param[in] k Number of top elements to look at for computing precision. */ @@ -48,7 +48,7 @@ public: /** Static function to check if given info will lead to a valid configuration of @ref CPPTopKVKernel * * @param[in] predictions A batch_size x classes tensor info. Data types supported: F16/S32/F32/QASYMM8/QASYMM8_SIGNED - * @param[in] targets A batch_size 1D tensor info of class ids. Data types supported: S32 + * @param[in] targets A batch_size 1D tensor info of class ids. Data types supported: U32 * @param[in] output Computed precision at @p k as a bool 1D tensor info. Data types supported: U8 * @param[in] k Number of top elements to look at for computing precision. * diff --git a/arm_compute/runtime/CPP/functions/CPPUpsample.h b/arm_compute/runtime/CPP/functions/CPPUpsample.h index 3e1852bc30..080e86b676 100644 --- a/arm_compute/runtime/CPP/functions/CPPUpsample.h +++ b/arm_compute/runtime/CPP/functions/CPPUpsample.h @@ -38,8 +38,8 @@ class CPPUpsample : public ICPPSimpleFunction public: /** Configure the upsample CPP kernel * - * @param[in] input The input tensor to upsample. Data types supported: F32/F16/QASYMM8/QASYMM8_SIGNED - * @param[out] output The output tensor. Data types supported: Same as @p input + * @param[in] input The input tensor to upsample. Data types supported: All. + * @param[out] output The output tensor. Data types supported: same as @p input * @param[in] info Padding information */ void configure(const ITensor *input, ITensor *output, const PadStrideInfo &info); -- cgit v1.2.1