aboutsummaryrefslogtreecommitdiff
path: root/src/core/gpu/cl/kernels
diff options
context:
space:
mode:
authorGeorgios Pinitas <georgios.pinitas@arm.com>2021-07-02 09:01:49 +0100
committerGeorgios Pinitas <georgios.pinitas@arm.com>2021-07-02 15:47:11 +0000
commit2eb5d16b839cbc28c6cb7f0de7a0bf15290b425a (patch)
tree523d495c3a4c07d87b337c45a81afa06c9b1f495 /src/core/gpu/cl/kernels
parent4dfc5538948c196def6d2e3305fe8051a5df3f15 (diff)
downloadComputeLibrary-2eb5d16b839cbc28c6cb7f0de7a0bf15290b425a.tar.gz
Align kernel/operator header layout
- Redirect validate documentation to configure - Align header names - Align class layout Signed-off-by: Georgios Pinitas <georgios.pinitas@arm.com> Change-Id: Ia40f67383826a66e9f9a33745d66805551e31a3a Reviewed-on: https://review.mlplatform.org/c/ml/ComputeLibrary/+/5897 Comments-Addressed: Arm Jenkins <bsgcomp@arm.com> Tested-by: Arm Jenkins <bsgcomp@arm.com> Reviewed-by: Michele Di Giorgio <michele.digiorgio@arm.com>
Diffstat (limited to 'src/core/gpu/cl/kernels')
-rw-r--r--src/core/gpu/cl/kernels/ClActivationKernel.cpp1
-rw-r--r--src/core/gpu/cl/kernels/ClActivationKernel.h9
-rw-r--r--src/core/gpu/cl/kernels/ClBatchConcatenateKernel.cpp1
-rw-r--r--src/core/gpu/cl/kernels/ClBatchConcatenateKernel.h9
-rw-r--r--src/core/gpu/cl/kernels/ClConvertFullyConnectedWeightsKernel.h17
-rw-r--r--src/core/gpu/cl/kernels/ClCopyKernel.h8
-rw-r--r--src/core/gpu/cl/kernels/ClCropKernel.cpp5
-rw-r--r--src/core/gpu/cl/kernels/ClCropKernel.h32
-rw-r--r--src/core/gpu/cl/kernels/ClDepthConcatenateKernel.h7
-rw-r--r--src/core/gpu/cl/kernels/ClDequantizeKernel.h1
-rw-r--r--src/core/gpu/cl/kernels/ClDirectConv2dKernel.h2
-rw-r--r--src/core/gpu/cl/kernels/ClElementwiseKernel.cpp9
-rw-r--r--src/core/gpu/cl/kernels/ClElementwiseKernel.h47
-rw-r--r--src/core/gpu/cl/kernels/ClElementwiseUnaryKernel.h6
-rw-r--r--src/core/gpu/cl/kernels/ClFillKernel.cpp7
-rw-r--r--src/core/gpu/cl/kernels/ClFillKernel.h15
-rw-r--r--src/core/gpu/cl/kernels/ClFloorKernel.h6
-rw-r--r--src/core/gpu/cl/kernels/ClGemmLowpMatrixMultiplyNativeKernel.h2
-rw-r--r--src/core/gpu/cl/kernels/ClGemmLowpMatrixMultiplyReshapedKernel.h2
-rw-r--r--src/core/gpu/cl/kernels/ClGemmMatrixMultiplyNativeKernel.h2
-rw-r--r--src/core/gpu/cl/kernels/ClHeightConcatenateKernel.h13
-rw-r--r--src/core/gpu/cl/kernels/ClMulKernel.h2
-rw-r--r--src/core/gpu/cl/kernels/ClPermuteKernel.h11
-rw-r--r--src/core/gpu/cl/kernels/ClPool2dKernel.cpp1
-rw-r--r--src/core/gpu/cl/kernels/ClPool2dKernel.h9
-rw-r--r--src/core/gpu/cl/kernels/ClQuantizeKernel.h1
-rw-r--r--src/core/gpu/cl/kernels/ClReshapeKernel.h8
-rw-r--r--src/core/gpu/cl/kernels/ClScaleKernel.h16
-rw-r--r--src/core/gpu/cl/kernels/ClSoftmaxKernel.h22
-rw-r--r--src/core/gpu/cl/kernels/ClTransposeKernel.h7
-rw-r--r--src/core/gpu/cl/kernels/ClWidthConcatenate2TensorsKernel.h7
-rw-r--r--src/core/gpu/cl/kernels/ClWidthConcatenate4TensorsKernel.h9
-rw-r--r--src/core/gpu/cl/kernels/ClWidthConcatenateKernel.h13
-rw-r--r--src/core/gpu/cl/kernels/ClWinogradFilterTransformKernel.h3
-rw-r--r--src/core/gpu/cl/kernels/ClWinogradInputTransformKernel.cpp1
-rw-r--r--src/core/gpu/cl/kernels/ClWinogradInputTransformKernel.h13
-rw-r--r--src/core/gpu/cl/kernels/ClWinogradOutputTransformKernel.h4
37 files changed, 91 insertions, 237 deletions
diff --git a/src/core/gpu/cl/kernels/ClActivationKernel.cpp b/src/core/gpu/cl/kernels/ClActivationKernel.cpp
index e892d6a778..21c05632f9 100644
--- a/src/core/gpu/cl/kernels/ClActivationKernel.cpp
+++ b/src/core/gpu/cl/kernels/ClActivationKernel.cpp
@@ -87,7 +87,6 @@ Status validate_arguments(const ITensorInfo *src, const ITensorInfo *dst, const
} // namespace
ClActivationKernel::ClActivationKernel()
- : _run_in_place(false)
{
_type = CLKernelType::ELEMENTWISE;
}
diff --git a/src/core/gpu/cl/kernels/ClActivationKernel.h b/src/core/gpu/cl/kernels/ClActivationKernel.h
index 68c309e9e7..720b16a691 100644
--- a/src/core/gpu/cl/kernels/ClActivationKernel.h
+++ b/src/core/gpu/cl/kernels/ClActivationKernel.h
@@ -51,12 +51,9 @@ public:
* @param[in] act_info Activation layer information.
*/
void configure(const ClCompileContext &compile_context, ITensorInfo *src, ITensorInfo *dst, ActivationLayerInfo act_info);
- /** Static function to check if given info will lead to a valid configuration of @ref ClActivationKernel
+ /** Static function to check if given info will lead to a valid configuration
*
- * @param[in] src Source tensor info. In case of @p dst tensor info = nullptr, this tensor will store the result
- * of the activation function. Data types supported: QASYMM8/QASYMM8_SIGNED/QSYMM16/F16/F32.
- * @param[in] dst Destination tensor info. Data type supported: same as @p src
- * @param[in] act_info Activation layer information.
+ * Similar to @ref ClActivationKernel::configure()
*
* @return a status
*/
@@ -66,7 +63,7 @@ public:
void run_op(ITensorPack &tensors, const Window &window, ::cl::CommandQueue &queue) override;
private:
- bool _run_in_place;
+ bool _run_in_place{ false };
};
} // namespace kernels
} // namespace opencl
diff --git a/src/core/gpu/cl/kernels/ClBatchConcatenateKernel.cpp b/src/core/gpu/cl/kernels/ClBatchConcatenateKernel.cpp
index dbc628d6d5..fba1b0e087 100644
--- a/src/core/gpu/cl/kernels/ClBatchConcatenateKernel.cpp
+++ b/src/core/gpu/cl/kernels/ClBatchConcatenateKernel.cpp
@@ -60,7 +60,6 @@ Status validate_arguments(const ITensorInfo *src, unsigned int batch_offset, con
} // namespace
ClBatchConcatenateKernel::ClBatchConcatenateKernel()
- : _batch_offset(0)
{
_type = CLKernelType::ELEMENTWISE;
}
diff --git a/src/core/gpu/cl/kernels/ClBatchConcatenateKernel.h b/src/core/gpu/cl/kernels/ClBatchConcatenateKernel.h
index d9fa905e8e..2963d7cdfd 100644
--- a/src/core/gpu/cl/kernels/ClBatchConcatenateKernel.h
+++ b/src/core/gpu/cl/kernels/ClBatchConcatenateKernel.h
@@ -40,7 +40,6 @@ namespace kernels
class ClBatchConcatenateKernel : public IClKernel
{
public:
- /** Default constructor */
ClBatchConcatenateKernel();
ARM_COMPUTE_DISALLOW_COPY_ALLOW_MOVE(ClBatchConcatenateKernel);
/** Initialise the kernel's source and destination
@@ -55,11 +54,9 @@ public:
*
*/
void configure(const CLCompileContext &compile_context, ITensorInfo *src, unsigned int batch_offset, ITensorInfo *dst);
- /** Static function to check if given info will lead to a valid configuration of @ref ClBatchConcatenateKernel
+ /** Static function to check if given info will lead to a valid configuration
*
- * @param[in] src Input tensor info. Data types supported: All.
- * @param[in] batch_offset The offset on axis # 3.
- * @param[in] dst Destination tensor info. Data types supported: Same as @p src.
+ * Similar to @ref ClBatchConcatenateKernel::configure()
*
* @return a status
*/
@@ -69,7 +66,7 @@ public:
void run_op(ITensorPack &tensors, const Window &window, ::cl::CommandQueue &queue) override;
private:
- unsigned int _batch_offset;
+ unsigned int _batch_offset{ 0 };
};
} // namespace kernels
} // namespace opencl
diff --git a/src/core/gpu/cl/kernels/ClConvertFullyConnectedWeightsKernel.h b/src/core/gpu/cl/kernels/ClConvertFullyConnectedWeightsKernel.h
index 6f4f09dc32..3976fd45db 100644
--- a/src/core/gpu/cl/kernels/ClConvertFullyConnectedWeightsKernel.h
+++ b/src/core/gpu/cl/kernels/ClConvertFullyConnectedWeightsKernel.h
@@ -21,8 +21,8 @@
* OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
* SOFTWARE.
*/
-#ifndef ARM_COMPUTE_CLCONVERTFULLYCONNECTEDWEIGHTSKERNEL_H
-#define ARM_COMPUTE_CLCONVERTFULLYCONNECTEDWEIGHTSKERNEL_H
+#ifndef ARM_COMPUTE_CL_CONVERT_FULLYCONNECTED_WEIGHTS_KERNEL_H
+#define ARM_COMPUTE_CL_CONVERT_FULLYCONNECTED_WEIGHTS_KERNEL_H
#include "src/core/common/Macros.h"
#include "src/core/gpu/cl/ClCompileContext.h"
@@ -30,8 +30,6 @@
namespace arm_compute
{
-class ICLTensor;
-
/** Interface to convert the 2D Fully Connected weights from NCHW to NHWC or vice versa.
*
* @note This function can be applied to the 2D weights used by a Fully Connected layer if:
@@ -58,12 +56,11 @@ public:
* @param[in] data_layout The data layout the weights have been trained in.
*/
void configure(const CLCompileContext &compile_context, const ITensorInfo *src, ITensorInfo *dst, const TensorShape &original_src_shape, DataLayout data_layout);
- /** Static function to check if given info will lead to a valid configuration of @ref ClConvertFullyConnectedWeightsKernel
+ /** Static function to check if given info will lead to a valid configuration
+ *
+ * Similar to @ref ClConvertFullyConnectedWeightsKernel::configure()
*
- * @param[in] src Source weights tensor info to convert. Must be 2 dimensional. Data types supported: All.
- * @param[in] dst The converted weights tensor info. Shape and Data Type: Same as @p src.
- * @param[in] original_src_shape Shape of the original src tensor (the one entering fully connected layer).
- * @param[in] data_layout The data layout the weights have been trained in.
+ * @return a status
*/
static Status validate(const ITensorInfo *src, const ITensorInfo *dst, const TensorShape &original_src_shape, DataLayout data_layout);
@@ -73,4 +70,4 @@ public:
} // namespace kernels
} // namespace opencl
} // namespace arm_compute
-#endif /*ARM_COMPUTE_CLCONVERTFULLYCONNECTEDWEIGHTSKERNEL_H */
+#endif /* ARM_COMPUTE_CL_CONVERT_FULLYCONNECTED_WEIGHTS_KERNEL_H */
diff --git a/src/core/gpu/cl/kernels/ClCopyKernel.h b/src/core/gpu/cl/kernels/ClCopyKernel.h
index f3eb0aab62..d2732c4e59 100644
--- a/src/core/gpu/cl/kernels/ClCopyKernel.h
+++ b/src/core/gpu/cl/kernels/ClCopyKernel.h
@@ -48,11 +48,9 @@ public:
* @param[in] dst_window (Optional) Window to be used in case only copying into part of a tensor. Default is nullptr.
*/
void configure(const CLCompileContext &compile_context, const ITensorInfo *src, ITensorInfo *dst, Window *dst_window = nullptr);
- /** Static function to check if given info will lead to a valid configuration of @ref ClCopyKernel
+ /** Static function to check if given info will lead to a valid configuration
*
- * @param[in] src Source tensor info. Data types supported: All.
- * @param[in] dst Destination tensor info. Data types supported: same as @p src.
- * @param[in] dst_window (Optional) Window to be used in case only copying into part of a tensor. Default is nullptr.
+ * Similar to @ref ClCopyKernel::configure()
*
* @return a status
*/
@@ -68,4 +66,4 @@ private:
} // namespace kernels
} // namespace opencl
} // namespace arm_compute
-#endif /*ARM_COMPUTE_CL_COPY_KERNEL_H */
+#endif /* ARM_COMPUTE_CL_COPY_KERNEL_H */
diff --git a/src/core/gpu/cl/kernels/ClCropKernel.cpp b/src/core/gpu/cl/kernels/ClCropKernel.cpp
index ef2e48b45d..a052ef53f9 100644
--- a/src/core/gpu/cl/kernels/ClCropKernel.cpp
+++ b/src/core/gpu/cl/kernels/ClCropKernel.cpp
@@ -41,11 +41,6 @@ namespace opencl
{
namespace kernels
{
-void ClCropKernel::configure(const ITensorInfo *src, ITensorInfo *dst, Coordinates2D start, Coordinates2D end, uint32_t batch_index, float extrapolation_value, Window *dst_window)
-{
- configure(CLKernelLibrary::get().get_compile_context(), src, dst, start, end, batch_index, extrapolation_value, dst_window);
-}
-
ClCropKernel::ClCropKernel()
{
_type = CLKernelType::ELEMENTWISE;
diff --git a/src/core/gpu/cl/kernels/ClCropKernel.h b/src/core/gpu/cl/kernels/ClCropKernel.h
index 7120dbbc81..d81912284e 100644
--- a/src/core/gpu/cl/kernels/ClCropKernel.h
+++ b/src/core/gpu/cl/kernels/ClCropKernel.h
@@ -21,8 +21,8 @@
* OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
* SOFTWARE.
*/
-#ifndef ARM_COMPUTE_CLCROPKERNEL_H
-#define ARM_COMPUTE_CLCROPKERNEL_H
+#ifndef ARM_COMPUTE_CL_CROP_KERNEL_H
+#define ARM_COMPUTE_CL_CROP_KERNEL_H
#include "src/core/common/Macros.h"
#include "src/core/gpu/cl/ClCompileContext.h"
@@ -44,19 +44,6 @@ public:
*
* @note Supported tensor rank: up to 4
*
- * @param[in] src Source tensor info. Data type supported: All. Data layouts supported: NHWC.
- * @param[out] dst Destination tensor info. Data type supported: F32
- * @param[in] start Coordinates of where to start cropping the image.
- * @param[in] end Coordinates of where to end cropping the image.
- * @param[in] batch_index Fourth dimension index of the 3D image to crop in @p src.
- * @param[in] extrapolation_value Value to be used for values outside of the image. Default is 0.
- * @param[in] dst_window Output window to be used in case cropped image is being copied into a tensor. Default is nullptr.
- */
- void configure(const ITensorInfo *src, ITensorInfo *dst, Coordinates2D start, Coordinates2D end, uint32_t batch_index, float extrapolation_value = 0, Window *dst_window = nullptr);
- /** Configure kernel
- *
- * @note Supported tensor rank: up to 4
- *
* @param[in] compile_context The compile context to be used.
* @param[in] src Source tensor info. Data type supported: All. Data layouts supported: NHWC.
* @param[out] dst Destination tensor info. Data type supported: F32
@@ -68,18 +55,11 @@ public:
*/
void configure(const CLCompileContext &compile_context, const ITensorInfo *src, ITensorInfo *dst, Coordinates2D start, Coordinates2D end, uint32_t batch_index, float extrapolation_value = 0,
Window *dst_window = nullptr);
-
- /** Static function to check if given info will lead to a valid configuration of @ref CLStridedSliceKernel
+ /** Static function to check if given info will lead to a valid configuration
*
- * @note Supported tensor rank: up to 4
+ * Similar to @ref ClCropKernel::configure()
*
- * @param[in] src Source tensor info. Data type supported: All. Data layouts supported: NHWC.
- * @param[in] dst Destination tensor info. Data type supported: F32
- * @param[in] start Coordinates of where to start cropping the image.
- * @param[in] end Coordinates of where to end cropping the image.
- * @param[in] batch_index Fourth dimension index of the 3D image to crop in @p src.
- * @param[in] extrapolation_value Value to be used for values outside of the image. Default is 0.
- * @param[in] dst_window Output window to be used in case cropped image is being copied into a tensor. Default is nullptr.
+ * @return a status
*/
static Status validate(const ITensorInfo *src, const ITensorInfo *dst, Coordinates2D start, Coordinates2D end, uint32_t batch_index, float extrapolation_value = 0,
Window *dst_window = nullptr);
@@ -95,4 +75,4 @@ private:
} // namespace kernels
} // namespace opencl
} // namespace arm_compute
-#endif /*ARM_COMPUTE_CLCROPKERNEL_H */
+#endif /* ARM_COMPUTE_CL_CROP_KERNEL_H */
diff --git a/src/core/gpu/cl/kernels/ClDepthConcatenateKernel.h b/src/core/gpu/cl/kernels/ClDepthConcatenateKernel.h
index 103ef00695..0f408477b1 100644
--- a/src/core/gpu/cl/kernels/ClDepthConcatenateKernel.h
+++ b/src/core/gpu/cl/kernels/ClDepthConcatenateKernel.h
@@ -40,7 +40,6 @@ namespace kernels
class ClDepthConcatenateKernel : public IClKernel
{
public:
- /** Default constructor */
ClDepthConcatenateKernel();
ARM_COMPUTE_DISALLOW_COPY_ALLOW_MOVE(ClDepthConcatenateKernel);
/** Initialise the kernel's source and destination
@@ -55,11 +54,9 @@ public:
*
*/
void configure(const CLCompileContext &compile_context, ITensorInfo *src, unsigned int depth_offset, ITensorInfo *dst);
- /** Static function to check if given info will lead to a valid configuration of @ref ClDepthConcatenateKernel
+ /** Static function to check if given info will lead to a valid configuration
*
- * @param[in] src Source tensor info. Data types supported: QASYMM8/QASYMM8_SIGNED/F16/F32
- * @param[in] depth_offset The offset on the Z axis.
- * @param[in] dst Destination tensor info. Data types supported: Same as @p src.
+ * Similar to @ref ClDepthConcatenateKernel::configure()
*
* @return a status
*/
diff --git a/src/core/gpu/cl/kernels/ClDequantizeKernel.h b/src/core/gpu/cl/kernels/ClDequantizeKernel.h
index 2460674067..0912e1b228 100644
--- a/src/core/gpu/cl/kernels/ClDequantizeKernel.h
+++ b/src/core/gpu/cl/kernels/ClDequantizeKernel.h
@@ -38,7 +38,6 @@ namespace kernels
class ClDequantizeKernel : public IClKernel
{
public:
- /** Default constructor */
ClDequantizeKernel();
ARM_COMPUTE_DISALLOW_COPY_ALLOW_MOVE(ClDequantizeKernel);
/** Initialise the kernel's input and output
diff --git a/src/core/gpu/cl/kernels/ClDirectConv2dKernel.h b/src/core/gpu/cl/kernels/ClDirectConv2dKernel.h
index b592a2191b..4880d4a668 100644
--- a/src/core/gpu/cl/kernels/ClDirectConv2dKernel.h
+++ b/src/core/gpu/cl/kernels/ClDirectConv2dKernel.h
@@ -86,4 +86,4 @@ public:
} // namespace kernels
} // namespace opencl
} // namespace arm_compute
-#endif /*ARM_COMPUTE_CL_DIRECT_CONV2D_KERNEL_H */
+#endif /* ARM_COMPUTE_CL_DIRECT_CONV2D_KERNEL_H */
diff --git a/src/core/gpu/cl/kernels/ClElementwiseKernel.cpp b/src/core/gpu/cl/kernels/ClElementwiseKernel.cpp
index 7bfdb9efdd..b645353dd6 100644
--- a/src/core/gpu/cl/kernels/ClElementwiseKernel.cpp
+++ b/src/core/gpu/cl/kernels/ClElementwiseKernel.cpp
@@ -276,21 +276,12 @@ ClElementwiseKernel::ClElementwiseKernel()
_type = CLKernelType::ELEMENTWISE;
}
-void ClElementwiseKernel::configure_common(ITensorInfo *src1, ITensorInfo *src2, ITensorInfo *dst)
-{
- configure_common(CLKernelLibrary::get().get_compile_context(), src1, src2, dst);
-}
-
void ClElementwiseKernel::configure_common(const ClCompileContext &compile_context, ITensorInfo *src1, ITensorInfo *src2, ITensorInfo *dst)
{
// Configure kernel window
auto win_config = validate_and_configure_window(*src1, *src2, *dst);
ARM_COMPUTE_ERROR_THROW_ON(win_config.first);
- _src1 = src1;
- _src2 = src2;
- _dst = dst;
-
std::string kernel_name = "elementwise_operation_" + name();
if(is_data_type_quantized(src1->data_type()))
{
diff --git a/src/core/gpu/cl/kernels/ClElementwiseKernel.h b/src/core/gpu/cl/kernels/ClElementwiseKernel.h
index 7f55151a87..ab5c777ae6 100644
--- a/src/core/gpu/cl/kernels/ClElementwiseKernel.h
+++ b/src/core/gpu/cl/kernels/ClElementwiseKernel.h
@@ -44,7 +44,6 @@ namespace kernels
class ClElementwiseKernel : public IClKernel
{
public:
- /** Default constructor */
ClElementwiseKernel();
ARM_COMPUTE_DISALLOW_COPY_ALLOW_MOVE(ClElementwiseKernel);
@@ -80,30 +79,14 @@ protected:
/** Commmon configure function for element-wise operators with no additional options (e.g., Div, Min, Max, SquaredDiff)
*
*/
- void configure_common(ITensorInfo *src1, ITensorInfo *src2, ITensorInfo *dst);
- /** Commmon configure function for element-wise operators with no additional options (e.g., Div, Min, Max, SquaredDiff)
- *
- */
void configure_common(const ClCompileContext &compile_context, ITensorInfo *src1, ITensorInfo *src2, ITensorInfo *dst);
ActivationLayerInfo _act_info{};
-
-private:
- const ITensorInfo *_src1
- {
- nullptr
- }; /**< Source tensor info 1 */
- const ITensorInfo *_src2
- {
- nullptr
- }; /**< Source tensor info 2 */
- ITensorInfo *_dst{ nullptr }; /**< Destination tensor info */
};
class ClLogicalBinaryKernel : public ClElementwiseKernel
{
public:
- /** Default constructor */
ClLogicalBinaryKernel() = default;
ARM_COMPUTE_DISALLOW_COPY_ALLOW_MOVE(ClLogicalBinaryKernel);
/** Function to configure kernel
@@ -115,12 +98,11 @@ public:
* @param[in] dst Destination tensor info. Data types supported: same as @p src1.
*/
void configure(const ClCompileContext &compile_context, LogicalOperation op, ITensorInfo *src1, ITensorInfo *src2, ITensorInfo *dst);
- /** Static function to check if the given configuration is valid for this kernel
+ /** Static function to check if given info will lead to a valid configuration
*
- * @param[in] op Logical binary operation to be executed.
- * @param[in] src1 First source tensor info. Data types supported: U8.
- * @param[in] src2 Second source tensor info. Data types supported: same as @p src1.
- * @param[in] dst Destination tensor info. Data types supported: same as @p src1.
+ * Similar to @ref ClLogicalBinaryKernel::configure()
+ *
+ * @return a status
*/
static Status validate(LogicalOperation op, const ITensorInfo *src1, const ITensorInfo *src2, const ITensorInfo *dst);
@@ -153,16 +135,11 @@ public:
void configure(const ClCompileContext &compile_context, ArithmeticOperation op, ITensorInfo *input1, ITensorInfo *input2, ITensorInfo *output, const ConvertPolicy &policy,
const ActivationLayerInfo &act_info = ActivationLayerInfo());
- /** Static function to check if given info will lead to a valid configuration of @ref ClSaturatedArithmeticKernel
+ /** Static function to check if given info will lead to a valid configuration
*
- * @param[in] op Arithmetic operation to be executed.
- * @param[in] input1 First tensor input info info. Data types supported: U8/QASYMM8/QASYMM8_SIGNED/S16/QSYMM16/F16/S32/F32.
- * @param[in] input2 Second tensor input info info. Data types supported: Same as @p input1.
- * @param[in] output Output tensor info info. Data types supported: Same as @p input1.
- * @param[in] policy Policy to use to handle overflow.
- * @param[in] act_info (Optional) Activation layer information in case of a fused activation.
+ * Similar to @ref ClSaturatedArithmeticKernel::configure()
*
- * @return a Status
+ * @return a status
*/
static Status validate(ArithmeticOperation op, const ITensorInfo *input1, const ITensorInfo *input2, const ITensorInfo *output, const ConvertPolicy &policy,
const ActivationLayerInfo &act_info = ActivationLayerInfo());
@@ -197,15 +174,11 @@ public:
void configure(const ClCompileContext &compile_context, ArithmeticOperation op, ITensorInfo *src1, ITensorInfo *src2, ITensorInfo *dst,
const ActivationLayerInfo &act_info = ActivationLayerInfo());
- /** Static function to check if given info will lead to a valid configuration of @ref ClArithmeticKernel
+ /** Static function to check if given info will lead to a valid configuration
*
- * @param[in] op Arithmetic operation to be executed.
- * @param[in] src1 First source tensor info. Data types supported: U8/QASYMM8/QASYMM8_SIGNED/S16/QSYMM16/F16/S32/F32.
- * @param[in] src2 Second source tensor info. Data types supported: same as @p src1.
- * @param[in] dst Destination tensor info. Data types supported: same as @p src1.
- * @param[in] act_info (Optional) Activation layer information in case of a fused activation.
+ * Similar to @ref ClArithmeticKernel::configure()
*
- * @return a Status
+ * @return a status
*/
static Status validate(ArithmeticOperation op, const ITensorInfo *src1, const ITensorInfo *src2, const ITensorInfo *dst, const ActivationLayerInfo &act_info = ActivationLayerInfo());
diff --git a/src/core/gpu/cl/kernels/ClElementwiseUnaryKernel.h b/src/core/gpu/cl/kernels/ClElementwiseUnaryKernel.h
index 225869b58b..64cc2f7afc 100644
--- a/src/core/gpu/cl/kernels/ClElementwiseUnaryKernel.h
+++ b/src/core/gpu/cl/kernels/ClElementwiseUnaryKernel.h
@@ -48,11 +48,9 @@ public:
* @param[in] op Element wise unary operation to perform.
*/
void configure(const CLCompileContext &compile_context, const ITensorInfo *src, ITensorInfo *dst, const ElementWiseUnary &op);
- /** Static function to check if given info will lead to a valid configuration of @ref ClElementWiseUnaryKernel
+ /** Static function to check if given info will lead to a valid configuration
*
- * @param[in] src First source tensor info. Data types supported: F16/F32.
- * @param[in] dst Destination tensor info. Data types supported: same as @p src.
- * @param[in] op Element wise unary operation to perform.
+ * Similar to @ref ClElementWiseUnaryKernel::configure()
*
* @return a status
*/
diff --git a/src/core/gpu/cl/kernels/ClFillKernel.cpp b/src/core/gpu/cl/kernels/ClFillKernel.cpp
index 526a466a00..f213bf8e6a 100644
--- a/src/core/gpu/cl/kernels/ClFillKernel.cpp
+++ b/src/core/gpu/cl/kernels/ClFillKernel.cpp
@@ -47,13 +47,6 @@ ClFillKernel::ClFillKernel()
_type = CLKernelType::ELEMENTWISE;
}
-void ClFillKernel::configure(ITensorInfo *tensor,
- const PixelValue &constant_value,
- Window *window)
-{
- configure(CLKernelLibrary::get().get_compile_context(), tensor, constant_value, window);
-}
-
void ClFillKernel::configure(const CLCompileContext &compile_context, ITensorInfo *tensor,
const PixelValue &constant_value,
Window *window)
diff --git a/src/core/gpu/cl/kernels/ClFillKernel.h b/src/core/gpu/cl/kernels/ClFillKernel.h
index 9542c20508..ecc2546e4a 100644
--- a/src/core/gpu/cl/kernels/ClFillKernel.h
+++ b/src/core/gpu/cl/kernels/ClFillKernel.h
@@ -42,24 +42,15 @@ public:
ARM_COMPUTE_DISALLOW_COPY_ALLOW_MOVE(ClFillKernel);
/** Initialise the kernel's tensor and filling value
*
- * @param[in,out] tensor Input tensor info. Supported data types: All.
- * @param[in] constant_value The value used to fill the planes of the tensor
- * @param[in] window Window to be used in case setting only part of a tensor. Default is nullptr.
- */
- void configure(ITensorInfo *tensor, const PixelValue &constant_value, Window *window = nullptr);
- /** Initialise the kernel's tensor and filling value
- *
* @param[in] compile_context The compile context to be used.
* @param[in,out] tensor Input tensor info. Supported data types: All.
* @param[in] constant_value The value used to fill the planes of the tensor
* @param[in] window Window to be used in case setting only part of a tensor. Default is nullptr.
*/
void configure(const CLCompileContext &compile_context, ITensorInfo *tensor, const PixelValue &constant_value, Window *window = nullptr);
- /** Static function to check if given info will lead to a valid configuration of @ref ClFillKernel
+ /** Static function to check if given info will lead to a valid configuration
*
- * @param[in] tensor Source tensor info. Data types supported: All.
- * @param[in] constant_value The value used to fill the planes of the tensor
- * @param[in] window Window to be used in case setting only part of a tensor. Default is nullptr.
+ * Similar to @ref ClFillKernel::configure()
*
* @return a status
*/
@@ -74,4 +65,4 @@ private:
} // namespace kernels
} // namespace opencl
} // namespace arm_compute
-#endif /*ARM_COMPUTE_CLMEMSETRKERNEL_H */
+#endif /* ARM_COMPUTE_CL_FILL_KERNEL_H */
diff --git a/src/core/gpu/cl/kernels/ClFloorKernel.h b/src/core/gpu/cl/kernels/ClFloorKernel.h
index 3bc648b7be..57c9906f2c 100644
--- a/src/core/gpu/cl/kernels/ClFloorKernel.h
+++ b/src/core/gpu/cl/kernels/ClFloorKernel.h
@@ -47,11 +47,9 @@ public:
* @param[out] dst Destination tensor info. Same as @p src
*/
void configure(const ClCompileContext &compile_context, const ITensorInfo *src, ITensorInfo *dst);
-
- /** Static function to check if given info will lead to a valid configuration of @ref ClFloorKernel
+ /** Static function to check if given info will lead to a valid configuration
*
- * @param[in] src Source tensor info. Data type supported: F16/F32.
- * @param[in] dst Destination tensor info. Same as @p src
+ * Similar to @ref ClFloorKernel::configure()
*
* @return a status
*/
diff --git a/src/core/gpu/cl/kernels/ClGemmLowpMatrixMultiplyNativeKernel.h b/src/core/gpu/cl/kernels/ClGemmLowpMatrixMultiplyNativeKernel.h
index 491c3e44df..eaa125fbf2 100644
--- a/src/core/gpu/cl/kernels/ClGemmLowpMatrixMultiplyNativeKernel.h
+++ b/src/core/gpu/cl/kernels/ClGemmLowpMatrixMultiplyNativeKernel.h
@@ -78,4 +78,4 @@ private:
} // namespace kernels
} // namespace opencl
} // namespace arm_compute
-#endif /* ARM_COMPUTE_CL_GEMMLOWP_MATRIXMULTIPLY_NATIVE_KERNEL_H*/
+#endif /* ARM_COMPUTE_CL_GEMMLOWP_MATRIXMULTIPLY_NATIVE_KERNEL_H */
diff --git a/src/core/gpu/cl/kernels/ClGemmLowpMatrixMultiplyReshapedKernel.h b/src/core/gpu/cl/kernels/ClGemmLowpMatrixMultiplyReshapedKernel.h
index b99dec33af..99cff011d1 100644
--- a/src/core/gpu/cl/kernels/ClGemmLowpMatrixMultiplyReshapedKernel.h
+++ b/src/core/gpu/cl/kernels/ClGemmLowpMatrixMultiplyReshapedKernel.h
@@ -87,4 +87,4 @@ private:
} // namespace kernels
} // namespace opencl
} // namespace arm_compute
-#endif /* ARM_COMPUTE_CL_GEMMLOWP_MATRIXMULTIPLY_RESHAPED_KERNEL_H*/
+#endif /* ARM_COMPUTE_CL_GEMMLOWP_MATRIXMULTIPLY_RESHAPED_KERNEL_H */
diff --git a/src/core/gpu/cl/kernels/ClGemmMatrixMultiplyNativeKernel.h b/src/core/gpu/cl/kernels/ClGemmMatrixMultiplyNativeKernel.h
index c3bdc7589e..cd7bf278c2 100644
--- a/src/core/gpu/cl/kernels/ClGemmMatrixMultiplyNativeKernel.h
+++ b/src/core/gpu/cl/kernels/ClGemmMatrixMultiplyNativeKernel.h
@@ -85,4 +85,4 @@ private:
} // namespace kernels
} // namespace opencl
} // namespace arm_compute
-#endif /*ARM_COMPUTE_CL_GEMM_MATRIXMULTIPLY_NATIVE_KERNEL_H*/
+#endif /* ARM_COMPUTE_CL_GEMM_MATRIXMULTIPLY_NATIVE_KERNEL_H */
diff --git a/src/core/gpu/cl/kernels/ClHeightConcatenateKernel.h b/src/core/gpu/cl/kernels/ClHeightConcatenateKernel.h
index 9a4380a5b7..0733078fc2 100644
--- a/src/core/gpu/cl/kernels/ClHeightConcatenateKernel.h
+++ b/src/core/gpu/cl/kernels/ClHeightConcatenateKernel.h
@@ -21,8 +21,8 @@
* OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
* SOFTWARE.
*/
-#ifndef ARM_COMPUTE_CL_HEIGHT_CONCATENATE_LAYER_KERNEL_H
-#define ARM_COMPUTE_CL_HEIGHT_CONCATENATE_LAYER_KERNEL_H
+#ifndef ARM_COMPUTE_CL_HEIGHT_CONCATENATE_KERNEL_H
+#define ARM_COMPUTE_CL_HEIGHT_CONCATENATE_KERNEL_H
#include "src/core/common/Macros.h"
#include "src/core/gpu/cl/ClCompileContext.h"
@@ -40,7 +40,6 @@ namespace kernels
class ClHeightConcatenateKernel : public IClKernel
{
public:
- /** Default constructor */
ClHeightConcatenateKernel();
ARM_COMPUTE_DISALLOW_COPY_ALLOW_MOVE(ClHeightConcatenateKernel);
/** Initialise the kernel's source and destination
@@ -52,11 +51,9 @@ public:
*
*/
void configure(const CLCompileContext &compile_context, ITensorInfo *src, unsigned int height_offset, ITensorInfo *dst);
- /** Static function to check if given info will lead to a valid configuration of @ref ClHeightConcatenateKernel
+ /** Static function to check if given info will lead to a valid configuration
*
- * @param[in] src Source tensor info. Data types supported: All.
- * @param[in] height_offset The starting offset on the Y axis for the dst tensor.
- * @param[in] dst Destination tensor info. Data types supported: same as @p src.
+ * Similar to @ref ClHeightConcatenateKernel::configure()
*
* @return a status
*/
@@ -71,4 +68,4 @@ private:
} // namespace kernels
} // namespace opencl
} // namespace arm_compute
-#endif /* ARM_COMPUTE_CL_HEIGHT_CONCATENATE_LAYER_KERNEL_H */
+#endif /* ARM_COMPUTE_CL_HEIGHT_CONCATENATE_KERNEL_H */
diff --git a/src/core/gpu/cl/kernels/ClMulKernel.h b/src/core/gpu/cl/kernels/ClMulKernel.h
index 9c70301d89..dec8dba61c 100644
--- a/src/core/gpu/cl/kernels/ClMulKernel.h
+++ b/src/core/gpu/cl/kernels/ClMulKernel.h
@@ -38,7 +38,6 @@ namespace kernels
class ClMulKernel : public IClKernel
{
public:
- /** Default constructor */
ClMulKernel();
ARM_COMPUTE_DISALLOW_COPY_ALLOW_MOVE(ClMulKernel);
/** Initialise the kernel's src and dst.
@@ -87,7 +86,6 @@ public:
class ClComplexMulKernel : public ICLKernel
{
public:
- /** Default constructor */
ClComplexMulKernel();
ARM_COMPUTE_DISALLOW_COPY_ALLOW_MOVE(ClComplexMulKernel);
/** Initialise the kernel's src and dst.
diff --git a/src/core/gpu/cl/kernels/ClPermuteKernel.h b/src/core/gpu/cl/kernels/ClPermuteKernel.h
index 326110a27c..839e224ee4 100644
--- a/src/core/gpu/cl/kernels/ClPermuteKernel.h
+++ b/src/core/gpu/cl/kernels/ClPermuteKernel.h
@@ -41,7 +41,6 @@ namespace kernels
class ClPermuteKernel : public IClKernel
{
public:
- /** Default constructor */
ClPermuteKernel();
ARM_COMPUTE_DISALLOW_COPY_ALLOW_MOVE(ClPermuteKernel);
/** Set the src and dst of the kernel.
@@ -54,13 +53,9 @@ public:
* @param[in] perm Permutation vector
*/
void configure(const CLCompileContext &compile_context, const ITensorInfo *src, ITensorInfo *dst, const PermutationVector &perm);
- /** Static function to check if given info will lead to a valid configuration of @ref ClPermuteKernel
+ /** Static function to check if given info will lead to a valid configuration
*
- * @note Arbitrary permutation vectors are supported with rank not greater than 4
- *
- * @param[in] src The src tensor info. Data types supported: All.
- * @param[in] dst The dst tensor info. Data types supported: same as @p src.
- * @param[in] perm Permutation vector
+ * Similar to @ref ClPermuteKernel::configure()
*
* @return a status
*/
@@ -75,4 +70,4 @@ private:
} // namespace kernels
} // namespace opencl
} // namespace arm_compute
-#endif /*ARM_COMPUTE_CL_PERMUTE_KERNEL_H */
+#endif /* ARM_COMPUTE_CL_PERMUTE_KERNEL_H */
diff --git a/src/core/gpu/cl/kernels/ClPool2dKernel.cpp b/src/core/gpu/cl/kernels/ClPool2dKernel.cpp
index 9d5a24fdf2..e522814b6d 100644
--- a/src/core/gpu/cl/kernels/ClPool2dKernel.cpp
+++ b/src/core/gpu/cl/kernels/ClPool2dKernel.cpp
@@ -203,7 +203,6 @@ std::tuple<Status, Window, ClPoolingConfig> validate_and_configure_window(ITenso
} // namespace
ClPool2dKernel::ClPool2dKernel()
- : _pool_info(), _data_layout(DataLayout::UNKNOWN), _border_size(0), _num_elems_processed_per_iteration(1)
{
_type = CLKernelType::POOL;
}
diff --git a/src/core/gpu/cl/kernels/ClPool2dKernel.h b/src/core/gpu/cl/kernels/ClPool2dKernel.h
index 8ecb8eb7b7..ab8c56a857 100644
--- a/src/core/gpu/cl/kernels/ClPool2dKernel.h
+++ b/src/core/gpu/cl/kernels/ClPool2dKernel.h
@@ -38,7 +38,6 @@ namespace kernels
class ClPool2dKernel : public IClKernel
{
public:
- /** Default constructor */
ClPool2dKernel();
ARM_COMPUTE_DISALLOW_COPY_ALLOW_MOVE(ClPool2dKernel);
@@ -65,10 +64,10 @@ public:
BorderSize border_size() const override;
public:
- PoolingLayerInfo _pool_info;
- DataLayout _data_layout;
- BorderSize _border_size;
- unsigned int _num_elems_processed_per_iteration;
+ PoolingLayerInfo _pool_info{};
+ DataLayout _data_layout{ DataLayout::UNKNOWN };
+ BorderSize _border_size{ 0 };
+ unsigned int _num_elems_processed_per_iteration{ 1 };
};
} // namespace kernels
} // namespace opencl
diff --git a/src/core/gpu/cl/kernels/ClQuantizeKernel.h b/src/core/gpu/cl/kernels/ClQuantizeKernel.h
index cd972987f5..1991a2fba8 100644
--- a/src/core/gpu/cl/kernels/ClQuantizeKernel.h
+++ b/src/core/gpu/cl/kernels/ClQuantizeKernel.h
@@ -41,7 +41,6 @@ namespace kernels
class ClQuantizeKernel : public IClKernel
{
public:
- /** Default constructor */
ClQuantizeKernel();
ARM_COMPUTE_DISALLOW_COPY_ALLOW_MOVE(ClQuantizeKernel);
/** Set the input, output.
diff --git a/src/core/gpu/cl/kernels/ClReshapeKernel.h b/src/core/gpu/cl/kernels/ClReshapeKernel.h
index 3cd8369012..01e1ee84b9 100644
--- a/src/core/gpu/cl/kernels/ClReshapeKernel.h
+++ b/src/core/gpu/cl/kernels/ClReshapeKernel.h
@@ -47,11 +47,9 @@ public:
* @param[out] dst Destination tensor info. Data type supported: Same as @p src
*/
void configure(const CLCompileContext &compile_context, const ITensorInfo *src, ITensorInfo *dst);
-
- /** Static function to check if given info will lead to a valid configuration of @ref ClReshapeKernel
+ /** Static function to check if given info will lead to a valid configuration
*
- * @param[in] src Source tensor info. Data type supported: All
- * @param[in] dst Destination tensor info. Data type supported: Same as @p src
+ * Similar to @ref ClReshapeKernel::configure()
*
* @return a status
*/
@@ -63,4 +61,4 @@ public:
} // namespace opencl
} // namespace kernels
} // namespace arm_compute
-#endif /*ARM_COMPUTE_CL_RESHAPE_KERNEL_H */
+#endif /* ARM_COMPUTE_CL_RESHAPE_KERNEL_H */
diff --git a/src/core/gpu/cl/kernels/ClScaleKernel.h b/src/core/gpu/cl/kernels/ClScaleKernel.h
index 826c4821b5..8333c7d6c0 100644
--- a/src/core/gpu/cl/kernels/ClScaleKernel.h
+++ b/src/core/gpu/cl/kernels/ClScaleKernel.h
@@ -31,8 +31,6 @@
namespace arm_compute
{
-class ICLTensor;
-
namespace opencl
{
namespace kernels
@@ -41,10 +39,8 @@ namespace kernels
class ClScaleKernel : public IClKernel
{
public:
- /** Default constructor */
ClScaleKernel();
ARM_COMPUTE_DISALLOW_COPY_ALLOW_MOVE(ClScaleKernel);
-
/** Initialise the kernel's inputs, output and interpolation policy
*
* @param[in] compile_context The compile context to be used.
@@ -54,13 +50,9 @@ public:
* @param[in] info @ref ScaleKernelInfo Kernel descriptor to be used to configure.
*/
void configure(const CLCompileContext &compile_context, ITensorInfo *src, ITensorInfo *dst, const ScaleKernelInfo &info);
-
- /** Static function to check if given info will lead to a valid configuration of @ref ClScaleKernel
+ /** Static function to check if given info will lead to a valid configuration
*
- * @param[in] src Source tensor info. Data types supported: U8/QASYMM8/QASYMM8_SIGNED/S16/F16/F32
- * @param[in] dst Destination tensor info. Data types supported: Same as @p src
- * All but the lowest two dimensions must be the same size as in the input tensor, i.e. scaling is only performed within the XY-plane.
- * @param[in] info @ref ScaleKernelInfo Kernel descriptor to be used to validate
+ * Similar to @ref ClScaleKernel::configure()
*
* @return a status
*/
@@ -71,9 +63,9 @@ public:
void run_op(ITensorPack &tensors, const Window &window, cl::CommandQueue &queue) override;
private:
- DataLayout _data_layout = DataLayout::UNKNOWN;
+ DataLayout _data_layout{ DataLayout::UNKNOWN };
};
} // namespace kernels
} // namespace opencl
} // namespace arm_compute
-#endif /*ARM_COMPUTE_CLSCALEKERNEL_H */
+#endif /* ARM_COMPUTE_CL_SCALE_KERNEL_H */
diff --git a/src/core/gpu/cl/kernels/ClSoftmaxKernel.h b/src/core/gpu/cl/kernels/ClSoftmaxKernel.h
index db1aca3b9b..a2ad02d6b7 100644
--- a/src/core/gpu/cl/kernels/ClSoftmaxKernel.h
+++ b/src/core/gpu/cl/kernels/ClSoftmaxKernel.h
@@ -21,8 +21,8 @@
* OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
* SOFTWARE.
*/
-#ifndef ARM_COMPUTE_CLSOFTMAXLAYERKERNEL_H
-#define ARM_COMPUTE_CLSOFTMAXLAYERKERNEL_H
+#ifndef ARM_COMPUTE_CL_SOFTMAX_KERNEL_H
+#define ARM_COMPUTE_CL_SOFTMAX_KERNEL_H
#include "arm_compute/core/Error.h"
#include "arm_compute/core/KernelDescriptors.h"
@@ -50,7 +50,6 @@ public:
/** Info for whether a parallel reduction will be run and the vector size of the execution. */
using ParallelReductionInfo = std::tuple<bool, unsigned int>;
- /** Default constructor */
ClLogits1DMaxShiftExpSumKernel();
ARM_COMPUTE_DISALLOW_COPY_ALLOW_MOVE(ClLogits1DMaxShiftExpSumKernel);
/** Configure the kernel using the given information about tensors
@@ -63,12 +62,9 @@ public:
* @param[in] info Contains information consumed by kernels for softmax described in @ref SoftmaxKernelInfo.
*/
void configure(const CLCompileContext &compile_context, const ITensorInfo &src, ITensorInfo &max, ITensorInfo &dst, ITensorInfo &sum, const SoftmaxKernelInfo &info);
- /** Static function to check if given info will lead to a valid configuration of @ref ClLogits1DMaxShiftExpSumKernel
+ /** Static function to check if given info will lead to a valid configuration
*
- * @param[in] src Source tensor. Data types supported: QASYMM8/QASYMM8_SIGNED/F16/F32
- * @param[in] max Max values tensor. Data types supported: same as @p src
- * @param[in] dst Destination tensor. Data types supported: same as @p src
- * @param[in] sum Sum of 1D logits tensor. Data types supported: same as @p src
+ * Similar to @ref ClLogits1DMaxShiftExpSumKernel::configure()
*
* @return a status
*/
@@ -93,7 +89,6 @@ public:
class ClLogits1DNormKernel : public IClKernel
{
public:
- /** Default constructor */
ClLogits1DNormKernel();
ARM_COMPUTE_DISALLOW_COPY_ALLOW_MOVE(ClLogits1DNormKernel);
@@ -106,12 +101,9 @@ public:
* @param[in] info Contains information consumed by kernels for softmax described in @ref SoftmaxKernelInfo.
*/
void configure(const CLCompileContext &compile_context, const ITensorInfo &src, const ITensorInfo &sum, ITensorInfo &dst, const SoftmaxKernelInfo &info);
- /** Static function to check if given info will lead to a valid configuration of @ref ClLogits1DNormKernel
+ /** Static function to check if given info will lead to a valid configuration
*
- * @param[in] src Source tensor. Data types supported: S32/F16/F32. If this kernel is used for log softmax, only F32/F16 is supported.
- * @param[in] sum Sum tensor. Dimensions should be dim(input)-1. Data types supported: same as @p input
- * @param[in] dst Destination tensor. Data types supported: QASYMM8 for S32 @p input, or same as @p input
- * @param[in] info Contains information consumed by kernels for softmax described in @ref SoftmaxKernelInfo.
+ * Similar to @ref ClLogits1DNormKernel::configure()
*
* @return a status
*/
@@ -123,4 +115,4 @@ public:
} // namespace kernels
} // namespace opencl
} // namespace arm_compute
-#endif /*ARM_COMPUTE_CLSOFTMAXLAYERKERNEL_H */
+#endif /* ARM_COMPUTE_CL_SOFTMAX_KERNEL_H */
diff --git a/src/core/gpu/cl/kernels/ClTransposeKernel.h b/src/core/gpu/cl/kernels/ClTransposeKernel.h
index 7d1226cc0d..c8379d44c7 100644
--- a/src/core/gpu/cl/kernels/ClTransposeKernel.h
+++ b/src/core/gpu/cl/kernels/ClTransposeKernel.h
@@ -47,10 +47,9 @@ public:
* @param[in] dst The dst tensor info. Data types supported: Same as @p src
*/
void configure(const CLCompileContext &compile_context, const ITensorInfo *src, ITensorInfo *dst);
- /** Static function to check if given info will lead to a valid configuration of @ref ClTransposeKernel
+ /** Static function to check if given info will lead to a valid configuration
*
- * @param[in] src The src tensor info. Data types supported: All.
- * @param[in] dst The dst tensor info. Data types supported: same as @p src.
+ * Similar to @ref ClTransposeKernel::configure()
*
* @return a status
*/
@@ -62,4 +61,4 @@ public:
} // namespace kernels
} // namespace opencl
} // namespace arm_compute
-#endif /*ARM_COMPUTE_CL_TRANSPOSE_KERNEL_H */
+#endif /* ARM_COMPUTE_CL_TRANSPOSE_KERNEL_H */
diff --git a/src/core/gpu/cl/kernels/ClWidthConcatenate2TensorsKernel.h b/src/core/gpu/cl/kernels/ClWidthConcatenate2TensorsKernel.h
index 56202bad2e..15e0757aec 100644
--- a/src/core/gpu/cl/kernels/ClWidthConcatenate2TensorsKernel.h
+++ b/src/core/gpu/cl/kernels/ClWidthConcatenate2TensorsKernel.h
@@ -40,7 +40,6 @@ namespace kernels
class ClWidthConcatenate2TensorsKernel : public IClKernel
{
public:
- /** Default constructor */
ClWidthConcatenate2TensorsKernel();
ARM_COMPUTE_DISALLOW_COPY_ALLOW_MOVE(ClWidthConcatenate2TensorsKernel);
/** Initialise the kernel's sources and destination
@@ -51,11 +50,9 @@ public:
* @param[out] dst Destination tensor info. Data types supported: Same as @p src1.
*/
void configure(const CLCompileContext &compile_context, ITensorInfo *src1, ITensorInfo *src2, ITensorInfo *dst);
- /** Static function to check if given info will lead to a valid configuration of @ref ClWidthConcatenate2TensorsKernel
+ /** Static function to check if given info will lead to a valid configuration
*
- * @param[in] src1 First tensor info. Data types supported: All.
- * @param[in] src2 Second tensor info. Data types supported: same as @p src1
- * @param[in] dst Destination tensor info. Data types supported: Same as @p src1.
+ * Similar to @ref ClWidthConcatenate2TensorsKernel::configure()
*
* @return a status
*/
diff --git a/src/core/gpu/cl/kernels/ClWidthConcatenate4TensorsKernel.h b/src/core/gpu/cl/kernels/ClWidthConcatenate4TensorsKernel.h
index 19bda65902..1e3f47f7fb 100644
--- a/src/core/gpu/cl/kernels/ClWidthConcatenate4TensorsKernel.h
+++ b/src/core/gpu/cl/kernels/ClWidthConcatenate4TensorsKernel.h
@@ -41,7 +41,6 @@ namespace kernels
class ClWidthConcatenate4TensorsKernel : public IClKernel
{
public:
- /** Default constructor */
ClWidthConcatenate4TensorsKernel();
ARM_COMPUTE_DISALLOW_COPY_ALLOW_MOVE(ClWidthConcatenate4TensorsKernel);
/** Initialise the kernel's sources and destination
@@ -54,13 +53,9 @@ public:
* @param[out] dst Destination tensor info. Data types supported: same as @p src1.
*/
void configure(const CLCompileContext &compile_context, ITensorInfo *src1, ITensorInfo *src2, ITensorInfo *src3, ITensorInfo *src4, ITensorInfo *dst);
- /** Static function to check if given info will lead to a valid configuration of @ref ClWidthConcatenate4TensorsKernel
+ /** Static function to check if given info will lead to a valid configuration
*
- * @param[in] src1 First tensor info. Data types supported: All.
- * @param[in] src2 Second tensor info. Data types supported: same as @p src1
- * @param[in] src3 Third tensor info. Data types supported: same as @p src1
- * @param[in] src4 Fourth tensor info. Data types supported: same as @p src1
- * @param[in] dst Destination tensor info. Data types supported: same as @p src1.
+ * Similar to @ref ClWidthConcatenate4TensorsKernel::configure()
*
* @return a status
*/
diff --git a/src/core/gpu/cl/kernels/ClWidthConcatenateKernel.h b/src/core/gpu/cl/kernels/ClWidthConcatenateKernel.h
index 6bc8e57a08..300c4beb30 100644
--- a/src/core/gpu/cl/kernels/ClWidthConcatenateKernel.h
+++ b/src/core/gpu/cl/kernels/ClWidthConcatenateKernel.h
@@ -21,8 +21,8 @@
* OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
* SOFTWARE.
*/
-#ifndef ARM_COMPUTE_CL_WIDTH_CONCATENATE_LAYER_KERNEL_H
-#define ARM_COMPUTE_CL_WIDTH_CONCATENATE_LAYER_KERNEL_H
+#ifndef ARM_COMPUTE_CL_WIDTH_CONCATENATE_KERNEL_H
+#define ARM_COMPUTE_CL_WIDTH_CONCATENATE_KERNEL_H
#include "src/core/common/Macros.h"
#include "src/core/gpu/cl/ClCompileContext.h"
@@ -40,7 +40,6 @@ namespace kernels
class ClWidthConcatenateKernel : public IClKernel
{
public:
- /** Default constructor */
ClWidthConcatenateKernel();
ARM_COMPUTE_DISALLOW_COPY_ALLOW_MOVE(ClWidthConcatenateKernel);
/** Initialise the kernel's source and destination
@@ -52,11 +51,9 @@ public:
*
*/
void configure(const CLCompileContext &compile_context, ITensorInfo *src, unsigned int width_offset, ITensorInfo *dst);
- /** Static function to check if given info will lead to a valid configuration of @ref ClWidthConcatenateKernel
+ /** Static function to check if given info will lead to a valid configuration
*
- * @param[in] src Source tensor info. Data types supported: All.
- * @param[in] width_offset The offset on the X axis.
- * @param[in] dst Destination tensor info. Data types supported: same as @p src.
+ * Similar to @ref ClWidthConcatenateKernel::configure()
*
* @return a status
*/
@@ -68,4 +65,4 @@ public:
} // namespace kernels
} // namespace opencl
} // namespace arm_compute
-#endif /* ARM_COMPUTE_CL_WIDTH_CONCATENATE_LAYER_KERNEL_H */
+#endif /* ARM_COMPUTE_CL_WIDTH_CONCATENATE_KERNEL_H */
diff --git a/src/core/gpu/cl/kernels/ClWinogradFilterTransformKernel.h b/src/core/gpu/cl/kernels/ClWinogradFilterTransformKernel.h
index 13200dc419..145954fbb1 100644
--- a/src/core/gpu/cl/kernels/ClWinogradFilterTransformKernel.h
+++ b/src/core/gpu/cl/kernels/ClWinogradFilterTransformKernel.h
@@ -39,7 +39,6 @@ namespace kernels
class ClWinogradFilterTransformKernel : public IClKernel
{
public:
- /** Default constructor */
ClWinogradFilterTransformKernel();
ARM_COMPUTE_DISALLOW_COPY_ALLOW_MOVE(ClWinogradFilterTransformKernel);
/** Set the input and output tensor.
@@ -75,4 +74,4 @@ public:
} // namespace kernels
} // namespace opencl
} // namespace arm_compute
-#endif /*ARM_COMPUTE_CL_WINOGRAD_FILTER_TRANSFORM_KERNEL_H */
+#endif /* ARM_COMPUTE_CL_WINOGRAD_FILTER_TRANSFORM_KERNEL_H */
diff --git a/src/core/gpu/cl/kernels/ClWinogradInputTransformKernel.cpp b/src/core/gpu/cl/kernels/ClWinogradInputTransformKernel.cpp
index 62db2282e0..538d8ae602 100644
--- a/src/core/gpu/cl/kernels/ClWinogradInputTransformKernel.cpp
+++ b/src/core/gpu/cl/kernels/ClWinogradInputTransformKernel.cpp
@@ -101,7 +101,6 @@ std::pair<Status, Window> validate_and_configure_window(ITensorInfo *input, ITen
} // namespace
ClWinogradInputTransformKernel::ClWinogradInputTransformKernel()
- : _border_size(0), _data_layout(DataLayout::UNKNOWN), _num_tiles_x(0), _num_tiles_y(0), _step_z(1)
{
_type = CLKernelType::WINOGRAD;
}
diff --git a/src/core/gpu/cl/kernels/ClWinogradInputTransformKernel.h b/src/core/gpu/cl/kernels/ClWinogradInputTransformKernel.h
index 76b45279a4..40fc2f387a 100644
--- a/src/core/gpu/cl/kernels/ClWinogradInputTransformKernel.h
+++ b/src/core/gpu/cl/kernels/ClWinogradInputTransformKernel.h
@@ -39,7 +39,6 @@ namespace kernels
class ClWinogradInputTransformKernel : public IClKernel
{
public:
- /** Default constructor */
ClWinogradInputTransformKernel();
ARM_COMPUTE_DISALLOW_COPY_ALLOW_MOVE(ClWinogradInputTransformKernel);
/** Set the input and output of the kernel.
@@ -76,13 +75,13 @@ public:
private:
using WinogradKey = std::pair<std::pair<int, int>, std::pair<int, int>>;
- BorderSize _border_size;
- DataLayout _data_layout;
- int _num_tiles_x;
- int _num_tiles_y;
- unsigned int _step_z;
+ BorderSize _border_size{ 0 };
+ DataLayout _data_layout{ DataLayout::UNKNOWN };
+ int _num_tiles_x{ 0 };
+ int _num_tiles_y{ 0 };
+ unsigned int _step_z{ 1 };
};
} // namespace kernels
} // namespace opencl
} // namespace arm_compute
-#endif /*ARM_COMPUTE_CL_WINOGRAD_INPUT_TRANSFORM_KERNEL_H */
+#endif /* ARM_COMPUTE_CL_WINOGRAD_INPUT_TRANSFORM_KERNEL_H */
diff --git a/src/core/gpu/cl/kernels/ClWinogradOutputTransformKernel.h b/src/core/gpu/cl/kernels/ClWinogradOutputTransformKernel.h
index 2948d3f181..22b7f079c1 100644
--- a/src/core/gpu/cl/kernels/ClWinogradOutputTransformKernel.h
+++ b/src/core/gpu/cl/kernels/ClWinogradOutputTransformKernel.h
@@ -39,7 +39,6 @@ namespace kernels
class ClWinogradOutputTransformKernel : public IClKernel
{
public:
- /** Default constructor */
ClWinogradOutputTransformKernel();
ARM_COMPUTE_DISALLOW_COPY_ALLOW_MOVE(ClWinogradOutputTransformKernel);
/** Set the input and output tensor.
@@ -64,7 +63,6 @@ public:
*/
void configure(const ClCompileContext &compile_context, ITensorInfo *src, ITensorInfo *bias, ITensorInfo *dst, const WinogradInfo &winograd_info,
const ActivationLayerInfo &act_info = ActivationLayerInfo());
-
/** Static function to check if given info will lead to a valid configuration
*
* Similar to ClWinogradOutputTransformKernel::configure()
@@ -84,4 +82,4 @@ private:
} // namespace kernels
} // namespace opencl
} // namespace arm_compute
-#endif /*ARM_COMPUTE_CL_WINOGRAD_OUTPUT_TRANSFORM_KERNEL_H */
+#endif /* ARM_COMPUTE_CL_WINOGRAD_OUTPUT_TRANSFORM_KERNEL_H */