diff options
author | SiCong Li <sicong.li@arm.com> | 2020-11-17 14:09:01 +0000 |
---|---|---|
committer | SiCong Li <sicong.li@arm.com> | 2020-11-17 16:38:52 +0000 |
commit | 04a0706dddc6ca24cb80e3e0789c6b0f54c48b28 (patch) | |
tree | eb136b5fd7bd4f3e9424cf5bcf4504e5a4f158d1 /src/core/CL/kernels/CLGEMMLowpQuantizeDownInt32ScaleByFixedPointKernel.cpp | |
parent | 1d1bca75f766625140ab0fdf000b6336f013ed83 (diff) | |
download | ComputeLibrary-04a0706dddc6ca24cb80e3e0789c6b0f54c48b28.tar.gz |
COMPMID-3979 Sanitise Padding Removal epic
* Add missing padding immutability asserts in all relevant CL kernels
* Remove unnecessary zero padding validation tests.
Change-Id: If93f9ccbc988e0286f5e7b135f812141476d5da0
Signed-off-by: SiCong Li <sicong.li@arm.com>
Reviewed-on: https://review.mlplatform.org/c/ml/ComputeLibrary/+/4446
Reviewed-by: Gian Marco Iodice <gianmarco.iodice@arm.com>
Comments-Addressed: Arm Jenkins <bsgcomp@arm.com>
Tested-by: Arm Jenkins <bsgcomp@arm.com>
Diffstat (limited to 'src/core/CL/kernels/CLGEMMLowpQuantizeDownInt32ScaleByFixedPointKernel.cpp')
-rw-r--r-- | src/core/CL/kernels/CLGEMMLowpQuantizeDownInt32ScaleByFixedPointKernel.cpp | 4 |
1 files changed, 4 insertions, 0 deletions
diff --git a/src/core/CL/kernels/CLGEMMLowpQuantizeDownInt32ScaleByFixedPointKernel.cpp b/src/core/CL/kernels/CLGEMMLowpQuantizeDownInt32ScaleByFixedPointKernel.cpp index d0f016879e..6a58d5e202 100644 --- a/src/core/CL/kernels/CLGEMMLowpQuantizeDownInt32ScaleByFixedPointKernel.cpp +++ b/src/core/CL/kernels/CLGEMMLowpQuantizeDownInt32ScaleByFixedPointKernel.cpp @@ -84,6 +84,8 @@ void CLGEMMLowpQuantizeDownInt32ScaleByFixedPointKernel::configure(const CLCompi ARM_COMPUTE_ERROR_ON_NULLPTR(input, output); ARM_COMPUTE_ERROR_THROW_ON(validate_arguments(input->info(), (bias != nullptr) ? bias->info() : nullptr, output->info(), info)); + auto padding_info = get_padding_info({ input, bias, output }); + // Output auto inizialitation if not yet initialized auto_init_if_empty(*output->info(), input->info()->clone()->set_data_type(info->output_data_type)); @@ -116,6 +118,8 @@ void CLGEMMLowpQuantizeDownInt32ScaleByFixedPointKernel::configure(const CLCompi // Configure kernel window auto win = calculate_max_window(*output->info(), Steps(num_elems_processed_per_iteration)); ICLKernel::configure_internal(win); + + ARM_COMPUTE_ERROR_ON(has_padding_changed(padding_info)); } void CLGEMMLowpQuantizeDownInt32ScaleByFixedPointKernel::run(const Window &window, cl::CommandQueue &queue) |