aboutsummaryrefslogtreecommitdiff
path: root/src/core/NEON/kernels/arm_conv/pooling/kernels/a64_u8_nhwc_avg_generic_depthfirst/generic.cpp
diff options
context:
space:
mode:
Diffstat (limited to 'src/core/NEON/kernels/arm_conv/pooling/kernels/a64_u8_nhwc_avg_generic_depthfirst/generic.cpp')
-rw-r--r--src/core/NEON/kernels/arm_conv/pooling/kernels/a64_u8_nhwc_avg_generic_depthfirst/generic.cpp424
1 files changed, 212 insertions, 212 deletions
diff --git a/src/core/NEON/kernels/arm_conv/pooling/kernels/a64_u8_nhwc_avg_generic_depthfirst/generic.cpp b/src/core/NEON/kernels/arm_conv/pooling/kernels/a64_u8_nhwc_avg_generic_depthfirst/generic.cpp
index 2c8a29248d..f8984c451c 100644
--- a/src/core/NEON/kernels/arm_conv/pooling/kernels/a64_u8_nhwc_avg_generic_depthfirst/generic.cpp
+++ b/src/core/NEON/kernels/arm_conv/pooling/kernels/a64_u8_nhwc_avg_generic_depthfirst/generic.cpp
@@ -1,5 +1,5 @@
/*
- * Copyright (c) 2021 Arm Limited.
+ * Copyright (c) 2021-2023 Arm Limited.
*
* SPDX-License-Identifier: MIT
*
@@ -23,6 +23,7 @@
*/
#include <cstdint>
+#include <cstddef>
#include <cstring>
#include <cmath>
@@ -84,26 +85,27 @@ void a64_u8_nhwc_avg_generic_depthfirst_impl(
f_rescale_value *= 2.0f;
}
- rescale_value = static_cast<int32_t>(round(f_rescale_value * static_cast<float>(1ll << 31)));
- if (static_cast<int64_t>(rescale_value) == (1ll << 31))
+ int64_t long_rescale_value = round(f_rescale_value * static_cast<float>(1ll << 31));
+ if (long_rescale_value == (1ll << 31))
{
shift_value++;
- rescale_value >>= 1;
+ long_rescale_value >>= 1;
}
+ rescale_value = static_cast<int32_t>(long_rescale_value);
}
__asm__ __volatile__(
- "mov x26, #0x0\n"
- "mov x25, #0x10\n" // cntb _, ALL, #1
- "mov x24, #0x20\n" // cntb _, ALL, #2
- "mov x23, #0x30\n" // cntb _, ALL, #3
"cmp %x[n_channels], #0x40\n"
+ "mov x27, #0x0\n"
+ "mov x26, #0x10\n" // cntb _, ALL, #1
+ "mov x25, #0x20\n" // cntb _, ALL, #2
+ "mov x24, #0x30\n" // cntb _, ALL, #3
"blt 7f\n"
"1:" // 4-vectors of channels
+ "lsr x23, %x[n_valid_cells], #0x1\n"
"movi v15.4s, #0x0\n"
- "mov x19, %x[inptrs]\n"
"movi v14.4s, #0x0\n"
- "lsr x22, %x[n_valid_cells], #0x1\n"
+ "mov x22, %x[inptrs]\n"
"movi v13.4s, #0x0\n"
"movi v12.4s, #0x0\n"
"movi v11.4s, #0x0\n"
@@ -118,43 +120,43 @@ void a64_u8_nhwc_avg_generic_depthfirst_impl(
"movi v2.4s, #0x0\n"
"movi v1.4s, #0x0\n"
"movi v0.4s, #0x0\n"
- "cbz x22, 4f\n"
- "ldp x21, x20, [x19, #0x0]\n"
- "ldr q31, [x21, x26]\n"
- "add x19, x19, #0x10\n"
- "ldr q30, [x20, x26]\n"
- "subs x22, x22, #0x1\n"
- "ldr q29, [x21, x25]\n"
- "ldr q28, [x20, x25]\n"
- "ldr q27, [x21, x24]\n"
- "ldr q26, [x20, x24]\n"
- "ldr q25, [x21, x23]\n"
- "ldr q24, [x20, x23]\n"
+ "cbz x23, 4f\n"
+ "ldp x21, x20, [x22, #0x0]\n"
+ "ldr q31, [x21, x27]\n"
+ "subs x23, x23, #0x1\n"
+ "add x22, x22, #0x10\n"
+ "ldr q30, [x20, x27]\n"
+ "ldr q29, [x21, x26]\n"
+ "ldr q28, [x20, x26]\n"
+ "ldr q27, [x21, x25]\n"
+ "ldr q26, [x20, x25]\n"
+ "ldr q25, [x21, x24]\n"
+ "ldr q24, [x20, x24]\n"
"beq 3f\n"
"2:" // 4-vectors of channels: 2 inputs loop
"uaddl v23.8h, v31.8b, v30.8b\n"
- "ldp x21, x20, [x19, #0x0]\n"
- "add x19, x19, #0x10\n"
"uaddl2 v22.8h, v31.16b, v30.16b\n"
- "ldr q31, [x21, x26]\n"
+ "ldp x21, x20, [x22, #0x0]\n"
+ "ldr q31, [x21, x27]\n"
+ "ldr q30, [x20, x27]\n"
"uaddl v21.8h, v29.8b, v28.8b\n"
- "subs x22, x22, #0x1\n"
"uaddl2 v20.8h, v29.16b, v28.16b\n"
- "ldr q30, [x20, x26]\n"
+ "ldr q29, [x21, x26]\n"
+ "ldr q28, [x20, x26]\n"
"uaddl v19.8h, v27.8b, v26.8b\n"
- "ldr q29, [x21, x25]\n"
"uaddl2 v18.8h, v27.16b, v26.16b\n"
- "ldr q28, [x20, x25]\n"
+ "ldr q27, [x21, x25]\n"
+ "ldr q26, [x20, x25]\n"
"uaddl v17.8h, v25.8b, v24.8b\n"
- "ldr q27, [x21, x24]\n"
"uaddl2 v16.8h, v25.16b, v24.16b\n"
- "ldr q26, [x20, x24]\n"
+ "ldr q25, [x21, x24]\n"
+ "ldr q24, [x20, x24]\n"
+ "subs x23, x23, #0x1\n"
"uaddw v15.4s, v15.4s, v23.4h\n"
- "ldr q25, [x21, x23]\n"
"uaddw2 v14.4s, v14.4s, v23.8h\n"
- "ldr q24, [x20, x23]\n"
"uaddw v13.4s, v13.4s, v22.4h\n"
"uaddw2 v12.4s, v12.4s, v22.8h\n"
+ "add x22, x22, #0x10\n"
"uaddw v11.4s, v11.4s, v21.4h\n"
"uaddw2 v10.4s, v10.4s, v21.8h\n"
"uaddw v9.4s, v9.4s, v20.4h\n"
@@ -194,23 +196,23 @@ void a64_u8_nhwc_avg_generic_depthfirst_impl(
"uaddw v1.4s, v1.4s, v16.4h\n"
"uaddw2 v0.4s, v0.4s, v16.8h\n"
"4:" // 4-vectors of channels: After loop
- "ands x20, %x[n_valid_cells], #0x1\n"
+ "ands x23, %x[n_valid_cells], #0x1\n"
"beq 6f\n"
"5:" // 4-vectors of channels: Single input loop
- "ldr x21, [x19], #0x8\n"
- "subs x20, x20, #0x1\n"
- "ldr q31, [x21, x26]\n"
- "uxtl v23.8h, v31.8b\n"
- "ldr q29, [x21, x25]\n"
- "uxtl2 v22.8h, v31.16b\n"
- "ldr q27, [x21, x24]\n"
- "ldr q25, [x21, x23]\n"
- "uxtl v21.8h, v29.8b\n"
- "uxtl2 v20.8h, v29.16b\n"
- "uxtl v19.8h, v27.8b\n"
- "uxtl2 v18.8h, v27.16b\n"
- "uxtl v17.8h, v25.8b\n"
- "uxtl2 v16.8h, v25.16b\n"
+ "ldr x20, [x22], #0x8\n"
+ "ldr q16, [x20, x27]\n"
+ "uxtl v23.8h, v16.8b\n"
+ "uxtl2 v22.8h, v16.16b\n"
+ "ldr q16, [x20, x26]\n"
+ "ldr q17, [x20, x25]\n"
+ "uxtl v21.8h, v16.8b\n"
+ "uxtl2 v20.8h, v16.16b\n"
+ "ldr q16, [x20, x24]\n"
+ "uxtl v19.8h, v17.8b\n"
+ "uxtl2 v18.8h, v17.16b\n"
+ "subs x23, x23, #0x1\n"
+ "uxtl v17.8h, v16.8b\n"
+ "uxtl2 v16.8h, v16.16b\n"
"uaddw v15.4s, v15.4s, v23.4h\n"
"uaddw2 v14.4s, v14.4s, v23.8h\n"
"uaddw v13.4s, v13.4s, v22.4h\n"
@@ -229,195 +231,195 @@ void a64_u8_nhwc_avg_generic_depthfirst_impl(
"uaddw2 v0.4s, v0.4s, v16.8h\n"
"bgt 5b\n"
"6:" // 4-vectors of channels: Single input loop: End
- "movi v19.4s, #0x0\n"
- "ld1r { v18.4s }, [%x[rescale_ptr]]\n"
- "sub %x[n_channels], %x[n_channels], #0x40\n"
- "movi v17.4s, #0xff\n"
+ "ld1r { v17.4s }, [%x[rescale_ptr]]\n"
"ld1r { v16.4s }, [%x[shift_ptr]]\n"
+ "sqdmulh v15.4s, v15.4s, v17.4s\n"
+ "sqdmulh v14.4s, v14.4s, v17.4s\n"
+ "sqdmulh v13.4s, v13.4s, v17.4s\n"
+ "sqdmulh v12.4s, v12.4s, v17.4s\n"
+ "sub %x[n_channels], %x[n_channels], #0x40\n"
"cmp %x[n_channels], #0x40\n"
- "sqdmulh v15.4s, v15.4s, v18.4s\n"
- "sqdmulh v14.4s, v14.4s, v18.4s\n"
- "sqdmulh v13.4s, v13.4s, v18.4s\n"
- "sqdmulh v12.4s, v12.4s, v18.4s\n"
- "sqdmulh v11.4s, v11.4s, v18.4s\n"
+ "sqdmulh v11.4s, v11.4s, v17.4s\n"
+ "sqdmulh v10.4s, v10.4s, v17.4s\n"
+ "sqdmulh v9.4s, v9.4s, v17.4s\n"
+ "sqdmulh v8.4s, v8.4s, v17.4s\n"
+ "sqdmulh v7.4s, v7.4s, v17.4s\n"
+ "sqdmulh v6.4s, v6.4s, v17.4s\n"
+ "sqdmulh v5.4s, v5.4s, v17.4s\n"
+ "sqdmulh v4.4s, v4.4s, v17.4s\n"
+ "sqdmulh v3.4s, v3.4s, v17.4s\n"
+ "sqdmulh v2.4s, v2.4s, v17.4s\n"
+ "sqdmulh v1.4s, v1.4s, v17.4s\n"
+ "sqdmulh v0.4s, v0.4s, v17.4s\n"
"srshl v15.4s, v15.4s, v16.4s\n"
"srshl v14.4s, v14.4s, v16.4s\n"
"srshl v13.4s, v13.4s, v16.4s\n"
"srshl v12.4s, v12.4s, v16.4s\n"
"srshl v11.4s, v11.4s, v16.4s\n"
- "sqdmulh v10.4s, v10.4s, v18.4s\n"
- "sqdmulh v9.4s, v9.4s, v18.4s\n"
- "sqdmulh v8.4s, v8.4s, v18.4s\n"
- "sqdmulh v7.4s, v7.4s, v18.4s\n"
"srshl v10.4s, v10.4s, v16.4s\n"
"srshl v9.4s, v9.4s, v16.4s\n"
"srshl v8.4s, v8.4s, v16.4s\n"
"srshl v7.4s, v7.4s, v16.4s\n"
- "sqdmulh v6.4s, v6.4s, v18.4s\n"
- "sqdmulh v5.4s, v5.4s, v18.4s\n"
- "sqdmulh v4.4s, v4.4s, v18.4s\n"
- "sqdmulh v3.4s, v3.4s, v18.4s\n"
"srshl v6.4s, v6.4s, v16.4s\n"
"srshl v5.4s, v5.4s, v16.4s\n"
"srshl v4.4s, v4.4s, v16.4s\n"
"srshl v3.4s, v3.4s, v16.4s\n"
- "sqdmulh v2.4s, v2.4s, v18.4s\n"
- "sqdmulh v1.4s, v1.4s, v18.4s\n"
- "sqdmulh v0.4s, v0.4s, v18.4s\n"
- "smax v15.4s, v15.4s, v19.4s\n"
"srshl v2.4s, v2.4s, v16.4s\n"
"srshl v1.4s, v1.4s, v16.4s\n"
"srshl v0.4s, v0.4s, v16.4s\n"
- "smin v15.4s, v15.4s, v17.4s\n"
- "smax v14.4s, v14.4s, v19.4s\n"
- "smax v13.4s, v13.4s, v19.4s\n"
- "smax v12.4s, v12.4s, v19.4s\n"
- "smin v14.4s, v14.4s, v17.4s\n"
- "smin v13.4s, v13.4s, v17.4s\n"
- "smin v12.4s, v12.4s, v17.4s\n"
- "smax v11.4s, v11.4s, v19.4s\n"
- "smax v10.4s, v10.4s, v19.4s\n"
- "smax v9.4s, v9.4s, v19.4s\n"
- "smin v11.4s, v11.4s, v17.4s\n"
- "smin v10.4s, v10.4s, v17.4s\n"
- "smin v9.4s, v9.4s, v17.4s\n"
- "smax v8.4s, v8.4s, v19.4s\n"
- "smax v7.4s, v7.4s, v19.4s\n"
- "smax v6.4s, v6.4s, v19.4s\n"
- "smin v8.4s, v8.4s, v17.4s\n"
- "smin v7.4s, v7.4s, v17.4s\n"
- "smin v6.4s, v6.4s, v17.4s\n"
- "smax v5.4s, v5.4s, v19.4s\n"
- "smax v4.4s, v4.4s, v19.4s\n"
- "smax v3.4s, v3.4s, v19.4s\n"
- "smin v5.4s, v5.4s, v17.4s\n"
- "smin v4.4s, v4.4s, v17.4s\n"
- "smin v3.4s, v3.4s, v17.4s\n"
- "smax v2.4s, v2.4s, v19.4s\n"
- "smax v1.4s, v1.4s, v19.4s\n"
- "smax v0.4s, v0.4s, v19.4s\n"
- "smin v2.4s, v2.4s, v17.4s\n"
- "smin v1.4s, v1.4s, v17.4s\n"
- "smin v0.4s, v0.4s, v17.4s\n"
+ "movi v16.4s, #0x0\n"
+ "smax v15.4s, v15.4s, v16.4s\n"
+ "smax v14.4s, v14.4s, v16.4s\n"
+ "smax v13.4s, v13.4s, v16.4s\n"
+ "smax v12.4s, v12.4s, v16.4s\n"
+ "smax v11.4s, v11.4s, v16.4s\n"
+ "smax v10.4s, v10.4s, v16.4s\n"
+ "smax v9.4s, v9.4s, v16.4s\n"
+ "smax v8.4s, v8.4s, v16.4s\n"
+ "smax v7.4s, v7.4s, v16.4s\n"
+ "smax v6.4s, v6.4s, v16.4s\n"
+ "smax v5.4s, v5.4s, v16.4s\n"
+ "smax v4.4s, v4.4s, v16.4s\n"
+ "smax v3.4s, v3.4s, v16.4s\n"
+ "smax v2.4s, v2.4s, v16.4s\n"
+ "smax v1.4s, v1.4s, v16.4s\n"
+ "smax v0.4s, v0.4s, v16.4s\n"
+ "movi v16.4s, #0xff\n"
+ "smin v15.4s, v15.4s, v16.4s\n"
+ "smin v14.4s, v14.4s, v16.4s\n"
+ "smin v13.4s, v13.4s, v16.4s\n"
+ "smin v12.4s, v12.4s, v16.4s\n"
+ "smin v11.4s, v11.4s, v16.4s\n"
+ "smin v10.4s, v10.4s, v16.4s\n"
+ "smin v9.4s, v9.4s, v16.4s\n"
+ "smin v8.4s, v8.4s, v16.4s\n"
+ "smin v7.4s, v7.4s, v16.4s\n"
+ "smin v6.4s, v6.4s, v16.4s\n"
+ "smin v5.4s, v5.4s, v16.4s\n"
+ "smin v4.4s, v4.4s, v16.4s\n"
+ "smin v3.4s, v3.4s, v16.4s\n"
+ "smin v2.4s, v2.4s, v16.4s\n"
+ "smin v1.4s, v1.4s, v16.4s\n"
+ "smin v0.4s, v0.4s, v16.4s\n"
"uzp1 v23.16b, v15.16b, v14.16b\n"
"uzp1 v16.16b, v13.16b, v12.16b\n"
"uzp1 v22.16b, v11.16b, v10.16b\n"
- "uzp1 v21.16b, v9.16b, v8.16b\n"
- "uzp1 v20.16b, v7.16b, v6.16b\n"
+ "uzp1 v18.16b, v9.16b, v8.16b\n"
+ "uzp1 v21.16b, v7.16b, v6.16b\n"
"uzp1 v17.16b, v5.16b, v4.16b\n"
- "uzp1 v19.16b, v3.16b, v2.16b\n"
- "uzp1 v18.16b, v1.16b, v0.16b\n"
+ "uzp1 v20.16b, v3.16b, v2.16b\n"
+ "uzp1 v19.16b, v1.16b, v0.16b\n"
"uzp1 v16.16b, v23.16b, v16.16b\n"
- "str q16, [%x[outptr], x26]\n"
- "uzp1 v16.16b, v22.16b, v21.16b\n"
+ "uzp1 v18.16b, v22.16b, v18.16b\n"
+ "str q16, [%x[outptr], x27]\n"
+ "add x27, x27, #0x40\n"
+ "uzp1 v17.16b, v21.16b, v17.16b\n"
+ "uzp1 v16.16b, v20.16b, v19.16b\n"
+ "str q18, [%x[outptr], x26]\n"
"add x26, x26, #0x40\n"
- "uzp1 v17.16b, v20.16b, v17.16b\n"
- "str q16, [%x[outptr], x25]\n"
- "uzp1 v16.16b, v19.16b, v18.16b\n"
+ "str q17, [%x[outptr], x25]\n"
"add x25, x25, #0x40\n"
- "str q17, [%x[outptr], x24]\n"
+ "str q16, [%x[outptr], x24]\n"
"add x24, x24, #0x40\n"
- "str q16, [%x[outptr], x23]\n"
- "add x23, x23, #0x40\n"
"bge 1b\n"
"cbz %x[n_channels], 43f\n"
"7:" // Single vector of channels
"cmp %x[n_channels], #0x10\n"
"blt 14f\n"
"8:" // Single vector of channels: Loop
+ "lsr x23, %x[n_valid_cells], #0x1\n"
"movi v15.4s, #0x0\n"
- "mov x19, %x[inptrs]\n"
"movi v14.4s, #0x0\n"
- "lsr x22, %x[n_valid_cells], #0x1\n"
+ "mov x22, %x[inptrs]\n"
"movi v13.4s, #0x0\n"
"movi v12.4s, #0x0\n"
- "cbz x22, 11f\n"
- "ldp x21, x20, [x19, #0x0]\n"
- "ldr q31, [x21, x26]\n"
- "add x19, x19, #0x10\n"
- "ldr q30, [x20, x26]\n"
- "subs x22, x22, #0x1\n"
+ "cbz x23, 11f\n"
+ "ldp x21, x20, [x22, #0x0]\n"
+ "ldr q31, [x21, x27]\n"
+ "subs x23, x23, #0x1\n"
+ "add x22, x22, #0x10\n"
+ "ldr q30, [x20, x27]\n"
"beq 10f\n"
"9:" // Single vector of channels: Loop: 2 inputs loop
- "uaddl v23.8h, v31.8b, v30.8b\n"
- "ldp x21, x20, [x19, #0x0]\n"
- "add x19, x19, #0x10\n"
- "uaddl2 v22.8h, v31.16b, v30.16b\n"
- "ldr q31, [x21, x26]\n"
- "uaddw v15.4s, v15.4s, v23.4h\n"
- "subs x22, x22, #0x1\n"
- "uaddw2 v14.4s, v14.4s, v23.8h\n"
- "ldr q30, [x20, x26]\n"
- "uaddw v13.4s, v13.4s, v22.4h\n"
- "uaddw2 v12.4s, v12.4s, v22.8h\n"
+ "uaddl v17.8h, v31.8b, v30.8b\n"
+ "uaddl2 v16.8h, v31.16b, v30.16b\n"
+ "ldp x21, x20, [x22, #0x0]\n"
+ "ldr q31, [x21, x27]\n"
+ "ldr q30, [x20, x27]\n"
+ "subs x23, x23, #0x1\n"
+ "uaddw v15.4s, v15.4s, v17.4h\n"
+ "uaddw2 v14.4s, v14.4s, v17.8h\n"
+ "uaddw v13.4s, v13.4s, v16.4h\n"
+ "uaddw2 v12.4s, v12.4s, v16.8h\n"
+ "add x22, x22, #0x10\n"
"bgt 9b\n"
"10:" // Single vector of channels: Loop: 2 inputs tail
- "uaddl v23.8h, v31.8b, v30.8b\n"
- "uaddl2 v22.8h, v31.16b, v30.16b\n"
- "uaddw v15.4s, v15.4s, v23.4h\n"
- "uaddw2 v14.4s, v14.4s, v23.8h\n"
- "uaddw v13.4s, v13.4s, v22.4h\n"
- "uaddw2 v12.4s, v12.4s, v22.8h\n"
+ "uaddl v17.8h, v31.8b, v30.8b\n"
+ "uaddl2 v16.8h, v31.16b, v30.16b\n"
+ "uaddw v15.4s, v15.4s, v17.4h\n"
+ "uaddw2 v14.4s, v14.4s, v17.8h\n"
+ "uaddw v13.4s, v13.4s, v16.4h\n"
+ "uaddw2 v12.4s, v12.4s, v16.8h\n"
"11:" // Single vector of channels: Loop: After loop
- "ands x20, %x[n_valid_cells], #0x1\n"
+ "ands x23, %x[n_valid_cells], #0x1\n"
"beq 13f\n"
"12:" // Single vector of channels: Loop: Single input loop
- "ldr x21, [x19], #0x8\n"
- "subs x20, x20, #0x1\n"
- "ldr q31, [x21, x26]\n"
- "uxtl v23.8h, v31.8b\n"
- "uxtl2 v22.8h, v31.16b\n"
- "uaddw v15.4s, v15.4s, v23.4h\n"
- "uaddw2 v14.4s, v14.4s, v23.8h\n"
- "uaddw v13.4s, v13.4s, v22.4h\n"
- "uaddw2 v12.4s, v12.4s, v22.8h\n"
+ "ldr x20, [x22], #0x8\n"
+ "ldr q16, [x20, x27]\n"
+ "uxtl v17.8h, v16.8b\n"
+ "uxtl2 v16.8h, v16.16b\n"
+ "subs x23, x23, #0x1\n"
+ "uaddw v15.4s, v15.4s, v17.4h\n"
+ "uaddw2 v14.4s, v14.4s, v17.8h\n"
+ "uaddw v13.4s, v13.4s, v16.4h\n"
+ "uaddw2 v12.4s, v12.4s, v16.8h\n"
"bgt 12b\n"
"13:" // Single vector of channels: Loop: Single input loop: End
- "movi v19.4s, #0x0\n"
- "ld1r { v18.4s }, [%x[rescale_ptr]]\n"
- "sub %x[n_channels], %x[n_channels], #0x10\n"
- "movi v17.4s, #0xff\n"
+ "ld1r { v17.4s }, [%x[rescale_ptr]]\n"
"ld1r { v16.4s }, [%x[shift_ptr]]\n"
+ "sqdmulh v15.4s, v15.4s, v17.4s\n"
+ "sqdmulh v14.4s, v14.4s, v17.4s\n"
+ "sqdmulh v13.4s, v13.4s, v17.4s\n"
+ "sqdmulh v12.4s, v12.4s, v17.4s\n"
+ "sub %x[n_channels], %x[n_channels], #0x10\n"
"cmp %x[n_channels], #0x10\n"
- "sqdmulh v15.4s, v15.4s, v18.4s\n"
- "sqdmulh v14.4s, v14.4s, v18.4s\n"
- "sqdmulh v13.4s, v13.4s, v18.4s\n"
- "sqdmulh v12.4s, v12.4s, v18.4s\n"
"srshl v15.4s, v15.4s, v16.4s\n"
"srshl v14.4s, v14.4s, v16.4s\n"
"srshl v13.4s, v13.4s, v16.4s\n"
"srshl v12.4s, v12.4s, v16.4s\n"
- "smax v15.4s, v15.4s, v19.4s\n"
- "smax v14.4s, v14.4s, v19.4s\n"
- "smax v13.4s, v13.4s, v19.4s\n"
- "smax v12.4s, v12.4s, v19.4s\n"
- "smin v15.4s, v15.4s, v17.4s\n"
- "smin v14.4s, v14.4s, v17.4s\n"
- "smin v13.4s, v13.4s, v17.4s\n"
- "smin v12.4s, v12.4s, v17.4s\n"
- "uzp1 v23.16b, v15.16b, v14.16b\n"
+ "movi v16.4s, #0x0\n"
+ "smax v15.4s, v15.4s, v16.4s\n"
+ "smax v14.4s, v14.4s, v16.4s\n"
+ "smax v13.4s, v13.4s, v16.4s\n"
+ "smax v12.4s, v12.4s, v16.4s\n"
+ "movi v16.4s, #0xff\n"
+ "smin v15.4s, v15.4s, v16.4s\n"
+ "smin v14.4s, v14.4s, v16.4s\n"
+ "smin v13.4s, v13.4s, v16.4s\n"
+ "smin v12.4s, v12.4s, v16.4s\n"
+ "uzp1 v17.16b, v15.16b, v14.16b\n"
"uzp1 v16.16b, v13.16b, v12.16b\n"
- "uzp1 v16.16b, v23.16b, v16.16b\n"
- "str q16, [%x[outptr], x26]\n"
- "add x26, x26, #0x10\n"
+ "uzp1 v16.16b, v17.16b, v16.16b\n"
+ "str q16, [%x[outptr], x27]\n"
+ "add x27, x27, #0x10\n"
"bge 8b\n"
"cbz %x[n_channels], 43f\n"
"14:" // Oddments
+ "lsr x23, %x[n_valid_cells], #0x1\n"
+ "add %x[outptr], %x[outptr], x27\n"
"movi v15.4s, #0x0\n"
- "add %x[outptr], %x[outptr], x26\n"
"movi v14.4s, #0x0\n"
- "mov x19, %x[inptrs]\n"
"movi v13.4s, #0x0\n"
- "lsr x22, %x[n_valid_cells], #0x1\n"
"movi v12.4s, #0x0\n"
- "cbz x22, 24f\n"
+ "mov x22, %x[inptrs]\n"
+ "cbz x23, 24f\n"
"15:" // Oddments: 2 inputs loop
+ "ldp x21, x20, [x22, #0x0]\n"
+ "add x22, x22, #0x10\n"
+ "add x21, x21, x27\n"
"movi v31.16b, #0x0\n"
- "ldp x21, x20, [x19, #0x0]\n"
- "add x19, x19, #0x10\n"
+ "add x20, x20, x27\n"
"movi v30.16b, #0x0\n"
- "add x21, x21, x26\n"
- "add x20, x20, x26\n"
"tbz %x[n_channels], #3, 19f\n"
"ldr d31, [x21], #0x8\n"
"ldr d30, [x20], #0x8\n"
@@ -478,21 +480,21 @@ void a64_u8_nhwc_avg_generic_depthfirst_impl(
"ldr b31, [x21], #0x1\n"
"ldr b30, [x20], #0x1\n"
"23:" // Oddments: 2 inputs loop: Load: Bit 3: End
- "uaddl v23.8h, v31.8b, v30.8b\n"
- "subs x22, x22, #0x1\n"
- "uaddl2 v22.8h, v31.16b, v30.16b\n"
- "uaddw v15.4s, v15.4s, v23.4h\n"
- "uaddw2 v14.4s, v14.4s, v23.8h\n"
- "uaddw v13.4s, v13.4s, v22.4h\n"
- "uaddw2 v12.4s, v12.4s, v22.8h\n"
+ "uaddl v17.8h, v31.8b, v30.8b\n"
+ "uaddl2 v16.8h, v31.16b, v30.16b\n"
+ "subs x23, x23, #0x1\n"
+ "uaddw v15.4s, v15.4s, v17.4h\n"
+ "uaddw2 v14.4s, v14.4s, v17.8h\n"
+ "uaddw v13.4s, v13.4s, v16.4h\n"
+ "uaddw2 v12.4s, v12.4s, v16.8h\n"
"bgt 15b\n"
"24:" // Oddments: After loop
- "ands x20, %x[n_valid_cells], #0x1\n"
+ "ands x23, %x[n_valid_cells], #0x1\n"
"beq 34f\n"
"25:" // Oddments: Single input loop
+ "ldr x21, [x22], #0x8\n"
+ "add x21, x21, x27\n"
"movi v31.16b, #0x0\n"
- "ldr x21, [x19], #0x8\n"
- "add x21, x21, x26\n"
"tbz %x[n_channels], #3, 29f\n"
"ldr d31, [x21], #0x8\n"
"tbz %x[n_channels], #2, 27f\n"
@@ -538,38 +540,38 @@ void a64_u8_nhwc_avg_generic_depthfirst_impl(
"tbz %x[n_channels], #0, 33f\n"
"ldr b31, [x21], #0x1\n"
"33:" // Oddments: Single input loop: Load: Bit 3: End
- "uxtl v23.8h, v31.8b\n"
- "subs x20, x20, #0x1\n"
- "uxtl2 v22.8h, v31.16b\n"
- "uaddw v15.4s, v15.4s, v23.4h\n"
- "uaddw2 v14.4s, v14.4s, v23.8h\n"
- "uaddw v13.4s, v13.4s, v22.4h\n"
- "uaddw2 v12.4s, v12.4s, v22.8h\n"
+ "uxtl v17.8h, v31.8b\n"
+ "uxtl2 v16.8h, v31.16b\n"
+ "subs x23, x23, #0x1\n"
+ "uaddw v15.4s, v15.4s, v17.4h\n"
+ "uaddw2 v14.4s, v14.4s, v17.8h\n"
+ "uaddw v13.4s, v13.4s, v16.4h\n"
+ "uaddw2 v12.4s, v12.4s, v16.8h\n"
"bgt 25b\n"
"34:" // Oddments: Single input loop: End
- "movi v19.4s, #0x0\n"
- "ld1r { v18.4s }, [%x[rescale_ptr]]\n"
- "movi v17.4s, #0xff\n"
+ "ld1r { v17.4s }, [%x[rescale_ptr]]\n"
"ld1r { v16.4s }, [%x[shift_ptr]]\n"
- "sqdmulh v15.4s, v15.4s, v18.4s\n"
- "sqdmulh v14.4s, v14.4s, v18.4s\n"
- "sqdmulh v13.4s, v13.4s, v18.4s\n"
- "sqdmulh v12.4s, v12.4s, v18.4s\n"
+ "sqdmulh v15.4s, v15.4s, v17.4s\n"
+ "sqdmulh v14.4s, v14.4s, v17.4s\n"
+ "sqdmulh v13.4s, v13.4s, v17.4s\n"
+ "sqdmulh v12.4s, v12.4s, v17.4s\n"
"srshl v15.4s, v15.4s, v16.4s\n"
"srshl v14.4s, v14.4s, v16.4s\n"
"srshl v13.4s, v13.4s, v16.4s\n"
"srshl v12.4s, v12.4s, v16.4s\n"
- "smax v15.4s, v15.4s, v19.4s\n"
- "smax v14.4s, v14.4s, v19.4s\n"
- "smax v13.4s, v13.4s, v19.4s\n"
- "smax v12.4s, v12.4s, v19.4s\n"
- "smin v15.4s, v15.4s, v17.4s\n"
- "smin v14.4s, v14.4s, v17.4s\n"
- "smin v13.4s, v13.4s, v17.4s\n"
- "smin v12.4s, v12.4s, v17.4s\n"
- "uzp1 v23.16b, v15.16b, v14.16b\n"
+ "movi v16.4s, #0x0\n"
+ "smax v15.4s, v15.4s, v16.4s\n"
+ "smax v14.4s, v14.4s, v16.4s\n"
+ "smax v13.4s, v13.4s, v16.4s\n"
+ "smax v12.4s, v12.4s, v16.4s\n"
+ "movi v16.4s, #0xff\n"
+ "smin v15.4s, v15.4s, v16.4s\n"
+ "smin v14.4s, v14.4s, v16.4s\n"
+ "smin v13.4s, v13.4s, v16.4s\n"
+ "smin v12.4s, v12.4s, v16.4s\n"
+ "uzp1 v17.16b, v15.16b, v14.16b\n"
"uzp1 v16.16b, v13.16b, v12.16b\n"
- "uzp1 v16.16b, v23.16b, v16.16b\n"
+ "uzp1 v16.16b, v17.16b, v16.16b\n"
"tbz %x[n_channels], #3, 38f\n"
"st1 { v16.d }[0], [%x[outptr]], #0x8\n"
"tbz %x[n_channels], #2, 36f\n"
@@ -615,12 +617,10 @@ void a64_u8_nhwc_avg_generic_depthfirst_impl(
"tbz %x[n_channels], #0, 42f\n"
"st1 { v16.b }[0], [%x[outptr]], #0x1\n"
"42:" // Oddments: Store: Bit 3: End
-
"43:" // End
-
: [n_channels] "+&r" (n_channels), [outptr] "+&r" (outptr)
: [inptrs] "r" (inptrs), [n_valid_cells] "r" (n_valid_cells), [rescale_ptr] "r" (&rescale_value), [shift_ptr] "r" (&shift_value)
- : "cc", "memory", "v0", "v1", "v2", "v3", "v4", "v5", "v6", "v7", "v8", "v9", "v10", "v11", "v12", "v13", "v14", "v15", "v16", "v17", "v18", "v19", "v20", "v21", "v22", "v23", "v24", "v25", "v26", "v27", "v28", "v29", "v30", "v31", "x19", "x20", "x21", "x22", "x23", "x24", "x25", "x26"
+ : "cc", "memory", "v0", "v1", "v2", "v3", "v4", "v5", "v6", "v7", "v8", "v9", "v10", "v11", "v12", "v13", "v14", "v15", "v16", "v17", "v18", "v19", "v20", "v21", "v22", "v23", "v24", "v25", "v26", "v27", "v28", "v29", "v30", "v31", "x20", "x21", "x22", "x23", "x24", "x25", "x26", "x27"
);
}