aboutsummaryrefslogtreecommitdiff
path: root/src/core/NEON/kernels/arm_gemm/kernels/a64_hybrid_fp32_mla_6x16/a55.cpp
diff options
context:
space:
mode:
authorGeorgios Pinitas <georgios.pinitas@arm.com>2021-07-16 16:16:43 +0100
committerGeorgios Pinitas <georgios.pinitas@arm.com>2021-07-22 02:25:50 +0000
commit4ee8b1599dbaf7634d25607fa5ac96ba3dc6b0f2 (patch)
tree2f8362d33cdad4212f4b96995681c68184c759e1 /src/core/NEON/kernels/arm_gemm/kernels/a64_hybrid_fp32_mla_6x16/a55.cpp
parent59fd7a722e5bc7e85309d6200bc37a772721a719 (diff)
downloadComputeLibrary-4ee8b1599dbaf7634d25607fa5ac96ba3dc6b0f2.tar.gz
Update GEMM assembly kernels
- Introduce Fp32 kernels with internal calculations in Bfloat16 when fast_mode is enabled - Improve kernel selection heuristics Signed-off-by: Georgios Pinitas <georgios.pinitas@arm.com> Change-Id: I68a9e7e862b6fd2721b46e0d7cc791091c4ab279 Reviewed-on: https://review.mlplatform.org/c/ml/ComputeLibrary/+/5965 Tested-by: Arm Jenkins <bsgcomp@arm.com> Comments-Addressed: Arm Jenkins <bsgcomp@arm.com>
Diffstat (limited to 'src/core/NEON/kernels/arm_gemm/kernels/a64_hybrid_fp32_mla_6x16/a55.cpp')
-rw-r--r--src/core/NEON/kernels/arm_gemm/kernels/a64_hybrid_fp32_mla_6x16/a55.cpp52
1 files changed, 26 insertions, 26 deletions
diff --git a/src/core/NEON/kernels/arm_gemm/kernels/a64_hybrid_fp32_mla_6x16/a55.cpp b/src/core/NEON/kernels/arm_gemm/kernels/a64_hybrid_fp32_mla_6x16/a55.cpp
index 184cfaf95c..e8b7db21bd 100644
--- a/src/core/NEON/kernels/arm_gemm/kernels/a64_hybrid_fp32_mla_6x16/a55.cpp
+++ b/src/core/NEON/kernels/arm_gemm/kernels/a64_hybrid_fp32_mla_6x16/a55.cpp
@@ -839,14 +839,14 @@ void a64_hybrid_fp32_mla_6x16_a55 (
"ldr q9, [x16, #0x10]\n"
"ldr q10, [x16, #0x20]\n"
"mov v12.16b, v8.16b\n"
- "mov v16.16b, v8.16b\n"
- "mov v13.16b, v9.16b\n"
- "mov v17.16b, v9.16b\n"
- "mov v14.16b, v10.16b\n"
- "mov v18.16b, v10.16b\n"
"ldr q11, [x16, #0x30]\n"
+ "mov v13.16b, v9.16b\n"
"add x16, x16, #0x40\n"
+ "mov v14.16b, v10.16b\n"
"mov v15.16b, v11.16b\n"
+ "mov v16.16b, v8.16b\n"
+ "mov v17.16b, v9.16b\n"
+ "mov v18.16b, v10.16b\n"
"mov v19.16b, v11.16b\n"
"b 80f\n"
"69:" // Height 3: no bias
@@ -1364,18 +1364,18 @@ void a64_hybrid_fp32_mla_6x16_a55 (
"ldr q9, [x16, #0x10]\n"
"ldr q10, [x16, #0x20]\n"
"mov v12.16b, v8.16b\n"
- "mov v16.16b, v8.16b\n"
+ "ldr q11, [x16, #0x30]\n"
"mov v13.16b, v9.16b\n"
- "mov v17.16b, v9.16b\n"
+ "add x16, x16, #0x40\n"
"mov v14.16b, v10.16b\n"
+ "mov v15.16b, v11.16b\n"
+ "mov v16.16b, v8.16b\n"
+ "mov v17.16b, v9.16b\n"
"mov v18.16b, v10.16b\n"
+ "mov v19.16b, v11.16b\n"
"mov v20.16b, v8.16b\n"
"mov v21.16b, v9.16b\n"
"mov v22.16b, v10.16b\n"
- "ldr q11, [x16, #0x30]\n"
- "add x16, x16, #0x40\n"
- "mov v15.16b, v11.16b\n"
- "mov v19.16b, v11.16b\n"
"mov v23.16b, v11.16b\n"
"b 113f\n"
"102:" // Height 4: no bias
@@ -1996,22 +1996,22 @@ void a64_hybrid_fp32_mla_6x16_a55 (
"ldr q9, [x16, #0x10]\n"
"ldr q10, [x16, #0x20]\n"
"mov v12.16b, v8.16b\n"
- "mov v16.16b, v8.16b\n"
+ "ldr q11, [x16, #0x30]\n"
"mov v13.16b, v9.16b\n"
- "mov v17.16b, v9.16b\n"
+ "add x16, x16, #0x40\n"
"mov v14.16b, v10.16b\n"
+ "mov v15.16b, v11.16b\n"
+ "mov v16.16b, v8.16b\n"
+ "mov v17.16b, v9.16b\n"
"mov v18.16b, v10.16b\n"
+ "mov v19.16b, v11.16b\n"
"mov v20.16b, v8.16b\n"
"mov v21.16b, v9.16b\n"
"mov v22.16b, v10.16b\n"
+ "mov v23.16b, v11.16b\n"
"mov v24.16b, v8.16b\n"
"mov v25.16b, v9.16b\n"
"mov v26.16b, v10.16b\n"
- "ldr q11, [x16, #0x30]\n"
- "add x16, x16, #0x40\n"
- "mov v15.16b, v11.16b\n"
- "mov v19.16b, v11.16b\n"
- "mov v23.16b, v11.16b\n"
"mov v27.16b, v11.16b\n"
"b 146f\n"
"135:" // Height 5: no bias
@@ -2738,26 +2738,26 @@ void a64_hybrid_fp32_mla_6x16_a55 (
"ldr q9, [x16, #0x10]\n"
"ldr q10, [x16, #0x20]\n"
"mov v12.16b, v8.16b\n"
- "mov v16.16b, v8.16b\n"
+ "ldr q11, [x16, #0x30]\n"
"mov v13.16b, v9.16b\n"
- "mov v17.16b, v9.16b\n"
+ "add x16, x16, #0x40\n"
"mov v14.16b, v10.16b\n"
+ "mov v15.16b, v11.16b\n"
+ "mov v16.16b, v8.16b\n"
+ "mov v17.16b, v9.16b\n"
"mov v18.16b, v10.16b\n"
+ "mov v19.16b, v11.16b\n"
"mov v20.16b, v8.16b\n"
"mov v21.16b, v9.16b\n"
"mov v22.16b, v10.16b\n"
+ "mov v23.16b, v11.16b\n"
"mov v24.16b, v8.16b\n"
"mov v25.16b, v9.16b\n"
"mov v26.16b, v10.16b\n"
+ "mov v27.16b, v11.16b\n"
"mov v28.16b, v8.16b\n"
"mov v29.16b, v9.16b\n"
"mov v30.16b, v10.16b\n"
- "ldr q11, [x16, #0x30]\n"
- "add x16, x16, #0x40\n"
- "mov v15.16b, v11.16b\n"
- "mov v19.16b, v11.16b\n"
- "mov v23.16b, v11.16b\n"
- "mov v27.16b, v11.16b\n"
"mov v31.16b, v11.16b\n"
"b 179f\n"
"168:" // Height 6: no bias