diff options
author | Georgios Pinitas <georgios.pinitas@arm.com> | 2019-10-14 19:03:09 +0100 |
---|---|---|
committer | Georgios Pinitas <georgios.pinitas@arm.com> | 2019-10-23 12:08:12 +0000 |
commit | 48b3ef89de5f21a0169d8416e3d54081f82c7bf8 (patch) | |
tree | f857d733ccf446c704823dc7ac796a96eb55095e /src/core/NEON/kernels/arm_gemm/kernels/a64_hybrid_fp32_mla_16x4/a55.cpp | |
parent | 1dce3101ef8d77c8cf0af7dfd4af6595a0136b91 (diff) | |
download | ComputeLibrary-48b3ef89de5f21a0169d8416e3d54081f82c7bf8.tar.gz |
COMPMID-2577: Fuse bias addition and activation in gemm assembly kernels
Change-Id: I7f52112d2d05b1ea3d3f3d4b19b8eafab05d6c44
Signed-off-by: Georgios Pinitas <georgios.pinitas@arm.com>
Reviewed-on: https://review.mlplatform.org/c/2141
Comments-Addressed: Arm Jenkins <bsgcomp@arm.com>
Tested-by: Arm Jenkins <bsgcomp@arm.com>
Reviewed-by: Pablo Marquez <pablo.tello@arm.com>
Diffstat (limited to 'src/core/NEON/kernels/arm_gemm/kernels/a64_hybrid_fp32_mla_16x4/a55.cpp')
-rw-r--r-- | src/core/NEON/kernels/arm_gemm/kernels/a64_hybrid_fp32_mla_16x4/a55.cpp | 452 |
1 files changed, 258 insertions, 194 deletions
diff --git a/src/core/NEON/kernels/arm_gemm/kernels/a64_hybrid_fp32_mla_16x4/a55.cpp b/src/core/NEON/kernels/arm_gemm/kernels/a64_hybrid_fp32_mla_16x4/a55.cpp index 7261761d7e..e3325f7728 100644 --- a/src/core/NEON/kernels/arm_gemm/kernels/a64_hybrid_fp32_mla_16x4/a55.cpp +++ b/src/core/NEON/kernels/arm_gemm/kernels/a64_hybrid_fp32_mla_16x4/a55.cpp @@ -25,20 +25,41 @@ #include <algorithm> +#include "arm_gemm.hpp" #include "../../asmlib.hpp" #include "../../utils.hpp" namespace arm_gemm { -void a64_hybrid_fp32_mla_16x4_a55(const float *A, int lda, const float *B, float *C, int ldc, float beta, int M, int N, int K) { - const long beta0 = (beta == 0.0f); +void a64_hybrid_fp32_mla_16x4_a55(const float *A, int lda, const float *B, float *C, int ldc, int M, int N, int K, const float *bias, Activation act, bool append) { const int K_stride = K; const long loops_count = ((K + 4) / 8) - 1; K -= loops_count * 8; const long regs_count = (K / 4) - 1; K -= (regs_count + 1) * 4; const long blocks_count = K / 1; + float nullbias[16]; + if (!append && !bias) { + memset(nullbias, 0, (16 * sizeof(float))); + } + float minval = - static_cast<float>(std::numeric_limits<float>::infinity()); + float maxval = static_cast<float>(std::numeric_limits<float>::infinity()); + const float * const minptr = &minval; + const float * const maxptr = &maxval; + + switch(act.type) + { + default: + case Activation::Type::None: + break; + case Activation::Type::BoundedReLU: + maxval = static_cast<float>(act.param1); + /* fall through */ + case Activation::Type::ReLU: + minval = 0.0f; + break; + } for (int y=0; y<M; y+=4) { const float * const a_ptr0_base = A + (y * lda); @@ -48,7 +69,6 @@ void a64_hybrid_fp32_mla_16x4_a55(const float *A, int lda, const float *B, float for (int x0=0; x0<N; x0+=16ul) { const long width = std::min((unsigned long)N-x0, 16ul); - const float *betaptr = β long loops = loops_count; long regs = regs_count; long blocks = blocks_count; @@ -58,7 +78,7 @@ void a64_hybrid_fp32_mla_16x4_a55(const float *A, int lda, const float *B, float float result_buffer[64]; const unsigned long ldcb = (use_result_buffer ? 16 : ldc) * sizeof(float); float *c_ptr_real = c_ptr0; - if (use_result_buffer && !beta0) { + if (use_result_buffer && append) { for(int cy=0; cy<std::min(M-y, 4); cy++) { for(unsigned int cx=0; cx<width; cx++) { result_buffer[cy * 16 + cx] = c_ptr_real[cy * ldc + cx]; @@ -68,6 +88,7 @@ void a64_hybrid_fp32_mla_16x4_a55(const float *A, int lda, const float *B, float if (use_result_buffer) { c_ptr0 = result_buffer; } + const float *biasptr = bias ? bias+x0 : nullbias; switch(M-y) { case 1: @@ -76,17 +97,17 @@ void a64_hybrid_fp32_mla_16x4_a55(const float *A, int lda, const float *B, float "temploadreg1 .req X1\n" "temploadreg2 .req X2\n" "temploadreg3 .req X3\n" - "cbz %[beta0], 1f\n" - "movi v16.4s, #0\n" + "cbnz %[append], 1f\n" + "ldr q16, [%[biasptr]]\n" + "ldr q17, [%[biasptr], #0x10]\n" + "ldr q18, [%[biasptr], #0x20]\n" + "ldr q19, [%[biasptr], #0x30]\n" "ldr q0, [%[a_ptr0]]\n" - "movi v17.4s, #0\n" + "add %[a_ptr0], %[a_ptr0], #0x10\n" "ldr q8, [%[b_ptr0]]\n" - "movi v18.4s, #0\n" "ldr q9, [%[b_ptr0], #0x10]\n" - "movi v19.4s, #0\n" "ldr q10, [%[b_ptr0], #0x20]\n" "ldr q11, [%[b_ptr0], #0x30]\n" - "add %[a_ptr0], %[a_ptr0], #0x10\n" "ldr q12, [%[b_ptr0], #0x40]\n" "ldr q13, [%[b_ptr0], #0x50]\n" "ldr d14, [%[b_ptr0], #0x60]\n" @@ -95,21 +116,16 @@ void a64_hybrid_fp32_mla_16x4_a55(const float *A, int lda, const float *B, float "cbz %[loops], 2f\n" "b 3f\n" "1:\n" - "ld1r {v15.4s}, [%[betaptr]]\n" "ldr q16, [%[c_ptr0]]\n" "ldr q17, [%[c_ptr0], #0x10]\n" "ldr q18, [%[c_ptr0], #0x20]\n" "ldr q19, [%[c_ptr0], #0x30]\n" - "fmul v16.4s, v16.4s, v15.4s\n" "ldr q0, [%[a_ptr0]]\n" - "fmul v17.4s, v17.4s, v15.4s\n" + "add %[a_ptr0], %[a_ptr0], #0x10\n" "ldr q8, [%[b_ptr0]]\n" - "fmul v18.4s, v18.4s, v15.4s\n" "ldr q9, [%[b_ptr0], #0x10]\n" - "fmul v19.4s, v19.4s, v15.4s\n" "ldr q10, [%[b_ptr0], #0x20]\n" "ldr q11, [%[b_ptr0], #0x30]\n" - "add %[a_ptr0], %[a_ptr0], #0x10\n" "ldr q12, [%[b_ptr0], #0x40]\n" "ldr q13, [%[b_ptr0], #0x50]\n" "ldr d14, [%[b_ptr0], #0x60]\n" @@ -434,6 +450,16 @@ void a64_hybrid_fp32_mla_16x4_a55(const float *A, int lda, const float *B, float "fmla v19.4s, v11.4s, v0.s[0]\n" "b.ne 7b\n" "6:\n" + "ld1r {v14.4s}, [%[minptr]]\n" + "ld1r {v15.4s}, [%[maxptr]]\n" + "fmax v16.4s, v16.4s, v14.4s\n" + "fmax v17.4s, v17.4s, v14.4s\n" + "fmax v18.4s, v18.4s, v14.4s\n" + "fmax v19.4s, v19.4s, v14.4s\n" + "fmin v16.4s, v16.4s, v15.4s\n" + "fmin v17.4s, v17.4s, v15.4s\n" + "fmin v18.4s, v18.4s, v15.4s\n" + "fmin v19.4s, v19.4s, v15.4s\n" "str q16, [%[c_ptr0]]\n" "str q17, [%[c_ptr0], #0x10]\n" "str q18, [%[c_ptr0], #0x20]\n" @@ -444,7 +470,7 @@ void a64_hybrid_fp32_mla_16x4_a55(const float *A, int lda, const float *B, float ".unreq temploadreg2\n" ".unreq temploadreg3\n" : [a_ptr0] "+r" (a_ptr0), [b_ptr0] "+r" (b_ptr0), [c_ptr0] "+r" (c_ptr0), [loops] "+r" (loops), [regs] "+r" (regs), [blocks] "+r" (blocks) - : [betaptr] "r" (betaptr), [width] "r" (width), [beta0] "r" (beta0), [lda] "r" (ldab), [ldc] "r" (ldcb) + : [width] "r" (width), [append] "r" (static_cast<uint64_t>(append)), [lda] "r" (ldab), [ldc] "r" (ldcb), [biasptr] "r" (biasptr), [minptr] "r" (minptr), [maxptr] "r" (maxptr) : "v0", "v1", "v2", "v3", "v4", "v5", "v6", "v7", "v8", "v9", "v10", "v11", "v12", "v13", "v14", "v15", "v16", "v17", "v18", "v19", "v20", "v21", "v22", "v23", "v24", "v25", "v26", "v27", "v28", "v29", "v30", "v31", "x0", "x1", "x2", "x3", "cc", "memory" ); break; @@ -458,56 +484,47 @@ void a64_hybrid_fp32_mla_16x4_a55(const float *A, int lda, const float *B, float "temploadreg3 .req X5\n" "add a_ptr1, %[a_ptr0], %[lda]\n" "add c_ptr1, %[c_ptr0], %[ldc]\n" - "cbz %[beta0], 1f\n" - "movi v16.4s, #0\n" + "cbnz %[append], 1f\n" + "ldr q16, [%[biasptr]]\n" + "ldr q17, [%[biasptr], #0x10]\n" + "ldr q18, [%[biasptr], #0x20]\n" + "ldr q19, [%[biasptr], #0x30]\n" + "mov v20.16b, v16.16b\n" "ldr q0, [%[a_ptr0]]\n" - "movi v17.4s, #0\n" + "mov v21.16b, v17.16b\n" "ldr q1, [a_ptr1]\n" - "movi v18.4s, #0\n" + "mov v22.16b, v18.16b\n" "ldr q8, [%[b_ptr0]]\n" - "movi v19.4s, #0\n" + "mov v23.16b, v19.16b\n" "ldr q9, [%[b_ptr0], #0x10]\n" - "movi v20.4s, #0\n" "ldr q10, [%[b_ptr0], #0x20]\n" - "movi v21.4s, #0\n" + "add %[a_ptr0], %[a_ptr0], #0x10\n" "ldr q11, [%[b_ptr0], #0x30]\n" - "movi v22.4s, #0\n" + "add a_ptr1, a_ptr1, #0x10\n" "ldr q12, [%[b_ptr0], #0x40]\n" - "movi v23.4s, #0\n" "ldr q13, [%[b_ptr0], #0x50]\n" "ldr d14, [%[b_ptr0], #0x60]\n" - "add %[a_ptr0], %[a_ptr0], #0x10\n" "ldr temploadreg2, [%[b_ptr0], #0x68]\n" - "add a_ptr1, a_ptr1, #0x10\n" "add %[b_ptr0], %[b_ptr0], #0x80\n" "cbz %[loops], 2f\n" "b 3f\n" "1:\n" - "ld1r {v15.4s}, [%[betaptr]]\n" "ldr q16, [%[c_ptr0]]\n" "ldr q17, [%[c_ptr0], #0x10]\n" "ldr q18, [%[c_ptr0], #0x20]\n" "ldr q19, [%[c_ptr0], #0x30]\n" - "fmul v16.4s, v16.4s, v15.4s\n" "ldr q20, [c_ptr1]\n" - "fmul v17.4s, v17.4s, v15.4s\n" "ldr q21, [c_ptr1, #0x10]\n" - "fmul v18.4s, v18.4s, v15.4s\n" "ldr q22, [c_ptr1, #0x20]\n" - "fmul v19.4s, v19.4s, v15.4s\n" "ldr q23, [c_ptr1, #0x30]\n" - "fmul v20.4s, v20.4s, v15.4s\n" "ldr q0, [%[a_ptr0]]\n" - "fmul v21.4s, v21.4s, v15.4s\n" + "add %[a_ptr0], %[a_ptr0], #0x10\n" "ldr q1, [a_ptr1]\n" - "fmul v22.4s, v22.4s, v15.4s\n" + "add a_ptr1, a_ptr1, #0x10\n" "ldr q8, [%[b_ptr0]]\n" - "fmul v23.4s, v23.4s, v15.4s\n" "ldr q9, [%[b_ptr0], #0x10]\n" "ldr q10, [%[b_ptr0], #0x20]\n" - "add %[a_ptr0], %[a_ptr0], #0x10\n" "ldr q11, [%[b_ptr0], #0x30]\n" - "add a_ptr1, a_ptr1, #0x10\n" "ldr q12, [%[b_ptr0], #0x40]\n" "ldr q13, [%[b_ptr0], #0x50]\n" "ldr d14, [%[b_ptr0], #0x60]\n" @@ -931,9 +948,27 @@ void a64_hybrid_fp32_mla_16x4_a55(const float *A, int lda, const float *B, float "fmla v23.4s, v11.4s, v1.s[0]\n" "b.ne 7b\n" "6:\n" + "ld1r {v14.4s}, [%[minptr]]\n" + "ld1r {v15.4s}, [%[maxptr]]\n" + "fmax v16.4s, v16.4s, v14.4s\n" + "fmax v17.4s, v17.4s, v14.4s\n" + "fmax v18.4s, v18.4s, v14.4s\n" + "fmax v19.4s, v19.4s, v14.4s\n" + "fmin v16.4s, v16.4s, v15.4s\n" + "fmin v17.4s, v17.4s, v15.4s\n" + "fmin v18.4s, v18.4s, v15.4s\n" + "fmin v19.4s, v19.4s, v15.4s\n" "str q16, [%[c_ptr0]]\n" + "fmax v20.4s, v20.4s, v14.4s\n" + "fmax v21.4s, v21.4s, v14.4s\n" + "fmax v22.4s, v22.4s, v14.4s\n" "str q17, [%[c_ptr0], #0x10]\n" + "fmax v23.4s, v23.4s, v14.4s\n" + "fmin v20.4s, v20.4s, v15.4s\n" + "fmin v21.4s, v21.4s, v15.4s\n" "str q18, [%[c_ptr0], #0x20]\n" + "fmin v22.4s, v22.4s, v15.4s\n" + "fmin v23.4s, v23.4s, v15.4s\n" "str q19, [%[c_ptr0], #0x30]\n" "add %[c_ptr0], %[c_ptr0], #0x40\n" "str q20, [c_ptr1]\n" @@ -947,7 +982,7 @@ void a64_hybrid_fp32_mla_16x4_a55(const float *A, int lda, const float *B, float ".unreq temploadreg2\n" ".unreq temploadreg3\n" : [a_ptr0] "+r" (a_ptr0), [b_ptr0] "+r" (b_ptr0), [c_ptr0] "+r" (c_ptr0), [loops] "+r" (loops), [regs] "+r" (regs), [blocks] "+r" (blocks) - : [betaptr] "r" (betaptr), [width] "r" (width), [beta0] "r" (beta0), [lda] "r" (ldab), [ldc] "r" (ldcb) + : [width] "r" (width), [append] "r" (static_cast<uint64_t>(append)), [lda] "r" (ldab), [ldc] "r" (ldcb), [biasptr] "r" (biasptr), [minptr] "r" (minptr), [maxptr] "r" (maxptr) : "v0", "v1", "v2", "v3", "v4", "v5", "v6", "v7", "v8", "v9", "v10", "v11", "v12", "v13", "v14", "v15", "v16", "v17", "v18", "v19", "v20", "v21", "v22", "v23", "v24", "v25", "v26", "v27", "v28", "v29", "v30", "v31", "x0", "x1", "x2", "x3", "x4", "x5", "cc", "memory" ); break; @@ -965,131 +1000,118 @@ void a64_hybrid_fp32_mla_16x4_a55(const float *A, int lda, const float *B, float "add c_ptr1, %[c_ptr0], %[ldc]\n" "add a_ptr2, a_ptr1, %[lda]\n" "add c_ptr2, c_ptr1, %[ldc]\n" - "cbz %[beta0], 1f\n" - "movi v16.4s, #0\n" + "cbnz %[append], 1f\n" + "ldr q16, [%[biasptr]]\n" + "ldr q17, [%[biasptr], #0x10]\n" + "ldr q18, [%[biasptr], #0x20]\n" + "ldr q19, [%[biasptr], #0x30]\n" + "mov v20.16b, v16.16b\n" "ldr q0, [%[a_ptr0]]\n" - "movi v17.4s, #0\n" + "mov v21.16b, v17.16b\n" "ldr q1, [a_ptr1]\n" - "movi v18.4s, #0\n" + "mov v22.16b, v18.16b\n" "ldr q2, [a_ptr2]\n" - "movi v19.4s, #0\n" + "mov v23.16b, v19.16b\n" "ldr q8, [%[b_ptr0]]\n" - "movi v20.4s, #0\n" + "mov v24.16b, v16.16b\n" "ldr q9, [%[b_ptr0], #0x10]\n" - "movi v21.4s, #0\n" + "mov v25.16b, v17.16b\n" "ldr q10, [%[b_ptr0], #0x20]\n" - "movi v22.4s, #0\n" + "mov v26.16b, v18.16b\n" "ldr q11, [%[b_ptr0], #0x30]\n" - "movi v23.4s, #0\n" + "mov v27.16b, v19.16b\n" "ldr q12, [%[b_ptr0], #0x40]\n" - "movi v24.4s, #0\n" "ldr q13, [%[b_ptr0], #0x50]\n" - "movi v25.4s, #0\n" - "ldr d14, [%[b_ptr0], #0x60]\n" - "movi v26.4s, #0\n" - "ldr temploadreg2, [%[b_ptr0], #0x68]\n" - "movi v27.4s, #0\n" "add %[a_ptr0], %[a_ptr0], #0x10\n" + "ldr d14, [%[b_ptr0], #0x60]\n" "add a_ptr1, a_ptr1, #0x10\n" - "ins v14.d[1], temploadreg2\n" + "ldr temploadreg2, [%[b_ptr0], #0x68]\n" "add a_ptr2, a_ptr2, #0x10\n" "add %[b_ptr0], %[b_ptr0], #0x80\n" "cbz %[loops], 2f\n" "b 3f\n" "1:\n" - "ld1r {v15.4s}, [%[betaptr]]\n" "ldr q16, [%[c_ptr0]]\n" "ldr q17, [%[c_ptr0], #0x10]\n" "ldr q18, [%[c_ptr0], #0x20]\n" "ldr q19, [%[c_ptr0], #0x30]\n" - "fmul v16.4s, v16.4s, v15.4s\n" "ldr q20, [c_ptr1]\n" - "fmul v17.4s, v17.4s, v15.4s\n" "ldr q21, [c_ptr1, #0x10]\n" - "fmul v18.4s, v18.4s, v15.4s\n" "ldr q22, [c_ptr1, #0x20]\n" - "fmul v19.4s, v19.4s, v15.4s\n" "ldr q23, [c_ptr1, #0x30]\n" - "fmul v20.4s, v20.4s, v15.4s\n" "ldr q24, [c_ptr2]\n" - "fmul v21.4s, v21.4s, v15.4s\n" "ldr q25, [c_ptr2, #0x10]\n" - "fmul v22.4s, v22.4s, v15.4s\n" "ldr q26, [c_ptr2, #0x20]\n" - "fmul v23.4s, v23.4s, v15.4s\n" "ldr q27, [c_ptr2, #0x30]\n" - "fmul v24.4s, v24.4s, v15.4s\n" "ldr q0, [%[a_ptr0]]\n" - "fmul v25.4s, v25.4s, v15.4s\n" + "add %[a_ptr0], %[a_ptr0], #0x10\n" "ldr q1, [a_ptr1]\n" - "fmul v26.4s, v26.4s, v15.4s\n" + "add a_ptr1, a_ptr1, #0x10\n" "ldr q2, [a_ptr2]\n" - "fmul v27.4s, v27.4s, v15.4s\n" + "add a_ptr2, a_ptr2, #0x10\n" "ldr q8, [%[b_ptr0]]\n" "ldr q9, [%[b_ptr0], #0x10]\n" - "add %[a_ptr0], %[a_ptr0], #0x10\n" "ldr q10, [%[b_ptr0], #0x20]\n" - "add a_ptr1, a_ptr1, #0x10\n" "ldr q11, [%[b_ptr0], #0x30]\n" - "add a_ptr2, a_ptr2, #0x10\n" "ldr q12, [%[b_ptr0], #0x40]\n" "ldr q13, [%[b_ptr0], #0x50]\n" "ldr d14, [%[b_ptr0], #0x60]\n" "ldr temploadreg2, [%[b_ptr0], #0x68]\n" "add %[b_ptr0], %[b_ptr0], #0x80\n" - "ins v14.d[1], temploadreg2\n" "cbz %[loops], 2f\n" "3:\n" "fmla v16.4s, v8.4s, v0.s[0]\n" - "ldr d15, [%[b_ptr0], #-0x10]\n" + "ins v14.d[1], temploadreg2\n" "fmla v20.4s, v8.4s, v1.s[0]\n" - "ldr temploadreg3, [%[b_ptr0], #-0x8]\n" + "ldr d15, [%[b_ptr0], #-0x10]\n" "fmla v24.4s, v8.4s, v2.s[0]\n" - "ldr d4, [%[a_ptr0]]\n" + "ldr temploadreg3, [%[b_ptr0], #-0x8]\n" "fmla v17.4s, v9.4s, v0.s[0]\n" - "ldr temploadreg0, [%[a_ptr0], #0x8]\n" + "ldr d4, [%[a_ptr0]]\n" "fmla v21.4s, v9.4s, v1.s[0]\n" - "ldr d5, [a_ptr1]\n" + "ldr temploadreg0, [%[a_ptr0], #0x8]\n" "fmla v25.4s, v9.4s, v2.s[0]\n" - "ldr temploadreg1, [a_ptr1, #0x8]\n" + "ldr d5, [a_ptr1]\n" "fmla v18.4s, v10.4s, v0.s[0]\n" - "ldr d6, [a_ptr2]\n" + "ldr temploadreg1, [a_ptr1, #0x8]\n" "fmla v22.4s, v10.4s, v1.s[0]\n" - "ldr temploadreg2, [a_ptr2, #0x8]\n" + "ldr d6, [a_ptr2]\n" "fmla v26.4s, v10.4s, v2.s[0]\n" - "ldr d8, [%[b_ptr0]]\n" + "ldr temploadreg2, [a_ptr2, #0x8]\n" "fmla v19.4s, v11.4s, v0.s[0]\n" - "ins v4.d[1], temploadreg0\n" + "ldr d8, [%[b_ptr0]]\n" "fmla v23.4s, v11.4s, v1.s[0]\n" - "ldr temploadreg0, [%[b_ptr0], #0x8]\n" + "ins v4.d[1], temploadreg0\n" "fmla v27.4s, v11.4s, v2.s[0]\n" - "ldr d9, [%[b_ptr0], #0x10]\n" + "ldr temploadreg0, [%[b_ptr0], #0x8]\n" "fmla v16.4s, v12.4s, v0.s[1]\n" - "ins v5.d[1], temploadreg1\n" + "ldr d9, [%[b_ptr0], #0x10]\n" "fmla v20.4s, v12.4s, v1.s[1]\n" - "ldr temploadreg1, [%[b_ptr0], #0x18]\n" + "ins v5.d[1], temploadreg1\n" "fmla v24.4s, v12.4s, v2.s[1]\n" - "ldr d10, [%[b_ptr0], #0x20]\n" + "ldr temploadreg1, [%[b_ptr0], #0x18]\n" "fmla v17.4s, v13.4s, v0.s[1]\n" - "ins v6.d[1], temploadreg2\n" + "ldr d10, [%[b_ptr0], #0x20]\n" "fmla v21.4s, v13.4s, v1.s[1]\n" - "ldr temploadreg2, [%[b_ptr0], #0x28]\n" + "ins v6.d[1], temploadreg2\n" "fmla v25.4s, v13.4s, v2.s[1]\n" - "ldr d11, [%[b_ptr0], #0x30]\n" + "ldr temploadreg2, [%[b_ptr0], #0x28]\n" "fmla v18.4s, v14.4s, v0.s[1]\n" - "ins v15.d[1], temploadreg3\n" + "ldr d11, [%[b_ptr0], #0x30]\n" "fmla v22.4s, v14.4s, v1.s[1]\n" - "ldr temploadreg3, [%[b_ptr0], #0x38]\n" + "ins v15.d[1], temploadreg3\n" "fmla v26.4s, v14.4s, v2.s[1]\n" + "ldr temploadreg3, [%[b_ptr0], #0x38]\n" "ldr d12, [%[b_ptr0], #0x40]\n" - "ins v8.d[1], temploadreg0\n" "subs %[loops], %[loops], #0x1\n" "fmla v19.4s, v15.4s, v0.s[1]\n" - "ldr temploadreg0, [%[b_ptr0], #0x48]\n" + "ins v8.d[1], temploadreg0\n" "fmla v23.4s, v15.4s, v1.s[1]\n" - "ldr d13, [%[b_ptr0], #0x50]\n" + "ldr temploadreg0, [%[b_ptr0], #0x48]\n" "fmla v27.4s, v15.4s, v2.s[1]\n" + "ldr d13, [%[b_ptr0], #0x50]\n" "ins v9.d[1], temploadreg1\n" + "prfm PLDL1KEEP, [%[a_ptr0], #0x40]\n" "fmla v16.4s, v8.4s, v0.s[2]\n" "ldr temploadreg1, [%[b_ptr0], #0x58]\n" "fmla v20.4s, v8.4s, v1.s[2]\n" @@ -1113,41 +1135,39 @@ void a64_hybrid_fp32_mla_16x4_a55(const float *A, int lda, const float *B, float "fmla v23.4s, v11.4s, v1.s[2]\n" "ins v15.d[1], temploadreg3\n" "fmla v27.4s, v11.4s, v2.s[2]\n" - "prfm PLDL1KEEP, [%[a_ptr0], #0x40]\n" - "fmla v16.4s, v12.4s, v0.s[3]\n" "add %[b_ptr0], %[b_ptr0], #0x100\n" - "fmla v20.4s, v12.4s, v1.s[3]\n" + "fmla v16.4s, v12.4s, v0.s[3]\n" "ldr d8, [%[b_ptr0], #-0x80]\n" - "fmla v24.4s, v12.4s, v2.s[3]\n" + "fmla v20.4s, v12.4s, v1.s[3]\n" "ldr temploadreg0, [%[b_ptr0], #-0x78]\n" - "fmla v17.4s, v13.4s, v0.s[3]\n" + "fmla v24.4s, v12.4s, v2.s[3]\n" "ldr d9, [%[b_ptr0], #-0x70]\n" - "fmla v21.4s, v13.4s, v1.s[3]\n" + "fmla v17.4s, v13.4s, v0.s[3]\n" "ldr temploadreg1, [%[b_ptr0], #-0x68]\n" - "fmla v25.4s, v13.4s, v2.s[3]\n" + "fmla v21.4s, v13.4s, v1.s[3]\n" "ldr d10, [%[b_ptr0], #-0x60]\n" - "fmla v18.4s, v14.4s, v0.s[3]\n" + "fmla v25.4s, v13.4s, v2.s[3]\n" "ldr temploadreg2, [%[b_ptr0], #-0x58]\n" - "fmla v22.4s, v14.4s, v1.s[3]\n" + "fmla v18.4s, v14.4s, v0.s[3]\n" "ldr d11, [%[b_ptr0], #-0x50]\n" - "fmla v26.4s, v14.4s, v2.s[3]\n" + "fmla v22.4s, v14.4s, v1.s[3]\n" "ldr temploadreg3, [%[b_ptr0], #-0x48]\n" - "fmla v19.4s, v15.4s, v0.s[3]\n" + "fmla v26.4s, v14.4s, v2.s[3]\n" "ldr d12, [%[b_ptr0], #-0x40]\n" - "fmla v23.4s, v15.4s, v1.s[3]\n" + "fmla v19.4s, v15.4s, v0.s[3]\n" "ins v8.d[1], temploadreg0\n" - "fmla v27.4s, v15.4s, v2.s[3]\n" + "fmla v23.4s, v15.4s, v1.s[3]\n" "ldr temploadreg0, [%[b_ptr0], #-0x38]\n" + "fmla v27.4s, v15.4s, v2.s[3]\n" "ldr d13, [%[b_ptr0], #-0x30]\n" + "ins v9.d[1], temploadreg1\n" "add %[a_ptr0], %[a_ptr0], #0x20\n" "fmla v16.4s, v8.4s, v4.s[0]\n" - "ins v9.d[1], temploadreg1\n" - "fmla v20.4s, v8.4s, v5.s[0]\n" "ldr temploadreg1, [%[b_ptr0], #-0x28]\n" - "fmla v24.4s, v8.4s, v6.s[0]\n" + "fmla v20.4s, v8.4s, v5.s[0]\n" "ldr d14, [%[b_ptr0], #-0x20]\n" + "fmla v24.4s, v8.4s, v6.s[0]\n" "ins v10.d[1], temploadreg2\n" - "add a_ptr1, a_ptr1, #0x20\n" "fmla v17.4s, v9.4s, v4.s[0]\n" "ldr temploadreg2, [%[b_ptr0], #-0x18]\n" "fmla v21.4s, v9.4s, v5.s[0]\n" @@ -1163,70 +1183,71 @@ void a64_hybrid_fp32_mla_16x4_a55(const float *A, int lda, const float *B, float "fmla v19.4s, v11.4s, v4.s[0]\n" "ldr temploadreg0, [%[a_ptr0], #-0x8]\n" "fmla v23.4s, v11.4s, v5.s[0]\n" - "ldr d1, [a_ptr1, #-0x10]\n" - "fmla v27.4s, v11.4s, v6.s[0]\n" "ins v13.d[1], temploadreg1\n" + "fmla v27.4s, v11.4s, v6.s[0]\n" + "ins v14.d[1], temploadreg2\n" "fmla v16.4s, v12.4s, v4.s[1]\n" - "ldr temploadreg1, [a_ptr1, #-0x8]\n" + "ldr d8, [%[b_ptr0]]\n" "fmla v20.4s, v12.4s, v5.s[1]\n" - "ins v14.d[1], temploadreg2\n" + "ins v0.d[1], temploadreg0\n" "fmla v24.4s, v12.4s, v6.s[1]\n" - "ldr d8, [%[b_ptr0]]\n" + "ldr temploadreg0, [%[b_ptr0], #0x8]\n" "fmla v17.4s, v13.4s, v4.s[1]\n" - "ins v0.d[1], temploadreg0\n" + "ldr d9, [%[b_ptr0], #0x10]\n" "fmla v21.4s, v13.4s, v5.s[1]\n" - "ldr temploadreg0, [%[b_ptr0], #0x8]\n" + "ldr d10, [%[b_ptr0], #0x20]\n" "fmla v25.4s, v13.4s, v6.s[1]\n" - "ldr d9, [%[b_ptr0], #0x10]\n" + "ldr d11, [%[b_ptr0], #0x30]\n" "fmla v18.4s, v14.4s, v4.s[1]\n" - "ins v1.d[1], temploadreg1\n" + "ins v15.d[1], temploadreg3\n" "fmla v22.4s, v14.4s, v5.s[1]\n" - "ldr temploadreg1, [%[b_ptr0], #0x18]\n" + "ldr temploadreg3, [%[b_ptr0], #0x38]\n" "fmla v26.4s, v14.4s, v6.s[1]\n" - "ldr d10, [%[b_ptr0], #0x20]\n" - "ldr d11, [%[b_ptr0], #0x30]\n" - "add a_ptr2, a_ptr2, #0x20\n" - "ins v15.d[1], temploadreg3\n" - "prfm PLDL1KEEP, [a_ptr1, #0x40]\n" - "ldr d2, [a_ptr2, #-0x10]\n" - "prfm PLDL1KEEP, [a_ptr2, #0x40]\n" + "ldr d12, [%[b_ptr0], #0x40]\n" + "ins v8.d[1], temploadreg0\n" + "add a_ptr1, a_ptr1, #0x20\n" "fmla v19.4s, v15.4s, v4.s[1]\n" - "ldr temploadreg2, [a_ptr2, #-0x8]\n" + "ldr d1, [a_ptr1, #-0x10]\n" "fmla v23.4s, v15.4s, v5.s[1]\n" - "ldr temploadreg3, [%[b_ptr0], #0x38]\n" + "ldr temploadreg1, [a_ptr1, #-0x8]\n" "fmla v27.4s, v15.4s, v6.s[1]\n" - "ldr d12, [%[b_ptr0], #0x40]\n" - "ins v8.d[1], temploadreg0\n" - "ins v2.d[1], temploadreg2\n" - "ldr temploadreg2, [%[b_ptr0], #0x28]\n" "ldr temploadreg0, [%[b_ptr0], #0x48]\n" "fmla v16.4s, v8.4s, v4.s[2]\n" "ldr d13, [%[b_ptr0], #0x50]\n" "fmla v20.4s, v8.4s, v5.s[2]\n" - "ins v9.d[1], temploadreg1\n" + "ins v1.d[1], temploadreg1\n" "fmla v24.4s, v8.4s, v6.s[2]\n" - "ldr temploadreg1, [%[b_ptr0], #0x58]\n" + "ldr temploadreg1, [%[b_ptr0], #0x18]\n" "ldr d14, [%[b_ptr0], #0x60]\n" - "ins v10.d[1], temploadreg2\n" - "fmla v17.4s, v9.4s, v4.s[2]\n" - "ldr temploadreg2, [%[b_ptr0], #0x68]\n" - "fmla v21.4s, v9.4s, v5.s[2]\n" + "add a_ptr2, a_ptr2, #0x20\n" "ldr d15, [%[b_ptr0], #0x70]\n" - "fmla v25.4s, v9.4s, v6.s[2]\n" + "prfm PLDL1KEEP, [a_ptr1, #0x40]\n" + "ldr d2, [a_ptr2, #-0x10]\n" + "prfm PLDL1KEEP, [a_ptr2, #0x40]\n" + "ldr temploadreg2, [a_ptr2, #-0x8]\n" + "ins v9.d[1], temploadreg1\n" + "ldr temploadreg1, [%[b_ptr0], #0x58]\n" "ins v11.d[1], temploadreg3\n" - "fmla v18.4s, v10.4s, v4.s[2]\n" + "ins v2.d[1], temploadreg2\n" + "fmla v17.4s, v9.4s, v4.s[2]\n" + "ldr temploadreg2, [%[b_ptr0], #0x28]\n" + "fmla v21.4s, v9.4s, v5.s[2]\n" "ldr temploadreg3, [%[b_ptr0], #0x78]\n" - "fmla v22.4s, v10.4s, v5.s[2]\n" + "fmla v25.4s, v9.4s, v6.s[2]\n" "ins v12.d[1], temploadreg0\n" - "fmla v26.4s, v10.4s, v6.s[2]\n" - "ins v13.d[1], temploadreg1\n" "fmla v19.4s, v11.4s, v4.s[2]\n" - "ins v14.d[1], temploadreg2\n" + "ins v10.d[1], temploadreg2\n" "fmla v23.4s, v11.4s, v5.s[2]\n" - "ins v15.d[1], temploadreg3\n" + "ldr temploadreg2, [%[b_ptr0], #0x68]\n" "fmla v27.4s, v11.4s, v6.s[2]\n" - "add %[b_ptr0], %[b_ptr0], #0x100\n" + "ins v13.d[1], temploadreg1\n" "fmla v16.4s, v12.4s, v4.s[3]\n" + "ins v15.d[1], temploadreg3\n" + "fmla v18.4s, v10.4s, v4.s[2]\n" + "ins v14.d[1], temploadreg2\n" + "fmla v22.4s, v10.4s, v5.s[2]\n" + "add %[b_ptr0], %[b_ptr0], #0x100\n" + "fmla v26.4s, v10.4s, v6.s[2]\n" "ldr d8, [%[b_ptr0], #-0x80]\n" "fmla v20.4s, v12.4s, v5.s[3]\n" "ldr temploadreg0, [%[b_ptr0], #-0x78]\n" @@ -1258,13 +1279,13 @@ void a64_hybrid_fp32_mla_16x4_a55(const float *A, int lda, const float *B, float "ins v11.d[1], temploadreg3\n" "ins v12.d[1], temploadreg0\n" "ins v13.d[1], temploadreg1\n" - "ins v14.d[1], temploadreg2\n" "b.ne 3b\n" "2:\n" - "ldr d15, [%[b_ptr0], #-0x10]\n" + "ins v14.d[1], temploadreg2\n" "prfm PSTL1KEEP, [%[c_ptr0]]\n" - "ldr temploadreg3, [%[b_ptr0], #-0x8]\n" + "ldr d15, [%[b_ptr0], #-0x10]\n" "prfm PSTL1KEEP, [c_ptr1]\n" + "ldr temploadreg3, [%[b_ptr0], #-0x8]\n" "prfm PSTL1KEEP, [c_ptr2]\n" "ins v15.d[1], temploadreg3\n" "cbz %[regs], 4f\n" @@ -1554,13 +1575,39 @@ void a64_hybrid_fp32_mla_16x4_a55(const float *A, int lda, const float *B, float "fmla v27.4s, v11.4s, v2.s[0]\n" "b.ne 7b\n" "6:\n" + "ld1r {v14.4s}, [%[minptr]]\n" + "ld1r {v15.4s}, [%[maxptr]]\n" + "fmax v16.4s, v16.4s, v14.4s\n" + "fmax v17.4s, v17.4s, v14.4s\n" + "fmax v18.4s, v18.4s, v14.4s\n" + "fmax v19.4s, v19.4s, v14.4s\n" + "fmin v16.4s, v16.4s, v15.4s\n" + "fmin v17.4s, v17.4s, v15.4s\n" + "fmin v18.4s, v18.4s, v15.4s\n" + "fmin v19.4s, v19.4s, v15.4s\n" "str q16, [%[c_ptr0]]\n" + "fmax v20.4s, v20.4s, v14.4s\n" + "fmax v21.4s, v21.4s, v14.4s\n" + "fmax v22.4s, v22.4s, v14.4s\n" "str q17, [%[c_ptr0], #0x10]\n" + "fmax v23.4s, v23.4s, v14.4s\n" + "fmin v20.4s, v20.4s, v15.4s\n" + "fmin v21.4s, v21.4s, v15.4s\n" "str q18, [%[c_ptr0], #0x20]\n" + "fmin v22.4s, v22.4s, v15.4s\n" + "fmin v23.4s, v23.4s, v15.4s\n" + "fmax v24.4s, v24.4s, v14.4s\n" "str q19, [%[c_ptr0], #0x30]\n" + "fmax v25.4s, v25.4s, v14.4s\n" "add %[c_ptr0], %[c_ptr0], #0x40\n" + "fmax v26.4s, v26.4s, v14.4s\n" "str q20, [c_ptr1]\n" + "fmin v24.4s, v24.4s, v15.4s\n" + "fmin v25.4s, v25.4s, v15.4s\n" + "fmax v27.4s, v27.4s, v14.4s\n" "str q21, [c_ptr1, #0x10]\n" + "fmin v26.4s, v26.4s, v15.4s\n" + "fmin v27.4s, v27.4s, v15.4s\n" "str q22, [c_ptr1, #0x20]\n" "str q23, [c_ptr1, #0x30]\n" "str q24, [c_ptr2]\n" @@ -1576,7 +1623,7 @@ void a64_hybrid_fp32_mla_16x4_a55(const float *A, int lda, const float *B, float ".unreq temploadreg2\n" ".unreq temploadreg3\n" : [a_ptr0] "+r" (a_ptr0), [b_ptr0] "+r" (b_ptr0), [c_ptr0] "+r" (c_ptr0), [loops] "+r" (loops), [regs] "+r" (regs), [blocks] "+r" (blocks) - : [betaptr] "r" (betaptr), [width] "r" (width), [beta0] "r" (beta0), [lda] "r" (ldab), [ldc] "r" (ldcb) + : [width] "r" (width), [append] "r" (static_cast<uint64_t>(append)), [lda] "r" (ldab), [ldc] "r" (ldcb), [biasptr] "r" (biasptr), [minptr] "r" (minptr), [maxptr] "r" (maxptr) : "v0", "v1", "v2", "v3", "v4", "v5", "v6", "v7", "v8", "v9", "v10", "v11", "v12", "v13", "v14", "v15", "v16", "v17", "v18", "v19", "v20", "v21", "v22", "v23", "v24", "v25", "v26", "v27", "v28", "v29", "v30", "v31", "x0", "x1", "x2", "x3", "x4", "x5", "x6", "x7", "cc", "memory" ); break; @@ -1599,89 +1646,72 @@ void a64_hybrid_fp32_mla_16x4_a55(const float *A, int lda, const float *B, float "add c_ptr2, c_ptr1, %[ldc]\n" "add a_ptr3, a_ptr2, %[lda]\n" "add c_ptr3, c_ptr2, %[ldc]\n" - "cbz %[beta0], 1f\n" - "movi v16.4s, #0\n" + "cbnz %[append], 1f\n" + "ldr q16, [%[biasptr]]\n" + "ldr q17, [%[biasptr], #0x10]\n" + "ldr q18, [%[biasptr], #0x20]\n" + "ldr q19, [%[biasptr], #0x30]\n" + "mov v20.16b, v16.16b\n" "ldr q0, [%[a_ptr0]]\n" - "movi v17.4s, #0\n" + "mov v21.16b, v17.16b\n" "ldr q1, [a_ptr1]\n" - "movi v18.4s, #0\n" + "mov v22.16b, v18.16b\n" "ldr q2, [a_ptr2]\n" - "movi v19.4s, #0\n" + "mov v23.16b, v19.16b\n" "ldr q3, [a_ptr3]\n" - "movi v20.4s, #0\n" + "mov v24.16b, v16.16b\n" "ldr q8, [%[b_ptr0]]\n" - "movi v21.4s, #0\n" + "mov v25.16b, v17.16b\n" "ldr q9, [%[b_ptr0], #0x10]\n" - "movi v22.4s, #0\n" + "mov v26.16b, v18.16b\n" "ldr q10, [%[b_ptr0], #0x20]\n" - "movi v23.4s, #0\n" + "mov v27.16b, v19.16b\n" "ldr q11, [%[b_ptr0], #0x30]\n" - "movi v24.4s, #0\n" + "mov v28.16b, v16.16b\n" "ldr q12, [%[b_ptr0], #0x40]\n" - "movi v25.4s, #0\n" + "mov v29.16b, v17.16b\n" "ldr q13, [%[b_ptr0], #0x50]\n" - "movi v26.4s, #0\n" + "mov v30.16b, v18.16b\n" "ldr d14, [%[b_ptr0], #0x60]\n" - "movi v27.4s, #0\n" + "mov v31.16b, v19.16b\n" "ldr temploadreg2, [%[b_ptr0], #0x68]\n" - "movi v28.4s, #0\n" "add %[a_ptr0], %[a_ptr0], #0x10\n" - "movi v29.4s, #0\n" - "ins v14.d[1], temploadreg2\n" - "movi v30.4s, #0\n" "add a_ptr1, a_ptr1, #0x10\n" - "movi v31.4s, #0\n" "add a_ptr2, a_ptr2, #0x10\n" + "ins v14.d[1], temploadreg2\n" "add a_ptr3, a_ptr3, #0x10\n" "add %[b_ptr0], %[b_ptr0], #0x80\n" "cbz %[loops], 2f\n" "b 3f\n" "1:\n" - "ld1r {v15.4s}, [%[betaptr]]\n" "ldr q16, [%[c_ptr0]]\n" "ldr q17, [%[c_ptr0], #0x10]\n" "ldr q18, [%[c_ptr0], #0x20]\n" "ldr q19, [%[c_ptr0], #0x30]\n" - "fmul v16.4s, v16.4s, v15.4s\n" "ldr q20, [c_ptr1]\n" - "fmul v17.4s, v17.4s, v15.4s\n" "ldr q21, [c_ptr1, #0x10]\n" - "fmul v18.4s, v18.4s, v15.4s\n" "ldr q22, [c_ptr1, #0x20]\n" - "fmul v19.4s, v19.4s, v15.4s\n" "ldr q23, [c_ptr1, #0x30]\n" - "fmul v20.4s, v20.4s, v15.4s\n" "ldr q24, [c_ptr2]\n" - "fmul v21.4s, v21.4s, v15.4s\n" "ldr q25, [c_ptr2, #0x10]\n" - "fmul v22.4s, v22.4s, v15.4s\n" "ldr q26, [c_ptr2, #0x20]\n" - "fmul v23.4s, v23.4s, v15.4s\n" "ldr q27, [c_ptr2, #0x30]\n" - "fmul v24.4s, v24.4s, v15.4s\n" "ldr q28, [c_ptr3]\n" - "fmul v25.4s, v25.4s, v15.4s\n" "ldr q29, [c_ptr3, #0x10]\n" - "fmul v26.4s, v26.4s, v15.4s\n" "ldr q30, [c_ptr3, #0x20]\n" - "fmul v27.4s, v27.4s, v15.4s\n" "ldr q31, [c_ptr3, #0x30]\n" - "fmul v28.4s, v28.4s, v15.4s\n" "ldr q0, [%[a_ptr0]]\n" - "fmul v29.4s, v29.4s, v15.4s\n" + "add %[a_ptr0], %[a_ptr0], #0x10\n" "ldr q1, [a_ptr1]\n" - "fmul v30.4s, v30.4s, v15.4s\n" + "add a_ptr1, a_ptr1, #0x10\n" "ldr q2, [a_ptr2]\n" - "fmul v31.4s, v31.4s, v15.4s\n" + "add a_ptr2, a_ptr2, #0x10\n" "ldr q3, [a_ptr3]\n" + "add a_ptr3, a_ptr3, #0x10\n" "ldr q8, [%[b_ptr0]]\n" - "add %[a_ptr0], %[a_ptr0], #0x10\n" "ldr q9, [%[b_ptr0], #0x10]\n" - "add a_ptr1, a_ptr1, #0x10\n" "ldr q10, [%[b_ptr0], #0x20]\n" - "add a_ptr2, a_ptr2, #0x10\n" "ldr q11, [%[b_ptr0], #0x30]\n" - "add a_ptr3, a_ptr3, #0x10\n" "ldr q12, [%[b_ptr0], #0x40]\n" "ldr q13, [%[b_ptr0], #0x50]\n" "ldr d14, [%[b_ptr0], #0x60]\n" @@ -2303,16 +2333,50 @@ void a64_hybrid_fp32_mla_16x4_a55(const float *A, int lda, const float *B, float "fmla v31.4s, v11.4s, v3.s[0]\n" "b.ne 7b\n" "6:\n" + "ld1r {v14.4s}, [%[minptr]]\n" + "ld1r {v15.4s}, [%[maxptr]]\n" + "fmax v16.4s, v16.4s, v14.4s\n" + "fmax v17.4s, v17.4s, v14.4s\n" + "fmax v18.4s, v18.4s, v14.4s\n" + "fmax v19.4s, v19.4s, v14.4s\n" + "fmin v16.4s, v16.4s, v15.4s\n" + "fmin v17.4s, v17.4s, v15.4s\n" + "fmin v18.4s, v18.4s, v15.4s\n" + "fmin v19.4s, v19.4s, v15.4s\n" "str q16, [%[c_ptr0]]\n" + "fmax v20.4s, v20.4s, v14.4s\n" + "fmax v21.4s, v21.4s, v14.4s\n" + "fmax v22.4s, v22.4s, v14.4s\n" "str q17, [%[c_ptr0], #0x10]\n" + "fmax v23.4s, v23.4s, v14.4s\n" + "fmin v20.4s, v20.4s, v15.4s\n" + "fmin v21.4s, v21.4s, v15.4s\n" "str q18, [%[c_ptr0], #0x20]\n" + "fmin v22.4s, v22.4s, v15.4s\n" + "fmin v23.4s, v23.4s, v15.4s\n" + "fmax v24.4s, v24.4s, v14.4s\n" "str q19, [%[c_ptr0], #0x30]\n" + "fmax v25.4s, v25.4s, v14.4s\n" "add %[c_ptr0], %[c_ptr0], #0x40\n" + "fmax v26.4s, v26.4s, v14.4s\n" "str q20, [c_ptr1]\n" + "fmin v24.4s, v24.4s, v15.4s\n" + "fmin v25.4s, v25.4s, v15.4s\n" + "fmax v27.4s, v27.4s, v14.4s\n" "str q21, [c_ptr1, #0x10]\n" + "fmin v26.4s, v26.4s, v15.4s\n" + "fmax v28.4s, v28.4s, v14.4s\n" + "fmax v29.4s, v29.4s, v14.4s\n" "str q22, [c_ptr1, #0x20]\n" + "fmin v27.4s, v27.4s, v15.4s\n" + "fmax v30.4s, v30.4s, v14.4s\n" + "fmin v28.4s, v28.4s, v15.4s\n" "str q23, [c_ptr1, #0x30]\n" + "fmin v29.4s, v29.4s, v15.4s\n" + "fmax v31.4s, v31.4s, v14.4s\n" + "fmin v30.4s, v30.4s, v15.4s\n" "str q24, [c_ptr2]\n" + "fmin v31.4s, v31.4s, v15.4s\n" "str q25, [c_ptr2, #0x10]\n" "str q26, [c_ptr2, #0x20]\n" "str q27, [c_ptr2, #0x30]\n" @@ -2331,7 +2395,7 @@ void a64_hybrid_fp32_mla_16x4_a55(const float *A, int lda, const float *B, float ".unreq temploadreg2\n" ".unreq temploadreg3\n" : [a_ptr0] "+r" (a_ptr0), [b_ptr0] "+r" (b_ptr0), [c_ptr0] "+r" (c_ptr0), [loops] "+r" (loops), [regs] "+r" (regs), [blocks] "+r" (blocks) - : [betaptr] "r" (betaptr), [width] "r" (width), [beta0] "r" (beta0), [lda] "r" (ldab), [ldc] "r" (ldcb) + : [width] "r" (width), [append] "r" (static_cast<uint64_t>(append)), [lda] "r" (ldab), [ldc] "r" (ldcb), [biasptr] "r" (biasptr), [minptr] "r" (minptr), [maxptr] "r" (maxptr) : "v0", "v1", "v2", "v3", "v4", "v5", "v6", "v7", "v8", "v9", "v10", "v11", "v12", "v13", "v14", "v15", "v16", "v17", "v18", "v19", "v20", "v21", "v22", "v23", "v24", "v25", "v26", "v27", "v28", "v29", "v30", "v31", "x0", "x1", "x2", "x3", "x4", "x5", "x6", "x7", "x8", "x9", "cc", "memory" ); break; |