test / src /f16-gemm /gen /f16-gemminc-6x8-minmax-asm-aarch64-neonfp16arith-ld64.S
Androidonnxfork's picture
Upload folder using huggingface_hub
8b7c501
raw
history blame
9.2 kB
// Auto-generated file. Do not edit!
// Template: src/f16-gemm/6x8-aarch64-neonfp16arith-ld64.S.in
// Generator: tools/xngen
//
// Copyright 2020 Google LLC
//
// This source code is licensed under the BSD-style license found in the
// LICENSE file in the root directory of this source tree.
#include <xnnpack/assembly.h>
# void xnn_f16_gemminc_minmax_ukernel_6x8__asm_aarch64_neonfp16arith_ld64(
# size_t mr, x0
# size_t nc, x1
# size_t kc, x2 / x0
# const void* restrict a, x3
# size_t a_stride, x4
# const void* restrict w, x5
# void* restrict c, x6
# size_t cm_stride, x7
# size_t cn_stride, [sp] -> (x8)
# const float* restrict acc, [sp + 8] -> x15
# const union xnn_f16_minmax_params params[restrict XNN_MIN_ELEMENTS(1)]) [sp + 16] -> (x8)
# d8-d15, x19-x30 need to be preserved if used. x18 is reserved by the OS.
# Register usage
# A0 x3 v0
# A1 x9 v1
# A2 x10 v2
# A3 x11 v3
# A4 x12 v4
# A5 x4 v5
# B x5 v16 v17 v18 v19
# C0 x6 v20
# C1 x16 v22
# C2 x17 v24
# C3 x14 v26
# C4 x13 v28
# C5 x7 v30
# Clamp v6, (v4), (v5)
# unused A v8 v9 v10 v11
# unused B v12 v13 v14 v15
BEGIN_FUNCTION xnn_f16_gemminc_minmax_ukernel_6x8__asm_aarch64_neonfp16arith_ld64
# Load acc, params pointer
LDP x15, x8, [sp, 8]
# Clamp A and C pointers
CMP x0, 2 // if mr < 2
ADD x9, x3, x4 // a1 = a0 + a_stride
ADD x16, x6, x7 // c1 = c0 + cm_stride
CSEL x9, x3, x9, LO // a1 = a0
CSEL x16, x6, x16, LO // c1 = c0
# Load params
LDR s6, [x8]
ADD x10, x9, x4 // a2 = a1 + a_stride
ADD x17, x16, x7 // c2 = c1 + cm_stride
// if mr <= 2
CSEL x10, x9, x10, LS // a2 = a1
CSEL x17, x16, x17, LS // c2 = c1
CMP x0, 4 // if mr < 4
ADD x11, x10, x4 // a3 = a2 + a_stride
ADD x14, x17, x7 // c3 = c2 + cm_stride
CSEL x11, x10, x11, LO // a3 = a2
CSEL x14, x17, x14, LO // c3 = c2
ADD x12, x11, x4 // a4 = a3 + a_stride
ADD x13, x14, x7 // c4 = c3 + cm_stride
// if mr <= 4
CSEL x12, x11, x12, LS // a4 = a3
CSEL x13, x14, x13, LS // c4 = c3
CMP x0, 6 // if mr < 6
ADD x4, x12, x4 // a5 = a4 + a_stride
ADD x7, x13, x7 // c5 = c4 + cm_stride
CSEL x4, x12, x4, LO // a5 = a4
CSEL x7, x13, x7, LO // c5 = c4
LDR x8, [sp] // load cn_stride
0:
# Load initial accumulators
LDP q20, q22, [x15], 32
LDP q24, q26, [x15], 32
LDP q28, q30, [x15], 32
# Is there at least 4 halffloats (8 bytes)?
SUBS x0, x2, 8 // k = kc - 8
B.LO 3f
# Main loop - 4 halffloats of A (8 bytes)
# 24 FMA + 6 ld64 A + 4 LDR B
1:
LDR d0, [x3], 8
LDR q16, [x5], 16
LDR q17, [x5], 16
LDR d1, [x9], 8
LDR d2, [x10], 8
LDR d3, [x11], 8
LDR d4, [x12], 8
LDR d5, [x4], 8
LDR q18, [x5], 16
LDR q19, [x5], 16
SUBS x0, x0, 8
FMLA v20.8h, v16.8h, v0.h[0]
FMLA v22.8h, v16.8h, v1.h[0]
FMLA v24.8h, v16.8h, v2.h[0]
FMLA v26.8h, v16.8h, v3.h[0]
FMLA v28.8h, v16.8h, v4.h[0]
FMLA v30.8h, v16.8h, v5.h[0]
FMLA v20.8h, v17.8h, v0.h[1]
FMLA v22.8h, v17.8h, v1.h[1]
FMLA v24.8h, v17.8h, v2.h[1]
FMLA v26.8h, v17.8h, v3.h[1]
FMLA v28.8h, v17.8h, v4.h[1]
FMLA v30.8h, v17.8h, v5.h[1]
FMLA v20.8h, v18.8h, v0.h[2]
FMLA v22.8h, v18.8h, v1.h[2]
FMLA v24.8h, v18.8h, v2.h[2]
FMLA v26.8h, v18.8h, v3.h[2]
FMLA v28.8h, v18.8h, v4.h[2]
FMLA v30.8h, v18.8h, v5.h[2]
FMLA v20.8h, v19.8h, v0.h[3]
FMLA v22.8h, v19.8h, v1.h[3]
FMLA v24.8h, v19.8h, v2.h[3]
FMLA v26.8h, v19.8h, v3.h[3]
FMLA v28.8h, v19.8h, v4.h[3]
FMLA v30.8h, v19.8h, v5.h[3]
B.HS 1b
# Is there a remainder?- 2 halffloats of A (4 bytes)
TBNZ x0, 2, 4f
# Is there a remainder?- 1 halffloat of A (2 bytes)
TBNZ x0, 1, 5f
2:
# Clamp
DUP v4.8h, v6.h[0]
DUP v5.8h, v6.h[1]
FMAX v20.8h, v20.8h, v4.8h
FMAX v22.8h, v22.8h, v4.8h
FMAX v24.8h, v24.8h, v4.8h
FMAX v26.8h, v26.8h, v4.8h
FMAX v28.8h, v28.8h, v4.8h
FMAX v30.8h, v30.8h, v4.8h
SUBS x1, x1, 8
FMIN v20.8h, v20.8h, v5.8h
FMIN v22.8h, v22.8h, v5.8h
FMIN v24.8h, v24.8h, v5.8h
FMIN v26.8h, v26.8h, v5.8h
FMIN v28.8h, v28.8h, v5.8h
FMIN v30.8h, v30.8h, v5.8h
# Store full 6 x 8
B.LO 6f
ST1 {v30.16b}, [x7], x8
SUB x3, x3, x2 // a0 -= kc
ST1 {v28.16b}, [x13], x8
SUB x9, x9, x2 // a1 -= kc
ST1 {v26.16b}, [x14], x8
SUB x10, x10, x2 // a2 -= kc
ST1 {v24.16b}, [x17], x8
SUB x11, x11, x2 // a3 -= kc
ST1 {v22.16b}, [x16], x8
SUB x12, x12, x2 // a4 -= kc
ST1 {v20.16b}, [x6], x8
SUB x4, x4, x2 // a5 -= kc
B.HI 0b
RET
3:
TBZ x0, 2, 5f
4:
# Remainder- 2 halffloats of A (4 bytes)
LDR s0, [x3], 4
LDR q16, [x5], 16
LDR q17, [x5], 16
LDR s1, [x9], 4
LDR s2, [x10], 4
LDR s3, [x11], 4
LDR s4, [x12], 4
LDR s5, [x4], 4
FMLA v20.8h, v16.8h, v0.h[0]
FMLA v22.8h, v16.8h, v1.h[0]
FMLA v24.8h, v16.8h, v2.h[0]
FMLA v26.8h, v16.8h, v3.h[0]
FMLA v28.8h, v16.8h, v4.h[0]
FMLA v30.8h, v16.8h, v5.h[0]
FMLA v20.8h, v17.8h, v0.h[1]
FMLA v22.8h, v17.8h, v1.h[1]
FMLA v24.8h, v17.8h, v2.h[1]
FMLA v26.8h, v17.8h, v3.h[1]
FMLA v28.8h, v17.8h, v4.h[1]
FMLA v30.8h, v17.8h, v5.h[1]
TBZ x0, 1, 2b
5:
# Remainder- 1 halffloat of A (2 bytes)
LDR h0, [x3], 2
LDR q16, [x5], 16
LDR h1, [x9], 2
LDR h2, [x10], 2
LDR h3, [x11], 2
LDR h4, [x12], 2
LDR h5, [x4], 2
FMLA v20.8h, v16.8h, v0.h[0]
FMLA v22.8h, v16.8h, v1.h[0]
FMLA v24.8h, v16.8h, v2.h[0]
FMLA v26.8h, v16.8h, v3.h[0]
FMLA v28.8h, v16.8h, v4.h[0]
FMLA v30.8h, v16.8h, v5.h[0]
B 2b
# Store odd width
6:
TBZ x1, 2, 7f
STR d30, [x7], 8
STR d28, [x13], 8
DUP d30, v30.d[1]
DUP d28, v28.d[1]
STR d26, [x14], 8
STR d24, [x17], 8
DUP d26, v26.d[1]
DUP d24, v24.d[1]
STR d22, [x16], 8
STR d20, [x6], 8
DUP d22, v22.d[1]
DUP d20, v20.d[1]
7:
TBZ x1, 1, 8f
STR s30, [x7], 4
STR s28, [x13], 4
DUP s30, v30.s[1]
DUP s28, v28.s[1]
STR s26, [x14], 4
STR s24, [x17], 4
DUP s26, v26.s[1]
DUP s24, v24.s[1]
STR s22, [x16], 4
STR s20, [x6], 4
DUP s22, v22.s[1]
DUP s20, v20.s[1]
8:
TBZ x1, 0, 9f
STR h30, [x7]
STR h28, [x13]
STR h26, [x14]
STR h24, [x17]
STR h22, [x16]
STR h20, [x6]
9:
RET
END_FUNCTION xnn_f16_gemminc_minmax_ukernel_6x8__asm_aarch64_neonfp16arith_ld64
#ifdef __ELF__
.section ".note.GNU-stack","",%progbits
#endif