LoongArch: DGEMM small matrix opt

This commit is contained in:
gxw 2023-09-16 11:19:12 +08:00
parent 3a3ff1ba5e
commit 8ab2e9ec65
7 changed files with 2279 additions and 0 deletions

View File

@ -269,6 +269,9 @@ else ifeq ($(ARCH), power)
SMALL_MATRIX_OPT = 1
BUILD_BFLOAT16 = 1
endif
ifeq ($(ARCH), loongarch64)
SMALL_MATRIX_OPT = 1
endif
ifeq ($(SMALL_MATRIX_OPT), 1)
CCOMMON_OPT += -DSMALL_MATRIX_OPT
endif

View File

@ -162,4 +162,14 @@ STRSMKERNEL_LN = ../generic/trsm_kernel_LN.c
STRSMKERNEL_LT = ../generic/trsm_kernel_LT.c
STRSMKERNEL_RN = ../generic/trsm_kernel_RN.c
STRSMKERNEL_RT = ../generic/trsm_kernel_RT.c
DGEMM_SMALL_M_PERMIT = dgemm_small_matrix_permit.c
DGEMM_SMALL_K_NN = dgemm_small_kernel_nn_lasx.S
DGEMM_SMALL_K_B0_NN = dgemm_small_kernel_nn_lasx.S
DGEMM_SMALL_K_NT = dgemm_small_kernel_nt_lasx.S
DGEMM_SMALL_K_B0_NT = dgemm_small_kernel_nt_lasx.S
DGEMM_SMALL_K_TN = dgemm_small_kernel_tn_lasx.S
DGEMM_SMALL_K_B0_TN = dgemm_small_kernel_tn_lasx.S
DGEMM_SMALL_K_TT = dgemm_small_kernel_tt_lasx.S
DGEMM_SMALL_K_B0_TT = dgemm_small_kernel_tt_lasx.S
endif

View File

@ -0,0 +1,549 @@
/***************************************************************************
Copyright (c) 2024, The OpenBLAS Project
All rights reserved.
Redistribution and use in source and binary forms, with or without
modification, are permitted provided that the following conditions are
met:
1. Redistributions of source code must retain the above copyright
notice, this list of conditions and the following disclaimer.
2. Redistributions in binary form must reproduce the above copyright
notice, this list of conditions and the following disclaimer in
the documentation and/or other materials provided with the
distribution.
3. Neither the name of the OpenBLAS project nor the names of
its contributors may be used to endorse or promote products
derived from this software without specific prior written permission.
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
ARE DISCLAIMED. IN NO EVENT SHALL THE OPENBLAS PROJECT OR CONTRIBUTORS BE
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE
USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*****************************************************************************/
#define ASSEMBLER
#include "common.h"
#include "loongarch64_asm.S"
#define M $a0
#define N $a1
#define K $a2
#define A $a3
#define LDA $a4
#define ALPHA $f0
#define B $a5
#define LDB $a6
#define C $a7
#define LDC $t0
#ifdef B0
#define BETA $f1
#endif
#undef ZERO
#define ZERO $r0
#define M16 $t1
#define M8 $t1
#define M4 $t1
#define M2 $t1
#define M1 $t1
#define N4 $t2
#define N2 $t2
#define N1 $t2
#define K8 $t3
#define A0 $t4
#define X0 $t5
#define B1 $t6
#define B2 $t7
#define B3 $t8
#define C0 $s0
#define C1 $s1
#define C2 $s2
#define C3 $s3
#define K1 $s4
#define VALPHA $xr0
#ifndef B0
#define VBETA $xr1
#endif
#define D0 $xr2
#define D1 $xr3
#define D2 $xr4
#define D3 $xr5
#define D4 $xr6
#define D5 $xr7
#define D6 $xr8
#define D7 $xr9
#define D8 $xr10
#define D9 $xr11
#define D10 $xr12
#define D11 $xr13
#define D12 $xr14
#define D13 $xr15
#define D14 $xr16
#define D15 $xr17
#define S0 $xr18
#define S1 $xr19
#define S2 $xr20
#define S3 $xr21
#define Z0 $xr22
#define Z1 $xr23
#define Z2 $xr24
#define Z3 $xr25
#define V0 $vr2
#define V1 $vr3
#define V2 $vr4
#define V3 $vr5
#define F0 $f2
#define F1 $f3
#define F2 $f4
#define F3 $f5
.macro DGEMM_SMALL_KERNEL_NN_TAIL M
PTR_SRAI N4, N, 2 // N >> 2
move A0, A // Restore A0
move X0, B // Restore X0
PTR_ADD B1, X0, LDB
PTR_ADD B2, B1, LDB
PTR_ADD B3, B2, LDB
move C0, C // Restore C0
PTR_ADD C1, C0, LDC
PTR_ADD C2, C1, LDC
PTR_ADD C3, C2, LDC
beqz N4, .L_M\M\()_N3
.L_M\M\()_N4:
GXOR xv, v, D0, D0, D0, D1, D1, D1, D2, D2, D2, D3, D3, D3
move K1, K // Restore K1
PTR_ADDI N4, N4, -1
bge ZERO, K, .L_M\M\()_N4_END
.L_M\M\()_N4_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00
GLDREPL xv, d, Z0, X0, 0x00, Z1, B1, 0x00, Z2, B2, 0x00, Z3, B3, 0x00
GMADD xvf, d, D0, S0, Z0, D0, D1, S0, Z1, D1, D2, S0, Z2, D2, D3, S0, Z3, D3
PTR_ADDI X0, X0, 0x08
PTR_ADDI B1, B1, 0x08
PTR_ADDI B2, B2, 0x08
PTR_ADDI B3, B3, 0x08
PTR_ADD A0, A0, LDA
bnez K1, .L_M\M\()_N4_K1
.L_M\M\()_N4_END:
GMUL xvf, d, D0, D0, VALPHA, D1, D1, VALPHA, D2, D2, VALPHA, D3, D3, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00
GMADD xvf, d, D0, S0, VBETA, D0
GLD xv, , S0, C1, 0x00
GMADD xvf, d, D1, S0, VBETA, D1
GLD xv, , S0, C2, 0x00
GMADD xvf, d, D2, S0, VBETA, D2
GLD xv, , S0, C3, 0x00
GMADD xvf, d, D3, S0, VBETA, D3
#endif
.if \M == 4
GST xv, , D0, C0, 0x00, D1, C1, 0x00, D2, C2, 0x00, D3, C3, 0x00
.elseif \M == 2
GST v, , V0, C0, 0x00, V1, C1, 0x00, V2, C2, 0x00, V3, C3, 0x00
.elseif \M == 1
GST f, d, F0, C0, 0x00, F1, C1, 0x00, F2, C2, 0x00, F3, C3, 0x00
.endif
// Update C0, C1, C2, C3
PTR_ALSL C0, LDC, C0, 2
PTR_ALSL C1, LDC, C1, 2
PTR_ALSL C2, LDC, C2, 2
PTR_ALSL C3, LDC, C3, 2
// Update X0, B1, B2, B3
PTR_SUB X0, X0, K8
PTR_SUB B1, B1, K8
PTR_SUB B2, B2, K8
PTR_SUB B3, B3, K8
PTR_ALSL X0, LDB, X0, 2
PTR_ALSL B1, LDB, B1, 2
PTR_ALSL B2, LDB, B2, 2
PTR_ALSL B3, LDB, B3, 2
// Restore A0
move A0, A
bnez N4, .L_M\M\()_N4
.L_M\M\()_N3:
andi N2, N, 0x02
beqz N2, .L_M\M\()_N1
.L_M\M\()_N2:
GXOR xv, v, D0, D0, D0, D1, D1, D1
move K1, K // Restore K1
bge ZERO, K, .L_M\M\()_N2_END
.L_M\M\()_N2_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00
GLDREPL xv, d, Z0, X0, 0x00, Z1, B1, 0x00
GMADD xvf, d, D0, S0, Z0, D0, D1, S0, Z1, D1
PTR_ADDI X0, X0, 0x08
PTR_ADDI B1, B1, 0x08
PTR_ADD A0, A0, LDA
bnez K1, .L_M\M\()_N2_K1
.L_M\M\()_N2_END:
GMUL xvf, d, D0, D0, VALPHA, D1, D1, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00
GMADD xvf, d, D0, S0, VBETA, D0
GLD xv, , S0, C1, 0x00
GMADD xvf, d, D1, S0, VBETA, D1
#endif
.if \M == 4
GST xv, , D0, C0, 0x00, D1, C1, 0x00
.elseif \M == 2
GST v, , V0, C0, 0x00, V1, C1, 0x00
.elseif \M == 1
GST f, d, F0, C0, 0x00, F1, C1, 0x00
.endif
// Update C0, C1
PTR_ALSL C0, LDC, C0, 1
PTR_ALSL C1, LDC, C1, 1
// Update X0, B1
PTR_SUB X0, X0, K8
PTR_SUB B1, B1, K8
PTR_ALSL X0, LDB, X0, 1
PTR_ALSL B1, LDB, B1, 1
// Restore A0
move A0, A
.L_M\M\()_N1:
andi N1, N, 0x01
beqz N1, .L_M\M\()_END
GXOR xv, v, D0, D0, D0
move K1, K // Restore K1
bge ZERO, K, .L_M\M\()_N1_END
.L_M\M\()_N1_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00
GLDREPL xv, d, Z0, X0, 0x00
GMADD xvf, d, D0, S0, Z0, D0
PTR_ADDI X0, X0, 0x08
PTR_ADD A0, A0, LDA
bnez K1, .L_M\M\()_N1_K1
.L_M\M\()_N1_END:
GMUL xvf, d, D0, D0, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00
GMADD xvf, d, D0, S0, VBETA, D0
#endif
.if \M == 4
GST xv, , D0, C0, 0x00
.elseif \M == 2
GST v, , V0, C0, 0x00
.elseif \M == 1
GST f, d, F0, C0, 0x00
.endif
.L_M\M\()_END:
.if \M == 4
PTR_ADDI A, A, 0x20
PTR_ADDI C, C, 0x20
.elseif \M == 2
PTR_ADDI A, A, 0x10
PTR_ADDI C, C, 0x10
.elseif \M == 1
.endif
.endm
PROLOGUE
PTR_LD LDC, $sp, 0
push_if_used 5, 2
xvreplve0.d VALPHA, VALPHA
#ifndef B0
xvreplve0.d VBETA, VBETA
#endif
PTR_SLLI LDA, LDA, 3
PTR_SLLI LDB, LDB, 3
PTR_SLLI LDC, LDC, 3
PTR_SLLI K8, K, 3
PTR_SRAI M16, M, 4 // M >> 4
beqz M16, .L_M15
.L_M16:
PTR_SRAI N4, N, 2 // N >> 2
move A0, A // Restore A0
move X0, B // Restore X0
PTR_ADD B1, X0, LDB
PTR_ADD B2, B1, LDB
PTR_ADD B3, B2, LDB
move C0, C // Restore C0
PTR_ADD C1, C0, LDC
PTR_ADD C2, C1, LDC
PTR_ADD C3, C2, LDC
beqz N4, .L_M16_N3
.L_M16_N4:
GXOR xv, v, D0, D0, D0, D1, D1, D1, D2, D2, D2, D3, D3, D3, \
D4, D4, D4, D5, D5, D5, D6, D6, D6, D7, D7, D7, \
D8, D8, D8, D9, D9, D9, D10, D10, D10, D11, D11, D11, \
D12, D12, D12, D13, D13, D13, D14, D14, D14, D15, D15, D15
move K1, K // Restore K1
PTR_ADDI N4, N4, -1
bge ZERO, K, .L_M16_N4_END
.L_M16_N4_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00, S1, A0, 0x20, S2, A0, 0x40, S3, A0, 0x60
GLDREPL xv, d, Z0, X0, 0x00, Z1, B1, 0x00, Z2, B2, 0x00, Z3, B3, 0x00
GMADD xvf, d, D0, S0, Z0, D0, D1, S1, Z0, D1, D2, S2, Z0, D2, D3, S3, Z0, D3, \
D4, S0, Z1, D4, D5, S1, Z1, D5, D6, S2, Z1, D6, D7, S3, Z1, D7, \
D8, S0, Z2, D8, D9, S1, Z2, D9, D10, S2, Z2, D10, D11, S3, Z2, D11, \
D12, S0, Z3, D12, D13, S1, Z3, D13, D14, S2, Z3, D14, D15, S3, Z3, D15
PTR_ADDI X0, X0, 0x08
PTR_ADDI B1, B1, 0x08
PTR_ADDI B2, B2, 0x08
PTR_ADDI B3, B3, 0x08
PTR_ADD A0, A0, LDA
bnez K1, .L_M16_N4_K1
.L_M16_N4_END:
GMUL xvf, d, D0, D0, VALPHA, D1, D1, VALPHA, D2, D2, VALPHA, D3, D3, VALPHA, \
D4, D4, VALPHA, D5, D5, VALPHA, D6, D6, VALPHA, D7, D7, VALPHA, \
D8, D8, VALPHA, D9, D9, VALPHA, D10, D10, VALPHA, D11, D11, VALPHA, \
D12, D12, VALPHA, D13, D13, VALPHA, D14, D14, VALPHA, D15, D15, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00, S1, C0, 0x20, S2, C0, 0x40, S3, C0, 0x60
GMADD xvf, d, D0, S0, VBETA, D0, D1, S1, VBETA, D1, D2, S2, VBETA, D2, D3, S3, VBETA, D3
GLD xv, , S0, C1, 0x00, S1, C1, 0x20, S2, C1, 0x40, S3, C1, 0x60
GMADD xvf, d, D4, S0, VBETA, D4, D5, S1, VBETA, D5, D6, S2, VBETA, D6, D7, S3, VBETA, D7
GLD xv, , S0, C2, 0x00, S1, C2, 0x20, S2, C2, 0x40, S3, C2, 0x60
GMADD xvf, d, D8, S0, VBETA, D8, D9, S1, VBETA, D9, D10, S2, VBETA, D10, D11, S3, VBETA, D11
GLD xv, , S0, C3, 0x00, S1, C3, 0x20, S2, C3, 0x40, S3, C3, 0x60
GMADD xvf, d, D12, S0, VBETA, D12, D13, S1, VBETA, D13, D14, S2, VBETA, D14, D15, S3, VBETA, D15
#endif
GST xv, , D12, C3, 0x00, D13, C3, 0x20, D14, C3, 0x40, D15, C3, 0x60, \
D8, C2, 0x00, D9, C2, 0x20, D10, C2, 0x40, D11, C2, 0x60, \
D4, C1, 0x00, D5, C1, 0x20, D6, C1, 0x40, D7, C1, 0x60, \
D0, C0, 0x00, D1, C0, 0x20, D2, C0, 0x40, D3, C0, 0x60
// Update C0, C1, C2, C3
PTR_ALSL C0, LDC, C0, 2
PTR_ALSL C1, LDC, C1, 2
PTR_ALSL C2, LDC, C2, 2
PTR_ALSL C3, LDC, C3, 2
// Update X0, B1, B2, B3
PTR_SUB X0, X0, K8
PTR_SUB B1, B1, K8
PTR_SUB B2, B2, K8
PTR_SUB B3, B3, K8
PTR_ALSL X0, LDB, X0, 2
PTR_ALSL B1, LDB, B1, 2
PTR_ALSL B2, LDB, B2, 2
PTR_ALSL B3, LDB, B3, 2
// Restore A0
move A0, A
bnez N4, .L_M16_N4
.L_M16_N3:
andi N2, N, 0x02
beqz N2, .L_M16_N1
.L_M16_N2:
GXOR xv, v, D0, D0, D0, D1, D1, D1, D2, D2, D2, D3, D3, D3, \
D4, D4, D4, D5, D5, D5, D6, D6, D6, D7, D7, D7
move K1, K // Restore K1
bge ZERO, K, .L_M16_N2_END
.L_M16_N2_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00, S1, A0, 0x20, S2, A0, 0x40, S3, A0, 0x60
GLDREPL xv, d, Z0, X0, 0x00, Z1, B1, 0x00
GMADD xvf, d, D0, S0, Z0, D0, D1, S1, Z0, D1, D2, S2, Z0, D2, D3, S3, Z0, D3, \
D4, S0, Z1, D4, D5, S1, Z1, D5, D6, S2, Z1, D6, D7, S3, Z1, D7
PTR_ADDI X0, X0, 0x08
PTR_ADDI B1, B1, 0x08
PTR_ADD A0, A0, LDA
bnez K1, .L_M16_N2_K1
.L_M16_N2_END:
GMUL xvf, d, D0, D0, VALPHA, D1, D1, VALPHA, D2, D2, VALPHA, D3, D3, VALPHA, \
D4, D4, VALPHA, D5, D5, VALPHA, D6, D6, VALPHA, D7, D7, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00, S1, C0, 0x20, S2, C0, 0x40, S3, C0, 0x60
GMADD xvf, d, D0, S0, VBETA, D0, D1, S1, VBETA, D1, D2, S2, VBETA, D2, D3, S3, VBETA, D3
GLD xv, , S0, C1, 0x00, S1, C1, 0x20, S2, C1, 0x40, S3, C1, 0x60
GMADD xvf, d, D4, S0, VBETA, D4, D5, S1, VBETA, D5, D6, S2, VBETA, D6, D7, S3, VBETA, D7
#endif
GST xv, , D4, C1, 0x00, D5, C1, 0x20, D6, C1, 0x40, D7, C1, 0x60, \
D0, C0, 0x00, D1, C0, 0x20, D2, C0, 0x40, D3, C0, 0x60
// Update C0, C1, C2, C3
PTR_ALSL C0, LDC, C0, 1
PTR_ALSL C1, LDC, C1, 1
// Update X0, B1, B2, B3
PTR_SUB X0, X0, K8
PTR_SUB B1, B1, K8
PTR_ALSL X0, LDB, X0, 1
PTR_ALSL B1, LDB, B1, 1
// Restore A0
move A0, A
.L_M16_N1:
andi N1, N, 0x01
beqz N1, .L_M16_END
GXOR xv, v, D0, D0, D0, D1, D1, D1, D2, D2, D2, D3, D3, D3
move K1, K // Restore K1
bge ZERO, K, .L_M16_N1_END
.L_M16_N1_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00, S1, A0, 0x20, S2, A0, 0x40, S3, A0, 0x60
GLDREPL xv, d, Z0, X0, 0x00
GMADD xvf, d, D0, S0, Z0, D0, D1, S1, Z0, D1, D2, S2, Z0, D2, D3, S3, Z0, D3
PTR_ADDI X0, X0, 0x08
PTR_ADD A0, A0, LDA
bnez K1, .L_M16_N1_K1
.L_M16_N1_END:
GMUL xvf, d, D0, D0, VALPHA, D1, D1, VALPHA, D2, D2, VALPHA, D3, D3, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00, S1, C0, 0x20, S2, C0, 0x40, S3, C0, 0x60
GMADD xvf, d, D0, S0, VBETA, D0, D1, S1, VBETA, D1, D2, S2, VBETA, D2, D3, S3, VBETA, D3
#endif
GST xv, , D0, C0, 0x00, D1, C0, 0x20, D2, C0, 0x40, D3, C0, 0x60
// Update C0, C1, C2, C3
PTR_ALSL C0, LDC, C0, 2
// Update X0, B1, B2, B3
PTR_SUB X0, X0, K8
PTR_ALSL X0, LDB, X0, 2
// Restore A0
move A0, A
.L_M16_END:
PTR_ADDI M16, M16, -1
PTR_ADDI A, A, 0x80
PTR_ADDI C, C, 0x80
bnez M16, .L_M16
.L_M15:
andi M8, M, 0x08
beqz M8, .L_M7
.L_M8:
PTR_SRAI N4, N, 2 // N >> 2
move A0, A // Restore A0
move X0, B // Restore X0
PTR_ADD B1, X0, LDB
PTR_ADD B2, B1, LDB
PTR_ADD B3, B2, LDB
move C0, C // Restore C0
PTR_ADD C1, C0, LDC
PTR_ADD C2, C1, LDC
PTR_ADD C3, C2, LDC
beqz N4, .L_M8_N3
.L_M8_N4:
GXOR xv, v, D0, D0, D0, D1, D1, D1, D2, D2, D2, D3, D3, D3, \
D4, D4, D4, D5, D5, D5, D6, D6, D6, D7, D7, D7
move K1, K // Restore K1
PTR_ADDI N4, N4, -1
bge ZERO, K, .L_M8_N4_END
.L_M8_N4_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00, S1, A0, 0x20
GLDREPL xv, d, Z0, X0, 0x00, Z1, B1, 0x00, Z2, B2, 0x00, Z3, B3, 0x00
GMADD xvf, d, D0, S0, Z0, D0, D1, S1, Z0, D1, \
D2, S0, Z1, D2, D3, S1, Z1, D3, \
D4, S0, Z2, D4, D5, S1, Z2, D5, \
D6, S0, Z3, D6, D7, S1, Z3, D7,
PTR_ADDI X0, X0, 0x08
PTR_ADDI B1, B1, 0x08
PTR_ADDI B2, B2, 0x08
PTR_ADDI B3, B3, 0x08
PTR_ADD A0, A0, LDA
bnez K1, .L_M8_N4_K1
.L_M8_N4_END:
GMUL xvf, d, D0, D0, VALPHA, D1, D1, VALPHA, D2, D2, VALPHA, D3, D3, VALPHA, \
D4, D4, VALPHA, D5, D5, VALPHA, D6, D6, VALPHA, D7, D7, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00, S1, C0, 0x20
GMADD xvf, d, D0, S0, VBETA, D0, D1, S1, VBETA, D1
GLD xv, , S0, C1, 0x00, S1, C1, 0x20
GMADD xvf, d, D2, S0, VBETA, D2, D3, S1, VBETA, D3
GLD xv, , S0, C2, 0x00, S1, C2, 0x20
GMADD xvf, d, D4, S0, VBETA, D4, D5, S1, VBETA, D5
GLD xv, , S0, C3, 0x00, S1, C3, 0x20
GMADD xvf, d, D6, S0, VBETA, D6, D7, S1, VBETA, D7
#endif
GST xv, , D4, C2, 0x00, D5, C2, 0x20, D6, C3, 0x00, D7, C3, 0x20, \
D0, C0, 0x00, D1, C0, 0x20, D2, C1, 0x00, D3, C1, 0x20
// Update C0, C1, C2, C3
PTR_ALSL C0, LDC, C0, 2
PTR_ALSL C1, LDC, C1, 2
PTR_ALSL C2, LDC, C2, 2
PTR_ALSL C3, LDC, C3, 2
// Update X0, B1, B2, B3
PTR_SUB X0, X0, K8
PTR_SUB B1, B1, K8
PTR_SUB B2, B2, K8
PTR_SUB B3, B3, K8
PTR_ALSL X0, LDB, X0, 2
PTR_ALSL B1, LDB, B1, 2
PTR_ALSL B2, LDB, B2, 2
PTR_ALSL B3, LDB, B3, 2
// Restore A0
move A0, A
bnez N4, .L_M8_N4
.L_M8_N3:
andi N2, N, 0x02
beqz N2, .L_M8_N1
.L_M8_N2:
GXOR xv, v, D0, D0, D0, D1, D1, D1, D2, D2, D2, D3, D3, D3
move K1, K // Restore K1
bge ZERO, K, .L_M8_N2_END
.L_M8_N2_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00, S1, A0, 0x20
GLDREPL xv, d, Z0, X0, 0x00, Z1, B1, 0x00
GMADD xvf, d, D0, S0, Z0, D0, D1, S1, Z0, D1, \
D2, S0, Z1, D2, D3, S1, Z1, D3
PTR_ADDI X0, X0, 0x08
PTR_ADDI B1, B1, 0x08
PTR_ADD A0, A0, LDA
bnez K1, .L_M8_N2_K1
.L_M8_N2_END:
GMUL xvf, d, D0, D0, VALPHA, D1, D1, VALPHA, D2, D2, VALPHA, D3, D3, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00, S1, C0, 0x20
GMADD xvf, d, D0, S0, VBETA, D0, D1, S1, VBETA, D1
GLD xv, , S0, C1, 0x00, S1, C1, 0x20
GMADD xvf, d, D2, S0, VBETA, D2, D3, S1, VBETA, D3
#endif
GST xv, , D0, C0, 0x00, D1, C0, 0x20, D2, C1, 0x00, D3, C1, 0x20
// Update C0, C1
PTR_ALSL C0, LDC, C0, 1
PTR_ALSL C1, LDC, C1, 1
// Update X0, B1
PTR_SUB X0, X0, K8
PTR_SUB B1, B1, K8
PTR_ALSL X0, LDB, X0, 1
PTR_ALSL B1, LDB, B1, 1
// Restore A0
move A0, A
.L_M8_N1:
andi N1, N, 0x01
beqz N1, .L_M8_END
GXOR xv, v, D0, D0, D0, D1, D1, D1
move K1, K // Restore K1
bge ZERO, K, .L_M8_N1_END
.L_M8_N1_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00, S1, A0, 0x20
GLDREPL xv, d, Z0, X0, 0x00
GMADD xvf, d, D0, S0, Z0, D0, D1, S1, Z0, D1
PTR_ADDI X0, X0, 0x08
PTR_ADD A0, A0, LDA
bnez K1, .L_M8_N1_K1
.L_M8_N1_END:
GMUL xvf, d, D0, D0, VALPHA, D1, D1, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00, S1, C0, 0x20
GMADD xvf, d, D0, S0, VBETA, D0, D1, S1, VBETA, D1
#endif
GST xv, , D0, C0, 0x00, D1, C0, 0x20
.L_M8_END:
PTR_ADDI A, A, 0x40
PTR_ADDI C, C, 0x40
.L_M7:
andi M4, M, 0x04
beqz M4, .L_M3
.L_M4:
DGEMM_SMALL_KERNEL_NN_TAIL 4
.L_M3:
andi M2, M, 0x02
beqz M2, .L_M1
.L_M2:
DGEMM_SMALL_KERNEL_NN_TAIL 2
.L_M1:
andi M1, M, 0x01
beqz M1, .L_M0
DGEMM_SMALL_KERNEL_NN_TAIL 1
.L_M0:
pop_if_used 5, 2
jirl $r0, $r1, 0x0
EPILOGUE

View File

@ -0,0 +1,500 @@
/***************************************************************************
Copyright (c) 2024 The OpenBLAS Project
All rights reserved.
Redistribution and use in source and binary forms, with or without
modification, are permitted provided that the following conditions are
met:
1. Redistributions of source code must retain the above copyright
notice, this list of conditions and the following disclaimer.
2. Redistributions in binary form must reproduce the above copyright
notice, this list of conditions and the following disclaimer in
the documentation and/or other materials provided with the
distribution.
3. Neither the name of the OpenBLAS project nor the names of
its contributors may be used to endorse or promote products
derived from this software without specific prior written permission.
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
ARE DISCLAIMED. IN NO EVENT SHALL THE OPENBLAS PROJECT OR CONTRIBUTORS BE
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE
USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*****************************************************************************/
#define ASSEMBLER
#include "common.h"
#include "loongarch64_asm.S"
#define M $a0
#define N $a1
#define K $a2
#define A $a3
#define LDA $a4
#define ALPHA $f0
#define B $a5
#define LDB $a6
#define C $a7
#define LDC $t0
#ifdef B0
#define BETA $f1
#endif
#undef ZERO
#define ZERO $r0
#define M16 $t1
#define M8 $t1
#define M4 $t1
#define M2 $t1
#define M1 $t1
#define N4 $t2
#define N2 $t2
#define N1 $t2
#define K_LDB $t3
#define A0 $t4
#define X0 $t5
#define C0 $t6
#define C1 $t7
#define C2 $t8
#define C3 $s0
#define K1 $s1
#define VALPHA $xr0
#ifndef B0
#define VBETA $xr1
#endif
#define D0 $xr2
#define D1 $xr3
#define D2 $xr4
#define D3 $xr5
#define D4 $xr6
#define D5 $xr7
#define D6 $xr8
#define D7 $xr9
#define D8 $xr10
#define D9 $xr11
#define D10 $xr12
#define D11 $xr13
#define D12 $xr14
#define D13 $xr15
#define D14 $xr16
#define D15 $xr17
#define S0 $xr18
#define S1 $xr19
#define S2 $xr20
#define S3 $xr21
#define Z0 $xr22
#define Z1 $xr23
#define Z2 $xr24
#define Z3 $xr25
#define V0 $vr2
#define V1 $vr3
#define V2 $vr4
#define V3 $vr5
#define F0 $f2
#define F1 $f3
#define F2 $f4
#define F3 $f5
.macro DGEMM_SMALL_KERNEL_NT_TAIL M
PTR_SRAI N4, N, 2 // N >> 2
move A0, A // Restore A0
move X0, B // Restore X0
move C0, C // Restore C0
PTR_ADD C1, C0, LDC
PTR_ADD C2, C1, LDC
PTR_ADD C3, C2, LDC
beqz N4, .L_M\M\()_N3
.L_M\M\()_N4:
GXOR xv, v, D0, D0, D0, D1, D1, D1, D2, D2, D2, D3, D3, D3
move K1, K // Restore K1
PTR_ADDI N4, N4, -1
bge ZERO, K, .L_M\M\()_N4_END
.L_M\M\()_N4_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00
GLDREPL xv, d, Z0, X0, 0x00, Z1, X0, 0x08, Z2, X0, 0x10, Z3, X0, 0x18
GMADD xvf, d, D0, S0, Z0, D0, D1, S0, Z1, D1, D2, S0, Z2, D2, D3, S0, Z3, D3
PTR_ADD X0, X0, LDB
PTR_ADD A0, A0, LDA
bnez K1, .L_M\M\()_N4_K1
.L_M\M\()_N4_END:
GMUL xvf, d, D0, D0, VALPHA, D1, D1, VALPHA, D2, D2, VALPHA, D3, D3, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00
GMADD xvf, d, D0, S0, VBETA, D0
GLD xv, , S0, C1, 0x00
GMADD xvf, d, D1, S0, VBETA, D1
GLD xv, , S0, C2, 0x00
GMADD xvf, d, D2, S0, VBETA, D2
GLD xv, , S0, C3, 0x00
GMADD xvf, d, D3, S0, VBETA, D3
#endif
.if \M == 4
GST xv, , D0, C0, 0x00, D1, C1, 0x00, D2, C2, 0x00, D3, C3, 0x00
.elseif \M == 2
GST v, , V0, C0, 0x00, V1, C1, 0x00, V2, C2, 0x00, V3, C3, 0x00
.elseif \M == 1
GST f, d, F0, C0, 0x00, F1, C1, 0x00, F2, C2, 0x00, F3, C3, 0x00
.endif
// Update C0, C1, C2, C3
PTR_ALSL C0, LDC, C0, 2
PTR_ALSL C1, LDC, C1, 2
PTR_ALSL C2, LDC, C2, 2
PTR_ALSL C3, LDC, C3, 2
// Update X0
PTR_SUB X0, X0, K_LDB
PTR_ADDI X0, X0, 0x20
// Restore A0
move A0, A
bnez N4, .L_M\M\()_N4
.L_M\M\()_N3:
andi N2, N, 0x02
beqz N2, .L_M\M\()_N1
.L_M\M\()_N2:
GXOR xv, v, D0, D0, D0, D1, D1, D1
move K1, K // Restore K1
bge ZERO, K, .L_M\M\()_N2_END
.L_M\M\()_N2_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00
GLDREPL xv, d, Z0, X0, 0x00, Z1, X0, 0x08
GMADD xvf, d, D0, S0, Z0, D0, D1, S0, Z1, D1
PTR_ADD X0, X0, LDB
PTR_ADD A0, A0, LDA
bnez K1, .L_M\M\()_N2_K1
.L_M\M\()_N2_END:
GMUL xvf, d, D0, D0, VALPHA, D1, D1, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00
GMADD xvf, d, D0, S0, VBETA, D0
GLD xv, , S0, C1, 0x00
GMADD xvf, d, D1, S0, VBETA, D1
#endif
.if \M == 4
GST xv, , D0, C0, 0x00, D1, C1, 0x00
.elseif \M == 2
GST v, , V0, C0, 0x00, V1, C1, 0x00
.elseif \M == 1
GST f, d, F0, C0, 0x00, F1, C1, 0x00
.endif
// Update C0, C1
PTR_ALSL C0, LDC, C0, 1
PTR_ALSL C1, LDC, C1, 1
// Update X0
PTR_SUB X0, X0, K_LDB
PTR_ADDI X0, X0, 0x10
// Restore A0
move A0, A
.L_M\M\()_N1:
andi N1, N, 0x01
beqz N1, .L_M\M\()_END
GXOR xv, v, D0, D0, D0
move K1, K // Restore K1
bge ZERO, K, .L_M\M\()_N1_END
.L_M\M\()_N1_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00
GLDREPL xv, d, Z0, X0, 0x00
GMADD xvf, d, D0, S0, Z0, D0
PTR_ADD X0, X0, LDB
PTR_ADD A0, A0, LDA
bnez K1, .L_M\M\()_N1_K1
.L_M\M\()_N1_END:
GMUL xvf, d, D0, D0, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00
GMADD xvf, d, D0, S0, VBETA, D0
#endif
.if \M == 4
GST xv, , D0, C0, 0x00
.elseif \M == 2
GST v, , V0, C0, 0x00
.elseif \M == 1
GST f, d, F0, C0, 0x00
.endif
.L_M\M\()_END:
.if \M == 4
PTR_ADDI A, A, 0x20
PTR_ADDI C, C, 0x20
.elseif \M == 2
PTR_ADDI A, A, 0x10
PTR_ADDI C, C, 0x10
.elseif \M == 1
.endif
.endm
PROLOGUE
PTR_LD LDC, $sp, 0
push_if_used 2, 2
xvreplve0.d VALPHA, VALPHA
#ifndef B0
xvreplve0.d VBETA, VBETA
#endif
PTR_SLLI LDA, LDA, 3
PTR_SLLI LDB, LDB, 3
PTR_SLLI LDC, LDC, 3
PTR_MUL K_LDB, K, LDB
PTR_SRAI M16, M, 4 // M >> 4
beqz M16, .L_M15
.L_M16:
PTR_SRAI N4, N, 2 // N >> 2
move A0, A // Restore A0
move X0, B // Restore X0
move C0, C // Restore C0
PTR_ADD C1, C0, LDC
PTR_ADD C2, C1, LDC
PTR_ADD C3, C2, LDC
beqz N4, .L_M16_N3
.L_M16_N4:
GXOR xv, v, D0, D0, D0, D1, D1, D1, D2, D2, D2, D3, D3, D3, \
D4, D4, D4, D5, D5, D5, D6, D6, D6, D7, D7, D7, \
D8, D8, D8, D9, D9, D9, D10, D10, D10, D11, D11, D11, \
D12, D12, D12, D13, D13, D13, D14, D14, D14, D15, D15, D15
move K1, K // Restore K1
PTR_ADDI N4, N4, -1
bge ZERO, K, .L_M16_N4_END
.L_M16_N4_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00, S1, A0, 0x20, S2, A0, 0x40, S3, A0, 0x60
GLDREPL xv, d, Z0, X0, 0x00, Z1, X0, 0x08, Z2, X0, 0x10, Z3, X0, 0x18
GMADD xvf, d, D0, S0, Z0, D0, D1, S1, Z0, D1, D2, S2, Z0, D2, D3, S3, Z0, D3, \
D4, S0, Z1, D4, D5, S1, Z1, D5, D6, S2, Z1, D6, D7, S3, Z1, D7, \
D8, S0, Z2, D8, D9, S1, Z2, D9, D10, S2, Z2, D10, D11, S3, Z2, D11, \
D12, S0, Z3, D12, D13, S1, Z3, D13, D14, S2, Z3, D14, D15, S3, Z3, D15
PTR_ADD X0, X0, LDB
PTR_ADD A0, A0, LDA
bnez K1, .L_M16_N4_K1
.L_M16_N4_END:
GMUL xvf, d, D0, D0, VALPHA, D1, D1, VALPHA, D2, D2, VALPHA, D3, D3, VALPHA, \
D4, D4, VALPHA, D5, D5, VALPHA, D6, D6, VALPHA, D7, D7, VALPHA, \
D8, D8, VALPHA, D9, D9, VALPHA, D10, D10, VALPHA, D11, D11, VALPHA, \
D12, D12, VALPHA, D13, D13, VALPHA, D14, D14, VALPHA, D15, D15, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00, S1, C0, 0x20, S2, C0, 0x40, S3, C0, 0x60
GMADD xvf, d, D0, S0, VBETA, D0, D1, S1, VBETA, D1, D2, S2, VBETA, D2, D3, S3, VBETA, D3
GLD xv, , S0, C1, 0x00, S1, C1, 0x20, S2, C1, 0x40, S3, C1, 0x60
GMADD xvf, d, D4, S0, VBETA, D4, D5, S1, VBETA, D5, D6, S2, VBETA, D6, D7, S3, VBETA, D7
GLD xv, , S0, C2, 0x00, S1, C2, 0x20, S2, C2, 0x40, S3, C2, 0x60
GMADD xvf, d, D8, S0, VBETA, D8, D9, S1, VBETA, D9, D10, S2, VBETA, D10, D11, S3, VBETA, D11
GLD xv, , S0, C3, 0x00, S1, C3, 0x20, S2, C3, 0x40, S3, C3, 0x60
GMADD xvf, d, D12, S0, VBETA, D12, D13, S1, VBETA, D13, D14, S2, VBETA, D14, D15, S3, VBETA, D15
#endif
GST xv, , D12, C3, 0x00, D13, C3, 0x20, D14, C3, 0x40, D15, C3, 0x60, \
D8, C2, 0x00, D9, C2, 0x20, D10, C2, 0x40, D11, C2, 0x60, \
D4, C1, 0x00, D5, C1, 0x20, D6, C1, 0x40, D7, C1, 0x60, \
D0, C0, 0x00, D1, C0, 0x20, D2, C0, 0x40, D3, C0, 0x60
// Update C0, C1, C2, C3
PTR_ALSL C0, LDC, C0, 2
PTR_ALSL C1, LDC, C1, 2
PTR_ALSL C2, LDC, C2, 2
PTR_ALSL C3, LDC, C3, 2
// Update X0
PTR_SUB X0, X0, K_LDB
PTR_ADDI X0, X0, 0x20
// Restore A0
move A0, A
bnez N4, .L_M16_N4
.L_M16_N3:
andi N2, N, 0x02
beqz N2, .L_M16_N1
.L_M16_N2:
GXOR xv, v, D0, D0, D0, D1, D1, D1, D2, D2, D2, D3, D3, D3, \
D4, D4, D4, D5, D5, D5, D6, D6, D6, D7, D7, D7
move K1, K // Restore K1
bge ZERO, K, .L_M16_N2_END
.L_M16_N2_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00, S1, A0, 0x20, S2, A0, 0x40, S3, A0, 0x60
GLDREPL xv, d, Z0, X0, 0x00, Z1, X0, 0x08
GMADD xvf, d, D0, S0, Z0, D0, D1, S1, Z0, D1, D2, S2, Z0, D2, D3, S3, Z0, D3, \
D4, S0, Z1, D4, D5, S1, Z1, D5, D6, S2, Z1, D6, D7, S3, Z1, D7
PTR_ADD X0, X0, LDB
PTR_ADD A0, A0, LDA
bnez K1, .L_M16_N2_K1
.L_M16_N2_END:
GMUL xvf, d, D0, D0, VALPHA, D1, D1, VALPHA, D2, D2, VALPHA, D3, D3, VALPHA, \
D4, D4, VALPHA, D5, D5, VALPHA, D6, D6, VALPHA, D7, D7, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00, S1, C0, 0x20, S2, C0, 0x40, S3, C0, 0x60
GMADD xvf, d, D0, S0, VBETA, D0, D1, S1, VBETA, D1, D2, S2, VBETA, D2, D3, S3, VBETA, D3
GLD xv, , S0, C1, 0x00, S1, C1, 0x20, S2, C1, 0x40, S3, C1, 0x60
GMADD xvf, d, D4, S0, VBETA, D4, D5, S1, VBETA, D5, D6, S2, VBETA, D6, D7, S3, VBETA, D7
#endif
GST xv, , D4, C1, 0x00, D5, C1, 0x20, D6, C1, 0x40, D7, C1, 0x60, \
D0, C0, 0x00, D1, C0, 0x20, D2, C0, 0x40, D3, C0, 0x60
// Update C0, C1
PTR_ALSL C0, LDC, C0, 1
PTR_ALSL C1, LDC, C1, 1
// Update X0
PTR_SUB X0, X0, K_LDB
PTR_ADDI X0, X0, 0x10
// Restore A0
move A0, A
.L_M16_N1:
andi N1, N, 0x01
beqz N1, .L_M16_END
GXOR xv, v, D0, D0, D0, D1, D1, D1, D2, D2, D2, D3, D3, D3
move K1, K // Restore K1
bge ZERO, K, .L_M16_N1_END
.L_M16_N1_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00, S1, A0, 0x20, S2, A0, 0x40, S3, A0, 0x60
GLDREPL xv, d, Z0, X0, 0x00
GMADD xvf, d, D0, S0, Z0, D0, D1, S1, Z0, D1, D2, S2, Z0, D2, D3, S3, Z0, D3
PTR_ADD X0, X0, LDB
PTR_ADD A0, A0, LDA
bnez K1, .L_M16_N1_K1
.L_M16_N1_END:
GMUL xvf, d, D0, D0, VALPHA, D1, D1, VALPHA, D2, D2, VALPHA, D3, D3, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00, S1, C0, 0x20, S2, C0, 0x40, S3, C0, 0x60
GMADD xvf, d, D0, S0, VBETA, D0, D1, S1, VBETA, D1, D2, S2, VBETA, D2, D3, S3, VBETA, D3
#endif
GST xv, , D0, C0, 0x00, D1, C0, 0x20, D2, C0, 0x40, D3, C0, 0x60
// Update C0
PTR_ALSL C0, LDC, C0, 2
// Update X0
PTR_SUB X0, X0, K_LDB
PTR_ADDI X0, X0, 0x08
// Restore A0
move A0, A
.L_M16_END:
PTR_ADDI M16, M16, -1
PTR_ADDI A, A, 0x80
PTR_ADDI C, C, 0x80
bnez M16, .L_M16
.L_M15:
andi M8, M, 0x08
beqz M8, .L_M7
.L_M8:
PTR_SRAI N4, N, 2 // N >> 2
move A0, A // Restore A0
move X0, B // Restore X0
move C0, C // Restore C0
PTR_ADD C1, C0, LDC
PTR_ADD C2, C1, LDC
PTR_ADD C3, C2, LDC
beqz N4, .L_M8_N3
.L_M8_N4:
GXOR xv, v, D0, D0, D0, D1, D1, D1, D2, D2, D2, D3, D3, D3, \
D4, D4, D4, D5, D5, D5, D6, D6, D6, D7, D7, D7
move K1, K // Restore K1
PTR_ADDI N4, N4, -1
bge ZERO, K, .L_M8_N4_END
.L_M8_N4_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00, S1, A0, 0x20
GLDREPL xv, d, Z0, X0, 0x00, Z1, X0, 0x08, Z2, X0, 0x10, Z3, X0, 0x18
GMADD xvf, d, D0, S0, Z0, D0, D1, S1, Z0, D1, \
D2, S0, Z1, D2, D3, S1, Z1, D3, \
D4, S0, Z2, D4, D5, S1, Z2, D5, \
D6, S0, Z3, D6, D7, S1, Z3, D7,
PTR_ADD X0, X0, LDB
PTR_ADD A0, A0, LDA
bnez K1, .L_M8_N4_K1
.L_M8_N4_END:
GMUL xvf, d, D0, D0, VALPHA, D1, D1, VALPHA, D2, D2, VALPHA, D3, D3, VALPHA, \
D4, D4, VALPHA, D5, D5, VALPHA, D6, D6, VALPHA, D7, D7, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00, S1, C0, 0x20
GMADD xvf, d, D0, S0, VBETA, D0, D1, S1, VBETA, D1
GLD xv, , S0, C1, 0x00, S1, C1, 0x20
GMADD xvf, d, D2, S0, VBETA, D2, D3, S1, VBETA, D3
GLD xv, , S0, C2, 0x00, S1, C2, 0x20
GMADD xvf, d, D4, S0, VBETA, D4, D5, S1, VBETA, D5
GLD xv, , S0, C3, 0x00, S1, C3, 0x20
GMADD xvf, d, D6, S0, VBETA, D6, D7, S1, VBETA, D7
#endif
GST xv, , D4, C2, 0x00, D5, C2, 0x20, D6, C3, 0x00, D7, C3, 0x20, \
D0, C0, 0x00, D1, C0, 0x20, D2, C1, 0x00, D3, C1, 0x20
// Update C0, C1, C2, C3
PTR_ALSL C0, LDC, C0, 2
PTR_ALSL C1, LDC, C1, 2
PTR_ALSL C2, LDC, C2, 2
PTR_ALSL C3, LDC, C3, 2
// Update X0
PTR_SUB X0, X0, K_LDB
PTR_ADDI X0, X0, 0x20
// Restore A0
move A0, A
bnez N4, .L_M8_N4
.L_M8_N3:
andi N2, N, 0x02
beqz N2, .L_M8_N1
.L_M8_N2:
GXOR xv, v, D0, D0, D0, D1, D1, D1, D2, D2, D2, D3, D3, D3
move K1, K // Restore K1
bge ZERO, K, .L_M8_N2_END
.L_M8_N2_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00, S1, A0, 0x20
GLDREPL xv, d, Z0, X0, 0x00, Z1, X0, 0x08
GMADD xvf, d, D0, S0, Z0, D0, D1, S1, Z0, D1, \
D2, S0, Z1, D2, D3, S1, Z1, D3
PTR_ADD X0, X0, LDB
PTR_ADD A0, A0, LDA
bnez K1, .L_M8_N2_K1
.L_M8_N2_END:
GMUL xvf, d, D0, D0, VALPHA, D1, D1, VALPHA, D2, D2, VALPHA, D3, D3, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00, S1, C0, 0x20
GMADD xvf, d, D0, S0, VBETA, D0, D1, S1, VBETA, D1
GLD xv, , S0, C1, 0x00, S1, C1, 0x20
GMADD xvf, d, D2, S0, VBETA, D2, D3, S1, VBETA, D3
#endif
GST xv, , D0, C0, 0x00, D1, C0, 0x20, D2, C1, 0x00, D3, C1, 0x20
// Update C0, C1
PTR_ALSL C0, LDC, C0, 1
PTR_ALSL C1, LDC, C1, 1
// Update X0
PTR_SUB X0, X0, K_LDB
PTR_ADDI X0, X0, 0x10
// Restore A0
move A0, A
.L_M8_N1:
andi N1, N, 0x01
beqz N1, .L_M8_END
GXOR xv, v, D0, D0, D0, D1, D1, D1
move K1, K // Restore K1
bge ZERO, K, .L_M8_N1_END
.L_M8_N1_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00, S1, A0, 0x20
GLDREPL xv, d, Z0, X0, 0x00
GMADD xvf, d, D0, S0, Z0, D0, D1, S1, Z0, D1
PTR_ADD X0, X0, LDB
PTR_ADD A0, A0, LDA
bnez K1, .L_M8_N1_K1
.L_M8_N1_END:
GMUL xvf, d, D0, D0, VALPHA, D1, D1, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00, S1, C0, 0x20
GMADD xvf, d, D0, S0, VBETA, D0, D1, S1, VBETA, D1
#endif
GST xv, , D0, C0, 0x00, D1, C0, 0x20
.L_M8_END:
PTR_ADDI A, A, 0x40
PTR_ADDI C, C, 0x40
.L_M7:
andi M4, M, 0x04
beqz M4, .L_M3
.L_M4:
DGEMM_SMALL_KERNEL_NT_TAIL 4
.L_M3:
andi M2, M, 0x02
beqz M2, .L_M1
.L_M2:
DGEMM_SMALL_KERNEL_NT_TAIL 2
.L_M1:
andi M1, M, 0x01
beqz M1, .L_M0
DGEMM_SMALL_KERNEL_NT_TAIL 1
.L_M0:
pop_if_used 2, 2
jirl $r0, $r1, 0x0
EPILOGUE

View File

@ -0,0 +1,639 @@
/***************************************************************************
Copyright (c) 2024, The OpenBLAS Project
All rights reserved.
Redistribution and use in source and binary forms, with or without
modification, are permitted provided that the following conditions are
met:
1. Redistributions of source code must retain the above copyright
notice, this list of conditions and the following disclaimer.
2. Redistributions in binary form must reproduce the above copyright
notice, this list of conditions and the following disclaimer in
the documentation and/or other materials provided with the
distribution.
3. Neither the name of the OpenBLAS project nor the names of
its contributors may be used to endorse or promote products
derived from this software without specific prior written permission.
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
ARE DISCLAIMED. IN NO EVENT SHALL THE OPENBLAS PROJECT OR CONTRIBUTORS BE
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE
USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*****************************************************************************/
#define ASSEMBLER
#include "common.h"
#include "loongarch64_asm.S"
#define M $a0
#define N $a1
#define K $a2
#define A $a3
#define LDA $a4
#define ALPHA $f0
#define B $a5
#define LDB $a6
#define C $a7
#define LDC $t0
#ifdef B0
#define BETA $f1
#endif
#undef ZERO
#define ZERO $r0
#define M4 $t1
#define M2 $t1
#define M1 $t1
#define N4 $t2
#define N2 $t2
#define N1 $t2
#define K8 $t3
#define A0 $t4
#define X0 $t5
#define B1 $t6
#define B2 $t7
#define B3 $t8
#define C0 $s0
#define C1 $s1
#define C2 $s2
#define C3 $s3
#define K1 $s4
#define A1 $s5
#define A2 $s6
#define A3 $s7
#define VALPHA $xr0
#ifndef B0
#define VBETA $xr1
#endif
#define D0 $xr2
#define D1 $xr3
#define D2 $xr4
#define D3 $xr5
#define T0 $xr6
#define T1 $xr7
#define T2 $xr8
#define T3 $xr9
#define Y0 $xr10
#define Y1 $xr11
#define Y2 $xr12
#define Y3 $xr13
#define G0 $xr14
#define G1 $xr15
#define G2 $xr16
#define G3 $xr17
#define S0 $xr18
#define S1 $xr19
#define S2 $xr20
#define S3 $xr21
#define Z0 $xr22
#define Z1 $xr23
#define Z2 $xr24
#define Z3 $xr25
#define V0 $vr2
#define V1 $vr3
#define V2 $vr4
#define V3 $vr5
#define F0 $f2
#define F1 $f3
#define F2 $f4
#define F3 $f5
PROLOGUE
PTR_LD LDC, $sp, 0
push_if_used 8, 2
xvreplve0.d VALPHA, VALPHA
#ifndef B0
xvreplve0.d VBETA, VBETA
#endif
PTR_SLLI LDA, LDA, 3
PTR_SLLI LDB, LDB, 3
PTR_SLLI LDC, LDC, 3
PTR_SLLI K8, K, 3
PTR_SRAI M4, M, 2 // M >> 2
beqz M4, .L_M3
.L_M4:
PTR_SRAI N4, N, 2 // N >> 2
move A0, A // Restore A0
PTR_ADD A1, A0, LDA
PTR_ADD A2, A1, LDA
PTR_ADD A3, A2, LDA
move X0, B // Restore X0
PTR_ADD B1, X0, LDB
PTR_ADD B2, B1, LDB
PTR_ADD B3, B2, LDB
move C0, C // Restore C0
PTR_ADD C1, C0, LDC
PTR_ADD C2, C1, LDC
PTR_ADD C3, C2, LDC
beqz N4, .L_M4_N3
.L_M4_N4:
GXOR xv, v, D0, D0, D0, D1, D1, D1, D2, D2, D2, D3, D3, D3
move K1, K // Restore K1
PTR_ADDI N4, N4, -1
bge ZERO, K, .L_M4_N4_END
PTR_SRAI K1, K1, 3
beq ZERO, K1, .L_M4_N4_K7
.L_M4_N4_K8:
PTR_ADDI K1, K1, -1
GLD xv, , T0, A0, 0x00, T1, A1, 0x00, T2, A2, 0x00, T3, A3, 0x00
GTRANSPOSE4x4_D T0, T1, T2, T3, S0, S1, S2, S3, Z0, Z1
GLDREPL xv, d, Z0, X0, 0x00, Z1, B1, 0x00, Z2, B2, 0x00, Z3, B3, 0x00
GLDREPL xv, d, T0, X0, 0x08, T1, B1, 0x08, T2, B2, 0x08, T3, B3, 0x08
GLDREPL xv, d, Y0, X0, 0x10, Y1, B1, 0x10, Y2, B2, 0x10, Y3, B3, 0x10
GLDREPL xv, d, G0, X0, 0x18, G1, B1, 0x18, G2, B2, 0x18, G3, B3, 0x18
GMADD xvf, d, D0, S0, Z0, D0, \
D1, S0, Z1, D1, \
D2, S0, Z2, D2, \
D3, S0, Z3, D3
GMADD xvf, d, D0, S1, T0, D0, \
D1, S1, T1, D1, \
D2, S1, T2, D2, \
D3, S1, T3, D3
GMADD xvf, d, D0, S2, Y0, D0, \
D1, S2, Y1, D1, \
D2, S2, Y2, D2, \
D3, S2, Y3, D3
GMADD xvf, d, D0, S3, G0, D0, \
D1, S3, G1, D1, \
D2, S3, G2, D2, \
D3, S3, G3, D3
GLD xv, , T0, A0, 0x20, T1, A1, 0x20, T2, A2, 0x20, T3, A3, 0x20
GTRANSPOSE4x4_D T0, T1, T2, T3, S0, S1, S2, S3, Z0, Z1
GLDREPL xv, d, Z0, X0, 0x20, Z1, B1, 0x20, Z2, B2, 0x20, Z3, B3, 0x20
GLDREPL xv, d, T0, X0, 0x28, T1, B1, 0x28, T2, B2, 0x28, T3, B3, 0x28
GLDREPL xv, d, Y0, X0, 0x30, Y1, B1, 0x30, Y2, B2, 0x30, Y3, B3, 0x30
GLDREPL xv, d, G0, X0, 0x38, G1, B1, 0x38, G2, B2, 0x38, G3, B3, 0x38
GMADD xvf, d, D0, S0, Z0, D0, \
D1, S0, Z1, D1, \
D2, S0, Z2, D2, \
D3, S0, Z3, D3
GMADD xvf, d, D0, S1, T0, D0, \
D1, S1, T1, D1, \
D2, S1, T2, D2, \
D3, S1, T3, D3
GMADD xvf, d, D0, S2, Y0, D0, \
D1, S2, Y1, D1, \
D2, S2, Y2, D2, \
D3, S2, Y3, D3
GMADD xvf, d, D0, S3, G0, D0, \
D1, S3, G1, D1, \
D2, S3, G2, D2, \
D3, S3, G3, D3
PTR_ADDI X0, X0, 0x40
PTR_ADDI B1, B1, 0x40
PTR_ADDI B2, B2, 0x40
PTR_ADDI B3, B3, 0x40
PTR_ADDI A0, A0, 0x40
PTR_ADDI A1, A1, 0x40
PTR_ADDI A2, A2, 0x40
PTR_ADDI A3, A3, 0x40
bnez K1, .L_M4_N4_K8
.L_M4_N4_K7:
andi K1, K, 4
beqz K1, .L_M4_N4_3
.L_M4_N4_K4:
GLD xv, , T0, A0, 0x00, T1, A1, 0x00, T2, A2, 0x00, T3, A3, 0x00
GTRANSPOSE4x4_D T0, T1, T2, T3, S0, S1, S2, S3, Z0, Z1
GLDREPL xv, d, Z0, X0, 0x00, Z1, B1, 0x00, Z2, B2, 0x00, Z3, B3, 0x00
GLDREPL xv, d, T0, X0, 0x08, T1, B1, 0x08, T2, B2, 0x08, T3, B3, 0x08
GLDREPL xv, d, Y0, X0, 0x10, Y1, B1, 0x10, Y2, B2, 0x10, Y3, B3, 0x10
GLDREPL xv, d, G0, X0, 0x18, G1, B1, 0x18, G2, B2, 0x18, G3, B3, 0x18
GMADD xvf, d, D0, S0, Z0, D0, \
D1, S0, Z1, D1, \
D2, S0, Z2, D2, \
D3, S0, Z3, D3
GMADD xvf, d, D0, S1, T0, D0, \
D1, S1, T1, D1, \
D2, S1, T2, D2, \
D3, S1, T3, D3
GMADD xvf, d, D0, S2, Y0, D0, \
D1, S2, Y1, D1, \
D2, S2, Y2, D2, \
D3, S2, Y3, D3
GMADD xvf, d, D0, S3, G0, D0, \
D1, S3, G1, D1, \
D2, S3, G2, D2, \
D3, S3, G3, D3
PTR_ADDI X0, X0, 0x20
PTR_ADDI B1, B1, 0x20
PTR_ADDI B2, B2, 0x20
PTR_ADDI B3, B3, 0x20
PTR_ADDI A0, A0, 0x20
PTR_ADDI A1, A1, 0x20
PTR_ADDI A2, A2, 0x20
PTR_ADDI A3, A3, 0x20
.L_M4_N4_3:
andi K1, K, 3
beqz K1, .L_M4_N4_END
.L_M4_N4_K1:
GLD xv, , S0, A0, 0x00, S1, A1, 0x00, S2, A2, 0x00, S3, A3, 0x00
GINSVE0 xv, d, S0, S1, 1, S0, S2, 2, S0, S3, 3
GLDREPL xv, d, Z0, X0, 0x00, Z1, B1, 0x00, Z2, B2, 0x00, Z3, B3, 0x00
GMADD xvf, d, D0, S0, Z0, D0, \
D1, S0, Z1, D1, \
D2, S0, Z2, D2, \
D3, S0, Z3, D3
PTR_ADDI K1, K1, -1
PTR_ADDI X0, X0, 0x08
PTR_ADDI B1, B1, 0x08
PTR_ADDI B2, B2, 0x08
PTR_ADDI B3, B3, 0x08
PTR_ADDI A0, A0, 0x08
PTR_ADDI A1, A1, 0x08
PTR_ADDI A2, A2, 0x08
PTR_ADDI A3, A3, 0x08
bnez K1, .L_M4_N4_K1
.L_M4_N4_END:
GMUL xvf, d, D0, D0, VALPHA, D1, D1, VALPHA, D2, D2, VALPHA, D3, D3, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00, S1, C1, 0x00, S2, C2, 0x00, S3, C3, 0x00
GMADD xvf, d, D0, S0, VBETA, D0, \
D1, S1, VBETA, D1, \
D2, S2, VBETA, D2, \
D3, S3, VBETA, D3
#endif
GST xv, , D3, C3, 0x00, \
D2, C2, 0x00, \
D1, C1, 0x00, \
D0, C0, 0x00
// Update C0, C1, C2, C3
PTR_ALSL C0, LDC, C0, 2
PTR_ALSL C1, LDC, C1, 2
PTR_ALSL C2, LDC, C2, 2
PTR_ALSL C3, LDC, C3, 2
// Update X0, B1, B2, B3
PTR_SUB X0, X0, K8
PTR_SUB B1, B1, K8
PTR_SUB B2, B2, K8
PTR_SUB B3, B3, K8
PTR_ALSL X0, LDB, X0, 2
PTR_ALSL B1, LDB, B1, 2
PTR_ALSL B2, LDB, B2, 2
PTR_ALSL B3, LDB, B3, 2
// Restore A0, A1, A2, A3
move A0, A
PTR_ADD A1, A0, LDA
PTR_ADD A2, A1, LDA
PTR_ADD A3, A2, LDA
bnez N4, .L_M4_N4
.L_M4_N3:
andi N2, N, 0x02
beqz N2, .L_M4_N1
.L_M4_N2:
GXOR xv, v, D0, D0, D0, D1, D1, D1
move K1, K // Restore K1
bge ZERO, K, .L_M4_N2_END
.L_M4_N2_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00, S1, A1, 0x00, S2, A2, 0x00, S3, A3, 0x00
GINSVE0 xv, d, S0, S1, 1, S0, S2, 2, S0, S3, 3
GLDREPL xv, d, Z0, X0, 0x00, Z1, B1, 0x00
GMADD xvf, d, D0, S0, Z0, D0, \
D1, S0, Z1, D1
PTR_ADDI X0, X0, 0x08
PTR_ADDI B1, B1, 0x08
PTR_ADDI A0, A0, 0x08
PTR_ADDI A1, A1, 0x08
PTR_ADDI A2, A2, 0x08
PTR_ADDI A3, A3, 0x08
bnez K1, .L_M4_N2_K1
.L_M4_N2_END:
GMUL xvf, d, D0, D0, VALPHA, D1, D1, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00, S1, C1, 0x00
GMADD xvf, d, D0, S0, VBETA, D0, D1, S1, VBETA, D1
#endif
GST xv, , D1, C1, 0x00, \
D0, C0, 0x00
// Update C0, C1
PTR_ALSL C0, LDC, C0, 1
PTR_ALSL C1, LDC, C1, 1
// Update X0, B1
PTR_SUB X0, X0, K8
PTR_SUB B1, B1, K8
PTR_ALSL X0, LDB, X0, 1
PTR_ALSL B1, LDB, B1, 1
// Restore A0
move A0, A
PTR_ADD A1, A0, LDA
PTR_ADD A2, A1, LDA
PTR_ADD A3, A2, LDA
.L_M4_N1:
andi N1, N, 0x01
beqz N1, .L_M4_END
GXOR xv, v, D0, D0, D0
move K1, K // Restore K1
bge ZERO, K, .L_M4_N1_END
.L_M4_N1_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00, S1, A1, 0x00, S2, A2, 0x00, S3, A3, 0x00
GINSVE0 xv, d, S0, S1, 1, S0, S2, 2, S0, S3, 3
GLDREPL xv, d, Z0, X0, 0x00
GMADD xvf, d, D0, S0, Z0, D0
PTR_ADDI X0, X0, 0x08
PTR_ADDI A0, A0, 0x08
PTR_ADDI A1, A1, 0x08
PTR_ADDI A2, A2, 0x08
PTR_ADDI A3, A3, 0x08
bnez K1, .L_M4_N1_K1
.L_M4_N1_END:
GMUL xvf, d, D0, D0, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00
GMADD xvf, d, D0, S0, VBETA, D0
#endif
GST xv, , D0, C0, 0x00
// Update C0
PTR_ALSL C0, LDC, C0, 2
// Update X0
PTR_SUB X0, X0, K8
PTR_ALSL X0, LDB, X0, 2
// Restore A0
move A0, A
PTR_ADD A1, A0, LDA
PTR_ADD A2, A1, LDA
PTR_ADD A3, A2, LDA
.L_M4_END:
PTR_ADDI M4, M4, -1
PTR_ALSL A, LDA, A, 2 // A += LDA << 2;
PTR_ADDI C, C, 0x20
bnez M4, .L_M4
.L_M3:
andi M2, M, 0x02
beqz M2, .L_M1
.L_M2:
PTR_SRAI N4, N, 2 // N >> 2
move A0, A // Restore A0
PTR_ADD A1, A0, LDA
move X0, B // Restore X0
PTR_ADD B1, X0, LDB
PTR_ADD B2, B1, LDB
PTR_ADD B3, B2, LDB
move C0, C // Restore C0
PTR_ADD C1, C0, LDC
PTR_ADD C2, C1, LDC
PTR_ADD C3, C2, LDC
beqz N4, .L_M2_N3
.L_M2_N4:
GXOR xv, v, D0, D0, D0, D1, D1, D1, D2, D2, D2, D3, D3, D3
move K1, K // Restore K1
PTR_ADDI N4, N4, -1
bge ZERO, K, .L_M2_N4_END
.L_M2_N4_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00, S1, A1, 0x00
GINSVE0 xv, d, S0, S1, 1
GLDREPL xv, d, Z0, X0, 0x00, Z1, B1, 0x00, Z2, B2, 0x00, Z3, B3, 0x00
GMADD xvf, d, D0, S0, Z0, D0, \
D1, S0, Z1, D1, \
D2, S0, Z2, D2, \
D3, S0, Z3, D3
PTR_ADDI X0, X0, 0x08
PTR_ADDI B1, B1, 0x08
PTR_ADDI B2, B2, 0x08
PTR_ADDI B3, B3, 0x08
PTR_ADDI A0, A0, 0x08
PTR_ADDI A1, A1, 0x08
bnez K1, .L_M2_N4_K1
.L_M2_N4_END:
GMUL xvf, d, D0, D0, VALPHA, D1, D1, VALPHA, D2, D2, VALPHA, D3, D3, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00, S1, C1, 0x00, S2, C2, 0x00, S3, C3, 0x00
GMADD xvf, d, D0, S0, VBETA, D0, D1, S1, VBETA, D1, D2, S2, VBETA, D2, D3, S3, VBETA, D3
#endif
GST v, , V3, C3, 0x00, \
V2, C2, 0x00, \
V1, C1, 0x00, \
V0, C0, 0x00
// Update C0, C1, C2, C3
PTR_ALSL C0, LDC, C0, 2
PTR_ALSL C1, LDC, C1, 2
PTR_ALSL C2, LDC, C2, 2
PTR_ALSL C3, LDC, C3, 2
// Update X0, B1, B2, B3
PTR_SUB X0, X0, K8
PTR_SUB B1, B1, K8
PTR_SUB B2, B2, K8
PTR_SUB B3, B3, K8
PTR_ALSL X0, LDB, X0, 2
PTR_ALSL B1, LDB, B1, 2
PTR_ALSL B2, LDB, B2, 2
PTR_ALSL B3, LDB, B3, 2
// Restore A0, A1
move A0, A
PTR_ADD A1, A0, LDA
bnez N4, .L_M2_N4
.L_M2_N3:
andi N2, N, 0x02
beqz N2, .L_M2_N1
.L_M2_N2:
GXOR xv, v, D0, D0, D0, D1, D1, D1
move K1, K // Restore K1
bge ZERO, K, .L_M2_N2_END
.L_M2_N2_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00, S1, A1, 0x00
GINSVE0 xv, d, S0, S1, 1
GLDREPL xv, d, Z0, X0, 0x00, Z1, B1, 0x00
GMADD xvf, d, D0, S0, Z0, D0, \
D1, S0, Z1, D1
PTR_ADDI X0, X0, 0x08
PTR_ADDI B1, B1, 0x08
PTR_ADDI A0, A0, 0x08
PTR_ADDI A1, A1, 0x08
bnez K1, .L_M2_N2_K1
.L_M2_N2_END:
GMUL xvf, d, D0, D0, VALPHA, D1, D1, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00, S1, C1, 0x00
GMADD xvf, d, D0, S0, VBETA, D0, D1, S1, VBETA, D1
#endif
GST v, , V1, C1, 0x00, \
V0, C0, 0x00
// Update C0, C1
PTR_ALSL C0, LDC, C0, 1
PTR_ALSL C1, LDC, C1, 1
// Update X0, B1
PTR_SUB X0, X0, K8
PTR_SUB B1, B1, K8
PTR_ALSL X0, LDB, X0, 1
PTR_ALSL B1, LDB, B1, 1
// Restore A0, A1
move A0, A
PTR_ADD A1, A0, LDA
.L_M2_N1:
andi N1, N, 0x01
beqz N1, .L_M2_END
GXOR xv, v, D0, D0, D0
move K1, K // Restore K1
bge ZERO, K, .L_M2_N1_END
.L_M2_N1_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00, S1, A1, 0x00
GINSVE0 xv, d, S0, S1, 1
GLDREPL xv, d, Z0, X0, 0x00
GMADD xvf, d, D0, S0, Z0, D0
PTR_ADDI X0, X0, 0x08
PTR_ADDI A0, A0, 0x08
PTR_ADDI A1, A1, 0x08
bnez K1, .L_M2_N1_K1
.L_M2_N1_END:
GMUL xvf, d, D0, D0, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00
GMADD xvf, d, D0, S0, VBETA, D0
#endif
GST v, , V0, C0, 0x00
// Update C0
PTR_ALSL C0, LDC, C0, 2
// Update X0
PTR_SUB X0, X0, K8
PTR_ALSL X0, LDB, X0, 2
// Restore A0, A1
move A0, A
PTR_ADD A1, A0, LDA
.L_M2_END:
PTR_ALSL A, LDA, A, 1 // A += LDA << 1;
PTR_ADDI C, C, 0x10
.L_M1:
andi M1, M, 0x01
beqz M1, .L_M0
PTR_SRAI N4, N, 2 // N >> 2
move A0, A // Restore A0
move X0, B // Restore X0
PTR_ADD B1, X0, LDB
PTR_ADD B2, B1, LDB
PTR_ADD B3, B2, LDB
move C0, C // Restore C0
PTR_ADD C1, C0, LDC
PTR_ADD C2, C1, LDC
PTR_ADD C3, C2, LDC
beqz N4, .L_M1_N3
.L_M1_N4:
GXOR xv, v, D0, D0, D0, D1, D1, D1, D2, D2, D2, D3, D3, D3
move K1, K // Restore K1
PTR_ADDI N4, N4, -1
bge ZERO, K, .L_M1_N4_END
.L_M1_N4_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00
GLDREPL xv, d, Z0, X0, 0x00, Z1, B1, 0x00, Z2, B2, 0x00, Z3, B3, 0x00
GMADD xvf, d, D0, S0, Z0, D0, \
D1, S0, Z1, D1, \
D2, S0, Z2, D2, \
D3, S0, Z3, D3
PTR_ADDI X0, X0, 0x08
PTR_ADDI B1, B1, 0x08
PTR_ADDI B2, B2, 0x08
PTR_ADDI B3, B3, 0x08
PTR_ADDI A0, A0, 0x08
bnez K1, .L_M1_N4_K1
.L_M1_N4_END:
GMUL xvf, d, D0, D0, VALPHA, D1, D1, VALPHA, D2, D2, VALPHA, D3, D3, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00, S1, C1, 0x00, S2, C2, 0x00, S3, C3, 0x00
GMADD xvf, d, D0, S0, VBETA, D0, D1, S1, VBETA, D1, D2, S2, VBETA, D2, D3, S3, VBETA, D3
#endif
GST f, d, F3, C3, 0x00, \
F2, C2, 0x00, \
F1, C1, 0x00, \
F0, C0, 0x00
// Update C0, C1, C2, C3
PTR_ALSL C0, LDC, C0, 2
PTR_ALSL C1, LDC, C1, 2
PTR_ALSL C2, LDC, C2, 2
PTR_ALSL C3, LDC, C3, 2
// Update X0, B1, B2, B3
PTR_SUB X0, X0, K8
PTR_SUB B1, B1, K8
PTR_SUB B2, B2, K8
PTR_SUB B3, B3, K8
PTR_ALSL X0, LDB, X0, 2
PTR_ALSL B1, LDB, B1, 2
PTR_ALSL B2, LDB, B2, 2
PTR_ALSL B3, LDB, B3, 2
// Restore A0, A1
move A0, A
bnez N4, .L_M1_N4
.L_M1_N3:
andi N2, N, 0x02
beqz N2, .L_M1_N1
.L_M1_N2:
GXOR xv, v, D0, D0, D0, D1, D1, D1
move K1, K // Restore K1
bge ZERO, K, .L_M1_N2_END
.L_M1_N2_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00
GLDREPL xv, d, Z0, X0, 0x00, Z1, B1, 0x00
GMADD xvf, d, D0, S0, Z0, D0, \
D1, S0, Z1, D1
PTR_ADDI X0, X0, 0x08
PTR_ADDI B1, B1, 0x08
PTR_ADDI A0, A0, 0x08
bnez K1, .L_M1_N2_K1
.L_M1_N2_END:
GMUL xvf, d, D0, D0, VALPHA, D1, D1, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00, S1, C1, 0x00
GMADD xvf, d, D0, S0, VBETA, D0, D1, S1, VBETA, D1
#endif
GST f, d, F1, C1, 0x00, \
F0, C0, 0x00
// Update C0, C1
PTR_ALSL C0, LDC, C0, 1
PTR_ALSL C1, LDC, C1, 1
// Update X0, B1
PTR_SUB X0, X0, K8
PTR_SUB B1, B1, K8
PTR_ALSL X0, LDB, X0, 1
PTR_ALSL B1, LDB, B1, 1
// Restore A0
move A0, A
.L_M1_N1:
andi N1, N, 0x01
beqz N1, .L_M0
GXOR xv, v, D0, D0, D0
move K1, K // Restore K1
bge ZERO, K, .L_M1_N1_END
.L_M1_N1_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00
GLDREPL xv, d, Z0, X0, 0x00
GMADD xvf, d, D0, S0, Z0, D0
PTR_ADDI X0, X0, 0x08
PTR_ADDI A0, A0, 0x08
bnez K1, .L_M1_N1_K1
.L_M1_N1_END:
GMUL xvf, d, D0, D0, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00
GMADD xvf, d, D0, S0, VBETA, D0
#endif
GST f, d, F0, C0, 0x00
// Update C0
PTR_ALSL C0, LDC, C0, 2
// Update X0
PTR_SUB X0, X0, K8
PTR_ALSL X0, LDB, X0, 2
// Restore A0
move A0, A
.L_M0:
pop_if_used 8, 2
jirl $r0, $r1, 0x0
EPILOGUE

View File

@ -0,0 +1,534 @@
/***************************************************************************
Copyright (c) 2024, The OpenBLAS Project
All rights reserved.
Redistribution and use in source and binary forms, with or without
modification, are permitted provided that the following conditions are
met:
1. Redistributions of source code must retain the above copyright
notice, this list of conditions and the following disclaimer.
2. Redistributions in binary form must reproduce the above copyright
notice, this list of conditions and the following disclaimer in
the documentation and/or other materials provided with the
distribution.
3. Neither the name of the OpenBLAS project nor the names of
its contributors may be used to endorse or promote products
derived from this software without specific prior written permission.
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
ARE DISCLAIMED. IN NO EVENT SHALL THE OPENBLAS PROJECT OR CONTRIBUTORS BE
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE
USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*****************************************************************************/
#define ASSEMBLER
#include "common.h"
#include "loongarch64_asm.S"
#define M $a0
#define N $a1
#define K $a2
#define A $a3
#define LDA $a4
#define ALPHA $f0
#define B $a5
#define LDB $a6
#define C $a7
#define LDC $t0
#ifdef B0
#define BETA $f1
#endif
#undef ZERO
#define ZERO $r0
#define M4 $t1
#define M2 $t1
#define M1 $t1
#define N4 $t2
#define N2 $t2
#define N1 $t2
#define K_LDB $t3
#define A0 $t4
#define X0 $t5
#define A1 $t6
#define A2 $t7
#define A3 $t8
#define C0 $s0
#define C1 $s1
#define C2 $s2
#define C3 $s3
#define K1 $s4
#define B1 $s5
#define B2 $s6
#define B3 $s7
#define VALPHA $xr0
#ifndef B0
#define VBETA $xr1
#endif
#define D0 $xr2
#define D1 $xr3
#define D2 $xr4
#define D3 $xr5
#define T0 $xr6
#define T1 $xr7
#define T2 $xr8
#define T3 $xr9
#define Y0 $xr10
#define Y1 $xr11
#define Y2 $xr12
#define Y3 $xr13
#define G0 $xr14
#define G1 $xr15
#define G2 $xr16
#define G3 $xr17
#define S0 $xr18
#define S1 $xr19
#define S2 $xr20
#define S3 $xr21
#define Z0 $xr22
#define Z1 $xr23
#define Z2 $xr24
#define Z3 $xr25
#define V0 $vr2
#define V1 $vr3
#define V2 $vr4
#define V3 $vr5
#define F0 $f2
#define F1 $f3
#define F2 $f4
#define F3 $f5
PROLOGUE
PTR_LD LDC, $sp, 0
push_if_used 8, 2
xvreplve0.d VALPHA, VALPHA
#ifndef B0
xvreplve0.d VBETA, VBETA
#endif
PTR_SLLI LDA, LDA, 3
PTR_SLLI LDB, LDB, 3
PTR_SLLI LDC, LDC, 3
PTR_MUL K_LDB, K, LDB
PTR_SRAI M4, M, 2 // M >> 2
beqz M4, .L_M3
.L_M4:
PTR_SRAI N4, N, 2 // N >> 2
move A0, A // Restore A0
PTR_ADD A1, A0, LDA
PTR_ADD A2, A1, LDA
PTR_ADD A3, A2, LDA
move X0, B // Restore X0
move C0, C // Restore C0
PTR_ADD C1, C0, LDC
PTR_ADD C2, C1, LDC
PTR_ADD C3, C2, LDC
beqz N4, .L_M4_N3
.L_M4_N4:
GXOR xv, v, D0, D0, D0, D1, D1, D1, D2, D2, D2, D3, D3, D3
move K1, K // Restore K1
PTR_ADDI N4, N4, -1
bge ZERO, K, .L_M4_N4_END
PTR_SRAI K1, K1, 2
beq ZERO, K1, .L_M4_N4_K3
PTR_ADD B1, X0, LDB
PTR_ADD B2, B1, LDB
PTR_ADD B3, B2, LDB
.L_M4_N4_K4:
PTR_ADDI K1, K1, -1
GLD xv, , T0, A0, 0x00, T1, A1, 0x00, T2, A2, 0x00, T3, A3, 0x00
GTRANSPOSE4x4_D T0, T1, T2, T3, S0, S1, S2, S3, Z0, Z1
GLDREPL xv, d, Z0, X0, 0x00, Z1, B1, 0x00, Z2, B2, 0x00, Z3, B3, 0x00
GLDREPL xv, d, T0, X0, 0x08, T1, B1, 0x08, T2, B2, 0x08, T3, B3, 0x08
GLDREPL xv, d, Y0, X0, 0x10, Y1, B1, 0x10, Y2, B2, 0x10, Y3, B3, 0x10
GLDREPL xv, d, G0, X0, 0x18, G1, B1, 0x18, G2, B2, 0x18, G3, B3, 0x18
GMADD xvf, d, D0, S0, Z0, D0, \
D1, S0, T0, D1, \
D2, S0, Y0, D2, \
D3, S0, G0, D3
GMADD xvf, d, D0, S1, Z1, D0, \
D1, S1, T1, D1, \
D2, S1, Y1, D2, \
D3, S1, G1, D3
GMADD xvf, d, D0, S2, Z2, D0, \
D1, S2, T2, D1, \
D2, S2, Y2, D2, \
D3, S2, G2, D3
GMADD xvf, d, D0, S3, Z3, D0, \
D1, S3, T3, D1, \
D2, S3, Y3, D2, \
D3, S3, G3, D3
PTR_ALSL X0, LDB, X0, 2
PTR_ALSL B1, LDB, B1, 2
PTR_ALSL B2, LDB, B2, 2
PTR_ALSL B3, LDB, B3, 2
PTR_ADDI A0, A0, 0x20
PTR_ADDI A1, A1, 0x20
PTR_ADDI A2, A2, 0x20
PTR_ADDI A3, A3, 0x20
bnez K1, .L_M4_N4_K4
.L_M4_N4_K3:
andi K1, K, 3
beqz K1, .L_M4_N4_END
.L_M4_N4_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00, S1, A1, 0x00, S2, A2, 0x00, S3, A3, 0x00
GINSVE0 xv, d, S0, S1, 1, S0, S2, 2, S0, S3, 3
GLDREPL xv, d, Z0, X0, 0x00, Z1, X0, 0x08, Z2, X0, 0x10, Z3, X0, 0x18
GMADD xvf, d, D0, S0, Z0, D0, \
D1, S0, Z1, D1, \
D2, S0, Z2, D2, \
D3, S0, Z3, D3
PTR_ADD X0, X0, LDB
PTR_ADDI A0, A0, 0x08
PTR_ADDI A1, A1, 0x08
PTR_ADDI A2, A2, 0x08
PTR_ADDI A3, A3, 0x08
bnez K1, .L_M4_N4_K1
.L_M4_N4_END:
GMUL xvf, d, D0, D0, VALPHA, D1, D1, VALPHA, D2, D2, VALPHA, D3, D3, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00, S1, C1, 0x00, S2, C2, 0x00, S3, C3, 0x00
GMADD xvf, d, D0, S0, VBETA, D0, \
D1, S1, VBETA, D1, \
D2, S2, VBETA, D2, \
D3, S3, VBETA, D3
#endif
GST xv, , D3, C3, 0x00, \
D2, C2, 0x00, \
D1, C1, 0x00, \
D0, C0, 0x00
// Update C0, C1, C2, C3
PTR_ALSL C0, LDC, C0, 2
PTR_ALSL C1, LDC, C1, 2
PTR_ALSL C2, LDC, C2, 2
PTR_ALSL C3, LDC, C3, 2
// Update X0
PTR_SUB X0, X0, K_LDB
PTR_ADDI X0, X0, 0x20
// Restore A0, A1, A2, A3
move A0, A
PTR_ADD A1, A0, LDA
PTR_ADD A2, A1, LDA
PTR_ADD A3, A2, LDA
bnez N4, .L_M4_N4
.L_M4_N3:
andi N2, N, 0x02
beqz N2, .L_M4_N1
.L_M4_N2:
GXOR xv, v, D0, D0, D0, D1, D1, D1
move K1, K // Restore K1
bge ZERO, K, .L_M4_N2_END
.L_M4_N2_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00, S1, A1, 0x00, S2, A2, 0x00, S3, A3, 0x00
GINSVE0 xv, d, S0, S1, 1, S0, S2, 2, S0, S3, 3
GLDREPL xv, d, Z0, X0, 0x00, Z1, X0, 0x08
GMADD xvf, d, D0, S0, Z0, D0, \
D1, S0, Z1, D1
PTR_ADD X0, X0, LDB
PTR_ADDI A0, A0, 0x08
PTR_ADDI A1, A1, 0x08
PTR_ADDI A2, A2, 0x08
PTR_ADDI A3, A3, 0x08
bnez K1, .L_M4_N2_K1
.L_M4_N2_END:
GMUL xvf, d, D0, D0, VALPHA, D1, D1, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00, S1, C1, 0x00
GMADD xvf, d, D0, S0, VBETA, D0, D1, S1, VBETA, D1
#endif
GST xv, , D1, C1, 0x00, \
D0, C0, 0x00
// Update C0, C1
PTR_ALSL C0, LDC, C0, 1
PTR_ALSL C1, LDC, C1, 1
// Update X0
PTR_SUB X0, X0, K_LDB
PTR_ADDI X0, X0, 0x10
// Restore A0
move A0, A
PTR_ADD A1, A0, LDA
PTR_ADD A2, A1, LDA
PTR_ADD A3, A2, LDA
.L_M4_N1:
andi N1, N, 0x01
beqz N1, .L_M4_END
GXOR xv, v, D0, D0, D0
move K1, K // Restore K1
bge ZERO, K, .L_M4_N1_END
.L_M4_N1_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00, S1, A1, 0x00, S2, A2, 0x00, S3, A3, 0x00
GINSVE0 xv, d, S0, S1, 1, S0, S2, 2, S0, S3, 3
GLDREPL xv, d, Z0, X0, 0x00
GMADD xvf, d, D0, S0, Z0, D0
PTR_ADD X0, X0, LDB
PTR_ADDI A0, A0, 0x08
PTR_ADDI A1, A1, 0x08
PTR_ADDI A2, A2, 0x08
PTR_ADDI A3, A3, 0x08
bnez K1, .L_M4_N1_K1
.L_M4_N1_END:
GMUL xvf, d, D0, D0, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00
GMADD xvf, d, D0, S0, VBETA, D0
#endif
GST xv, , D0, C0, 0x00
// Update C0
PTR_ALSL C0, LDC, C0, 2
// Update X0
PTR_SUB X0, X0, K_LDB
PTR_ADDI X0, X0, 0x08
// Restore A0
move A0, A
PTR_ADD A1, A0, LDA
PTR_ADD A2, A1, LDA
PTR_ADD A3, A2, LDA
.L_M4_END:
PTR_ADDI M4, M4, -1
PTR_ALSL A, LDA, A, 2 // A += LDA << 2;
PTR_ADDI C, C, 0x20
bnez M4, .L_M4
.L_M3:
andi M2, M, 0x02
beqz M2, .L_M1
.L_M2:
PTR_SRAI N4, N, 2 // N >> 2
move A0, A // Restore A0
PTR_ADD A1, A0, LDA
move X0, B // Restore X0
move C0, C // Restore C0
PTR_ADD C1, C0, LDC
PTR_ADD C2, C1, LDC
PTR_ADD C3, C2, LDC
beqz N4, .L_M2_N3
.L_M2_N4:
GXOR xv, v, D0, D0, D0, D1, D1, D1, D2, D2, D2, D3, D3, D3
move K1, K // Restore K1
PTR_ADDI N4, N4, -1
bge ZERO, K, .L_M2_N4_END
.L_M2_N4_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00, S1, A1, 0x00
GINSVE0 xv, d, S0, S1, 1
GLDREPL xv, d, Z0, X0, 0x00, Z1, X0, 0x08, Z2, X0, 0x10, Z3, X0, 0x18
GMADD xvf, d, D0, S0, Z0, D0, \
D1, S0, Z1, D1, \
D2, S0, Z2, D2, \
D3, S0, Z3, D3
PTR_ADD X0, X0, LDB
PTR_ADDI A0, A0, 0x08
PTR_ADDI A1, A1, 0x08
bnez K1, .L_M2_N4_K1
.L_M2_N4_END:
GMUL xvf, d, D0, D0, VALPHA, D1, D1, VALPHA, D2, D2, VALPHA, D3, D3, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00, S1, C1, 0x00, S2, C2, 0x00, S3, C3, 0x00
GMADD xvf, d, D0, S0, VBETA, D0, D1, S1, VBETA, D1, D2, S2, VBETA, D2, D3, S3, VBETA, D3
#endif
GST v, , V3, C3, 0x00, \
V2, C2, 0x00, \
V1, C1, 0x00, \
V0, C0, 0x00
// Update C0, C1, C2, C3
PTR_ALSL C0, LDC, C0, 2
PTR_ALSL C1, LDC, C1, 2
PTR_ALSL C2, LDC, C2, 2
PTR_ALSL C3, LDC, C3, 2
// Update X0
PTR_SUB X0, X0, K_LDB
PTR_ADDI X0, X0, 0x20
// Restore A0, A1
move A0, A
PTR_ADD A1, A0, LDA
bnez N4, .L_M2_N4
.L_M2_N3:
andi N2, N, 0x02
beqz N2, .L_M2_N1
.L_M2_N2:
GXOR xv, v, D0, D0, D0, D1, D1, D1
move K1, K // Restore K1
bge ZERO, K, .L_M2_N2_END
.L_M2_N2_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00, S1, A1, 0x00
GINSVE0 xv, d, S0, S1, 1
GLDREPL xv, d, Z0, X0, 0x00, Z1, X0, 0x08
GMADD xvf, d, D0, S0, Z0, D0, \
D1, S0, Z1, D1
PTR_ADD X0, X0, LDB
PTR_ADDI A0, A0, 0x08
PTR_ADDI A1, A1, 0x08
bnez K1, .L_M2_N2_K1
.L_M2_N2_END:
GMUL xvf, d, D0, D0, VALPHA, D1, D1, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00, S1, C1, 0x00
GMADD xvf, d, D0, S0, VBETA, D0, D1, S1, VBETA, D1
#endif
GST v, , V1, C1, 0x00, \
V0, C0, 0x00
// Update C0, C1
PTR_ALSL C0, LDC, C0, 1
PTR_ALSL C1, LDC, C1, 1
// Update X0
PTR_SUB X0, X0, K_LDB
PTR_ADDI X0, X0, 0x10
// Restore A0, A1
move A0, A
PTR_ADD A1, A0, LDA
.L_M2_N1:
andi N1, N, 0x01
beqz N1, .L_M2_END
GXOR xv, v, D0, D0, D0
move K1, K // Restore K1
bge ZERO, K, .L_M2_N1_END
.L_M2_N1_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00, S1, A1, 0x00
GINSVE0 xv, d, S0, S1, 1
GLDREPL xv, d, Z0, X0, 0x00
GMADD xvf, d, D0, S0, Z0, D0
PTR_ADD X0, X0, LDB
PTR_ADDI A0, A0, 0x08
PTR_ADDI A1, A1, 0x08
bnez K1, .L_M2_N1_K1
.L_M2_N1_END:
GMUL xvf, d, D0, D0, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00
GMADD xvf, d, D0, S0, VBETA, D0
#endif
GST v, , V0, C0, 0x00
// Update C0
PTR_ALSL C0, LDC, C0, 2
// Update X0
PTR_SUB X0, X0, K_LDB
PTR_ADDI X0, X0, 0x08
// Restore A0, A1
move A0, A
PTR_ADD A1, A0, LDA
.L_M2_END:
PTR_ALSL A, LDA, A, 1 // A += LDA << 1;
PTR_ADDI C, C, 0x10
.L_M1:
andi M1, M, 0x01
beqz M1, .L_M0
PTR_SRAI N4, N, 2 // N >> 2
move A0, A // Restore A0
move X0, B // Restore X0
move C0, C // Restore C0
PTR_ADD C1, C0, LDC
PTR_ADD C2, C1, LDC
PTR_ADD C3, C2, LDC
beqz N4, .L_M1_N3
.L_M1_N4:
GXOR xv, v, D0, D0, D0, D1, D1, D1, D2, D2, D2, D3, D3, D3
move K1, K // Restore K1
PTR_ADDI N4, N4, -1
bge ZERO, K, .L_M1_N4_END
.L_M1_N4_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00
GLDREPL xv, d, Z0, X0, 0x00, Z1, X0, 0x08, Z2, X0, 0x10, Z3, X0, 0x18
GMADD xvf, d, D0, S0, Z0, D0, \
D1, S0, Z1, D1, \
D2, S0, Z2, D2, \
D3, S0, Z3, D3
PTR_ADD X0, X0, LDB
PTR_ADDI A0, A0, 0x08
bnez K1, .L_M1_N4_K1
.L_M1_N4_END:
GMUL xvf, d, D0, D0, VALPHA, D1, D1, VALPHA, D2, D2, VALPHA, D3, D3, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00, S1, C1, 0x00, S2, C2, 0x00, S3, C3, 0x00
GMADD xvf, d, D0, S0, VBETA, D0, D1, S1, VBETA, D1, D2, S2, VBETA, D2, D3, S3, VBETA, D3
#endif
GST f, d, F3, C3, 0x00, \
F2, C2, 0x00, \
F1, C1, 0x00, \
F0, C0, 0x00
// Update C0, C1, C2, C3
PTR_ALSL C0, LDC, C0, 2
PTR_ALSL C1, LDC, C1, 2
PTR_ALSL C2, LDC, C2, 2
PTR_ALSL C3, LDC, C3, 2
// Update X0
PTR_SUB X0, X0, K_LDB
PTR_ADDI X0, X0, 0x20
// Restore A0, A1
move A0, A
bnez N4, .L_M1_N4
.L_M1_N3:
andi N2, N, 0x02
beqz N2, .L_M1_N1
.L_M1_N2:
GXOR xv, v, D0, D0, D0, D1, D1, D1
move K1, K // Restore K1
bge ZERO, K, .L_M1_N2_END
.L_M1_N2_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00
GLDREPL xv, d, Z0, X0, 0x00, Z1, X0, 0x08
GMADD xvf, d, D0, S0, Z0, D0, \
D1, S0, Z1, D1
PTR_ADD X0, X0, LDB
PTR_ADDI A0, A0, 0x08
bnez K1, .L_M1_N2_K1
.L_M1_N2_END:
GMUL xvf, d, D0, D0, VALPHA, D1, D1, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00, S1, C1, 0x00
GMADD xvf, d, D0, S0, VBETA, D0, D1, S1, VBETA, D1
#endif
GST f, d, F1, C1, 0x00, \
F0, C0, 0x00
// Update C0, C1
PTR_ALSL C0, LDC, C0, 1
PTR_ALSL C1, LDC, C1, 1
// Update X0
PTR_SUB X0, X0, K_LDB
PTR_ADDI X0, X0, 0x10
// Restore A0
move A0, A
.L_M1_N1:
andi N1, N, 0x01
beqz N1, .L_M0
GXOR xv, v, D0, D0, D0
move K1, K // Restore K1
bge ZERO, K, .L_M1_N1_END
.L_M1_N1_K1:
PTR_ADDI K1, K1, -1
GLD xv, , S0, A0, 0x00
GLDREPL xv, d, Z0, X0, 0x00
GMADD xvf, d, D0, S0, Z0, D0
PTR_ADD X0, X0, LDB
PTR_ADDI A0, A0, 0x08
bnez K1, .L_M1_N1_K1
.L_M1_N1_END:
GMUL xvf, d, D0, D0, VALPHA
#ifndef B0
GLD xv, , S0, C0, 0x00
GMADD xvf, d, D0, S0, VBETA, D0
#endif
GST f, d, F0, C0, 0x00
// Update C0
PTR_ALSL C0, LDC, C0, 2
// Update X0
PTR_SUB X0, X0, K_LDB
PTR_ADDI X0, X0, 0x08
// Restore A0
move A0, A
.L_M0:
pop_if_used 8, 2
jirl $r0, $r1, 0x0
EPILOGUE

View File

@ -0,0 +1,44 @@
/***************************************************************************
Copyright (c) 2024, The OpenBLAS Project
All rights reserved.
Redistribution and use in source and binary forms, with or without
modification, are permitted provided that the following conditions are
met:
1. Redistributions of source code must retain the above copyright
notice, this list of conditions and the following disclaimer.
2. Redistributions in binary form must reproduce the above copyright
notice, this list of conditions and the following disclaimer in
the documentation and/or other materials provided with the
distribution.
3. Neither the name of the OpenBLAS project nor the names of
its contributors may be used to endorse or promote products
derived from this software without specific prior written permission.
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
ARE DISCLAIMED. IN NO EVENT SHALL THE OPENBLAS PROJECT OR CONTRIBUTORS BE
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE
USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*****************************************************************************/
#include "common.h"
int CNAME(int transa, int transb, BLASLONG M, BLASLONG N, BLASLONG K, FLOAT alpha, FLOAT beta)
{
double MNK = (double) M * (double) N * (double) K;
if (transa) {
if (MNK <= 24.0 * 24.0 * 24.0)
return 1;
} else {
if (MNK <= 64.0 * 64.0 * 64.0)
return 1;
}
return 0;
}