812 lines
27 KiB
C
812 lines
27 KiB
C
/***************************************************************************
|
|
Copyright (c) 2017, The OpenBLAS Project
|
|
All rights reserved.
|
|
Redistribution and use in source and binary forms, with or without
|
|
modification, are permitted provided that the following conditions are
|
|
met:
|
|
1. Redistributions of source code must retain the above copyright
|
|
notice, this list of conditions and the following disclaimer.
|
|
2. Redistributions in binary form must reproduce the above copyright
|
|
notice, this list of conditions and the following disclaimer in
|
|
the documentation and/or other materials provided with the
|
|
distribution.
|
|
3. Neither the name of the OpenBLAS project nor the names of
|
|
its contributors may be used to endorse or promote products
|
|
derived from this software without specific prior written permission.
|
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
|
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
ARE DISCLAIMED. IN NO EVENT SHALL THE OPENBLAS PROJECT OR CONTRIBUTORS BE
|
|
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
|
|
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
|
|
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
|
|
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE
|
|
USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*****************************************************************************/
|
|
|
|
#include "common.h"
|
|
|
|
#define NBMAX 2048
|
|
|
|
static void sgemv_kernel_4x4(BLASLONG n, FLOAT **ap, FLOAT *x, FLOAT *y)
|
|
{
|
|
__asm__ volatile (
|
|
"vzero %%v0 \n\t"
|
|
"vzero %%v1 \n\t"
|
|
"vzero %%v2 \n\t"
|
|
"vzero %%v3 \n\t"
|
|
"xgr %%r1,%%r1 \n\t"
|
|
|
|
"lghi %%r0,-32 \n\t"
|
|
"ngr %%r0,%0 \n\t"
|
|
"ltgr %%r0,%%r0 \n\t"
|
|
"jz 1f \n\t"
|
|
|
|
"srlg %%r0,%%r0,5 \n\t"
|
|
"0: \n\t"
|
|
"pfd 1,1024(%%r1,%1) \n\t"
|
|
"pfd 1,1024(%%r1,%2) \n\t"
|
|
"pfd 1,1024(%%r1,%3) \n\t"
|
|
"pfd 1,1024(%%r1,%4) \n\t"
|
|
"pfd 1,1024(%%r1,%5) \n\t"
|
|
|
|
"vl %%v16,0(%%r1,%5) \n\t"
|
|
"vl %%v17,16(%%r1,%5) \n\t"
|
|
"vl %%v18,32(%%r1,%5) \n\t"
|
|
"vl %%v19,48(%%r1,%5) \n\t"
|
|
"vl %%v20,64(%%r1,%5) \n\t"
|
|
"vl %%v21,80(%%r1,%5) \n\t"
|
|
"vl %%v22,96(%%r1,%5) \n\t"
|
|
"vl %%v23,112(%%r1,%5) \n\t"
|
|
|
|
"vl %%v24,0(%%r1,%1) \n\t"
|
|
"vfmasb %%v0,%%v16,%%v24,%%v0 \n\t"
|
|
"vl %%v25,0(%%r1,%2) \n\t"
|
|
"vfmasb %%v1,%%v16,%%v25,%%v1 \n\t"
|
|
"vl %%v26,0(%%r1,%3) \n\t"
|
|
"vfmasb %%v2,%%v16,%%v26,%%v2 \n\t"
|
|
"vl %%v27,0(%%r1,%4) \n\t"
|
|
"vfmasb %%v3,%%v16,%%v27,%%v3 \n\t"
|
|
|
|
"vl %%v28,16(%%r1,%1) \n\t"
|
|
"vfmasb %%v0,%%v17,%%v28,%%v0 \n\t"
|
|
"vl %%v29,16(%%r1,%2) \n\t"
|
|
"vfmasb %%v1,%%v17,%%v29,%%v1 \n\t"
|
|
"vl %%v30,16(%%r1,%3) \n\t"
|
|
"vfmasb %%v2,%%v17,%%v30,%%v2 \n\t"
|
|
"vl %%v31,16(%%r1,%4) \n\t"
|
|
"vfmasb %%v3,%%v17,%%v31,%%v3 \n\t"
|
|
|
|
"vl %%v24,32(%%r1,%1) \n\t"
|
|
"vfmasb %%v0,%%v18,%%v24,%%v0 \n\t"
|
|
"vl %%v25,32(%%r1,%2) \n\t"
|
|
"vfmasb %%v1,%%v18,%%v25,%%v1 \n\t"
|
|
"vl %%v26,32(%%r1,%3) \n\t"
|
|
"vfmasb %%v2,%%v18,%%v26,%%v2 \n\t"
|
|
"vl %%v27,32(%%r1,%4) \n\t"
|
|
"vfmasb %%v3,%%v18,%%v27,%%v3 \n\t"
|
|
|
|
"vl %%v28,48(%%r1,%1) \n\t"
|
|
"vfmasb %%v0,%%v19,%%v28,%%v0 \n\t"
|
|
"vl %%v29,48(%%r1,%2) \n\t"
|
|
"vfmasb %%v1,%%v19,%%v29,%%v1 \n\t"
|
|
"vl %%v30,48(%%r1,%3) \n\t"
|
|
"vfmasb %%v2,%%v19,%%v30,%%v2 \n\t"
|
|
"vl %%v31,48(%%r1,%4) \n\t"
|
|
"vfmasb %%v3,%%v19,%%v31,%%v3 \n\t"
|
|
|
|
"vl %%v24,64(%%r1,%1) \n\t"
|
|
"vfmasb %%v0,%%v20,%%v24,%%v0 \n\t"
|
|
"vl %%v25,64(%%r1,%2) \n\t"
|
|
"vfmasb %%v1,%%v20,%%v25,%%v1 \n\t"
|
|
"vl %%v26,64(%%r1,%3) \n\t"
|
|
"vfmasb %%v2,%%v20,%%v26,%%v2 \n\t"
|
|
"vl %%v27,64(%%r1,%4) \n\t"
|
|
"vfmasb %%v3,%%v20,%%v27,%%v3 \n\t"
|
|
|
|
"vl %%v28,80(%%r1,%1) \n\t"
|
|
"vfmasb %%v0,%%v21,%%v28,%%v0 \n\t"
|
|
"vl %%v29,80(%%r1,%2) \n\t"
|
|
"vfmasb %%v1,%%v21,%%v29,%%v1 \n\t"
|
|
"vl %%v30,80(%%r1,%3) \n\t"
|
|
"vfmasb %%v2,%%v21,%%v30,%%v2 \n\t"
|
|
"vl %%v31,80(%%r1,%4) \n\t"
|
|
"vfmasb %%v3,%%v21,%%v31,%%v3 \n\t"
|
|
|
|
"vl %%v24,96(%%r1,%1) \n\t"
|
|
"vfmasb %%v0,%%v22,%%v24,%%v0 \n\t"
|
|
"vl %%v25,96(%%r1,%2) \n\t"
|
|
"vfmasb %%v1,%%v22,%%v25,%%v1 \n\t"
|
|
"vl %%v26,96(%%r1,%3) \n\t"
|
|
"vfmasb %%v2,%%v22,%%v26,%%v2 \n\t"
|
|
"vl %%v27,96(%%r1,%4) \n\t"
|
|
"vfmasb %%v3,%%v22,%%v27,%%v3 \n\t"
|
|
|
|
"vl %%v28,112(%%r1,%1) \n\t"
|
|
"vfmasb %%v0,%%v23,%%v28,%%v0 \n\t"
|
|
"vl %%v29,112(%%r1,%2) \n\t"
|
|
"vfmasb %%v1,%%v23,%%v29,%%v1 \n\t"
|
|
"vl %%v30,112(%%r1,%3) \n\t"
|
|
"vfmasb %%v2,%%v23,%%v30,%%v2 \n\t"
|
|
"vl %%v31,112(%%r1,%4) \n\t"
|
|
"vfmasb %%v3,%%v23,%%v31,%%v3 \n\t"
|
|
|
|
"agfi %%r1,128 \n\t"
|
|
"brctg %%r0,0b \n\t"
|
|
|
|
"1: \n\t"
|
|
"lghi %%r0,28 \n\t"
|
|
"ngr %%r0,%0 \n\t"
|
|
"ltgr %%r0,%%r0 \n\t"
|
|
"jz 3f \n\t"
|
|
|
|
"srlg %%r0,%%r0,2 \n\t"
|
|
"2: \n\t"
|
|
"vl %%v16,0(%%r1,%5) \n\t"
|
|
|
|
"vl %%v24,0(%%r1,%1) \n\t"
|
|
"vfmasb %%v0,%%v16,%%v24,%%v0 \n\t"
|
|
"vl %%v25,0(%%r1,%2) \n\t"
|
|
"vfmasb %%v1,%%v16,%%v25,%%v1 \n\t"
|
|
"vl %%v26,0(%%r1,%3) \n\t"
|
|
"vfmasb %%v2,%%v16,%%v26,%%v2 \n\t"
|
|
"vl %%v27,0(%%r1,%4) \n\t"
|
|
"vfmasb %%v3,%%v16,%%v27,%%v3 \n\t"
|
|
|
|
"agfi %%r1,16 \n\t"
|
|
"brctg %%r0,2b \n\t"
|
|
|
|
"3: \n\t"
|
|
"veslg %%v4,%%v0,32 \n\t"
|
|
"vfasb %%v0,%%v0,%%v4 \n\t"
|
|
"vrepg %%v4,%%v0,1 \n\t"
|
|
"aebr %%f0,%%f4 \n\t"
|
|
"ste %%f0,0(%6) \n\t"
|
|
"veslg %%v4,%%v1,32 \n\t"
|
|
"vfasb %%v1,%%v1,%%v4 \n\t"
|
|
"vrepg %%v4,%%v1,1 \n\t"
|
|
"aebr %%f1,%%f4 \n\t"
|
|
"ste %%f1,4(%6) \n\t"
|
|
"veslg %%v4,%%v2,32 \n\t"
|
|
"vfasb %%v2,%%v2,%%v4 \n\t"
|
|
"vrepg %%v4,%%v2,1 \n\t"
|
|
"aebr %%f2,%%f4 \n\t"
|
|
"ste %%f2,8(%6) \n\t"
|
|
"veslg %%v4,%%v3,32 \n\t"
|
|
"vfasb %%v3,%%v3,%%v4 \n\t"
|
|
"vrepg %%v4,%%v3,1 \n\t"
|
|
"aebr %%f3,%%f4 \n\t"
|
|
"ste %%f3,12(%6) "
|
|
:
|
|
:"r"(n),"ZR"((const FLOAT (*)[n])ap[0]),"ZR"((const FLOAT (*)[n])ap[1]),"ZR"((const FLOAT (*)[n])ap[2]),"ZR"((const FLOAT (*)[n])ap[3]),"ZR"((const FLOAT (*)[n])x),"ZQ"((FLOAT (*)[4])y)
|
|
:"memory","cc","r0","r1","v0","v1","v2","v3","v4","v16","v17","v18","v19","v20","v21","v22","v23","v24","v25","v26","v27","v28","v29","v30","v31"
|
|
);
|
|
}
|
|
|
|
static void sgemv_kernel_4x2(BLASLONG n, FLOAT **ap, FLOAT *x, FLOAT *y)
|
|
{
|
|
__asm__ volatile (
|
|
"vzero %%v0 \n\t"
|
|
"vzero %%v1 \n\t"
|
|
"xgr %%r1,%%r1 \n\t"
|
|
|
|
"lghi %%r0,-32 \n\t"
|
|
"ngr %%r0,%0 \n\t"
|
|
"ltgr %%r0,%%r0 \n\t"
|
|
"jz 1f \n\t"
|
|
|
|
"srlg %%r0,%%r0,5 \n\t"
|
|
"0: \n\t"
|
|
"pfd 1,1024(%%r1,%1) \n\t"
|
|
"pfd 1,1024(%%r1,%2) \n\t"
|
|
"pfd 1,1024(%%r1,%3) \n\t"
|
|
|
|
"vl %%v16,0(%%r1,%3) \n\t"
|
|
"vl %%v17,16(%%r1,%3) \n\t"
|
|
"vl %%v18,32(%%r1,%3) \n\t"
|
|
"vl %%v19,48(%%r1,%3) \n\t"
|
|
"vl %%v20,64(%%r1,%3) \n\t"
|
|
"vl %%v21,80(%%r1,%3) \n\t"
|
|
"vl %%v22,96(%%r1,%3) \n\t"
|
|
"vl %%v23,112(%%r1,%3) \n\t"
|
|
|
|
"vl %%v24,0(%%r1,%1) \n\t"
|
|
"vfmasb %%v0,%%v16,%%v24,%%v0 \n\t"
|
|
"vl %%v25,0(%%r1,%2) \n\t"
|
|
"vfmasb %%v1,%%v16,%%v25,%%v1 \n\t"
|
|
|
|
"vl %%v26,16(%%r1,%1) \n\t"
|
|
"vfmasb %%v0,%%v17,%%v26,%%v0 \n\t"
|
|
"vl %%v27,16(%%r1,%2) \n\t"
|
|
"vfmasb %%v1,%%v17,%%v27,%%v1 \n\t"
|
|
|
|
"vl %%v28,32(%%r1,%1) \n\t"
|
|
"vfmasb %%v0,%%v18,%%v28,%%v0 \n\t"
|
|
"vl %%v29,32(%%r1,%2) \n\t"
|
|
"vfmasb %%v1,%%v18,%%v29,%%v1 \n\t"
|
|
|
|
"vl %%v30,48(%%r1,%1) \n\t"
|
|
"vfmasb %%v0,%%v19,%%v30,%%v0 \n\t"
|
|
"vl %%v31,48(%%r1,%2) \n\t"
|
|
"vfmasb %%v1,%%v19,%%v31,%%v1 \n\t"
|
|
|
|
"vl %%v24,64(%%r1,%1) \n\t"
|
|
"vfmasb %%v0,%%v20,%%v24,%%v0 \n\t"
|
|
"vl %%v25,64(%%r1,%2) \n\t"
|
|
"vfmasb %%v1,%%v20,%%v25,%%v1 \n\t"
|
|
|
|
"vl %%v26,80(%%r1,%1) \n\t"
|
|
"vfmasb %%v0,%%v21,%%v26,%%v0 \n\t"
|
|
"vl %%v27,80(%%r1,%2) \n\t"
|
|
"vfmasb %%v1,%%v21,%%v27,%%v1 \n\t"
|
|
|
|
"vl %%v28,96(%%r1,%1) \n\t"
|
|
"vfmasb %%v0,%%v22,%%v28,%%v0 \n\t"
|
|
"vl %%v29,96(%%r1,%2) \n\t"
|
|
"vfmasb %%v1,%%v22,%%v29,%%v1 \n\t"
|
|
|
|
"vl %%v30,112(%%r1,%1) \n\t"
|
|
"vfmasb %%v0,%%v23,%%v30,%%v0 \n\t"
|
|
"vl %%v31,112(%%r1,%2) \n\t"
|
|
"vfmasb %%v1,%%v23,%%v31,%%v1 \n\t"
|
|
|
|
"agfi %%r1,128 \n\t"
|
|
"brctg %%r0,0b \n\t"
|
|
|
|
"1: \n\t"
|
|
"lghi %%r0,28 \n\t"
|
|
"ngr %%r0,%0 \n\t"
|
|
"ltgr %%r0,%%r0 \n\t"
|
|
"jz 3f \n\t"
|
|
|
|
"srlg %%r0,%%r0,2 \n\t"
|
|
"2: \n\t"
|
|
"vl %%v16,0(%%r1,%3) \n\t"
|
|
|
|
"vl %%v24,0(%%r1,%1) \n\t"
|
|
"vfmasb %%v0,%%v16,%%v24,%%v0 \n\t"
|
|
"vl %%v25,0(%%r1,%2) \n\t"
|
|
"vfmasb %%v1,%%v16,%%v25,%%v1 \n\t"
|
|
|
|
"agfi %%r1,16 \n\t"
|
|
"brctg %%r0,2b \n\t"
|
|
|
|
"3: \n\t"
|
|
"veslg %%v2,%%v0,32 \n\t"
|
|
"vfasb %%v0,%%v0,%%v2 \n\t"
|
|
"vrepg %%v2,%%v0,1 \n\t"
|
|
"aebr %%f0,%%f2 \n\t"
|
|
"ste %%f0,0(%4) \n\t"
|
|
"veslg %%v2,%%v1,32 \n\t"
|
|
"vfasb %%v1,%%v1,%%v2 \n\t"
|
|
"vrepg %%v2,%%v1,1 \n\t"
|
|
"aebr %%f1,%%f2 \n\t"
|
|
"ste %%f1,4(%4) "
|
|
:
|
|
:"r"(n),"ZR"((const FLOAT (*)[n])ap[0]),"ZR"((const FLOAT (*)[n])ap[1]),"ZR"((const FLOAT (*)[n])x),"ZQ"((FLOAT (*)[2])y)
|
|
:"memory","cc","r0","r1","v0","v1","v2","v16","v17","v18","v19","v20","v21","v22","v23","v24","v25","v26","v27","v28","v29","v30","v31"
|
|
);
|
|
}
|
|
|
|
static void sgemv_kernel_4x1(BLASLONG n, FLOAT *a0, FLOAT *x, FLOAT *y)
|
|
{
|
|
__asm__ volatile (
|
|
"vzero %%v0 \n\t"
|
|
"xgr %%r1,%%r1 \n\t"
|
|
|
|
"lghi %%r0,-32 \n\t"
|
|
"ngr %%r0,%0 \n\t"
|
|
"ltgr %%r0,%%r0 \n\t"
|
|
"jz 1f \n\t"
|
|
|
|
"srlg %%r0,%%r0,5 \n\t"
|
|
"0: \n\t"
|
|
"pfd 1,1024(%%r1,%1) \n\t"
|
|
"pfd 1,1024(%%r1,%2) \n\t"
|
|
|
|
"vl %%v16,0(%%r1,%2) \n\t"
|
|
"vl %%v17,16(%%r1,%2) \n\t"
|
|
"vl %%v18,32(%%r1,%2) \n\t"
|
|
"vl %%v19,48(%%r1,%2) \n\t"
|
|
"vl %%v20,64(%%r1,%2) \n\t"
|
|
"vl %%v21,80(%%r1,%2) \n\t"
|
|
"vl %%v22,96(%%r1,%2) \n\t"
|
|
"vl %%v23,112(%%r1,%2) \n\t"
|
|
|
|
"vl %%v24,0(%%r1,%1) \n\t"
|
|
"vfmasb %%v0,%%v16,%%v24,%%v0 \n\t"
|
|
|
|
"vl %%v25,16(%%r1,%1) \n\t"
|
|
"vfmasb %%v0,%%v17,%%v25,%%v0 \n\t"
|
|
|
|
"vl %%v26,32(%%r1,%1) \n\t"
|
|
"vfmasb %%v0,%%v18,%%v26,%%v0 \n\t"
|
|
|
|
"vl %%v27,48(%%r1,%1) \n\t"
|
|
"vfmasb %%v0,%%v19,%%v27,%%v0 \n\t"
|
|
|
|
"vl %%v28,64(%%r1,%1) \n\t"
|
|
"vfmasb %%v0,%%v20,%%v28,%%v0 \n\t"
|
|
|
|
"vl %%v29,80(%%r1,%1) \n\t"
|
|
"vfmasb %%v0,%%v21,%%v29,%%v0 \n\t"
|
|
|
|
"vl %%v30,96(%%r1,%1) \n\t"
|
|
"vfmasb %%v0,%%v22,%%v30,%%v0 \n\t"
|
|
|
|
"vl %%v31,112(%%r1,%1) \n\t"
|
|
"vfmasb %%v0,%%v23,%%v31,%%v0 \n\t"
|
|
|
|
"agfi %%r1,128 \n\t"
|
|
"brctg %%r0,0b \n\t"
|
|
|
|
"1: \n\t"
|
|
"lghi %%r0,28 \n\t"
|
|
"ngr %%r0,%0 \n\t"
|
|
"ltgr %%r0,%%r0 \n\t"
|
|
"jz 3f \n\t"
|
|
|
|
"srlg %%r0,%%r0,2 \n\t"
|
|
"2: \n\t"
|
|
"vl %%v16,0(%%r1,%2) \n\t"
|
|
|
|
"vl %%v24,0(%%r1,%1) \n\t"
|
|
"vfmasb %%v0,%%v16,%%v24,%%v0 \n\t"
|
|
|
|
"agfi %%r1,16 \n\t"
|
|
"brctg %%r0,2b \n\t"
|
|
|
|
"3: \n\t"
|
|
"veslg %%v1,%%v0,32 \n\t"
|
|
"vfasb %%v0,%%v0,%%v1 \n\t"
|
|
"vrepg %%v1,%%v0,1 \n\t"
|
|
"aebr %%f0,%%f1 \n\t"
|
|
"ste %%f0,0(%3) "
|
|
:
|
|
:"r"(n),"ZR"((const FLOAT (*)[n])a0),"ZR"((const FLOAT (*)[n])x),"ZQ"((FLOAT (*)[1])y)
|
|
:"memory","cc","r0","r1","v0","v1","v16","v17","v18","v19","v20","v21","v22","v23","v24","v25","v26","v27","v28","v29","v30","v31"
|
|
);
|
|
}
|
|
|
|
static void copy_x(BLASLONG n, FLOAT *src, FLOAT *dest, BLASLONG inc_src)
|
|
{
|
|
BLASLONG i;
|
|
for (i = 0; i < n; i++)
|
|
{
|
|
dest[i] = *src;
|
|
src += inc_src;
|
|
}
|
|
}
|
|
|
|
static void add_y_kernel_4(BLASLONG n, FLOAT da, FLOAT *src, FLOAT *dest)
|
|
{
|
|
__asm__ volatile (
|
|
"vlrepf %%v0,%1 \n\t"
|
|
"xgr %%r1,%%r1 \n\t"
|
|
|
|
"lghi %%r0,-32 \n\t"
|
|
"ngr %%r0,%0 \n\t"
|
|
"ltgr %%r0,%%r0 \n\t"
|
|
"jz 1f \n\t"
|
|
|
|
"srlg %%r0,%%r0,5 \n\t"
|
|
"0: \n\t"
|
|
"pfd 1,1024(%%r1,%2) \n\t"
|
|
"pfd 2,1024(%%r1,%3) \n\t"
|
|
|
|
"vl %%v16,0(%%r1,%2) \n\t"
|
|
"vl %%v17,16(%%r1,%2) \n\t"
|
|
"vl %%v18,32(%%r1,%2) \n\t"
|
|
"vl %%v19,48(%%r1,%2) \n\t"
|
|
"vl %%v20,64(%%r1,%2) \n\t"
|
|
"vl %%v21,80(%%r1,%2) \n\t"
|
|
"vl %%v22,96(%%r1,%2) \n\t"
|
|
"vl %%v23,112(%%r1,%2) \n\t"
|
|
|
|
"vl %%v24, 0(%%r1,%3) \n\t"
|
|
"vfmasb %%v24,%%v16,%%v0,%%v24 \n\t"
|
|
"vst %%v24, 0(%%r1,%3) \n\t"
|
|
"vl %%v25, 16(%%r1,%3) \n\t"
|
|
"vfmasb %%v25,%%v17,%%v0,%%v25 \n\t"
|
|
"vst %%v25, 16(%%r1,%3) \n\t"
|
|
"vl %%v26, 32(%%r1,%3) \n\t"
|
|
"vfmasb %%v26,%%v18,%%v0,%%v26 \n\t"
|
|
"vst %%v26, 32(%%r1,%3) \n\t"
|
|
"vl %%v27, 48(%%r1,%3) \n\t"
|
|
"vfmasb %%v27,%%v19,%%v0,%%v27 \n\t"
|
|
"vst %%v27, 48(%%r1,%3) \n\t"
|
|
"vl %%v28, 64(%%r1,%3) \n\t"
|
|
"vfmasb %%v28,%%v20,%%v0,%%v28 \n\t"
|
|
"vst %%v28, 64(%%r1,%3) \n\t"
|
|
"vl %%v29, 80(%%r1,%3) \n\t"
|
|
"vfmasb %%v29,%%v21,%%v0,%%v29 \n\t"
|
|
"vst %%v29, 80(%%r1,%3) \n\t"
|
|
"vl %%v30, 96(%%r1,%3) \n\t"
|
|
"vfmasb %%v30,%%v22,%%v0,%%v30 \n\t"
|
|
"vst %%v30, 96(%%r1,%3) \n\t"
|
|
"vl %%v31, 112(%%r1,%3) \n\t"
|
|
"vfmasb %%v31,%%v23,%%v0,%%v31 \n\t"
|
|
"vst %%v31, 112(%%r1,%3) \n\t"
|
|
|
|
"agfi %%r1,128 \n\t"
|
|
"brctg %%r0,0b \n\t"
|
|
|
|
"1: \n\t"
|
|
"lghi %%r0,28 \n\t"
|
|
"ngr %%r0,%0 \n\t"
|
|
"ltgr %%r0,%%r0 \n\t"
|
|
"jz 3f \n\t"
|
|
|
|
"srlg %%r0,%%r0,2 \n\t"
|
|
"2: \n\t"
|
|
"vl %%v16,0(%%r1,%2) \n\t"
|
|
|
|
"vl %%v24, 0(%%r1,%3) \n\t"
|
|
"vfmasb %%v24,%%v16,%%v0,%%v24 \n\t"
|
|
"vst %%v24, 0(%%r1,%3) \n\t"
|
|
|
|
"agfi %%r1,16 \n\t"
|
|
"brctg %%r0,2b \n\t"
|
|
|
|
"3: \n\t"
|
|
"nop "
|
|
:
|
|
:"r"(n),"m"(da),"ZR"((const FLOAT (*)[n])src),"ZR"((FLOAT (*)[n])dest)
|
|
:"memory","cc","r0","r1","v0","v16","v17","v18","v19","v20","v21","v22","v23","v24","v25","v26","v27","v28","v29","v30","v31"
|
|
);
|
|
}
|
|
static void add_y(BLASLONG n, FLOAT da, FLOAT *src, FLOAT *dest, BLASLONG inc_dest)
|
|
{
|
|
if (inc_dest == 1)
|
|
add_y_kernel_4(n, da, src, dest);
|
|
else
|
|
{
|
|
BLASLONG i;
|
|
for (i = 0; i < n; i++)
|
|
{
|
|
*dest += src[i] * da;
|
|
dest += inc_dest;
|
|
}
|
|
}
|
|
}
|
|
|
|
int CNAME(BLASLONG m, BLASLONG n, BLASLONG dummy1, FLOAT alpha, FLOAT *a, BLASLONG lda, FLOAT *x, BLASLONG inc_x, FLOAT *y, BLASLONG inc_y, FLOAT *buffer)
|
|
{
|
|
BLASLONG register i;
|
|
BLASLONG register j;
|
|
FLOAT *a_ptr;
|
|
FLOAT *x_ptr;
|
|
FLOAT *y_ptr;
|
|
BLASLONG n0;
|
|
BLASLONG n1;
|
|
BLASLONG m1;
|
|
BLASLONG m2;
|
|
BLASLONG m3;
|
|
BLASLONG n2;
|
|
FLOAT ybuffer[2] __attribute__ ((aligned(16)));
|
|
FLOAT *xbuffer;
|
|
FLOAT *ytemp;
|
|
|
|
if ( m < 1 ) return(0);
|
|
if ( n < 1 ) return(0);
|
|
|
|
xbuffer = buffer;
|
|
ytemp = buffer + (m < NBMAX ? m : NBMAX);
|
|
|
|
n0 = n / NBMAX;
|
|
n1 = (n % NBMAX) >> 2 ;
|
|
n2 = n & 3 ;
|
|
|
|
m3 = m & 3 ;
|
|
m1 = m & -4 ;
|
|
m2 = (m & (NBMAX-1)) - m3 ;
|
|
|
|
|
|
BLASLONG NB = NBMAX;
|
|
|
|
while ( NB == NBMAX )
|
|
{
|
|
m1 -= NB;
|
|
if ( m1 < 0)
|
|
{
|
|
if ( m2 == 0 ) break;
|
|
NB = m2;
|
|
}
|
|
|
|
y_ptr = y;
|
|
a_ptr = a;
|
|
x_ptr = x;
|
|
|
|
if ( inc_x == 1 )
|
|
xbuffer = x_ptr;
|
|
else
|
|
copy_x(NB,x_ptr,xbuffer,inc_x);
|
|
|
|
|
|
FLOAT *ap[4];
|
|
FLOAT *yp;
|
|
BLASLONG register lda4 = 4 * lda;
|
|
ap[0] = a_ptr;
|
|
ap[1] = a_ptr + lda;
|
|
ap[2] = ap[1] + lda;
|
|
ap[3] = ap[2] + lda;
|
|
|
|
if ( n0 > 0 )
|
|
{
|
|
BLASLONG nb1 = NBMAX / 4;
|
|
for( j=0; j<n0; j++)
|
|
{
|
|
|
|
yp = ytemp;
|
|
for( i = 0; i < nb1 ; i++)
|
|
{
|
|
sgemv_kernel_4x4(NB,ap,xbuffer,yp);
|
|
ap[0] += lda4 ;
|
|
ap[1] += lda4 ;
|
|
ap[2] += lda4 ;
|
|
ap[3] += lda4 ;
|
|
yp += 4;
|
|
}
|
|
add_y(nb1*4, alpha, ytemp, y_ptr, inc_y );
|
|
y_ptr += nb1 * inc_y * 4;
|
|
a_ptr += nb1 * lda4 ;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
yp = ytemp;
|
|
|
|
for( i = 0; i < n1 ; i++)
|
|
{
|
|
sgemv_kernel_4x4(NB,ap,xbuffer,yp);
|
|
ap[0] += lda4 ;
|
|
ap[1] += lda4 ;
|
|
ap[2] += lda4 ;
|
|
ap[3] += lda4 ;
|
|
yp += 4;
|
|
}
|
|
if ( n1 > 0 )
|
|
{
|
|
add_y(n1*4, alpha, ytemp, y_ptr, inc_y );
|
|
y_ptr += n1 * inc_y * 4;
|
|
a_ptr += n1 * lda4 ;
|
|
}
|
|
|
|
if ( n2 & 2 )
|
|
{
|
|
|
|
sgemv_kernel_4x2(NB,ap,xbuffer,ybuffer);
|
|
a_ptr += lda * 2;
|
|
*y_ptr += ybuffer[0] * alpha;
|
|
y_ptr += inc_y;
|
|
*y_ptr += ybuffer[1] * alpha;
|
|
y_ptr += inc_y;
|
|
|
|
}
|
|
|
|
if ( n2 & 1 )
|
|
{
|
|
|
|
sgemv_kernel_4x1(NB,a_ptr,xbuffer,ybuffer);
|
|
// a_ptr += lda;
|
|
*y_ptr += ybuffer[0] * alpha;
|
|
// y_ptr += inc_y;
|
|
|
|
}
|
|
a += NB;
|
|
x += NB * inc_x;
|
|
}
|
|
|
|
if ( m3 == 0 ) return(0);
|
|
|
|
x_ptr = x;
|
|
a_ptr = a;
|
|
if ( m3 == 3 )
|
|
{
|
|
FLOAT xtemp0 = *x_ptr * alpha;
|
|
x_ptr += inc_x;
|
|
FLOAT xtemp1 = *x_ptr * alpha;
|
|
x_ptr += inc_x;
|
|
FLOAT xtemp2 = *x_ptr * alpha;
|
|
|
|
FLOAT *aj = a_ptr;
|
|
y_ptr = y;
|
|
|
|
if ( lda == 3 && inc_y == 1 )
|
|
{
|
|
|
|
for ( j=0; j< ( n & -4) ; j+=4 )
|
|
{
|
|
|
|
y_ptr[j] += aj[0] * xtemp0 + aj[1] * xtemp1 + aj[2] * xtemp2;
|
|
y_ptr[j+1] += aj[3] * xtemp0 + aj[4] * xtemp1 + aj[5] * xtemp2;
|
|
y_ptr[j+2] += aj[6] * xtemp0 + aj[7] * xtemp1 + aj[8] * xtemp2;
|
|
y_ptr[j+3] += aj[9] * xtemp0 + aj[10] * xtemp1 + aj[11] * xtemp2;
|
|
aj += 12;
|
|
}
|
|
|
|
for ( ; j<n; j++ )
|
|
{
|
|
y_ptr[j] += aj[0] * xtemp0 + aj[1] * xtemp1 + aj[2] * xtemp2;
|
|
aj += 3;
|
|
}
|
|
|
|
}
|
|
else
|
|
{
|
|
|
|
if ( inc_y == 1 )
|
|
{
|
|
|
|
BLASLONG register lda2 = lda << 1;
|
|
BLASLONG register lda4 = lda << 2;
|
|
BLASLONG register lda3 = lda2 + lda;
|
|
|
|
for ( j=0; j< ( n & -4 ); j+=4 )
|
|
{
|
|
|
|
y_ptr[j] += *aj * xtemp0 + *(aj+1) * xtemp1 + *(aj+2) * xtemp2;
|
|
y_ptr[j+1] += *(aj+lda) * xtemp0 + *(aj+lda+1) * xtemp1 + *(aj+lda+2) * xtemp2;
|
|
y_ptr[j+2] += *(aj+lda2) * xtemp0 + *(aj+lda2+1) * xtemp1 + *(aj+lda2+2) * xtemp2;
|
|
y_ptr[j+3] += *(aj+lda3) * xtemp0 + *(aj+lda3+1) * xtemp1 + *(aj+lda3+2) * xtemp2;
|
|
aj += lda4;
|
|
}
|
|
|
|
for ( ; j< n ; j++ )
|
|
{
|
|
|
|
y_ptr[j] += *aj * xtemp0 + *(aj+1) * xtemp1 + *(aj+2) * xtemp2 ;
|
|
aj += lda;
|
|
}
|
|
|
|
}
|
|
else
|
|
{
|
|
|
|
for ( j=0; j<n; j++ )
|
|
{
|
|
*y_ptr += *aj * xtemp0 + *(aj+1) * xtemp1 + *(aj+2) * xtemp2;
|
|
y_ptr += inc_y;
|
|
aj += lda;
|
|
}
|
|
|
|
|
|
}
|
|
|
|
}
|
|
return(0);
|
|
}
|
|
|
|
if ( m3 == 2 )
|
|
{
|
|
FLOAT xtemp0 = *x_ptr * alpha;
|
|
x_ptr += inc_x;
|
|
FLOAT xtemp1 = *x_ptr * alpha;
|
|
|
|
FLOAT *aj = a_ptr;
|
|
y_ptr = y;
|
|
|
|
if ( lda == 2 && inc_y == 1 )
|
|
{
|
|
|
|
for ( j=0; j< ( n & -4) ; j+=4 )
|
|
{
|
|
y_ptr[j] += aj[0] * xtemp0 + aj[1] * xtemp1 ;
|
|
y_ptr[j+1] += aj[2] * xtemp0 + aj[3] * xtemp1 ;
|
|
y_ptr[j+2] += aj[4] * xtemp0 + aj[5] * xtemp1 ;
|
|
y_ptr[j+3] += aj[6] * xtemp0 + aj[7] * xtemp1 ;
|
|
aj += 8;
|
|
|
|
}
|
|
|
|
for ( ; j<n; j++ )
|
|
{
|
|
y_ptr[j] += aj[0] * xtemp0 + aj[1] * xtemp1 ;
|
|
aj += 2;
|
|
}
|
|
|
|
}
|
|
else
|
|
{
|
|
if ( inc_y == 1 )
|
|
{
|
|
|
|
BLASLONG register lda2 = lda << 1;
|
|
BLASLONG register lda4 = lda << 2;
|
|
BLASLONG register lda3 = lda2 + lda;
|
|
|
|
for ( j=0; j< ( n & -4 ); j+=4 )
|
|
{
|
|
|
|
y_ptr[j] += *aj * xtemp0 + *(aj+1) * xtemp1 ;
|
|
y_ptr[j+1] += *(aj+lda) * xtemp0 + *(aj+lda+1) * xtemp1 ;
|
|
y_ptr[j+2] += *(aj+lda2) * xtemp0 + *(aj+lda2+1) * xtemp1 ;
|
|
y_ptr[j+3] += *(aj+lda3) * xtemp0 + *(aj+lda3+1) * xtemp1 ;
|
|
aj += lda4;
|
|
}
|
|
|
|
for ( ; j< n ; j++ )
|
|
{
|
|
|
|
y_ptr[j] += *aj * xtemp0 + *(aj+1) * xtemp1 ;
|
|
aj += lda;
|
|
}
|
|
|
|
}
|
|
else
|
|
{
|
|
for ( j=0; j<n; j++ )
|
|
{
|
|
*y_ptr += *aj * xtemp0 + *(aj+1) * xtemp1 ;
|
|
y_ptr += inc_y;
|
|
aj += lda;
|
|
}
|
|
}
|
|
|
|
}
|
|
return(0);
|
|
|
|
}
|
|
|
|
FLOAT xtemp = *x_ptr * alpha;
|
|
FLOAT *aj = a_ptr;
|
|
y_ptr = y;
|
|
if ( lda == 1 && inc_y == 1 )
|
|
{
|
|
for ( j=0; j< ( n & -4) ; j+=4 )
|
|
{
|
|
y_ptr[j] += aj[j] * xtemp;
|
|
y_ptr[j+1] += aj[j+1] * xtemp;
|
|
y_ptr[j+2] += aj[j+2] * xtemp;
|
|
y_ptr[j+3] += aj[j+3] * xtemp;
|
|
}
|
|
for ( ; j<n ; j++ )
|
|
{
|
|
y_ptr[j] += aj[j] * xtemp;
|
|
}
|
|
|
|
|
|
|
|
}
|
|
else
|
|
{
|
|
if ( inc_y == 1 )
|
|
{
|
|
|
|
BLASLONG register lda2 = lda << 1;
|
|
BLASLONG register lda4 = lda << 2;
|
|
BLASLONG register lda3 = lda2 + lda;
|
|
for ( j=0; j< ( n & -4 ); j+=4 )
|
|
{
|
|
y_ptr[j] += *aj * xtemp;
|
|
y_ptr[j+1] += *(aj+lda) * xtemp;
|
|
y_ptr[j+2] += *(aj+lda2) * xtemp;
|
|
y_ptr[j+3] += *(aj+lda3) * xtemp;
|
|
aj += lda4 ;
|
|
}
|
|
|
|
for ( ; j<n; j++ )
|
|
{
|
|
y_ptr[j] += *aj * xtemp;
|
|
aj += lda;
|
|
}
|
|
|
|
}
|
|
else
|
|
{
|
|
for ( j=0; j<n; j++ )
|
|
{
|
|
*y_ptr += *aj * xtemp;
|
|
y_ptr += inc_y;
|
|
aj += lda;
|
|
}
|
|
|
|
}
|
|
}
|
|
|
|
return(0);
|
|
}
|