260 lines
		
	
	
		
			8.3 KiB
		
	
	
	
		
			C
		
	
	
	
			
		
		
	
	
			260 lines
		
	
	
		
			8.3 KiB
		
	
	
	
		
			C
		
	
	
	
/***************************************************************************
 | 
						|
Copyright (c) 2013-2016, The OpenBLAS Project
 | 
						|
All rights reserved.
 | 
						|
Redistribution and use in source and binary forms, with or without
 | 
						|
modification, are permitted provided that the following conditions are
 | 
						|
met:
 | 
						|
1. Redistributions of source code must retain the above copyright
 | 
						|
notice, this list of conditions and the following disclaimer.
 | 
						|
2. Redistributions in binary form must reproduce the above copyright
 | 
						|
notice, this list of conditions and the following disclaimer in
 | 
						|
the documentation and/or other materials provided with the
 | 
						|
distribution.
 | 
						|
3. Neither the name of the OpenBLAS project nor the names of
 | 
						|
its contributors may be used to endorse or promote products
 | 
						|
derived from this software without specific prior written permission.
 | 
						|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
 | 
						|
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
 | 
						|
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
 | 
						|
ARE DISCLAIMED. IN NO EVENT SHALL THE OPENBLAS PROJECT OR CONTRIBUTORS BE
 | 
						|
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
 | 
						|
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
 | 
						|
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
 | 
						|
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
 | 
						|
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE
 | 
						|
USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
 | 
						|
*****************************************************************************/
 | 
						|
 | 
						|
/**************************************************************************************
 | 
						|
* 2016/03/23 Werner Saar (wernsaar@googlemail.com)
 | 
						|
* 	 BLASTEST 		: OK
 | 
						|
* 	 CTEST			: OK
 | 
						|
* 	 TEST			: OK
 | 
						|
*	 LAPACK-TEST		: OK
 | 
						|
**************************************************************************************/
 | 
						|
 | 
						|
 | 
						|
#define HAVE_KERNEL_4 1
 | 
						|
static void zaxpy_kernel_4 (long n, double *x, double *y,
 | 
						|
			    double alpha_r, double alpha_i)
 | 
						|
{
 | 
						|
#if !defined(CONJ)
 | 
						|
  static const double mvec[2] = { -1.0, 1.0 };
 | 
						|
#else
 | 
						|
  static const double mvec[2] = { 1.0, -1.0 };
 | 
						|
#endif
 | 
						|
  const double *mvecp = mvec;
 | 
						|
 | 
						|
  __vector double t0;
 | 
						|
  __vector double t1;
 | 
						|
  __vector double t2;
 | 
						|
  __vector double t3;
 | 
						|
  __vector double t4;
 | 
						|
  __vector double t5;
 | 
						|
  __vector double t6;
 | 
						|
  __vector double t7;
 | 
						|
  __vector double t8;
 | 
						|
  __vector double t9;
 | 
						|
  __vector double t10;
 | 
						|
  __vector double t11;
 | 
						|
  long ytmp;
 | 
						|
 | 
						|
  __asm__
 | 
						|
    (
 | 
						|
       XXSPLTD_S(32,%x19,0)	// alpha_r
 | 
						|
       XXSPLTD_S(33,%x20,0)	// alpha_i
 | 
						|
 | 
						|
       "lxvd2x		36, 0, %21	\n\t"	// mvec
 | 
						|
 | 
						|
#if !defined(CONJ)
 | 
						|
       "xvmuldp		33, 33, 36	\n\t"	// alpha_i * mvec
 | 
						|
#else
 | 
						|
       "xvmuldp		32, 32, 36	\n\t"	// alpha_r * mvec
 | 
						|
#endif
 | 
						|
 | 
						|
       "mr		%16, %3		\n\t"
 | 
						|
       "dcbt		0, %2		\n\t"
 | 
						|
       "dcbt		0, %3		\n\t"
 | 
						|
 | 
						|
 | 
						|
       "lxvd2x		40, 0, %2	\n\t"	// x0
 | 
						|
       "lxvd2x		41, %22, %2	\n\t"	// x1
 | 
						|
       "lxvd2x		42, %23, %2	\n\t"	// x2
 | 
						|
       "lxvd2x		43, %24, %2	\n\t"	// x3
 | 
						|
 | 
						|
       "lxvd2x		48, 0, %3	\n\t"	// y0
 | 
						|
       "lxvd2x		49, %22, %3	\n\t"	// y1
 | 
						|
       "lxvd2x		50, %23, %3	\n\t"	// y2
 | 
						|
       "lxvd2x		51, %24, %3	\n\t"	// y3
 | 
						|
 | 
						|
       XXSWAPD_S(%x8,40)	// exchange real and imag part
 | 
						|
       XXSWAPD_S(%x9,41)	// exchange real and imag part
 | 
						|
       XXSWAPD_S(%x10,42)	// exchange real and imag part
 | 
						|
       XXSWAPD_S(%x11,43)	// exchange real and imag part
 | 
						|
 | 
						|
       "addi		%2, %2, 64	\n\t"
 | 
						|
       "addi		%3, %3, 64	\n\t"
 | 
						|
 | 
						|
       "lxvd2x		44, 0, %2	\n\t"	// x4
 | 
						|
       "lxvd2x		45, %22, %2	\n\t"	// x5
 | 
						|
       "lxvd2x		46, %23, %2	\n\t"	// x6
 | 
						|
       "lxvd2x		47, %24, %2	\n\t"	// x7
 | 
						|
 | 
						|
       "lxvd2x		%x4, 0, %3	\n\t"	// y4
 | 
						|
       "lxvd2x		%x5, %22, %3	\n\t"	// y5
 | 
						|
       "lxvd2x		%x6, %23, %3	\n\t"	// y6
 | 
						|
       "lxvd2x		%x7, %24, %3	\n\t"	// y7
 | 
						|
 | 
						|
       XXSWAPD_S(%x12,44)	// exchange real and imag part
 | 
						|
       XXSWAPD_S(%x13,45)	// exchange real and imag part
 | 
						|
       XXSWAPD_S(%x14,46)	// exchange real and imag part
 | 
						|
       XXSWAPD_S(%x15,47)	// exchange real and imag part
 | 
						|
 | 
						|
       "addi		%2, %2, 64	\n\t"
 | 
						|
       "addi		%3, %3, 64	\n\t"
 | 
						|
 | 
						|
       "addic.		%1, %1, -8	\n\t"
 | 
						|
       "ble		two%=		\n\t"
 | 
						|
 | 
						|
       ".align	5		\n"
 | 
						|
       "one%=:				\n\t"
 | 
						|
 | 
						|
       "xvmaddadp	48, 40, 32	\n\t"	// alpha_r * x0_r , alpha_r * x0_i
 | 
						|
       "xvmaddadp	49, 41, 32	\n\t"
 | 
						|
       "lxvd2x		40, 0, %2	\n\t"	// x0
 | 
						|
       "lxvd2x		41, %22, %2	\n\t"	// x1
 | 
						|
       "xvmaddadp	50, 42, 32	\n\t"
 | 
						|
       "xvmaddadp	51, 43, 32	\n\t"
 | 
						|
       "lxvd2x		42, %23, %2	\n\t"	// x2
 | 
						|
       "lxvd2x		43, %24, %2	\n\t"	// x3
 | 
						|
 | 
						|
       "xvmaddadp	%x4, 44, 32	\n\t"
 | 
						|
       "addi		%2, %2, 64	\n\t"
 | 
						|
       "xvmaddadp	%x5, 45, 32	\n\t"
 | 
						|
       "lxvd2x		44, 0, %2	\n\t"	// x4
 | 
						|
       "lxvd2x		45, %22, %2	\n\t"	// x5
 | 
						|
       "xvmaddadp	%x6, 46, 32	\n\t"
 | 
						|
       "xvmaddadp	%x7, 47, 32	\n\t"
 | 
						|
       "lxvd2x		46, %23, %2	\n\t"	// x6
 | 
						|
       "lxvd2x		47, %24, %2	\n\t"	// x7
 | 
						|
 | 
						|
       "xvmaddadp	48, %x8, 33	\n\t"	// alpha_i * x0_i , alpha_i * x0_r
 | 
						|
       "addi		%2, %2, 64	\n\t"
 | 
						|
       "xvmaddadp	49, %x9, 33	\n\t"
 | 
						|
       "xvmaddadp	50, %x10, 33	\n\t"
 | 
						|
       "xvmaddadp	51, %x11, 33	\n\t"
 | 
						|
 | 
						|
       "xvmaddadp	%x4, %x12, 33	\n\t"
 | 
						|
       "xvmaddadp	%x5, %x13, 33	\n\t"
 | 
						|
       "xvmaddadp	%x6, %x14, 33	\n\t"
 | 
						|
       "xvmaddadp	%x7, %x15, 33	\n\t"
 | 
						|
 | 
						|
       "stxvd2x		48, 0, %16	\n\t"
 | 
						|
       "stxvd2x		49, %22, %16	\n\t"
 | 
						|
       "stxvd2x		50, %23, %16	\n\t"
 | 
						|
       "stxvd2x		51, %24, %16	\n\t"
 | 
						|
 | 
						|
       "addi		%16, %16, 64	\n\t"
 | 
						|
 | 
						|
       "stxvd2x		%x4, 0, %16	\n\t"
 | 
						|
       "stxvd2x		%x5, %22, %16	\n\t"
 | 
						|
       "stxvd2x		%x6, %23, %16	\n\t"
 | 
						|
       "stxvd2x		%x7, %24, %16	\n\t"
 | 
						|
 | 
						|
       "addi		%16, %16, 64	\n\t"
 | 
						|
 | 
						|
       XXSWAPD_S(%x8,40)	// exchange real and imag part
 | 
						|
       XXSWAPD_S(%x9,41)	// exchange real and imag part
 | 
						|
       "lxvd2x		48, 0, %3	\n\t"	// y0
 | 
						|
       "lxvd2x		49, %22, %3	\n\t"	// y1
 | 
						|
       XXSWAPD_S(%x10,42)	// exchange real and imag part
 | 
						|
       XXSWAPD_S(%x11,43)	// exchange real and imag part
 | 
						|
       "lxvd2x		50, %23, %3	\n\t"	// y2
 | 
						|
       "lxvd2x		51, %24, %3	\n\t"	// y3
 | 
						|
 | 
						|
       XXSWAPD_S(%x12,44)	// exchange real and imag part
 | 
						|
       "addi		%3, %3, 64	\n\t"
 | 
						|
       XXSWAPD_S(%x13,45)	// exchange real and imag part
 | 
						|
       "lxvd2x		%x4, 0, %3	\n\t"	// y4
 | 
						|
       "lxvd2x		%x5, %22, %3	\n\t"	// y5
 | 
						|
       XXSWAPD_S(%x14,46)	// exchange real and imag part
 | 
						|
       XXSWAPD_S(%x15,47)	// exchange real and imag part
 | 
						|
       "lxvd2x		%x6, %23, %3	\n\t"	// y6
 | 
						|
       "lxvd2x		%x7, %24, %3	\n\t"	// y7
 | 
						|
 | 
						|
       "addi		%3, %3, 64	\n\t"
 | 
						|
 | 
						|
       "addic.		%1, %1, -8	\n\t"
 | 
						|
       "bgt		one%=		\n"
 | 
						|
 | 
						|
       "two%=:				\n\t"
 | 
						|
 | 
						|
       "xvmaddadp	48, 40, 32	\n\t"	// alpha_r * x0_r , alpha_r * x0_i
 | 
						|
       "xvmaddadp	49, 41, 32	\n\t"
 | 
						|
       "xvmaddadp	50, 42, 32	\n\t"
 | 
						|
       "xvmaddadp	51, 43, 32	\n\t"
 | 
						|
 | 
						|
       "xvmaddadp	%x4, 44, 32	\n\t"
 | 
						|
       "xvmaddadp	%x5, 45, 32	\n\t"
 | 
						|
       "xvmaddadp	%x6, 46, 32	\n\t"
 | 
						|
       "xvmaddadp	%x7, 47, 32	\n\t"
 | 
						|
 | 
						|
       "xvmaddadp	48, %x8, 33	\n\t"	// alpha_i * x0_i , alpha_i * x0_r
 | 
						|
       "xvmaddadp	49, %x9, 33	\n\t"
 | 
						|
       "xvmaddadp	50, %x10, 33	\n\t"
 | 
						|
       "xvmaddadp	51, %x11, 33	\n\t"
 | 
						|
 | 
						|
       "xvmaddadp	%x4, %x12, 33	\n\t"
 | 
						|
       "xvmaddadp	%x5, %x13, 33	\n\t"
 | 
						|
       "xvmaddadp	%x6, %x14, 33	\n\t"
 | 
						|
       "xvmaddadp	%x7, %x15, 33	\n\t"
 | 
						|
 | 
						|
       "stxvd2x		48, 0, %16	\n\t"
 | 
						|
       "stxvd2x		49, %22, %16	\n\t"
 | 
						|
       "stxvd2x		50, %23, %16	\n\t"
 | 
						|
       "stxvd2x		51, %24, %16	\n\t"
 | 
						|
 | 
						|
       "addi		%16, %16, 64	\n\t"
 | 
						|
 | 
						|
       "stxvd2x		%x4, 0, %16	\n\t"
 | 
						|
       "stxvd2x		%x5, %22, %16	\n\t"
 | 
						|
       "stxvd2x		%x6, %23, %16	\n\t"
 | 
						|
       "stxvd2x		%x7, %24, %16	\n"
 | 
						|
 | 
						|
     "#n=%1 x=%17=%2 y=%0=%3 alpha=(%19,%20) mvecp=%18=%16 o16=%22 o32=%23 o48=%24 ytmp=%16\n"
 | 
						|
     "#t0=%x4 t1=%x5 t2=%x6 t3=%x7 t4=%x8 t5=%x9 t6=%x10 t7=%x11 t8=%x12 t9=%x13 t10=%x14 t11=%x15"
 | 
						|
     :
 | 
						|
       "+m" (*y),
 | 
						|
       "+r" (n),	// 1
 | 
						|
       "+b" (x),	// 2
 | 
						|
       "+b" (y),	// 3
 | 
						|
       "=wa" (t0),	// 4
 | 
						|
       "=wa" (t1),	// 5
 | 
						|
       "=wa" (t2),	// 6
 | 
						|
       "=wa" (t3),	// 7
 | 
						|
       "=wa" (t4),	// 8
 | 
						|
       "=wa" (t5),	// 9
 | 
						|
       "=wa" (t6),	// 10
 | 
						|
       "=wa" (t7),	// 11
 | 
						|
       "=wa" (t8),	// 12
 | 
						|
       "=wa" (t9),	// 13
 | 
						|
       "=wa" (t10),	// 14
 | 
						|
       "=wa" (t11),	// 15
 | 
						|
       "=b" (ytmp)	// 16
 | 
						|
     :
 | 
						|
       "m" (*x),
 | 
						|
       "m" (*mvecp),
 | 
						|
       "d" (alpha_r),	// 19
 | 
						|
       "d" (alpha_i),	// 20
 | 
						|
       "16" (mvecp),	// 21
 | 
						|
       "b" (16),	// 22
 | 
						|
       "b" (32),	// 23
 | 
						|
       "b" (48)		// 24
 | 
						|
     :
 | 
						|
       "cr0",
 | 
						|
       "vs32","vs33","vs34","vs35","vs36","vs37","vs38","vs39",
 | 
						|
       "vs40","vs41","vs42","vs43","vs44","vs45","vs46","vs47",
 | 
						|
       "vs48","vs49","vs50","vs51"
 | 
						|
     );
 | 
						|
}
 |