674 lines
		
	
	
		
			20 KiB
		
	
	
	
		
			C
		
	
	
	
			
		
		
	
	
			674 lines
		
	
	
		
			20 KiB
		
	
	
	
		
			C
		
	
	
	
/*********************************************************************/
 | 
						|
/* Copyright 2009, 2010 The University of Texas at Austin.           */
 | 
						|
/* All rights reserved.                                              */
 | 
						|
/*                                                                   */
 | 
						|
/* Redistribution and use in source and binary forms, with or        */
 | 
						|
/* without modification, are permitted provided that the following   */
 | 
						|
/* conditions are met:                                               */
 | 
						|
/*                                                                   */
 | 
						|
/*   1. Redistributions of source code must retain the above         */
 | 
						|
/*      copyright notice, this list of conditions and the following  */
 | 
						|
/*      disclaimer.                                                  */
 | 
						|
/*                                                                   */
 | 
						|
/*   2. Redistributions in binary form must reproduce the above      */
 | 
						|
/*      copyright notice, this list of conditions and the following  */
 | 
						|
/*      disclaimer in the documentation and/or other materials       */
 | 
						|
/*      provided with the distribution.                              */
 | 
						|
/*                                                                   */
 | 
						|
/*    THIS  SOFTWARE IS PROVIDED  BY THE  UNIVERSITY OF  TEXAS AT    */
 | 
						|
/*    AUSTIN  ``AS IS''  AND ANY  EXPRESS OR  IMPLIED WARRANTIES,    */
 | 
						|
/*    INCLUDING, BUT  NOT LIMITED  TO, THE IMPLIED  WARRANTIES OF    */
 | 
						|
/*    MERCHANTABILITY  AND FITNESS FOR  A PARTICULAR  PURPOSE ARE    */
 | 
						|
/*    DISCLAIMED.  IN  NO EVENT SHALL THE UNIVERSITY  OF TEXAS AT    */
 | 
						|
/*    AUSTIN OR CONTRIBUTORS BE  LIABLE FOR ANY DIRECT, INDIRECT,    */
 | 
						|
/*    INCIDENTAL,  SPECIAL, EXEMPLARY,  OR  CONSEQUENTIAL DAMAGES    */
 | 
						|
/*    (INCLUDING, BUT  NOT LIMITED TO,  PROCUREMENT OF SUBSTITUTE    */
 | 
						|
/*    GOODS  OR  SERVICES; LOSS  OF  USE,  DATA,  OR PROFITS;  OR    */
 | 
						|
/*    BUSINESS INTERRUPTION) HOWEVER CAUSED  AND ON ANY THEORY OF    */
 | 
						|
/*    LIABILITY, WHETHER  IN CONTRACT, STRICT  LIABILITY, OR TORT    */
 | 
						|
/*    (INCLUDING NEGLIGENCE OR OTHERWISE)  ARISING IN ANY WAY OUT    */
 | 
						|
/*    OF  THE  USE OF  THIS  SOFTWARE,  EVEN  IF ADVISED  OF  THE    */
 | 
						|
/*    POSSIBILITY OF SUCH DAMAGE.                                    */
 | 
						|
/*                                                                   */
 | 
						|
/* The views and conclusions contained in the software and           */
 | 
						|
/* documentation are those of the authors and should not be          */
 | 
						|
/* interpreted as representing official policies, either expressed   */
 | 
						|
/* or implied, of The University of Texas at Austin.                 */
 | 
						|
/*********************************************************************/
 | 
						|
 | 
						|
#include "common.h"
 | 
						|
 | 
						|
static FLOAT dm1 = -1.;
 | 
						|
 | 
						|
#ifdef CONJ
 | 
						|
#define GEMM_KERNEL   GEMM_KERNEL_R
 | 
						|
#else
 | 
						|
#define GEMM_KERNEL   GEMM_KERNEL_N
 | 
						|
#endif
 | 
						|
 | 
						|
#if GEMM_DEFAULT_UNROLL_M == 1
 | 
						|
#define GEMM_UNROLL_M_SHIFT 0
 | 
						|
#endif
 | 
						|
 | 
						|
#if GEMM_DEFAULT_UNROLL_M == 2
 | 
						|
#define GEMM_UNROLL_M_SHIFT 1
 | 
						|
#endif
 | 
						|
 | 
						|
#if GEMM_DEFAULT_UNROLL_M == 4
 | 
						|
#define GEMM_UNROLL_M_SHIFT 2
 | 
						|
#endif
 | 
						|
 | 
						|
#if GEMM_DEFAULT_UNROLL_M == 6
 | 
						|
#define GEMM_UNROLL_M_SHIFT 2
 | 
						|
#endif
 | 
						|
 | 
						|
#if GEMM_DEFAULT_UNROLL_M == 8
 | 
						|
#define GEMM_UNROLL_M_SHIFT 3
 | 
						|
#endif
 | 
						|
 | 
						|
#if GEMM_DEFAULT_UNROLL_M == 16
 | 
						|
#define GEMM_UNROLL_M_SHIFT 4
 | 
						|
#endif
 | 
						|
 | 
						|
#if GEMM_DEFAULT_UNROLL_N == 1
 | 
						|
#define GEMM_UNROLL_N_SHIFT 0
 | 
						|
#endif
 | 
						|
 | 
						|
#if GEMM_DEFAULT_UNROLL_N == 2
 | 
						|
#define GEMM_UNROLL_N_SHIFT 1
 | 
						|
#endif
 | 
						|
 | 
						|
#if GEMM_DEFAULT_UNROLL_N == 4
 | 
						|
#define GEMM_UNROLL_N_SHIFT 2
 | 
						|
#endif
 | 
						|
 | 
						|
#if GEMM_DEFAULT_UNROLL_N == 8
 | 
						|
#define GEMM_UNROLL_N_SHIFT 3
 | 
						|
#endif
 | 
						|
 | 
						|
#if GEMM_DEFAULT_UNROLL_N == 16
 | 
						|
#define GEMM_UNROLL_N_SHIFT 4
 | 
						|
#endif
 | 
						|
 | 
						|
 | 
						|
 | 
						|
static void dtrsm_RN_solve_opt(BLASLONG n, FLOAT *a, FLOAT *b, FLOAT *c, BLASLONG ldc, FLOAT *as, FLOAT *bs)  __attribute__ ((noinline));
 | 
						|
 | 
						|
static void dtrsm_RN_solve_opt(BLASLONG n, FLOAT *a, FLOAT *b, FLOAT *c, BLASLONG ldc, FLOAT *as, FLOAT *bs)
 | 
						|
{
 | 
						|
 | 
						|
	FLOAT *c3 = c + ldc + ldc*2 ;
 | 
						|
	FLOAT *c6 = c + ldc*4 + ldc*2 ;
 | 
						|
	ldc = ldc *8;
 | 
						|
	BLASLONG n1 = n * 8;
 | 
						|
	BLASLONG i=0;
 | 
						|
 | 
						|
        __asm__  __volatile__
 | 
						|
        (
 | 
						|
	"	vzeroupper							\n\t"
 | 
						|
 | 
						|
	"	vxorpd	%%ymm8 , %%ymm8 , %%ymm8				\n\t"
 | 
						|
	"	vxorpd	%%ymm9 , %%ymm9 , %%ymm9				\n\t"
 | 
						|
	"	vxorpd	%%ymm10, %%ymm10, %%ymm10				\n\t"
 | 
						|
	"	vxorpd	%%ymm11, %%ymm11, %%ymm11				\n\t"
 | 
						|
	"	vxorpd	%%ymm12, %%ymm12, %%ymm12				\n\t"
 | 
						|
	"	vxorpd	%%ymm13, %%ymm13, %%ymm13				\n\t"
 | 
						|
	"	vxorpd	%%ymm14, %%ymm14, %%ymm14				\n\t"
 | 
						|
	"	vxorpd	%%ymm15, %%ymm15, %%ymm15				\n\t"
 | 
						|
 | 
						|
	"	cmpq	       $0, %0						\n\t"
 | 
						|
	"	je	       4f						\n\t"
 | 
						|
 | 
						|
	"	vmovups         (%8,%1,4), %%ymm0				\n\t"	// read a
 | 
						|
	"	vmovups         (%9,%1,8), %%ymm1				\n\t"	// read b0
 | 
						|
	"	vmovups       32(%9,%1,8), %%ymm2				\n\t"	// read b1
 | 
						|
 | 
						|
 | 
						|
	"	addq		$8, %1						\n\t"
 | 
						|
	"	cmpq		%1, %0						\n\t"
 | 
						|
	"	je	       21f						\n\t"
 | 
						|
 | 
						|
	"	.p2align 4							\n\t"
 | 
						|
	"1:									\n\t"
 | 
						|
 | 
						|
	"	vmovups         (%8,%1,4), %%ymm4				\n\t"	// read a
 | 
						|
        "       vpermilpd         $0x05  , %%ymm0 , %%ymm3                	\n\t"   // was vpermpd 0xb1
 | 
						|
 | 
						|
	"	vfmadd231pd	%%ymm0 , %%ymm1 , %%ymm8			\n\t"
 | 
						|
	"	vfmadd231pd	%%ymm0 , %%ymm2 , %%ymm12			\n\t"
 | 
						|
 | 
						|
	"	vmovups         (%9,%1,8), %%ymm5				\n\t"	// read b0
 | 
						|
	"	vfmadd231pd	%%ymm3 , %%ymm1 , %%ymm9			\n\t"
 | 
						|
	"	vfmadd231pd	%%ymm3 , %%ymm2 , %%ymm13			\n\t"
 | 
						|
 | 
						|
        "       vpermpd         $0x1b  , %%ymm3 , %%ymm0                	\n\t"
 | 
						|
	"	vmovups       32(%9,%1,8), %%ymm6				\n\t"	// read b1
 | 
						|
        "       vpermilpd         $0x05  , %%ymm0 , %%ymm3                	\n\t"
 | 
						|
	"	vfmadd231pd	%%ymm0 , %%ymm1 , %%ymm10			\n\t"
 | 
						|
	"	vfmadd231pd	%%ymm0 , %%ymm2 , %%ymm14			\n\t"
 | 
						|
 | 
						|
	"	addq		$8, %1						\n\t"
 | 
						|
	"	vfmadd231pd	%%ymm3 , %%ymm1 , %%ymm11			\n\t"
 | 
						|
	"	vfmadd231pd	%%ymm3 , %%ymm2 , %%ymm15			\n\t"
 | 
						|
 | 
						|
	"	cmpq		%1, %0						\n\t"
 | 
						|
 | 
						|
	"	jz		22f						\n\t"
 | 
						|
 | 
						|
	"	vmovups         (%8,%1,4), %%ymm0				\n\t"	// read a
 | 
						|
 | 
						|
	"	vfmadd231pd	%%ymm4 , %%ymm5 , %%ymm8			\n\t"
 | 
						|
	"	vfmadd231pd	%%ymm4 , %%ymm6 , %%ymm12			\n\t"
 | 
						|
 | 
						|
        "       vpermilpd         $0x05  , %%ymm4 , %%ymm4                	\n\t"
 | 
						|
	"	vmovups         (%9,%1,8), %%ymm1				\n\t"	// read b0
 | 
						|
	"	vfmadd231pd	%%ymm4 , %%ymm5 , %%ymm9			\n\t"
 | 
						|
	"	vfmadd231pd	%%ymm4 , %%ymm6 , %%ymm13			\n\t"
 | 
						|
 | 
						|
        "       vpermpd         $0x1b  , %%ymm4 , %%ymm4                	\n\t"
 | 
						|
	"	vmovups       32(%9,%1,8), %%ymm2				\n\t"	// read b1
 | 
						|
	"	vfmadd231pd	%%ymm4 , %%ymm5 , %%ymm10			\n\t"
 | 
						|
	"	vfmadd231pd	%%ymm4 , %%ymm6 , %%ymm14			\n\t"
 | 
						|
 | 
						|
        "       vpermilpd         $0x05  , %%ymm4 , %%ymm4                	\n\t"
 | 
						|
	"	addq		$8, %1						\n\t"
 | 
						|
	"	vfmadd231pd	%%ymm4 , %%ymm5 , %%ymm11			\n\t"
 | 
						|
	"	vfmadd231pd	%%ymm4 , %%ymm6 , %%ymm15			\n\t"
 | 
						|
 | 
						|
	"	cmpq		%1, %0						\n\t"
 | 
						|
 | 
						|
	"	jnz		1b						\n\t"
 | 
						|
 | 
						|
 | 
						|
	"21:									\n\t"
 | 
						|
 | 
						|
	"	vfmadd231pd	%%ymm0 , %%ymm1 , %%ymm8			\n\t"
 | 
						|
	"	vfmadd231pd	%%ymm0 , %%ymm2 , %%ymm12			\n\t"
 | 
						|
 | 
						|
        "       vpermilpd         $0x05  , %%ymm0 , %%ymm0                	\n\t"
 | 
						|
	"	vfmadd231pd	%%ymm0 , %%ymm1 , %%ymm9			\n\t"
 | 
						|
	"	vfmadd231pd	%%ymm0 , %%ymm2 , %%ymm13			\n\t"
 | 
						|
 | 
						|
        "       vpermpd         $0x1b  , %%ymm0 , %%ymm0                	\n\t"
 | 
						|
	"	vfmadd231pd	%%ymm0 , %%ymm1 , %%ymm10			\n\t"
 | 
						|
	"	vfmadd231pd	%%ymm0 , %%ymm2 , %%ymm14			\n\t"
 | 
						|
 | 
						|
        "       vpermilpd         $0x05  , %%ymm0 , %%ymm0                	\n\t"
 | 
						|
	"	vfmadd231pd	%%ymm0 , %%ymm1 , %%ymm11			\n\t"
 | 
						|
	"	vfmadd231pd	%%ymm0 , %%ymm2 , %%ymm15			\n\t"
 | 
						|
 | 
						|
	"	jmp	3f							\n\t"
 | 
						|
 | 
						|
	"22:									\n\t"
 | 
						|
	
 | 
						|
	"	vfmadd231pd	%%ymm4 , %%ymm5 , %%ymm8			\n\t"
 | 
						|
	"	vfmadd231pd	%%ymm4 , %%ymm6 , %%ymm12			\n\t"
 | 
						|
 | 
						|
        "       vpermilpd         $0x05  , %%ymm4 , %%ymm4                	\n\t"
 | 
						|
	"	vfmadd231pd	%%ymm4 , %%ymm5 , %%ymm9			\n\t"
 | 
						|
	"	vfmadd231pd	%%ymm4 , %%ymm6 , %%ymm13			\n\t"
 | 
						|
 | 
						|
        "       vpermpd         $0x1b  , %%ymm4 , %%ymm4                	\n\t"
 | 
						|
	"	vfmadd231pd	%%ymm4 , %%ymm5 , %%ymm10			\n\t"
 | 
						|
	"	vfmadd231pd	%%ymm4 , %%ymm6 , %%ymm14			\n\t"
 | 
						|
 | 
						|
        "       vpermilpd         $0x05  , %%ymm4 , %%ymm4                	\n\t"
 | 
						|
	"	vfmadd231pd	%%ymm4 , %%ymm5 , %%ymm11			\n\t"
 | 
						|
	"	vfmadd231pd	%%ymm4 , %%ymm6 , %%ymm15			\n\t"
 | 
						|
 | 
						|
	"3:								\n\t"	
 | 
						|
 | 
						|
        "       vpermilpd         $0x05  , %%ymm9 , %%ymm9                \n\t"
 | 
						|
        "       vpermilpd         $0x05  , %%ymm11, %%ymm11               \n\t"
 | 
						|
 | 
						|
        "       vblendpd        $0x0a  , %%ymm9 , %%ymm8 , %%ymm0       \n\t"
 | 
						|
        "       vblendpd        $0x05  , %%ymm9 , %%ymm8 , %%ymm1       \n\t"
 | 
						|
        "       vblendpd        $0x0a  , %%ymm11, %%ymm10, %%ymm2       \n\t"
 | 
						|
        "       vblendpd        $0x05  , %%ymm11, %%ymm10, %%ymm3       \n\t"
 | 
						|
 | 
						|
        "       vperm2f128         $0x01  , %%ymm2 , %%ymm2 , %%ymm2    \n\t"
 | 
						|
        "       vperm2f128         $0x01  , %%ymm3 , %%ymm3 , %%ymm3    \n\t"
 | 
						|
 | 
						|
        "       vblendpd        $0x03  , %%ymm0 , %%ymm2 , %%ymm8       \n\t"
 | 
						|
        "       vblendpd        $0x03  , %%ymm1 , %%ymm3 , %%ymm9       \n\t"
 | 
						|
        "       vblendpd        $0x03  , %%ymm2 , %%ymm0 , %%ymm10      \n\t"
 | 
						|
        "       vblendpd        $0x03  , %%ymm3 , %%ymm1 , %%ymm11      \n\t"
 | 
						|
 | 
						|
        "       vpermilpd         $0x05  , %%ymm13, %%ymm13               \n\t"
 | 
						|
        "       vpermilpd         $0x05  , %%ymm15, %%ymm15               \n\t"
 | 
						|
 | 
						|
        "       vblendpd        $0x0a  , %%ymm13, %%ymm12, %%ymm0       \n\t"
 | 
						|
        "       vblendpd        $0x05  , %%ymm13, %%ymm12, %%ymm1       \n\t"
 | 
						|
        "       vblendpd        $0x0a  , %%ymm15, %%ymm14, %%ymm2       \n\t"
 | 
						|
        "       vblendpd        $0x05  , %%ymm15, %%ymm14, %%ymm3       \n\t"
 | 
						|
 | 
						|
        "       vperm2f128         $0x01  , %%ymm2 , %%ymm2 , %%ymm2       \n\t"
 | 
						|
        "       vperm2f128         $0x01  , %%ymm3 , %%ymm3 , %%ymm3       \n\t"
 | 
						|
 | 
						|
        "       vblendpd        $0x03  , %%ymm0 , %%ymm2 , %%ymm12      \n\t"
 | 
						|
        "       vblendpd        $0x03  , %%ymm1 , %%ymm3 , %%ymm13      \n\t"
 | 
						|
        "       vblendpd        $0x03  , %%ymm2 , %%ymm0 , %%ymm14      \n\t"
 | 
						|
        "       vblendpd        $0x03  , %%ymm3 , %%ymm1 , %%ymm15      \n\t"
 | 
						|
 | 
						|
 | 
						|
	"4:								\n\t"	
 | 
						|
 | 
						|
	"	vmovups		  (%4)      , %%ymm0			\n\t"	// read c0
 | 
						|
	"	vmovups		  (%4,%7,1) , %%ymm1			\n\t"	// read c1
 | 
						|
	"	vmovups		  (%4,%7,2) , %%ymm2			\n\t"	// read c2
 | 
						|
	"	vmovups		  (%5)      , %%ymm3			\n\t"	// read c3
 | 
						|
 | 
						|
	"	vmovups		  (%5,%7,1) , %%ymm4			\n\t"	// read c4
 | 
						|
	"	vmovups		  (%5,%7,2) , %%ymm5			\n\t"	// read c5
 | 
						|
	"	vmovups		  (%6)      , %%ymm6			\n\t"	// read c6
 | 
						|
	"	vmovups		  (%6,%7,1) , %%ymm7			\n\t"	// read c7
 | 
						|
 | 
						|
	"	vsubpd		%%ymm8 , %%ymm0 , %%ymm8		\n\t"
 | 
						|
	"	vmovups           (%3),  %%ymm0				\n\t"
 | 
						|
	"	vsubpd		%%ymm9 , %%ymm1 , %%ymm9		\n\t"
 | 
						|
	"	vpermpd		$0x55 ,  %%ymm0 , %%ymm1		\n\t"
 | 
						|
	"	vsubpd		%%ymm10, %%ymm2 , %%ymm10		\n\t"
 | 
						|
	"	vpermpd		$0xaa ,  %%ymm0 , %%ymm2		\n\t"
 | 
						|
	"	vsubpd		%%ymm11, %%ymm3 , %%ymm11		\n\t"
 | 
						|
	"	vpermpd		$0xff ,  %%ymm0 , %%ymm3		\n\t"
 | 
						|
	"	vpermpd		$0x00 ,  %%ymm0 , %%ymm0		\n\t"
 | 
						|
 | 
						|
	"	vsubpd		%%ymm12, %%ymm4 , %%ymm12		\n\t"
 | 
						|
	"	vmovups         32(%3),  %%ymm4				\n\t"
 | 
						|
	"	vsubpd		%%ymm13, %%ymm5 , %%ymm13		\n\t"
 | 
						|
	"	vpermpd		$0x55 ,  %%ymm4 , %%ymm5		\n\t"
 | 
						|
	"	vsubpd		%%ymm14, %%ymm6 , %%ymm14		\n\t"
 | 
						|
	"	vpermpd		$0xaa ,  %%ymm4 , %%ymm6		\n\t"
 | 
						|
	"	vsubpd		%%ymm15, %%ymm7 , %%ymm15		\n\t"
 | 
						|
	"	vpermpd		$0xff ,  %%ymm4 , %%ymm7		\n\t"
 | 
						|
	"	vpermpd		$0x00 ,  %%ymm4 , %%ymm4		\n\t"
 | 
						|
 | 
						|
 | 
						|
	"5:								\n\t"	// i = 0
 | 
						|
 | 
						|
	"	addq	$64, %3						\n\t"	// b=b+8
 | 
						|
 | 
						|
	"	vmulpd		%%ymm8 , %%ymm0, %%ymm8			\n\t"	// a *bb
 | 
						|
	"	vmovups           (%3),  %%ymm0				\n\t"
 | 
						|
	"	vmovups		%%ymm8 , (%2)				\n\t"	// write a
 | 
						|
	"	vmovups		%%ymm8 , (%4)				\n\t"	// write c
 | 
						|
 | 
						|
	"	vfnmadd231pd	%%ymm8 , %%ymm1 , %%ymm9		\n\t"
 | 
						|
	"	vmovups         32(%3),  %%ymm1				\n\t"
 | 
						|
	"	vfnmadd231pd	%%ymm8 , %%ymm2 , %%ymm10		\n\t"
 | 
						|
	"	vpermpd		$0xaa ,  %%ymm0 , %%ymm2		\n\t"
 | 
						|
	"	vfnmadd231pd	%%ymm8 , %%ymm3 , %%ymm11		\n\t"
 | 
						|
	"	vpermpd		$0xff ,  %%ymm0 , %%ymm3		\n\t"
 | 
						|
	"	vfnmadd231pd	%%ymm8 , %%ymm4 , %%ymm12		\n\t"
 | 
						|
	"	vpermpd		$0x55 ,  %%ymm0 , %%ymm0		\n\t"
 | 
						|
	"	vfnmadd231pd	%%ymm8 , %%ymm5 , %%ymm13		\n\t"
 | 
						|
	"	vpermpd		$0x55 ,  %%ymm1 , %%ymm5		\n\t"
 | 
						|
	"	vfnmadd231pd	%%ymm8 , %%ymm6 , %%ymm14		\n\t"
 | 
						|
	"	vpermpd		$0xaa ,  %%ymm1 , %%ymm6		\n\t"
 | 
						|
	"	vfnmadd231pd	%%ymm8 , %%ymm7 , %%ymm15		\n\t"
 | 
						|
	"	vpermpd		$0xff ,  %%ymm1 , %%ymm7		\n\t"
 | 
						|
	"	vpermpd		$0x00 ,  %%ymm1 , %%ymm4		\n\t"
 | 
						|
 | 
						|
	"	addq	$64, %3						\n\t"	// b=b+8
 | 
						|
	"	addq	$32, %2						\n\t"	// a=a+8
 | 
						|
 | 
						|
 | 
						|
 | 
						|
	"	vmulpd		%%ymm9 , %%ymm0, %%ymm9			\n\t"	// a *bb
 | 
						|
	"	vmovups           (%3),  %%ymm0				\n\t"
 | 
						|
	"	vmovups         32(%3),  %%ymm1				\n\t"
 | 
						|
	"	vmovups		%%ymm9 , (%2)				\n\t"	// write a
 | 
						|
	"	vmovups		%%ymm9 , (%4,%7,1)			\n\t"	// write c
 | 
						|
 | 
						|
	"	vfnmadd231pd	%%ymm9 , %%ymm2 , %%ymm10		\n\t"
 | 
						|
	"	vfnmadd231pd	%%ymm9 , %%ymm3 , %%ymm11		\n\t"
 | 
						|
	"	vpermpd		$0xff ,  %%ymm0 , %%ymm3		\n\t"
 | 
						|
	"	vfnmadd231pd	%%ymm9 , %%ymm4 , %%ymm12		\n\t"
 | 
						|
	"	vpermpd		$0xaa ,  %%ymm0 , %%ymm0		\n\t"
 | 
						|
	"	vfnmadd231pd	%%ymm9 , %%ymm5 , %%ymm13		\n\t"
 | 
						|
	"	vpermpd		$0x55 ,  %%ymm1 , %%ymm5		\n\t"
 | 
						|
	"	vfnmadd231pd	%%ymm9 , %%ymm6 , %%ymm14		\n\t"
 | 
						|
	"	vpermpd		$0xaa ,  %%ymm1 , %%ymm6		\n\t"
 | 
						|
	"	vfnmadd231pd	%%ymm9 , %%ymm7 , %%ymm15		\n\t"
 | 
						|
	"	vpermpd		$0xff ,  %%ymm1 , %%ymm7		\n\t"
 | 
						|
	"	vpermpd		$0x00 ,  %%ymm1 , %%ymm4		\n\t"
 | 
						|
 | 
						|
	"	addq	$64, %3						\n\t"	// b=b+8
 | 
						|
	"	addq	$32, %2						\n\t"	// a=a+8
 | 
						|
 | 
						|
	"	vmulpd		%%ymm10, %%ymm0, %%ymm10		\n\t"	// a *bb
 | 
						|
	"	vmovups           (%3),  %%ymm0				\n\t"
 | 
						|
	"	vmovups         32(%3),  %%ymm1				\n\t"
 | 
						|
	"	vmovups		%%ymm10, (%2)				\n\t"	// write a
 | 
						|
	"	vmovups		%%ymm10, (%4,%7,2)			\n\t"	// write c
 | 
						|
 | 
						|
	"	vfnmadd231pd	%%ymm10, %%ymm3 , %%ymm11		\n\t"
 | 
						|
	"	vpermpd		$0xff ,  %%ymm0 , %%ymm0		\n\t"
 | 
						|
	"	vfnmadd231pd	%%ymm10, %%ymm4 , %%ymm12		\n\t"
 | 
						|
	"	vfnmadd231pd	%%ymm10, %%ymm5 , %%ymm13		\n\t"
 | 
						|
	"	vpermpd		$0x55 ,  %%ymm1 , %%ymm5		\n\t"
 | 
						|
	"	vfnmadd231pd	%%ymm10, %%ymm6 , %%ymm14		\n\t"
 | 
						|
	"	vpermpd		$0xaa ,  %%ymm1 , %%ymm6		\n\t"
 | 
						|
	"	vfnmadd231pd	%%ymm10, %%ymm7 , %%ymm15		\n\t"
 | 
						|
	"	vpermpd		$0xff ,  %%ymm1 , %%ymm7		\n\t"
 | 
						|
	"	vpermpd		$0x00 ,  %%ymm1 , %%ymm4		\n\t"
 | 
						|
 | 
						|
 | 
						|
	"	addq	$64, %3						\n\t"	// b=b+8
 | 
						|
	"	addq	$32, %2						\n\t"	// a=a+8
 | 
						|
 | 
						|
 | 
						|
 | 
						|
	"	vmulpd		%%ymm11, %%ymm0, %%ymm11		\n\t"	// a *bb
 | 
						|
	"	vmovups         32(%3),  %%ymm1				\n\t"
 | 
						|
	"	vmovups		%%ymm11, (%2)				\n\t"	// write a
 | 
						|
	"	vmovups		%%ymm11, (%5)     			\n\t"	// write c
 | 
						|
 | 
						|
	"	vfnmadd231pd	%%ymm11, %%ymm4 , %%ymm12		\n\t"
 | 
						|
	"	vfnmadd231pd	%%ymm11, %%ymm5 , %%ymm13		\n\t"
 | 
						|
	"	vpermpd		$0x55 ,  %%ymm1 , %%ymm5		\n\t"
 | 
						|
	"	vfnmadd231pd	%%ymm11, %%ymm6 , %%ymm14		\n\t"
 | 
						|
	"	vpermpd		$0xaa ,  %%ymm1 , %%ymm6		\n\t"
 | 
						|
	"	vfnmadd231pd	%%ymm11, %%ymm7 , %%ymm15		\n\t"
 | 
						|
	"	vpermpd		$0xff ,  %%ymm1 , %%ymm7		\n\t"
 | 
						|
	"	vpermpd		$0x00 ,  %%ymm1 , %%ymm0		\n\t"
 | 
						|
 | 
						|
 | 
						|
	"	addq	$64, %3						\n\t"	// b=b+8
 | 
						|
	"	addq	$32, %2						\n\t"	// a=a+8
 | 
						|
 | 
						|
 | 
						|
	"	vmulpd		%%ymm12, %%ymm0, %%ymm12		\n\t"	// a *bb
 | 
						|
	"	vmovups         32(%3),  %%ymm1				\n\t"
 | 
						|
	"	vmovups		%%ymm12, (%2)				\n\t"	// write a
 | 
						|
	"	vmovups		%%ymm12, (%5,%7,1)			\n\t"	// write c
 | 
						|
 | 
						|
	"	vfnmadd231pd	%%ymm12, %%ymm5 , %%ymm13		\n\t"
 | 
						|
	"	vfnmadd231pd	%%ymm12, %%ymm6 , %%ymm14		\n\t"
 | 
						|
	"	vpermpd		$0xaa ,  %%ymm1 , %%ymm6		\n\t"
 | 
						|
	"	vfnmadd231pd	%%ymm12, %%ymm7 , %%ymm15		\n\t"
 | 
						|
	"	vpermpd		$0xff ,  %%ymm1 , %%ymm7		\n\t"
 | 
						|
	"	vpermpd		$0x55 ,  %%ymm1 , %%ymm0		\n\t"
 | 
						|
 | 
						|
	"	addq	$64, %3						\n\t"	// b=b+8
 | 
						|
	"	addq	$32, %2						\n\t"	// a=a+8
 | 
						|
 | 
						|
	"	vmulpd		%%ymm13, %%ymm0, %%ymm13		\n\t"	// a *bb
 | 
						|
	"	vmovups         32(%3),  %%ymm1				\n\t"
 | 
						|
	"	vmovups		%%ymm13, (%2)				\n\t"	// write a
 | 
						|
	"	vmovups		%%ymm13, (%5,%7,2)			\n\t"	// write c
 | 
						|
 | 
						|
	"	vfnmadd231pd	%%ymm13, %%ymm6 , %%ymm14		\n\t"
 | 
						|
	"	vfnmadd231pd	%%ymm13, %%ymm7 , %%ymm15		\n\t"
 | 
						|
	"	vpermpd		$0xff ,  %%ymm1 , %%ymm7		\n\t"
 | 
						|
	"	vpermpd		$0xaa ,  %%ymm1 , %%ymm0		\n\t"
 | 
						|
 | 
						|
 | 
						|
	"	addq	$64, %3						\n\t"	// b=b+8
 | 
						|
	"	addq	$32, %2						\n\t"	// a=a+8
 | 
						|
 | 
						|
 | 
						|
	"	vmulpd		%%ymm14, %%ymm0, %%ymm14		\n\t"	// a *bb
 | 
						|
	"	vmovups         32(%3),  %%ymm1				\n\t"
 | 
						|
	"	vmovups		%%ymm14, (%2)				\n\t"	// write a
 | 
						|
	"	vmovups		%%ymm14, (%6)     			\n\t"	// write c
 | 
						|
 | 
						|
	"	vfnmadd231pd	%%ymm14, %%ymm7 , %%ymm15		\n\t"
 | 
						|
 | 
						|
	"	vpermpd		$0xff ,  %%ymm1 , %%ymm0		\n\t"
 | 
						|
 | 
						|
	"	addq	$32, %2						\n\t"	// a=a+8
 | 
						|
 | 
						|
	"	vmulpd		%%ymm15, %%ymm0, %%ymm15		\n\t"	// a *bb
 | 
						|
	"	vmovups		%%ymm15, (%2)				\n\t"	// write a
 | 
						|
	"	vmovups		%%ymm15, (%6,%7,1)			\n\t"	// write c
 | 
						|
 | 
						|
	"	vzeroupper						\n\t"
 | 
						|
 | 
						|
        :
 | 
						|
          "+r" (n1),     // 0    
 | 
						|
          "+a" (i),      // 1    
 | 
						|
          "+r" (as),     // 2
 | 
						|
          "+r" (bs)      // 3
 | 
						|
        :
 | 
						|
          "r" (c),      // 4
 | 
						|
          "r" (c3),     // 5
 | 
						|
          "r" (c6),     // 6
 | 
						|
          "r" (ldc),    // 7
 | 
						|
          "r" (a),     // 8
 | 
						|
          "r" (b)      // 9
 | 
						|
        : "cc",
 | 
						|
          "%xmm0", "%xmm1", "%xmm2", "%xmm3",
 | 
						|
          "%xmm4", "%xmm5", "%xmm6", "%xmm7",
 | 
						|
          "%xmm8", "%xmm9", "%xmm10", "%xmm11",
 | 
						|
          "%xmm12", "%xmm13", "%xmm14", "%xmm15",
 | 
						|
          "memory"
 | 
						|
        );
 | 
						|
 | 
						|
}
 | 
						|
 | 
						|
 | 
						|
 | 
						|
#ifndef COMPLEX
 | 
						|
 | 
						|
static inline void solve(BLASLONG m, BLASLONG n, FLOAT *a, FLOAT *b, FLOAT *c, BLASLONG ldc) {
 | 
						|
 | 
						|
  FLOAT aa, bb;
 | 
						|
 | 
						|
  int i, j, k;
 | 
						|
 | 
						|
  for (i = 0; i < n; i++) {
 | 
						|
 | 
						|
    bb = *(b + i);
 | 
						|
 | 
						|
    for (j = 0; j < m; j ++) {
 | 
						|
      aa = *(c + j + i * ldc);
 | 
						|
      aa *= bb;
 | 
						|
      *a  = aa;
 | 
						|
      *(c + j + i * ldc) = aa;
 | 
						|
      a ++;
 | 
						|
 | 
						|
      for (k = i + 1; k < n; k ++){
 | 
						|
	*(c + j + k * ldc) -= aa * *(b + k);
 | 
						|
      }
 | 
						|
 | 
						|
    }
 | 
						|
    b += n;
 | 
						|
  }
 | 
						|
}
 | 
						|
 | 
						|
#else
 | 
						|
 | 
						|
static inline void solve(BLASLONG m, BLASLONG n, FLOAT *a, FLOAT *b, FLOAT *c, BLASLONG ldc) {
 | 
						|
 | 
						|
  FLOAT aa1, aa2;
 | 
						|
  FLOAT bb1, bb2;
 | 
						|
  FLOAT cc1, cc2;
 | 
						|
 | 
						|
  int i, j, k;
 | 
						|
 | 
						|
  ldc *= 2;
 | 
						|
 | 
						|
  for (i = 0; i < n; i++) {
 | 
						|
 | 
						|
    bb1 = *(b + i * 2 + 0);
 | 
						|
    bb2 = *(b + i * 2 + 1);
 | 
						|
 | 
						|
    for (j = 0; j < m; j ++) {
 | 
						|
      aa1 = *(c + j * 2 + 0 + i * ldc);
 | 
						|
      aa2 = *(c + j * 2 + 1 + i * ldc);
 | 
						|
 | 
						|
#ifndef CONJ
 | 
						|
      cc1 = aa1 * bb1 - aa2 * bb2;
 | 
						|
      cc2 = aa1 * bb2 + aa2 * bb1;
 | 
						|
#else
 | 
						|
      cc1 =  aa1 * bb1 + aa2 * bb2;
 | 
						|
      cc2 = -aa1 * bb2 + aa2 * bb1;
 | 
						|
#endif
 | 
						|
 | 
						|
      *(a + 0) = cc1;
 | 
						|
      *(a + 1) = cc2;
 | 
						|
      *(c + j * 2 + 0 + i * ldc) = cc1;
 | 
						|
      *(c + j * 2 + 1 + i * ldc) = cc2;
 | 
						|
      a += 2;
 | 
						|
 | 
						|
      for (k = i + 1; k < n; k ++){
 | 
						|
#ifndef CONJ
 | 
						|
	*(c + j * 2 + 0 + k * ldc) -= cc1 * *(b + k * 2 + 0) - cc2 * *(b + k * 2 + 1);
 | 
						|
	*(c + j * 2 + 1 + k * ldc) -= cc1 * *(b + k * 2 + 1) + cc2 * *(b + k * 2 + 0);
 | 
						|
#else
 | 
						|
	*(c + j * 2 + 0 + k * ldc) -=   cc1 * *(b + k * 2 + 0) + cc2 * *(b + k * 2 + 1);
 | 
						|
	*(c + j * 2 + 1 + k * ldc) -= - cc1 * *(b + k * 2 + 1) + cc2 * *(b + k * 2 + 0);
 | 
						|
#endif
 | 
						|
      }
 | 
						|
 | 
						|
    }
 | 
						|
    b += n * 2;
 | 
						|
  }
 | 
						|
}
 | 
						|
 | 
						|
#endif
 | 
						|
 | 
						|
 | 
						|
int CNAME(BLASLONG m, BLASLONG n, BLASLONG k, FLOAT dummy1,
 | 
						|
#ifdef COMPLEX
 | 
						|
	   FLOAT dummy2,
 | 
						|
#endif
 | 
						|
	   FLOAT *a, FLOAT *b, FLOAT *c, BLASLONG ldc, BLASLONG offset){
 | 
						|
 | 
						|
  FLOAT *aa, *cc;
 | 
						|
  BLASLONG  kk;
 | 
						|
  BLASLONG i, j, jj;
 | 
						|
 | 
						|
#if 0
 | 
						|
  fprintf(stderr, "TRSM RN KERNEL m = %3ld  n = %3ld  k = %3ld offset = %3ld\n",
 | 
						|
	  m, n, k, offset);
 | 
						|
#endif
 | 
						|
 | 
						|
  jj = 0;
 | 
						|
  j = (n >> GEMM_UNROLL_N_SHIFT);
 | 
						|
  kk = -offset;
 | 
						|
 | 
						|
  while (j > 0) {
 | 
						|
 | 
						|
    aa = a;
 | 
						|
    cc = c;
 | 
						|
 | 
						|
    i = (m >> GEMM_UNROLL_M_SHIFT);
 | 
						|
 | 
						|
    if (i > 0) {
 | 
						|
      do {
 | 
						|
 | 
						|
	dtrsm_RN_solve_opt(kk, aa, b, cc, ldc, aa + kk * GEMM_UNROLL_M * COMPSIZE, b + kk * GEMM_UNROLL_N * COMPSIZE);
 | 
						|
/* 
 | 
						|
        solve(GEMM_UNROLL_M, GEMM_UNROLL_N,
 | 
						|
              aa + kk * GEMM_UNROLL_M * COMPSIZE,
 | 
						|
              b  + kk * GEMM_UNROLL_N * COMPSIZE,
 | 
						|
              cc, ldc);
 | 
						|
*/
 | 
						|
	aa += GEMM_UNROLL_M * k * COMPSIZE;
 | 
						|
	cc += GEMM_UNROLL_M     * COMPSIZE;
 | 
						|
	i --;
 | 
						|
      } while (i > 0);
 | 
						|
    }
 | 
						|
 | 
						|
 | 
						|
    if (m & (GEMM_UNROLL_M - 1)) {
 | 
						|
      i = (GEMM_UNROLL_M >> 1);
 | 
						|
      while (i > 0) {
 | 
						|
	if (m & i) {
 | 
						|
	    if (kk > 0) {
 | 
						|
	      GEMM_KERNEL(i, GEMM_UNROLL_N, kk, dm1,
 | 
						|
#ifdef COMPLEX
 | 
						|
			  ZERO,
 | 
						|
#endif
 | 
						|
			  aa, b, cc, ldc);
 | 
						|
	    }
 | 
						|
	  solve(i, GEMM_UNROLL_N,
 | 
						|
		aa + kk * i             * COMPSIZE,
 | 
						|
		b  + kk * GEMM_UNROLL_N * COMPSIZE,
 | 
						|
		cc, ldc);
 | 
						|
 | 
						|
	  aa += i * k * COMPSIZE;
 | 
						|
	  cc += i     * COMPSIZE;
 | 
						|
	}
 | 
						|
	i >>= 1;
 | 
						|
      }
 | 
						|
    }
 | 
						|
 | 
						|
    kk += GEMM_UNROLL_N;
 | 
						|
    b += GEMM_UNROLL_N * k   * COMPSIZE;
 | 
						|
    c += GEMM_UNROLL_N * ldc * COMPSIZE;
 | 
						|
    j --;
 | 
						|
    jj += GEMM_UNROLL_M;
 | 
						|
  }
 | 
						|
 | 
						|
  if (n & (GEMM_UNROLL_N - 1)) {
 | 
						|
 | 
						|
    j = (GEMM_UNROLL_N >> 1);
 | 
						|
    while (j > 0) {
 | 
						|
      if (n & j) {
 | 
						|
 | 
						|
	aa = a;
 | 
						|
	cc = c;
 | 
						|
 | 
						|
	i = (m >> GEMM_UNROLL_M_SHIFT);
 | 
						|
 | 
						|
	while (i > 0) {
 | 
						|
	  if (kk > 0) {
 | 
						|
	    GEMM_KERNEL(GEMM_UNROLL_M, j, kk, dm1,
 | 
						|
#ifdef COMPLEX
 | 
						|
			ZERO,
 | 
						|
#endif
 | 
						|
			aa,
 | 
						|
			b,
 | 
						|
			cc,
 | 
						|
			ldc);
 | 
						|
	  }
 | 
						|
 | 
						|
	  solve(GEMM_UNROLL_M, j,
 | 
						|
		aa + kk * GEMM_UNROLL_M * COMPSIZE,
 | 
						|
		b  + kk * j             * COMPSIZE, cc, ldc);
 | 
						|
 | 
						|
	  aa += GEMM_UNROLL_M * k * COMPSIZE;
 | 
						|
	  cc += GEMM_UNROLL_M     * COMPSIZE;
 | 
						|
	  i --;
 | 
						|
	}
 | 
						|
 | 
						|
	if (m & (GEMM_UNROLL_M - 1)) {
 | 
						|
	  i = (GEMM_UNROLL_M >> 1);
 | 
						|
	  while (i > 0) {
 | 
						|
	    if (m & i) {
 | 
						|
	      if (kk > 0) {
 | 
						|
		GEMM_KERNEL(i, j, kk, dm1,
 | 
						|
#ifdef COMPLEX
 | 
						|
			    ZERO,
 | 
						|
#endif
 | 
						|
			    aa,
 | 
						|
			    b,
 | 
						|
			    cc,
 | 
						|
			    ldc);
 | 
						|
	      }
 | 
						|
 | 
						|
	      solve(i, j,
 | 
						|
		    aa + kk * i * COMPSIZE,
 | 
						|
		    b  + kk * j * COMPSIZE, cc, ldc);
 | 
						|
 | 
						|
	      aa += i * k * COMPSIZE;
 | 
						|
	      cc += i     * COMPSIZE;
 | 
						|
	      }
 | 
						|
	    i >>= 1;
 | 
						|
	  }
 | 
						|
	}
 | 
						|
 | 
						|
	b += j * k   * COMPSIZE;
 | 
						|
	c += j * ldc * COMPSIZE;
 | 
						|
	kk += j;
 | 
						|
      }
 | 
						|
      j >>= 1;
 | 
						|
    }
 | 
						|
  }
 | 
						|
 | 
						|
  return 0;
 | 
						|
}
 |