751 lines
		
	
	
		
			12 KiB
		
	
	
	
		
			ArmAsm
		
	
	
	
			
		
		
	
	
			751 lines
		
	
	
		
			12 KiB
		
	
	
	
		
			ArmAsm
		
	
	
	
/***************************************************************************
 | 
						|
Copyright (c) 2013, The OpenBLAS Project
 | 
						|
All rights reserved.
 | 
						|
Redistribution and use in source and binary forms, with or without
 | 
						|
modification, are permitted provided that the following conditions are
 | 
						|
met:
 | 
						|
1. Redistributions of source code must retain the above copyright
 | 
						|
notice, this list of conditions and the following disclaimer.
 | 
						|
2. Redistributions in binary form must reproduce the above copyright
 | 
						|
notice, this list of conditions and the following disclaimer in
 | 
						|
the documentation and/or other materials provided with the
 | 
						|
distribution.
 | 
						|
3. Neither the name of the OpenBLAS project nor the names of
 | 
						|
its contributors may be used to endorse or promote products
 | 
						|
derived from this software without specific prior written permission.
 | 
						|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
 | 
						|
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
 | 
						|
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
 | 
						|
ARE DISCLAIMED. IN NO EVENT SHALL THE OPENBLAS PROJECT OR CONTRIBUTORS BE
 | 
						|
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
 | 
						|
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
 | 
						|
SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
 | 
						|
CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
 | 
						|
OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE
 | 
						|
USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
 | 
						|
*****************************************************************************/
 | 
						|
 | 
						|
/**************************************************************************************
 | 
						|
* 2013/11/25 Saar
 | 
						|
* 	 BLASTEST 		: OK
 | 
						|
* 	 CTEST			: OK
 | 
						|
* 	 TEST			: OK
 | 
						|
*
 | 
						|
**************************************************************************************/
 | 
						|
 | 
						|
#define ASSEMBLER
 | 
						|
#include "common.h"
 | 
						|
 | 
						|
#define STACKSIZE 256
 | 
						|
 | 
						|
#define	OLD_LDA		[fp, #0 ]
 | 
						|
#define	X		[fp, #4 ]
 | 
						|
#define	OLD_INC_X	[fp, #8 ]
 | 
						|
#define	Y		[fp, #12 ]
 | 
						|
#define	OLD_INC_Y	[fp, #16 ]
 | 
						|
#define OLD_A		r3
 | 
						|
#define	OLD_N		r1
 | 
						|
 | 
						|
#define M	r0
 | 
						|
#define AO1	r1
 | 
						|
#define J	r2
 | 
						|
 | 
						|
#define AO2	r4
 | 
						|
#define XO	r5
 | 
						|
#define YO	r6
 | 
						|
#define LDA	r7
 | 
						|
#define INC_X	r8
 | 
						|
#define INC_Y	r9
 | 
						|
 | 
						|
#define I	r12
 | 
						|
 | 
						|
#define N	[fp, #-252 ]
 | 
						|
#define A	[fp, #-256 ]
 | 
						|
 | 
						|
 | 
						|
#define X_PRE	512
 | 
						|
#define A_PRE	512
 | 
						|
 | 
						|
/**************************************************************************************
 | 
						|
* Macro definitions
 | 
						|
**************************************************************************************/
 | 
						|
 | 
						|
 | 
						|
#if	defined(DOUBLE)
 | 
						|
 | 
						|
.macro INIT_F2
 | 
						|
 | 
						|
	vsub.f64	d2 , d2 , d2
 | 
						|
	vsub.f64	d3 , d3 , d3
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
.macro KERNEL_F2X4
 | 
						|
 | 
						|
	pld	[ XO  , #X_PRE ]
 | 
						|
	fldmiad	XO! ,  { d12 - d15 }
 | 
						|
	pld	[ AO1 , #A_PRE ]
 | 
						|
	fldmiad	AO1!,  { d8 - d9   }
 | 
						|
	pld	[ AO2 , #A_PRE ]
 | 
						|
	fldmiad	AO2!,  { d4 - d5 }
 | 
						|
	fldmiad	AO1!,  { d10 - d11 }
 | 
						|
	fldmiad	AO2!,  { d6 - d7 }
 | 
						|
 | 
						|
	vmla.f64	d2 , d12 , d8
 | 
						|
	vmla.f64	d3 , d12 , d4
 | 
						|
	vmla.f64	d2 , d13 , d9
 | 
						|
	vmla.f64	d3 , d13 , d5
 | 
						|
	vmla.f64	d2 , d14, d10
 | 
						|
	vmla.f64	d3 , d14, d6
 | 
						|
	vmla.f64	d2 , d15, d11
 | 
						|
	vmla.f64	d3 , d15, d7
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
.macro KERNEL_F2X1
 | 
						|
 | 
						|
	fldmiad	XO! ,  { d1 }
 | 
						|
	fldmiad	AO1!,  { d8 }
 | 
						|
	fldmiad	AO2!,  { d4 }
 | 
						|
	vmla.f64	d2 , d1 , d8
 | 
						|
	vmla.f64	d3 , d1 , d4
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
.macro	SAVE_F2
 | 
						|
 | 
						|
	fldmiad	YO,  { d4 - d5 }
 | 
						|
	vmla.f64	d4, d0, d2
 | 
						|
	vmla.f64	d5, d0, d3
 | 
						|
	fstmiad	YO!, { d4 - d5 }
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
.macro INIT_F1
 | 
						|
 | 
						|
	vsub.f64	d2 , d2 , d2
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
.macro KERNEL_F1X4
 | 
						|
 | 
						|
	pld	[ XO  , #X_PRE ]
 | 
						|
	fldmiad	XO! ,  { d12 - d15 }
 | 
						|
	pld	[ AO1 , #A_PRE ]
 | 
						|
	fldmiad	AO1!,  { d8 - d9   }
 | 
						|
	fldmiad	AO1!,  { d10 - d11 }
 | 
						|
	vmla.f64	d2 , d12 , d8
 | 
						|
	vmla.f64	d2 , d13 , d9
 | 
						|
	vmla.f64	d2 , d14, d10
 | 
						|
	vmla.f64	d2 , d15, d11
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
.macro KERNEL_F1X1
 | 
						|
 | 
						|
	fldmiad	XO! ,  { d1 }
 | 
						|
	fldmiad	AO1!,  { d8 }
 | 
						|
	vmla.f64	d2 , d1 , d8
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
.macro	SAVE_F1
 | 
						|
 | 
						|
	fldmiad	YO,  { d4 }
 | 
						|
	vmla.f64	d4, d0, d2
 | 
						|
	fstmiad	YO!, { d4 }
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
 | 
						|
.macro INIT_S2
 | 
						|
 | 
						|
	vsub.f64	d2 , d2 , d2
 | 
						|
	vsub.f64	d3 , d3 , d3
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
.macro KERNEL_S2X4
 | 
						|
 | 
						|
	fldmiad	XO ,  { d12 }
 | 
						|
	add	XO, XO, INC_X
 | 
						|
 | 
						|
	pld	[ AO1 , #A_PRE ]
 | 
						|
	fldmiad	AO1!,  { d8 - d9   }
 | 
						|
	pld	[ AO2 , #A_PRE ]
 | 
						|
	fldmiad	AO2!,  { d4 - d5 }
 | 
						|
 | 
						|
	fldmiad	XO ,  { d13 }
 | 
						|
	add	XO, XO, INC_X
 | 
						|
	fldmiad	AO1!,  { d10 - d11 }
 | 
						|
	fldmiad	AO2!,  { d6 - d7 }
 | 
						|
 | 
						|
	fldmiad	XO ,  { d14 }
 | 
						|
	add	XO, XO, INC_X
 | 
						|
 | 
						|
	fldmiad	XO ,  { d15 }
 | 
						|
	add	XO, XO, INC_X
 | 
						|
 | 
						|
	vmla.f64	d2 , d12 , d8
 | 
						|
	vmla.f64	d3 , d12 , d4
 | 
						|
	vmla.f64	d2 , d13 , d9
 | 
						|
	vmla.f64	d3 , d13 , d5
 | 
						|
	vmla.f64	d2 , d14, d10
 | 
						|
	vmla.f64	d3 , d14, d6
 | 
						|
	vmla.f64	d2 , d15, d11
 | 
						|
	vmla.f64	d3 , d15, d7
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
.macro KERNEL_S2X1
 | 
						|
 | 
						|
	fldmiad	XO ,  { d1 }
 | 
						|
	fldmiad	AO1!,  { d8 }
 | 
						|
	fldmiad	AO2!,  { d4 }
 | 
						|
	vmla.f64	d2 , d1 , d8
 | 
						|
	add	XO, XO, INC_X
 | 
						|
	vmla.f64	d3 , d1 , d4
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
.macro	SAVE_S2
 | 
						|
 | 
						|
	fldmiad	YO,  { d4 }
 | 
						|
	vmla.f64	d4, d0, d2
 | 
						|
	fstmiad	YO, { d4  }
 | 
						|
	add	YO, YO, INC_Y
 | 
						|
 | 
						|
	fldmiad	YO,  { d5 }
 | 
						|
	vmla.f64	d5, d0, d3
 | 
						|
	fstmiad	YO, { d5  }
 | 
						|
	add	YO, YO, INC_Y
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
.macro INIT_S1
 | 
						|
 | 
						|
	vsub.f64	d2 , d2 , d2
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
.macro KERNEL_S1X4
 | 
						|
 | 
						|
	fldmiad	XO ,  { d12 }
 | 
						|
	add	XO, XO, INC_X
 | 
						|
 | 
						|
	pld	[ AO1 , #A_PRE ]
 | 
						|
	fldmiad	AO1!,  { d8 - d9   }
 | 
						|
 | 
						|
	fldmiad	XO ,  { d13 }
 | 
						|
	add	XO, XO, INC_X
 | 
						|
	fldmiad	AO1!,  { d10 - d11 }
 | 
						|
 | 
						|
	fldmiad	XO ,  { d14 }
 | 
						|
	add	XO, XO, INC_X
 | 
						|
 | 
						|
	fldmiad	XO ,  { d15 }
 | 
						|
	add	XO, XO, INC_X
 | 
						|
 | 
						|
	vmla.f64	d2 , d12 , d8
 | 
						|
	vmla.f64	d2 , d13 , d9
 | 
						|
	vmla.f64	d2 , d14, d10
 | 
						|
	vmla.f64	d2 , d15, d11
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
.macro KERNEL_S1X1
 | 
						|
 | 
						|
	fldmiad	XO ,  { d1 }
 | 
						|
	fldmiad	AO1!,  { d8 }
 | 
						|
	vmla.f64	d2 , d1 , d8
 | 
						|
	add	XO, XO, INC_X
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
.macro	SAVE_S1
 | 
						|
 | 
						|
	fldmiad	YO,  { d4 }
 | 
						|
	vmla.f64	d4, d0, d2
 | 
						|
	fstmiad	YO, { d4  }
 | 
						|
	add	YO, YO, INC_Y
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
 | 
						|
#else	/************************* SINGLE PRECISION *****************************************/
 | 
						|
 | 
						|
.macro INIT_F2
 | 
						|
 | 
						|
	vsub.f32	s2 , s2 , s2
 | 
						|
	vsub.f32	s3 , s3 , s3
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
.macro KERNEL_F2X4
 | 
						|
 | 
						|
	fldmias	XO! ,  { s12 - s15 }
 | 
						|
	fldmias	AO1!,  { s8 - s9   }
 | 
						|
	fldmias	AO2!,  { s4 - s5 }
 | 
						|
	fldmias	AO1!,  { s10 - s11 }
 | 
						|
	fldmias	AO2!,  { s6 - s7 }
 | 
						|
 | 
						|
	vmla.f32	s2 , s12 , s8
 | 
						|
	vmla.f32	s3 , s12 , s4
 | 
						|
	vmla.f32	s2 , s13 , s9
 | 
						|
	vmla.f32	s3 , s13 , s5
 | 
						|
	vmla.f32	s2 , s14, s10
 | 
						|
	vmla.f32	s3 , s14, s6
 | 
						|
	vmla.f32	s2 , s15, s11
 | 
						|
	vmla.f32	s3 , s15, s7
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
.macro KERNEL_F2X1
 | 
						|
 | 
						|
	fldmias	XO! ,  { s1 }
 | 
						|
	fldmias	AO1!,  { s8 }
 | 
						|
	fldmias	AO2!,  { s4 }
 | 
						|
	vmla.f32	s2 , s1 , s8
 | 
						|
	vmla.f32	s3 , s1 , s4
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
.macro	SAVE_F2
 | 
						|
 | 
						|
	fldmias	YO,  { s4 - s5 }
 | 
						|
	vmla.f32	s4, s0, s2
 | 
						|
	vmla.f32	s5, s0, s3
 | 
						|
	fstmias	YO!, { s4 - s5 }
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
.macro INIT_F1
 | 
						|
 | 
						|
	vsub.f32	s2 , s2 , s2
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
.macro KERNEL_F1X4
 | 
						|
 | 
						|
	fldmias	XO! ,  { s12 - s15 }
 | 
						|
	fldmias	AO1!,  { s8 - s9   }
 | 
						|
	fldmias	AO1!,  { s10 - s11 }
 | 
						|
	vmla.f32	s2 , s12 , s8
 | 
						|
	vmla.f32	s2 , s13 , s9
 | 
						|
	vmla.f32	s2 , s14, s10
 | 
						|
	vmla.f32	s2 , s15, s11
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
.macro KERNEL_F1X1
 | 
						|
 | 
						|
	fldmias	XO! ,  { s1 }
 | 
						|
	fldmias	AO1!,  { s8 }
 | 
						|
	vmla.f32	s2 , s1 , s8
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
.macro	SAVE_F1
 | 
						|
 | 
						|
	fldmias	YO,  { s4 }
 | 
						|
	vmla.f32	s4, s0, s2
 | 
						|
	fstmias	YO!, { s4 }
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
 | 
						|
.macro INIT_S2
 | 
						|
 | 
						|
	vsub.f32	s2 , s2 , s2
 | 
						|
	vsub.f32	s3 , s3 , s3
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
.macro KERNEL_S2X4
 | 
						|
 | 
						|
	fldmias	XO ,  { s12 }
 | 
						|
	add	XO, XO, INC_X
 | 
						|
 | 
						|
	fldmias	AO1!,  { s8 - s9   }
 | 
						|
	fldmias	AO2!,  { s4 - s5 }
 | 
						|
 | 
						|
	fldmias	XO ,  { s13 }
 | 
						|
	add	XO, XO, INC_X
 | 
						|
	fldmias	AO1!,  { s10 - s11 }
 | 
						|
	fldmias	AO2!,  { s6 - s7 }
 | 
						|
 | 
						|
	fldmias	XO ,  { s14 }
 | 
						|
	add	XO, XO, INC_X
 | 
						|
 | 
						|
	fldmias	XO ,  { s15 }
 | 
						|
	add	XO, XO, INC_X
 | 
						|
 | 
						|
	vmla.f32	s2 , s12 , s8
 | 
						|
	vmla.f32	s3 , s12 , s4
 | 
						|
	vmla.f32	s2 , s13 , s9
 | 
						|
	vmla.f32	s3 , s13 , s5
 | 
						|
	vmla.f32	s2 , s14, s10
 | 
						|
	vmla.f32	s3 , s14, s6
 | 
						|
	vmla.f32	s2 , s15, s11
 | 
						|
	vmla.f32	s3 , s15, s7
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
.macro KERNEL_S2X1
 | 
						|
 | 
						|
	fldmias	XO ,  { s1 }
 | 
						|
	fldmias	AO1!,  { s8 }
 | 
						|
	fldmias	AO2!,  { s4 }
 | 
						|
	vmla.f32	s2 , s1 , s8
 | 
						|
	add	XO, XO, INC_X
 | 
						|
	vmla.f32	s3 , s1 , s4
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
.macro	SAVE_S2
 | 
						|
 | 
						|
	fldmias	YO,  { s4 }
 | 
						|
	vmla.f32	s4, s0, s2
 | 
						|
	fstmias	YO, { s4  }
 | 
						|
	add	YO, YO, INC_Y
 | 
						|
 | 
						|
	fldmias	YO,  { s5 }
 | 
						|
	vmla.f32	s5, s0, s3
 | 
						|
	fstmias	YO, { s5  }
 | 
						|
	add	YO, YO, INC_Y
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
.macro INIT_S1
 | 
						|
 | 
						|
	vsub.f32	s2 , s2 , s2
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
.macro KERNEL_S1X4
 | 
						|
 | 
						|
	fldmias	XO ,  { s12 }
 | 
						|
	add	XO, XO, INC_X
 | 
						|
 | 
						|
	pld	[ AO1 , #A_PRE ]
 | 
						|
	fldmias	AO1!,  { s8 - s9   }
 | 
						|
 | 
						|
	fldmias	XO ,  { s13 }
 | 
						|
	add	XO, XO, INC_X
 | 
						|
	fldmias	AO1!,  { s10 - s11 }
 | 
						|
 | 
						|
	fldmias	XO ,  { s14 }
 | 
						|
	add	XO, XO, INC_X
 | 
						|
 | 
						|
	fldmias	XO ,  { s15 }
 | 
						|
	add	XO, XO, INC_X
 | 
						|
 | 
						|
	vmla.f32	s2 , s12 , s8
 | 
						|
	vmla.f32	s2 , s13 , s9
 | 
						|
	vmla.f32	s2 , s14, s10
 | 
						|
	vmla.f32	s2 , s15, s11
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
.macro KERNEL_S1X1
 | 
						|
 | 
						|
	fldmias	XO ,  { s1 }
 | 
						|
	fldmias	AO1!,  { s8 }
 | 
						|
	vmla.f32	s2 , s1 , s8
 | 
						|
	add	XO, XO, INC_X
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
.macro	SAVE_S1
 | 
						|
 | 
						|
	fldmias	YO,  { s4 }
 | 
						|
	vmla.f32	s4, s0, s2
 | 
						|
	fstmias	YO, { s4  }
 | 
						|
	add	YO, YO, INC_Y
 | 
						|
 | 
						|
.endm
 | 
						|
 | 
						|
 | 
						|
 | 
						|
#endif
 | 
						|
 | 
						|
/**************************************************************************************
 | 
						|
* End of macro definitions
 | 
						|
**************************************************************************************/
 | 
						|
 | 
						|
	PROLOGUE
 | 
						|
 | 
						|
	.align 5
 | 
						|
	push    {r4 - r9 , fp}
 | 
						|
        add     fp, sp, #28
 | 
						|
	sub     sp, sp, #STACKSIZE                              // reserve stack
 | 
						|
 | 
						|
        sub     r12, fp, #192
 | 
						|
 | 
						|
#if	defined(DOUBLE)
 | 
						|
        vstm    r12, { d8 - d15 }                                 // store floating point registers
 | 
						|
#else
 | 
						|
        vstm    r12, { s8 - s15 }                                 // store floating point registers
 | 
						|
#endif
 | 
						|
 | 
						|
	cmp	M, #0
 | 
						|
	ble	gemvt_kernel_L999
 | 
						|
 | 
						|
	cmp	OLD_N, #0
 | 
						|
	ble	gemvt_kernel_L999
 | 
						|
 | 
						|
	str	OLD_A, A
 | 
						|
	str	OLD_N, N
 | 
						|
 | 
						|
	ldr    INC_X , OLD_INC_X
 | 
						|
	ldr    INC_Y , OLD_INC_Y
 | 
						|
 | 
						|
	cmp	INC_X, #0
 | 
						|
	beq	gemvt_kernel_L999
 | 
						|
 | 
						|
	cmp	INC_Y, #0
 | 
						|
	beq	gemvt_kernel_L999
 | 
						|
 | 
						|
	ldr	LDA, OLD_LDA
 | 
						|
 | 
						|
 | 
						|
#if defined(DOUBLE)
 | 
						|
	lsl	LDA, LDA, #3				// LDA * SIZE
 | 
						|
#else
 | 
						|
	lsl	LDA, LDA, #2				// LDA * SIZE
 | 
						|
#endif
 | 
						|
 | 
						|
	cmp	INC_X, #1
 | 
						|
	bne	gemvt_kernel_S2_BEGIN
 | 
						|
 | 
						|
	cmp	INC_Y, #1
 | 
						|
	bne	gemvt_kernel_S2_BEGIN
 | 
						|
 | 
						|
 | 
						|
gemvt_kernel_F2_BEGIN:
 | 
						|
 | 
						|
	ldr	YO , Y
 | 
						|
 | 
						|
	ldr	J, N
 | 
						|
	asrs	J, J, #1					// J = N / 2
 | 
						|
	ble	gemvt_kernel_F1_BEGIN
 | 
						|
 | 
						|
gemvt_kernel_F2X4:
 | 
						|
 | 
						|
	ldr	AO1, A
 | 
						|
	add	AO2, AO1, LDA
 | 
						|
	add	r3 , AO2, LDA
 | 
						|
	str	r3 , A
 | 
						|
 | 
						|
	ldr	XO , X
 | 
						|
 | 
						|
	INIT_F2
 | 
						|
 | 
						|
	asrs	I, M, #2					// I = M / 4
 | 
						|
	ble	gemvt_kernel_F2X1
 | 
						|
 | 
						|
 | 
						|
gemvt_kernel_F2X4_10:
 | 
						|
 | 
						|
	KERNEL_F2X4
 | 
						|
 | 
						|
	subs	I, I, #1
 | 
						|
	bne	gemvt_kernel_F2X4_10
 | 
						|
 | 
						|
 | 
						|
gemvt_kernel_F2X1:
 | 
						|
 | 
						|
	ands	I, M , #3
 | 
						|
	ble	gemvt_kernel_F2_END
 | 
						|
 | 
						|
gemvt_kernel_F2X1_10:
 | 
						|
 | 
						|
	KERNEL_F2X1
 | 
						|
 | 
						|
	subs	I, I, #1
 | 
						|
	bne	gemvt_kernel_F2X1_10
 | 
						|
 | 
						|
 | 
						|
gemvt_kernel_F2_END:
 | 
						|
 | 
						|
	SAVE_F2
 | 
						|
 | 
						|
	subs	J , J , #1
 | 
						|
	bne	gemvt_kernel_F2X4
 | 
						|
 | 
						|
 | 
						|
gemvt_kernel_F1_BEGIN:
 | 
						|
 | 
						|
	ldr	J, N
 | 
						|
	ands	J, J, #1
 | 
						|
	ble	gemvt_kernel_L999
 | 
						|
 | 
						|
gemvt_kernel_F1X4:
 | 
						|
 | 
						|
	ldr	AO1, A
 | 
						|
 | 
						|
	ldr	XO , X
 | 
						|
 | 
						|
	INIT_F1
 | 
						|
 | 
						|
	asrs	I, M, #2					// I = M / 4
 | 
						|
	ble	gemvt_kernel_F1X1
 | 
						|
 | 
						|
 | 
						|
gemvt_kernel_F1X4_10:
 | 
						|
 | 
						|
	KERNEL_F1X4
 | 
						|
 | 
						|
	subs	I, I, #1
 | 
						|
	bne	gemvt_kernel_F1X4_10
 | 
						|
 | 
						|
 | 
						|
gemvt_kernel_F1X1:
 | 
						|
 | 
						|
	ands	I, M , #3
 | 
						|
	ble	gemvt_kernel_F1_END
 | 
						|
 | 
						|
gemvt_kernel_F1X1_10:
 | 
						|
 | 
						|
	KERNEL_F1X1
 | 
						|
 | 
						|
	subs	I, I, #1
 | 
						|
	bne	gemvt_kernel_F1X1_10
 | 
						|
 | 
						|
 | 
						|
gemvt_kernel_F1_END:
 | 
						|
 | 
						|
	SAVE_F1
 | 
						|
 | 
						|
	b	gemvt_kernel_L999
 | 
						|
 | 
						|
 | 
						|
 | 
						|
/*************************************************************************************************************/
 | 
						|
 | 
						|
gemvt_kernel_S2_BEGIN:
 | 
						|
 | 
						|
#if defined(DOUBLE)
 | 
						|
	lsl	INC_X, INC_X, #3				// INC_X * SIZE
 | 
						|
	lsl	INC_Y, INC_Y, #3				// INC_Y * SIZE
 | 
						|
#else
 | 
						|
	lsl	INC_X, INC_X, #2				// INC_X * SIZE
 | 
						|
	lsl	INC_Y, INC_Y, #2				// INC_Y * SIZE
 | 
						|
#endif
 | 
						|
 | 
						|
	ldr	YO , Y
 | 
						|
 | 
						|
	ldr	J, N
 | 
						|
	asrs	J, J, #1					// J = N / 2
 | 
						|
	ble	gemvt_kernel_S1_BEGIN
 | 
						|
 | 
						|
gemvt_kernel_S2X4:
 | 
						|
 | 
						|
	ldr	AO1, A
 | 
						|
	add	AO2, AO1, LDA
 | 
						|
	add	r3 , AO2, LDA
 | 
						|
	str	r3 , A
 | 
						|
 | 
						|
	ldr	XO , X
 | 
						|
 | 
						|
	INIT_S2
 | 
						|
 | 
						|
	asrs	I, M, #2					// I = M / 4
 | 
						|
	ble	gemvt_kernel_S2X1
 | 
						|
 | 
						|
 | 
						|
gemvt_kernel_S2X4_10:
 | 
						|
 | 
						|
	KERNEL_S2X4
 | 
						|
 | 
						|
	subs	I, I, #1
 | 
						|
	bne	gemvt_kernel_S2X4_10
 | 
						|
 | 
						|
 | 
						|
gemvt_kernel_S2X1:
 | 
						|
 | 
						|
	ands	I, M , #3
 | 
						|
	ble	gemvt_kernel_S2_END
 | 
						|
 | 
						|
gemvt_kernel_S2X1_10:
 | 
						|
 | 
						|
	KERNEL_S2X1
 | 
						|
 | 
						|
	subs	I, I, #1
 | 
						|
	bne	gemvt_kernel_S2X1_10
 | 
						|
 | 
						|
 | 
						|
gemvt_kernel_S2_END:
 | 
						|
 | 
						|
	SAVE_S2
 | 
						|
 | 
						|
	subs	J , J , #1
 | 
						|
	bne	gemvt_kernel_S2X4
 | 
						|
 | 
						|
 | 
						|
gemvt_kernel_S1_BEGIN:
 | 
						|
 | 
						|
	ldr	J, N
 | 
						|
	ands	J, J, #1
 | 
						|
	ble	gemvt_kernel_L999
 | 
						|
 | 
						|
gemvt_kernel_S1X4:
 | 
						|
 | 
						|
	ldr	AO1, A
 | 
						|
 | 
						|
	ldr	XO , X
 | 
						|
 | 
						|
	INIT_S1
 | 
						|
 | 
						|
	asrs	I, M, #2					// I = M / 4
 | 
						|
	ble	gemvt_kernel_S1X1
 | 
						|
 | 
						|
 | 
						|
gemvt_kernel_S1X4_10:
 | 
						|
 | 
						|
	KERNEL_S1X4
 | 
						|
 | 
						|
	subs	I, I, #1
 | 
						|
	bne	gemvt_kernel_S1X4_10
 | 
						|
 | 
						|
 | 
						|
gemvt_kernel_S1X1:
 | 
						|
 | 
						|
	ands	I, M , #3
 | 
						|
	ble	gemvt_kernel_S1_END
 | 
						|
 | 
						|
gemvt_kernel_S1X1_10:
 | 
						|
 | 
						|
	KERNEL_S1X1
 | 
						|
 | 
						|
	subs	I, I, #1
 | 
						|
	bne	gemvt_kernel_S1X1_10
 | 
						|
 | 
						|
 | 
						|
gemvt_kernel_S1_END:
 | 
						|
 | 
						|
	SAVE_S1
 | 
						|
 | 
						|
 | 
						|
 | 
						|
/*************************************************************************************************************/
 | 
						|
 | 
						|
gemvt_kernel_L999:
 | 
						|
 | 
						|
        sub     r3, fp, #192
 | 
						|
 | 
						|
#if	defined(DOUBLE)
 | 
						|
        vldm    r3, { d8 - d15 }                                 // restore floating point registers
 | 
						|
#else
 | 
						|
        vldm    r3, { s8 - s15 }                                 // restore floating point registers
 | 
						|
#endif
 | 
						|
 | 
						|
	mov	r0, #0		// set return value
 | 
						|
 | 
						|
	sub     sp, fp, #28
 | 
						|
	pop     {r4 -r9 ,fp}
 | 
						|
	bx	lr
 | 
						|
 | 
						|
	EPILOGUE
 | 
						|
 |