519 lines
		
	
	
		
			10 KiB
		
	
	
	
		
			ArmAsm
		
	
	
	
			
		
		
	
	
			519 lines
		
	
	
		
			10 KiB
		
	
	
	
		
			ArmAsm
		
	
	
	
/*********************************************************************/
 | 
						|
/* Copyright 2009, 2010 The University of Texas at Austin.           */
 | 
						|
/* All rights reserved.                                              */
 | 
						|
/*                                                                   */
 | 
						|
/* Redistribution and use in source and binary forms, with or        */
 | 
						|
/* without modification, are permitted provided that the following   */
 | 
						|
/* conditions are met:                                               */
 | 
						|
/*                                                                   */
 | 
						|
/*   1. Redistributions of source code must retain the above         */
 | 
						|
/*      copyright notice, this list of conditions and the following  */
 | 
						|
/*      disclaimer.                                                  */
 | 
						|
/*                                                                   */
 | 
						|
/*   2. Redistributions in binary form must reproduce the above      */
 | 
						|
/*      copyright notice, this list of conditions and the following  */
 | 
						|
/*      disclaimer in the documentation and/or other materials       */
 | 
						|
/*      provided with the distribution.                              */
 | 
						|
/*                                                                   */
 | 
						|
/*    THIS  SOFTWARE IS PROVIDED  BY THE  UNIVERSITY OF  TEXAS AT    */
 | 
						|
/*    AUSTIN  ``AS IS''  AND ANY  EXPRESS OR  IMPLIED WARRANTIES,    */
 | 
						|
/*    INCLUDING, BUT  NOT LIMITED  TO, THE IMPLIED  WARRANTIES OF    */
 | 
						|
/*    MERCHANTABILITY  AND FITNESS FOR  A PARTICULAR  PURPOSE ARE    */
 | 
						|
/*    DISCLAIMED.  IN  NO EVENT SHALL THE UNIVERSITY  OF TEXAS AT    */
 | 
						|
/*    AUSTIN OR CONTRIBUTORS BE  LIABLE FOR ANY DIRECT, INDIRECT,    */
 | 
						|
/*    INCIDENTAL,  SPECIAL, EXEMPLARY,  OR  CONSEQUENTIAL DAMAGES    */
 | 
						|
/*    (INCLUDING, BUT  NOT LIMITED TO,  PROCUREMENT OF SUBSTITUTE    */
 | 
						|
/*    GOODS  OR  SERVICES; LOSS  OF  USE,  DATA,  OR PROFITS;  OR    */
 | 
						|
/*    BUSINESS INTERRUPTION) HOWEVER CAUSED  AND ON ANY THEORY OF    */
 | 
						|
/*    LIABILITY, WHETHER  IN CONTRACT, STRICT  LIABILITY, OR TORT    */
 | 
						|
/*    (INCLUDING NEGLIGENCE OR OTHERWISE)  ARISING IN ANY WAY OUT    */
 | 
						|
/*    OF  THE  USE OF  THIS  SOFTWARE,  EVEN  IF ADVISED  OF  THE    */
 | 
						|
/*    POSSIBILITY OF SUCH DAMAGE.                                    */
 | 
						|
/*                                                                   */
 | 
						|
/* The views and conclusions contained in the software and           */
 | 
						|
/* documentation are those of the authors and should not be          */
 | 
						|
/* interpreted as representing official policies, either expressed   */
 | 
						|
/* or implied, of The University of Texas at Austin.                 */
 | 
						|
/*********************************************************************/
 | 
						|
 | 
						|
#define ASSEMBLER
 | 
						|
#include "common.h"
 | 
						|
 | 
						|
#define N	%i0
 | 
						|
 | 
						|
#if defined(DOUBLE) && !defined(__64BIT__)
 | 
						|
#define X	%i3
 | 
						|
#define INCX	%i4
 | 
						|
#else
 | 
						|
#define X	%i5
 | 
						|
#define INCX	%i3
 | 
						|
#endif
 | 
						|
 | 
						|
#define I	%i1
 | 
						|
#define XX	%i2
 | 
						|
 | 
						|
#ifdef DOUBLE
 | 
						|
#define c1	%f0
 | 
						|
#define c2	%f2
 | 
						|
#define c3	%f4
 | 
						|
#define c4	%f6
 | 
						|
#define c5	%f8
 | 
						|
#define c6	%f10
 | 
						|
#define c7	%f12
 | 
						|
#define c8	%f14
 | 
						|
 | 
						|
#define t1	%f16
 | 
						|
#define t2	%f18
 | 
						|
#define t3	%f20
 | 
						|
#define t4	%f22
 | 
						|
#define t5	%f24
 | 
						|
#define t6	%f26
 | 
						|
#define t7	%f28
 | 
						|
#define t8	%f30
 | 
						|
 | 
						|
#define c9	%f32
 | 
						|
#define c10	%f34
 | 
						|
#define c11	%f36
 | 
						|
#define c12	%f38
 | 
						|
#define c13	%f40
 | 
						|
#define c14	%f42
 | 
						|
#define c15	%f44
 | 
						|
#define c16	%f46
 | 
						|
 | 
						|
#define s1	%f32
 | 
						|
#define s2	%f34
 | 
						|
#define s3	%f36
 | 
						|
#define s4	%f38
 | 
						|
#define s5	%f40
 | 
						|
#define s6	%f42
 | 
						|
#define s7	%f44
 | 
						|
#define s8	%f46
 | 
						|
 | 
						|
#define FZERO	%f48
 | 
						|
#define ALPHA_R	%f50
 | 
						|
#define ALPHA_I	%f52
 | 
						|
#else
 | 
						|
#define c1	%f0
 | 
						|
#define c2	%f1
 | 
						|
#define c3	%f2
 | 
						|
#define c4	%f3
 | 
						|
#define c5	%f4
 | 
						|
#define c6	%f5
 | 
						|
#define c7	%f6
 | 
						|
#define c8	%f7
 | 
						|
 | 
						|
#define c9	%f8
 | 
						|
#define c10	%f9
 | 
						|
#define c11	%f10
 | 
						|
#define c12	%f11
 | 
						|
#define c13	%f12
 | 
						|
#define c14	%f13
 | 
						|
#define c15	%f14
 | 
						|
#define c16	%f15
 | 
						|
 | 
						|
#define s1	%f8
 | 
						|
#define s2	%f9
 | 
						|
#define s3	%f10
 | 
						|
#define s4	%f11
 | 
						|
#define s5	%f12
 | 
						|
#define s6	%f13
 | 
						|
#define s7	%f14
 | 
						|
#define s8	%f15
 | 
						|
 | 
						|
#define t1	%f16
 | 
						|
#define t2	%f17
 | 
						|
#define t3	%f18
 | 
						|
#define t4	%f19
 | 
						|
#define t5	%f20
 | 
						|
#define t6	%f21
 | 
						|
#define t7	%f22
 | 
						|
#define t8	%f23
 | 
						|
 | 
						|
#define FZERO	%f24
 | 
						|
#define ALPHA_R	%f25
 | 
						|
#define ALPHA_I	%f26
 | 
						|
#endif
 | 
						|
 | 
						|
#define PREFETCHSIZE 128
 | 
						|
 | 
						|
	PROLOGUE
 | 
						|
	SAVESP
 | 
						|
 | 
						|
#ifndef __64BIT__
 | 
						|
#ifdef DOUBLE
 | 
						|
	st	%i3, [%sp + STACK_START + 16]
 | 
						|
	st	%i4, [%sp + STACK_START + 20]
 | 
						|
	st	%i5, [%sp + STACK_START + 24]
 | 
						|
 | 
						|
	ld	[%sp+ STACK_START + 32], X
 | 
						|
	ld	[%sp+ STACK_START + 36], INCX
 | 
						|
#else
 | 
						|
	st	%i3, [%sp + STACK_START + 16]
 | 
						|
	st	%i4, [%sp + STACK_START + 24]
 | 
						|
	ld	[%sp+  STACK_START + 28], INCX
 | 
						|
#endif
 | 
						|
	LDF	[%sp + STACK_START + 16], ALPHA_R
 | 
						|
	LDF	[%sp + STACK_START + 24], ALPHA_I
 | 
						|
#else
 | 
						|
	ldx	[%sp + STACK_START + 56], INCX
 | 
						|
#ifdef DOUBLE
 | 
						|
	FMOV	%f6, ALPHA_R
 | 
						|
	FMOV	%f8, ALPHA_I
 | 
						|
#else
 | 
						|
	FMOV	%f7, ALPHA_R
 | 
						|
	FMOV	%f9, ALPHA_I
 | 
						|
#endif
 | 
						|
#endif
 | 
						|
 | 
						|
#ifdef DOUBLE
 | 
						|
	FCLR(17)
 | 
						|
#else
 | 
						|
	FCLR(24)
 | 
						|
#endif
 | 
						|
 | 
						|
	FCMP	ALPHA_R, FZERO
 | 
						|
	fbne	.LL100
 | 
						|
	sll	INCX, ZBASE_SHIFT, INCX
 | 
						|
 | 
						|
	FCMP	ALPHA_I, FZERO
 | 
						|
	fbne	.LL100
 | 
						|
	nop
 | 
						|
	cmp	INCX, 2 * SIZE
 | 
						|
	bne	.LL50
 | 
						|
	nop
 | 
						|
	sra	N, 2, I
 | 
						|
	cmp	I, 0
 | 
						|
	ble,pn	%icc, .LL15
 | 
						|
	nop
 | 
						|
 | 
						|
.LL11:
 | 
						|
	prefetch [X  + PREFETCHSIZE * SIZE], 0
 | 
						|
 | 
						|
	STF	FZERO, [X +  0 * SIZE]
 | 
						|
	add	I, -1, I
 | 
						|
	STF	FZERO, [X +  1 * SIZE]
 | 
						|
	cmp	I, 0
 | 
						|
	STF	FZERO, [X +  2 * SIZE]
 | 
						|
	STF	FZERO, [X +  3 * SIZE]
 | 
						|
	STF	FZERO, [X +  4 * SIZE]
 | 
						|
	STF	FZERO, [X +  5 * SIZE]
 | 
						|
	add	X, 8 * SIZE, X
 | 
						|
	STF	FZERO, [X -  2 * SIZE]
 | 
						|
	bg,pt	%icc, .LL11
 | 
						|
	STF	FZERO, [X -  1 * SIZE]
 | 
						|
 | 
						|
.LL15:
 | 
						|
	and	N, 3, I
 | 
						|
	cmp	I,  0
 | 
						|
	ble,a,pn %icc, .LL19
 | 
						|
	nop
 | 
						|
 | 
						|
.LL16:
 | 
						|
	STF	FZERO, [X +  0 * SIZE]
 | 
						|
	STF	FZERO, [X +  1 * SIZE]
 | 
						|
	add	I, -1, I
 | 
						|
	cmp	I, 0
 | 
						|
	bg,pt	%icc, .LL16
 | 
						|
	add	X, 2 * SIZE, X
 | 
						|
 | 
						|
.LL19:
 | 
						|
	return	%i7 + 8
 | 
						|
	clr	%o0
 | 
						|
 | 
						|
.LL50:
 | 
						|
	sra	N, 2, I
 | 
						|
	cmp	I, 0
 | 
						|
	ble,pn	%icc, .LL55
 | 
						|
	nop
 | 
						|
 | 
						|
.LL51:
 | 
						|
	STF	FZERO, [X +  0 * SIZE]
 | 
						|
	add	I, -1, I
 | 
						|
	STF	FZERO, [X +  1 * SIZE]
 | 
						|
	add	X, INCX, X
 | 
						|
	STF	FZERO, [X +  0 * SIZE]
 | 
						|
	cmp	I, 0
 | 
						|
	STF	FZERO, [X +  1 * SIZE]
 | 
						|
	add	X, INCX, X
 | 
						|
	STF	FZERO, [X +  0 * SIZE]
 | 
						|
	STF	FZERO, [X +  1 * SIZE]
 | 
						|
	add	X, INCX, X
 | 
						|
	STF	FZERO, [X +  0 * SIZE]
 | 
						|
	STF	FZERO, [X +  1 * SIZE]
 | 
						|
	bg,pt	%icc, .LL51
 | 
						|
	add	X, INCX, X
 | 
						|
 | 
						|
.LL55:
 | 
						|
	and	N, 3, I
 | 
						|
	cmp	I,  0
 | 
						|
	ble,a,pn %icc, .LL59
 | 
						|
	nop
 | 
						|
 | 
						|
.LL56:
 | 
						|
	STF	FZERO, [X +  0 * SIZE]
 | 
						|
	add	I, -1, I
 | 
						|
	STF	FZERO, [X +  1 * SIZE]
 | 
						|
	cmp	I, 0
 | 
						|
	bg,pt	%icc, .LL56
 | 
						|
	add	X, INCX, X
 | 
						|
 | 
						|
.LL59:
 | 
						|
	return	%i7 + 8
 | 
						|
	clr	%o0
 | 
						|
 | 
						|
.LL100:
 | 
						|
	cmp	INCX, 2 * SIZE
 | 
						|
	bne	.LL150
 | 
						|
	sra	N, 2, I
 | 
						|
 | 
						|
	cmp	I, 0
 | 
						|
	ble,pn	%icc, .LL115
 | 
						|
	nop
 | 
						|
 | 
						|
	LDF	[X +  0 * SIZE], c1
 | 
						|
	LDF	[X +  1 * SIZE], c2
 | 
						|
	LDF	[X +  2 * SIZE], c3
 | 
						|
	LDF	[X +  3 * SIZE], c4
 | 
						|
	LDF	[X +  4 * SIZE], c5
 | 
						|
	LDF	[X +  5 * SIZE], c6
 | 
						|
	LDF	[X +  6 * SIZE], c7
 | 
						|
	LDF	[X +  7 * SIZE], c8
 | 
						|
 | 
						|
	FMUL	ALPHA_R, c1, t1
 | 
						|
	FMUL	ALPHA_I, c2, t3
 | 
						|
 | 
						|
	FMUL	ALPHA_I, c1, t2
 | 
						|
	LDF	[X +  8 * SIZE], c1
 | 
						|
	FMUL	ALPHA_R, c2, t4
 | 
						|
	LDF	[X +  9 * SIZE], c2
 | 
						|
 | 
						|
	FMUL	ALPHA_R, c3, t5
 | 
						|
	deccc	I
 | 
						|
	FMUL	ALPHA_I, c4, t7
 | 
						|
	FSUB	t1,  t3,  s1
 | 
						|
 | 
						|
	FMUL	ALPHA_I, c3, t6
 | 
						|
	LDF	[X + 10 * SIZE], c3
 | 
						|
	FMUL	ALPHA_R, c4, t8
 | 
						|
	LDF	[X + 11 * SIZE], c4
 | 
						|
	FADD	t4,  t2,  s2
 | 
						|
 | 
						|
	ble,pn	%icc, .LL112
 | 
						|
	nop
 | 
						|
 | 
						|
.LL111:
 | 
						|
	prefetch [X  + PREFETCHSIZE * SIZE], 0
 | 
						|
 | 
						|
	FMUL	ALPHA_R, c5, t1
 | 
						|
	FMUL	ALPHA_I, c6, t3
 | 
						|
	FSUB	t5,  t7,  s3
 | 
						|
	STF	s1, [X +  0 * SIZE]
 | 
						|
 | 
						|
	FMUL	ALPHA_I, c5, t2
 | 
						|
	LDF	[X + 12 * SIZE], c5
 | 
						|
	FMUL	ALPHA_R, c6, t4
 | 
						|
	LDF	[X + 13 * SIZE], c6
 | 
						|
 | 
						|
	FADD	t8,  t6,  s4
 | 
						|
	STF	s2, [X +  1 * SIZE]
 | 
						|
 | 
						|
	FMUL	ALPHA_R, c7, t5
 | 
						|
	FMUL	ALPHA_I, c8, t7
 | 
						|
	FSUB	t1,  t3,  s5
 | 
						|
	STF	s3, [X +  2 * SIZE]
 | 
						|
 | 
						|
	FMUL	ALPHA_I, c7, t6
 | 
						|
	LDF	[X + 14 * SIZE], c7
 | 
						|
	FMUL	ALPHA_R, c8, t8
 | 
						|
	LDF	[X + 15 * SIZE], c8
 | 
						|
 | 
						|
	FADD	t4,  t2,  s6
 | 
						|
	STF	s4, [X +  3 * SIZE]
 | 
						|
 | 
						|
	FMUL	ALPHA_R, c1, t1
 | 
						|
	FMUL	ALPHA_I, c2, t3
 | 
						|
	FSUB	t5,  t7,  s7
 | 
						|
	STF	s5, [X +  4 * SIZE]
 | 
						|
 | 
						|
	FMUL	ALPHA_I, c1, t2
 | 
						|
	LDF	[X + 16 * SIZE], c1
 | 
						|
	FMUL	ALPHA_R, c2, t4
 | 
						|
	LDF	[X + 17 * SIZE], c2
 | 
						|
 | 
						|
	FADD	t8,  t6,  s8
 | 
						|
	STF	s6, [X +  5 * SIZE]
 | 
						|
 | 
						|
	FMUL	ALPHA_R, c3, t5
 | 
						|
	deccc	I
 | 
						|
	FMUL	ALPHA_I, c4, t7
 | 
						|
	FSUB	t1,  t3,  s1
 | 
						|
	STF	s7, [X +  6 * SIZE]
 | 
						|
 | 
						|
	FMUL	ALPHA_I, c3, t6
 | 
						|
	LDF	[X + 18 * SIZE], c3
 | 
						|
	FMUL	ALPHA_R, c4, t8
 | 
						|
	LDF	[X + 19 * SIZE], c4
 | 
						|
 | 
						|
	FADD	t4,  t2,  s2
 | 
						|
	STF	s8, [X +  7 * SIZE]
 | 
						|
 | 
						|
	bg,pt	%icc, .LL111
 | 
						|
	add	X, 8 * SIZE, X
 | 
						|
 | 
						|
 | 
						|
.LL112:
 | 
						|
	FMUL	ALPHA_R, c5, t1
 | 
						|
	FMUL	ALPHA_I, c6, t3
 | 
						|
	FSUB	t5,  t7,  s3
 | 
						|
	STF	s1, [X +  0 * SIZE]
 | 
						|
 | 
						|
	FMUL	ALPHA_I, c5, t2
 | 
						|
	FMUL	ALPHA_R, c6, t4
 | 
						|
	FADD	t8,  t6,  s4
 | 
						|
	STF	s2, [X +  1 * SIZE]
 | 
						|
 | 
						|
	FMUL	ALPHA_R, c7, t5
 | 
						|
	FMUL	ALPHA_I, c8, t7
 | 
						|
	FSUB	t1,  t3,  s5
 | 
						|
	STF	s3, [X +  2 * SIZE]
 | 
						|
 | 
						|
	FMUL	ALPHA_I, c7, t6
 | 
						|
	FMUL	ALPHA_R, c8, t8
 | 
						|
	FADD	t4,  t2,  s6
 | 
						|
	STF	s4, [X +  3 * SIZE]
 | 
						|
 | 
						|
	FSUB	t5,  t7,  s7
 | 
						|
	FADD	t8,  t6,  s8
 | 
						|
 | 
						|
	STF	s5, [X +  4 * SIZE]
 | 
						|
	STF	s6, [X +  5 * SIZE]
 | 
						|
	STF	s7, [X +  6 * SIZE]
 | 
						|
	STF	s8, [X +  7 * SIZE]
 | 
						|
	add	X, 8 * SIZE, X
 | 
						|
 | 
						|
.LL115:
 | 
						|
	and	N, 3, I
 | 
						|
	cmp	I,  0
 | 
						|
	ble,a,pn %icc, .LL119
 | 
						|
	nop
 | 
						|
 | 
						|
.LL116:
 | 
						|
	LDF	[X +  0 * SIZE], c1
 | 
						|
	LDF	[X +  1 * SIZE], c2
 | 
						|
 | 
						|
	FMUL	ALPHA_R, c1, c3
 | 
						|
	FMUL	ALPHA_I, c1, c4
 | 
						|
	FMUL	ALPHA_I, c2, c1
 | 
						|
	FMUL	ALPHA_R, c2, c2
 | 
						|
 | 
						|
	FSUB	c3, c1, c1
 | 
						|
	FADD	c2, c4, c2
 | 
						|
 | 
						|
	STF	c1, [X +  0 * SIZE]
 | 
						|
	STF	c2, [X +  1 * SIZE]
 | 
						|
 | 
						|
	add	I, -1, I
 | 
						|
	cmp	I, 0
 | 
						|
	bg,pt	%icc, .LL116
 | 
						|
	add	X, 2 * SIZE, X
 | 
						|
 | 
						|
.LL119:
 | 
						|
	return	%i7 + 8
 | 
						|
	clr	%o0
 | 
						|
 | 
						|
.LL150:
 | 
						|
	sra	N, 2, I
 | 
						|
	cmp	I, 0
 | 
						|
	ble,pn	%icc, .LL155
 | 
						|
	mov	X, XX
 | 
						|
 | 
						|
.LL151:
 | 
						|
	LDF	[X +  0 * SIZE], c1
 | 
						|
	LDF	[X +  1 * SIZE], c2
 | 
						|
	add	X, INCX, X
 | 
						|
	LDF	[X +  0 * SIZE], c3
 | 
						|
	FMUL	ALPHA_R, c1, c9
 | 
						|
	LDF	[X +  1 * SIZE], c4
 | 
						|
	FMUL	ALPHA_I, c1, c10
 | 
						|
	add	X, INCX, X
 | 
						|
	LDF	[X +  0 * SIZE], c5
 | 
						|
	FMUL	ALPHA_I, c2, c1
 | 
						|
	LDF	[X +  1 * SIZE], c6
 | 
						|
	FMUL	ALPHA_R, c2, c2
 | 
						|
	add	X, INCX, X
 | 
						|
	LDF	[X +  0 * SIZE], c7
 | 
						|
	FMUL	ALPHA_R, c3, c11
 | 
						|
	LDF	[X +  1 * SIZE], c8
 | 
						|
	FMUL	ALPHA_I, c3, c12
 | 
						|
	add	X, INCX, X
 | 
						|
 | 
						|
	FMUL	ALPHA_I, c4, c3
 | 
						|
	FMUL	ALPHA_R, c4, c4
 | 
						|
 | 
						|
	FMUL	ALPHA_R, c5, c13
 | 
						|
	FMUL	ALPHA_I, c5, c14
 | 
						|
	FMUL	ALPHA_I, c6, c5
 | 
						|
	FMUL	ALPHA_R, c6, c6
 | 
						|
 | 
						|
	FMUL	ALPHA_R, c7, c15
 | 
						|
	FSUB	c9,  c1,  c1
 | 
						|
	FMUL	ALPHA_I, c7, c16
 | 
						|
	FADD	c2,  c10, c2
 | 
						|
	FMUL	ALPHA_I, c8, c7
 | 
						|
	FSUB	c11, c3,  c3
 | 
						|
	FMUL	ALPHA_R, c8, c8
 | 
						|
	FADD	c4,  c12, c4
 | 
						|
 | 
						|
	STF	c1, [XX +  0 * SIZE]
 | 
						|
	FSUB	c13, c5,  c5
 | 
						|
	add	I, -1, I
 | 
						|
	STF	c2, [XX +  1 * SIZE]
 | 
						|
	FADD	c6,  c14, c6
 | 
						|
	add	XX, INCX, XX
 | 
						|
	STF	c3, [XX +  0 * SIZE]
 | 
						|
	FSUB	c15, c7,  c7
 | 
						|
	cmp	I, 0
 | 
						|
	STF	c4, [XX +  1 * SIZE]
 | 
						|
	FADD	c8,  c16, c8
 | 
						|
	add	XX, INCX, XX
 | 
						|
	STF	c5, [XX +  0 * SIZE]
 | 
						|
	STF	c6, [XX +  1 * SIZE]
 | 
						|
	add	XX, INCX, XX
 | 
						|
	STF	c7, [XX +  0 * SIZE]
 | 
						|
	STF	c8, [XX +  1 * SIZE]
 | 
						|
	bg,pt	%icc, .LL151
 | 
						|
	add	XX, INCX, XX
 | 
						|
 | 
						|
.LL155:
 | 
						|
	and	N, 3, I
 | 
						|
	cmp	I,  0
 | 
						|
	ble,a,pn %icc, .LL159
 | 
						|
	nop
 | 
						|
 | 
						|
.LL156:
 | 
						|
	LDF	[X +  0 * SIZE], c1
 | 
						|
	LDF	[X +  1 * SIZE], c2
 | 
						|
 | 
						|
	FMUL	ALPHA_R, c1, c3
 | 
						|
	FMUL	ALPHA_I, c1, c4
 | 
						|
	FMUL	ALPHA_I, c2, c1
 | 
						|
	FMUL	ALPHA_R, c2, c2
 | 
						|
 | 
						|
	FSUB	c3, c1, c1
 | 
						|
	FADD	c2, c4, c2
 | 
						|
 | 
						|
	STF	c1, [X +  0 * SIZE]
 | 
						|
	STF	c2, [X +  1 * SIZE]
 | 
						|
 | 
						|
	add	I, -1, I
 | 
						|
	cmp	I, 0
 | 
						|
	bg,pt	%icc, .LL156
 | 
						|
	add	X, INCX, X
 | 
						|
 | 
						|
.LL159:
 | 
						|
	return	%i7 + 8
 | 
						|
	clr	%o0
 | 
						|
 | 
						|
 | 
						|
	EPILOGUE
 |