401 lines
7.9 KiB
ArmAsm
401 lines
7.9 KiB
ArmAsm
/*********************************************************************/
|
|
/* Copyright 2009, 2010 The University of Texas at Austin. */
|
|
/* All rights reserved. */
|
|
/* */
|
|
/* Redistribution and use in source and binary forms, with or */
|
|
/* without modification, are permitted provided that the following */
|
|
/* conditions are met: */
|
|
/* */
|
|
/* 1. Redistributions of source code must retain the above */
|
|
/* copyright notice, this list of conditions and the following */
|
|
/* disclaimer. */
|
|
/* */
|
|
/* 2. Redistributions in binary form must reproduce the above */
|
|
/* copyright notice, this list of conditions and the following */
|
|
/* disclaimer in the documentation and/or other materials */
|
|
/* provided with the distribution. */
|
|
/* */
|
|
/* THIS SOFTWARE IS PROVIDED BY THE UNIVERSITY OF TEXAS AT */
|
|
/* AUSTIN ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, */
|
|
/* INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF */
|
|
/* MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE */
|
|
/* DISCLAIMED. IN NO EVENT SHALL THE UNIVERSITY OF TEXAS AT */
|
|
/* AUSTIN OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, */
|
|
/* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES */
|
|
/* (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE */
|
|
/* GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR */
|
|
/* BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF */
|
|
/* LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT */
|
|
/* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT */
|
|
/* OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE */
|
|
/* POSSIBILITY OF SUCH DAMAGE. */
|
|
/* */
|
|
/* The views and conclusions contained in the software and */
|
|
/* documentation are those of the authors and should not be */
|
|
/* interpreted as representing official policies, either expressed */
|
|
/* or implied, of The University of Texas at Austin. */
|
|
/*********************************************************************/
|
|
|
|
#define ASSEMBLER
|
|
#include "common.h"
|
|
|
|
#define N %i0
|
|
#if defined(DOUBLE) && !defined(__64BIT__)
|
|
#define X %i5
|
|
#define INCX %i1
|
|
#else
|
|
#define X %i4
|
|
#define INCX %i5
|
|
#endif
|
|
|
|
#define I %i2
|
|
#define XX %i3
|
|
|
|
#ifdef DOUBLE
|
|
#define c1 %f0
|
|
#define c2 %f2
|
|
#define c3 %f4
|
|
#define c4 %f6
|
|
#define c5 %f8
|
|
#define c6 %f10
|
|
#define c7 %f12
|
|
#define c8 %f14
|
|
|
|
#define t1 %f16
|
|
#define t2 %f18
|
|
#define t3 %f20
|
|
#define t4 %f22
|
|
#define t5 %f24
|
|
#define t6 %f26
|
|
#define t7 %f28
|
|
#define t8 %f30
|
|
|
|
#define FZERO %f60
|
|
#define ALPHA %f62
|
|
#else
|
|
#define c1 %f0
|
|
#define c2 %f1
|
|
#define c3 %f2
|
|
#define c4 %f3
|
|
#define c5 %f4
|
|
#define c6 %f5
|
|
#define c7 %f6
|
|
#define c8 %f7
|
|
|
|
#define t1 %f8
|
|
#define t2 %f9
|
|
#define t3 %f10
|
|
#define t4 %f11
|
|
#define t5 %f12
|
|
#define t6 %f13
|
|
#define t7 %f14
|
|
#define t8 %f15
|
|
|
|
#define FZERO %f29
|
|
#define ALPHA %f30
|
|
#endif
|
|
|
|
#define PREFETCHSIZE 168
|
|
|
|
PROLOGUE
|
|
SAVESP
|
|
|
|
#ifndef __64BIT__
|
|
#ifdef DOUBLE
|
|
st %i3, [%sp + STACK_START + 16]
|
|
st %i4, [%sp + STACK_START + 20]
|
|
ld [%sp + STACK_START + 28], INCX
|
|
#else
|
|
st %i3, [%sp + STACK_START + 16]
|
|
#endif
|
|
|
|
LDF [%sp + STACK_START + 16], ALPHA
|
|
#else
|
|
#ifdef DOUBLE
|
|
FMOV %f6, ALPHA
|
|
#else
|
|
FMOV %f7, ALPHA
|
|
#endif
|
|
#endif
|
|
|
|
FCLR(29)
|
|
|
|
// FCMP ALPHA, FZERO
|
|
// fbne .LL100
|
|
b .LL100
|
|
|
|
sll INCX, BASE_SHIFT, INCX
|
|
|
|
cmp INCX, SIZE
|
|
bne .LL50
|
|
nop
|
|
|
|
sra N, 3, I
|
|
cmp I, 0
|
|
ble,pn %icc, .LL15
|
|
nop
|
|
|
|
.LL11:
|
|
prefetch [X + PREFETCHSIZE * SIZE], 0
|
|
|
|
STF FZERO, [X + 0 * SIZE]
|
|
add I, -1, I
|
|
STF FZERO, [X + 1 * SIZE]
|
|
cmp I, 0
|
|
STF FZERO, [X + 2 * SIZE]
|
|
STF FZERO, [X + 3 * SIZE]
|
|
STF FZERO, [X + 4 * SIZE]
|
|
STF FZERO, [X + 5 * SIZE]
|
|
add X, 8 * SIZE, X
|
|
STF FZERO, [X - 2 * SIZE]
|
|
bg,pt %icc, .LL11
|
|
STF FZERO, [X - 1 * SIZE]
|
|
|
|
.LL15:
|
|
and N, 7, I
|
|
cmp I, 0
|
|
ble,a,pn %icc, .LL19
|
|
nop
|
|
|
|
.LL16:
|
|
STF FZERO, [X + 0 * SIZE]
|
|
add I, -1, I
|
|
cmp I, 0
|
|
bg,pt %icc, .LL16
|
|
add X, 1 * SIZE, X
|
|
|
|
.LL19:
|
|
return %i7 + 8
|
|
clr %o0
|
|
|
|
.LL50:
|
|
sra N, 3, I
|
|
cmp I, 0
|
|
ble,pn %icc, .LL55
|
|
nop
|
|
|
|
.LL51:
|
|
STF FZERO, [X + 0 * SIZE]
|
|
add X, INCX, X
|
|
add I, -1, I
|
|
STF FZERO, [X + 0 * SIZE]
|
|
add X, INCX, X
|
|
cmp I, 0
|
|
STF FZERO, [X + 0 * SIZE]
|
|
add X, INCX, X
|
|
STF FZERO, [X + 0 * SIZE]
|
|
add X, INCX, X
|
|
STF FZERO, [X + 0 * SIZE]
|
|
add X, INCX, X
|
|
STF FZERO, [X + 0 * SIZE]
|
|
add X, INCX, X
|
|
STF FZERO, [X + 0 * SIZE]
|
|
add X, INCX, X
|
|
STF FZERO, [X + 0 * SIZE]
|
|
bg,pt %icc, .LL51
|
|
add X, INCX, X
|
|
|
|
.LL55:
|
|
and N, 7, I
|
|
cmp I, 0
|
|
ble,a,pn %icc, .LL59
|
|
nop
|
|
|
|
.LL56:
|
|
STF FZERO, [X + 0 * SIZE]
|
|
add I, -1, I
|
|
cmp I, 0
|
|
bg,pt %icc, .LL56
|
|
add X, INCX, X
|
|
|
|
.LL59:
|
|
return %i7 + 8
|
|
clr %o0
|
|
|
|
.LL100:
|
|
cmp INCX, SIZE
|
|
bne .LL150
|
|
sra N, 3, I
|
|
|
|
cmp I, 0
|
|
ble,pn %icc, .LL115
|
|
nop
|
|
|
|
LDF [X + 0 * SIZE], c1
|
|
LDF [X + 1 * SIZE], c2
|
|
LDF [X + 2 * SIZE], c3
|
|
LDF [X + 3 * SIZE], c4
|
|
LDF [X + 4 * SIZE], c5
|
|
LDF [X + 5 * SIZE], c6
|
|
LDF [X + 6 * SIZE], c7
|
|
LDF [X + 7 * SIZE], c8
|
|
FMUL ALPHA, c1, t1
|
|
LDF [X + 8 * SIZE], c1
|
|
FMUL ALPHA, c2, t2
|
|
LDF [X + 9 * SIZE], c2
|
|
|
|
deccc I
|
|
ble,pt %icc, .LL112
|
|
nop
|
|
|
|
.LL111:
|
|
prefetch [X + PREFETCHSIZE * SIZE], 0
|
|
deccc I
|
|
|
|
FMUL ALPHA, c3, t3
|
|
LDF [X + 10 * SIZE], c3
|
|
nop
|
|
STF t1, [X + 0 * SIZE]
|
|
|
|
FMUL ALPHA, c4, t4
|
|
LDF [X + 11 * SIZE], c4
|
|
nop
|
|
STF t2, [X + 1 * SIZE]
|
|
|
|
FMUL ALPHA, c5, t5
|
|
LDF [X + 12 * SIZE], c5
|
|
nop
|
|
STF t3, [X + 2 * SIZE]
|
|
|
|
FMUL ALPHA, c6, t6
|
|
LDF [X + 13 * SIZE], c6
|
|
nop
|
|
STF t4, [X + 3 * SIZE]
|
|
|
|
FMUL ALPHA, c7, t7
|
|
LDF [X + 14 * SIZE], c7
|
|
nop
|
|
STF t5, [X + 4 * SIZE]
|
|
|
|
FMUL ALPHA, c8, t8
|
|
LDF [X + 15 * SIZE], c8
|
|
nop
|
|
STF t6, [X + 5 * SIZE]
|
|
|
|
FMUL ALPHA, c1, t1
|
|
STF t7, [X + 6 * SIZE]
|
|
nop
|
|
LDF [X + 16 * SIZE], c1
|
|
|
|
FMUL ALPHA, c2, t2
|
|
STF t8, [X + 7 * SIZE]
|
|
nop
|
|
LDF [X + 17 * SIZE], c2
|
|
|
|
bg,pt %icc, .LL111
|
|
add X, 8 * SIZE, X
|
|
|
|
.LL112:
|
|
FMUL ALPHA, c3, t3
|
|
STF t1, [X + 0 * SIZE]
|
|
FMUL ALPHA, c4, t4
|
|
STF t2, [X + 1 * SIZE]
|
|
|
|
FMUL ALPHA, c5, t5
|
|
STF t3, [X + 2 * SIZE]
|
|
FMUL ALPHA, c6, t6
|
|
STF t4, [X + 3 * SIZE]
|
|
|
|
FMUL ALPHA, c7, t7
|
|
STF t5, [X + 4 * SIZE]
|
|
FMUL ALPHA, c8, t8
|
|
STF t6, [X + 5 * SIZE]
|
|
STF t7, [X + 6 * SIZE]
|
|
STF t8, [X + 7 * SIZE]
|
|
|
|
add X, 8 * SIZE, X
|
|
|
|
.LL115:
|
|
and N, 7, I
|
|
cmp I, 0
|
|
ble,a,pn %icc, .LL119
|
|
nop
|
|
|
|
.LL116:
|
|
LDF [X + 0 * SIZE], c1
|
|
add I, -1, I
|
|
FMUL ALPHA, c1, c1
|
|
cmp I, 0
|
|
STF c1, [X + 0 * SIZE]
|
|
bg,pt %icc, .LL116
|
|
add X, 1 * SIZE, X
|
|
|
|
.LL119:
|
|
return %i7 + 8
|
|
clr %o0
|
|
|
|
.LL150:
|
|
sra N, 3, I
|
|
cmp I, 0
|
|
ble,pn %icc, .LL155
|
|
mov X, XX
|
|
|
|
.LL151:
|
|
LDF [X + 0 * SIZE], c1
|
|
add X, INCX, X
|
|
LDF [X + 0 * SIZE], c2
|
|
add X, INCX, X
|
|
LDF [X + 0 * SIZE], c3
|
|
add X, INCX, X
|
|
LDF [X + 0 * SIZE], c4
|
|
add X, INCX, X
|
|
LDF [X + 0 * SIZE], c5
|
|
FMUL ALPHA, c1, c1
|
|
add X, INCX, X
|
|
LDF [X + 0 * SIZE], c6
|
|
FMUL ALPHA, c2, c2
|
|
add X, INCX, X
|
|
LDF [X + 0 * SIZE], c7
|
|
FMUL ALPHA, c3, c3
|
|
add X, INCX, X
|
|
LDF [X + 0 * SIZE], c8
|
|
FMUL ALPHA, c4, c4
|
|
add X, INCX, X
|
|
|
|
STF c1, [XX + 0 * SIZE]
|
|
FMUL ALPHA, c5, c5
|
|
add XX, INCX, XX
|
|
STF c2, [XX + 0 * SIZE]
|
|
FMUL ALPHA, c6, c6
|
|
add XX, INCX, XX
|
|
STF c3, [XX + 0 * SIZE]
|
|
FMUL ALPHA, c7, c7
|
|
add XX, INCX, XX
|
|
STF c4, [XX + 0 * SIZE]
|
|
FMUL ALPHA, c8, c8
|
|
add XX, INCX, XX
|
|
STF c5, [XX + 0 * SIZE]
|
|
add XX, INCX, XX
|
|
add I, -1, I
|
|
STF c6, [XX + 0 * SIZE]
|
|
add XX, INCX, XX
|
|
cmp I, 0
|
|
STF c7, [XX + 0 * SIZE]
|
|
add XX, INCX, XX
|
|
STF c8, [XX + 0 * SIZE]
|
|
|
|
bg,pt %icc, .LL151
|
|
add XX, INCX, XX
|
|
|
|
.LL155:
|
|
and N, 7, I
|
|
cmp I, 0
|
|
ble,a,pn %icc, .LL159
|
|
nop
|
|
|
|
.LL156:
|
|
LDF [X + 0 * SIZE], c1
|
|
add I, -1, I
|
|
FMUL ALPHA, c1, c1
|
|
cmp I, 0
|
|
STF c1, [X + 0 * SIZE]
|
|
bg,pt %icc, .LL156
|
|
add X, INCX, X
|
|
|
|
.LL159:
|
|
return %i7 + 8
|
|
clr %o0
|
|
|
|
|
|
EPILOGUE
|