474 lines
17 KiB
C
474 lines
17 KiB
C
/***************************************************************************
|
|
Copyright (c) 2024, The OpenBLAS Project
|
|
All rights reserved.
|
|
Redistribution and use in source and binary forms, with or without
|
|
modification, are permitted provided that the following conditions are
|
|
met:
|
|
1. Redistributions of source code must retain the above copyright
|
|
notice, this list of conditions and the following disclaimer.
|
|
2. Redistributions in binary form must reproduce the above copyright
|
|
notice, this list of conditions and the following disclaimer in
|
|
the documentation and/or other materials provided with the
|
|
distribution.
|
|
3. Neither the name of the OpenBLAS project nor the names of
|
|
its contributors may be used to endorse or promote products
|
|
derived from this software without specific prior written permission.
|
|
THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
|
AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
ARE DISCLAIMED. IN NO EVENT SHALL THE OPENBLAS PROJECT OR CONTRIBUTORS BE
|
|
LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
|
CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE
|
|
GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF
|
|
THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
*****************************************************************************/
|
|
|
|
#include "common.h"
|
|
|
|
#include <arm_neon.h>
|
|
#include <arm_sve.h>
|
|
#if defined(__ARM_NEON_SVE_BRIDGE) && defined(__has_include) && \
|
|
__has_include(<arm_neon_sve_bridge.h>)
|
|
#include <arm_neon_sve_bridge.h>
|
|
#else
|
|
#define svdup_neonq_f32(fixed_reg) \
|
|
({ \
|
|
svfloat32_t scalable_reg; \
|
|
asm("mov %0.q, %q1" : "=w"(scalable_reg) : "w"(fixed_reg) :); \
|
|
scalable_reg; \
|
|
})
|
|
#define svdup_neonq_f64(fixed_reg) \
|
|
({ \
|
|
svfloat64_t scalable_reg; \
|
|
asm("mov %0.q, %q1" : "=w"(scalable_reg) : "w"(fixed_reg) :); \
|
|
scalable_reg; \
|
|
})
|
|
#endif
|
|
|
|
#define RESET_A_POINTER() a_offset = A;
|
|
|
|
#define CREATE_A_POINTER(m, scale) FLOAT* a_offset##m = a_offset + scale;
|
|
#define UPDATE_A_POINTER(scale) a_offset = a_offset + scale;
|
|
#define A_ELEMENT_K(m, offset_k) *(a_offset##m + (k + offset_k) * lda)
|
|
#define A_ELEMENT(m) A_ELEMENT_K(m, 0)
|
|
|
|
#define RESET_B_POINTER() b_offset = B;
|
|
|
|
#define CREATE_B_POINTER(n, scale) FLOAT* b_offset##n = b_offset + scale;
|
|
#define UPDATE_B_POINTER(scale) b_offset = b_offset + scale;
|
|
#define B_ELEMENT_K(n, offset_k) *(b_offset##n + (k + offset_k) * ldb)
|
|
#define B_ELEMENT(n) B_ELEMENT_K(n, 0)
|
|
|
|
#define CREATE_C_POINTER(n, scale) FLOAT* c_offset##n = c_offset + scale * ldc;
|
|
#define INCR_C_POINTER(m, incr) // c_offset ## m += incr;
|
|
#define UPDATE_C_POINTER(scale) c_offset = c_offset + scale * ldc;
|
|
#define C_ELEMENT(m, n) *(c_offset##n + ((m * v_size) + i))
|
|
|
|
// #undef C_ELEMENT
|
|
// #define C_ELEMENT(m, n) C[(i+(m))+(j+(n))*ldc]
|
|
|
|
#define PACK_ELEMENT_K(n, offset_k) packed_b[(k + offset_k) * 4 + n]
|
|
#define PACK_ELEMENT(n) PACK_ELEMENT_K(n, 0)
|
|
|
|
// ASIMD
|
|
#define DECLARE_RESULT_VECTOR2(m, n) \
|
|
float64x2_t result##m##n = vdupq_n_f64(0.0);
|
|
#define DECLARE_RESULT(m, n) float64_t result##m##n = 0.0;
|
|
#define BROADCAST_LOAD_A2(m, offset_k) \
|
|
float64x2_t a##m##_k##offset_k = vld1q_dup_f64(&A_ELEMENT_K(m, offset_k));
|
|
#define LOAD_A1(m, offset_k) \
|
|
float64_t a##m##_k##offset_k = A_ELEMENT_K(m, offset_k);
|
|
#define VECTOR_LOAD_B2(n, offset_k) \
|
|
float64x2_t b##n##_k##offset_k = vld1q_f64(&B_ELEMENT_K(n, offset_k));
|
|
#define GATHER_LOAD_B2(n, offset_k) \
|
|
float64x2_t b##n##_k##offset_k = vdupq_n_f64(B_ELEMENT_K(n, offset_k)); \
|
|
b##n##_k##offset_k = \
|
|
vsetq_lane_f64(B_ELEMENT_K(n + 1, offset_k), b##n##_k##offset_k, 1);
|
|
#define UPDATE_RESULT_VECTOR2(m, n, offset_k) \
|
|
result##m##n = \
|
|
vfmaq_f64(result##m##n, a##m##_k##offset_k, b##n##_k##offset_k);
|
|
#define UPDATE_RESULT(m, n, offset_k) \
|
|
result##m##n = result##m##n + a##m##_k##offset_k * b##n##_k##offset_k;
|
|
#ifdef B0
|
|
#define SCATTER_STORE2(m, n) \
|
|
result##m##n = vmulq_f64(result##m##n, vdupq_n_f64(alpha)); \
|
|
C_ELEMENT(m, n + 0) = vgetq_lane_f64(result##m##n, 0); \
|
|
C_ELEMENT(m, n + 1) = vgetq_lane_f64(result##m##n, 1);
|
|
#else
|
|
#define SCATTER_STORE2(m, n) \
|
|
result##m##n = vmulq_f64(result##m##n, vdupq_n_f64(alpha)); \
|
|
C_ELEMENT(m, n + 0) = \
|
|
C_ELEMENT(m, n + 0) * beta + vgetq_lane_f64(result##m##n, 0); \
|
|
C_ELEMENT(m, n + 1) = \
|
|
C_ELEMENT(m, n + 1) * beta + vgetq_lane_f64(result##m##n, 1);
|
|
#endif
|
|
|
|
// SVE
|
|
#define DECLARE_RESULT_VECTOR(m, n) svfloat64_t result##m##n = svdup_f64(0.0);
|
|
#define BROADCAST_LOAD_A(m, offset_k) \
|
|
svfloat64_t a##s##m##_k##offset_k = svdup_f64(A_ELEMENT_K(m, offset_k));
|
|
#define BROADCAST_LOAD_B(n, offset_k) \
|
|
svfloat64_t b##s##n##_k##offset_k = svdup_f64(B_ELEMENT_K(n, offset_k));
|
|
#define VECTOR_LOAD_A(pg, m, offset_k) \
|
|
svfloat64_t a##s##m##_k##offset_k = svld1(pg, &A_ELEMENT_K(m, offset_k));
|
|
#define QUADWORD_LOAD_B(n, offset_k) \
|
|
svfloat64_t b##s##n##_k##offset_k = \
|
|
svld1rq(pg_true, &B_ELEMENT_K(n, offset_k));
|
|
#define UPDATE_RESULT_VECTOR(pg, m, n, offset_k) \
|
|
result##m##n = \
|
|
svmla_m(pg, result##m##n, a##s##m##_k##offset_k, b##s##n##_k##offset_k);
|
|
#define UPDATE_RESULT_VECTOR_QUADWORD(m, n, outer, lane, offset_k) \
|
|
result##m##n = svmla_lane( \
|
|
result##m##n, a##s##m##_k##offset_k, b##s##outer##_k##offset_k, lane);
|
|
#ifdef B0
|
|
#define VECTOR_STORE(pg, m, n) \
|
|
result##m##n = svmul_m(pg, result##m##n, alpha_vec); \
|
|
svst1(pg, &C_ELEMENT(m, n), result##m##n);
|
|
#define SCATTER_STORE(pg, m, n) \
|
|
result##m##n = svmul_m(pg, result##m##n, alpha_vec); \
|
|
svst1_scatter_index(pg, &C_ELEMENT(m, n), ldc_vec, result##m##n);
|
|
#else
|
|
#define VECTOR_STORE(pg, m, n) \
|
|
result##m##n = svmul_m(pg, result##m##n, alpha_vec); \
|
|
result##m##n = \
|
|
svmla_m(pg, result##m##n, svld1(pg, &C_ELEMENT(m, n)), beta_vec); \
|
|
svst1(pg, &C_ELEMENT(m, n), result##m##n);
|
|
#define SCATTER_STORE(pg, m, n) \
|
|
result##m##n = svmul_m(pg, result##m##n, alpha_vec); \
|
|
result##m##n = svmla_m(pg, \
|
|
result##m##n, \
|
|
svld1_gather_index(pg, &C_ELEMENT(m, n), ldc_vec), \
|
|
beta_vec); \
|
|
svst1_scatter_index(pg, &C_ELEMENT(m, n), ldc_vec, result##m##n);
|
|
#endif
|
|
|
|
#ifndef LIKELY
|
|
#ifdef __GNUC__
|
|
#define LIKELY(x) __builtin_expect(!!(x), 1)
|
|
#else
|
|
#define LIKELY(x) (x)
|
|
#endif
|
|
#endif
|
|
|
|
#ifdef B0
|
|
int
|
|
CNAME(BLASLONG M,
|
|
BLASLONG N,
|
|
BLASLONG K,
|
|
IFLOAT* A,
|
|
BLASLONG lda,
|
|
FLOAT alpha,
|
|
IFLOAT* B,
|
|
BLASLONG ldb,
|
|
FLOAT* C,
|
|
BLASLONG ldc)
|
|
#else
|
|
int
|
|
CNAME(BLASLONG M,
|
|
BLASLONG N,
|
|
BLASLONG K,
|
|
IFLOAT* A,
|
|
BLASLONG lda,
|
|
FLOAT alpha,
|
|
IFLOAT* B,
|
|
BLASLONG ldb,
|
|
FLOAT beta,
|
|
FLOAT* C,
|
|
BLASLONG ldc)
|
|
#endif
|
|
{
|
|
const uint64_t v_size = svcntd();
|
|
const uint64_t v_size2 = v_size * 2;
|
|
const svbool_t pg_true = svptrue_b64();
|
|
const svbool_t pg_quad = svwhilelt_b64(0, 2);
|
|
const svfloat64_t alpha_vec = svdup_f64(alpha);
|
|
#ifndef B0
|
|
const svfloat64_t beta_vec = svdup_f64(beta);
|
|
#endif
|
|
const BLASLONG n4 = N & -4;
|
|
const BLASLONG n2 = N & -2;
|
|
const BLASLONG v_m2 = M & -v_size2;
|
|
const BLASLONG v_m1 = M & -v_size;
|
|
|
|
FLOAT* b_offset = B;
|
|
FLOAT* a_offset = A;
|
|
FLOAT* c_offset = C;
|
|
|
|
BLASLONG j = 0;
|
|
for (; j < n4; j += 4) {
|
|
|
|
CREATE_C_POINTER(0, 0);
|
|
CREATE_C_POINTER(1, 1);
|
|
CREATE_C_POINTER(2, 2);
|
|
CREATE_C_POINTER(3, 3);
|
|
CREATE_B_POINTER(0, 0);
|
|
CREATE_B_POINTER(1, 1);
|
|
CREATE_B_POINTER(2, 2);
|
|
CREATE_B_POINTER(3, 3);
|
|
|
|
BLASLONG i = 0;
|
|
for (; i < v_m2; i += v_size2) {
|
|
|
|
CREATE_A_POINTER(0, 0);
|
|
CREATE_A_POINTER(1, v_size);
|
|
UPDATE_A_POINTER(v_size2);
|
|
|
|
BLASLONG k = 0;
|
|
DECLARE_RESULT_VECTOR(0, 0);
|
|
DECLARE_RESULT_VECTOR(0, 1);
|
|
DECLARE_RESULT_VECTOR(0, 2);
|
|
DECLARE_RESULT_VECTOR(0, 3);
|
|
DECLARE_RESULT_VECTOR(1, 0);
|
|
DECLARE_RESULT_VECTOR(1, 1);
|
|
DECLARE_RESULT_VECTOR(1, 2);
|
|
DECLARE_RESULT_VECTOR(1, 3);
|
|
|
|
for (; k < K; k++) {
|
|
|
|
QUADWORD_LOAD_B(0, 0);
|
|
VECTOR_LOAD_A(pg_true, 0, 0);
|
|
UPDATE_RESULT_VECTOR_QUADWORD(0, 0, 0, 0, 0);
|
|
UPDATE_RESULT_VECTOR_QUADWORD(0, 1, 0, 1, 0);
|
|
QUADWORD_LOAD_B(2, 0);
|
|
UPDATE_RESULT_VECTOR_QUADWORD(0, 2, 2, 0, 0);
|
|
UPDATE_RESULT_VECTOR_QUADWORD(0, 3, 2, 1, 0);
|
|
VECTOR_LOAD_A(pg_true, 1, 0);
|
|
UPDATE_RESULT_VECTOR_QUADWORD(1, 0, 0, 0, 0);
|
|
UPDATE_RESULT_VECTOR_QUADWORD(1, 1, 0, 1, 0);
|
|
UPDATE_RESULT_VECTOR_QUADWORD(1, 2, 2, 0, 0);
|
|
UPDATE_RESULT_VECTOR_QUADWORD(1, 3, 2, 1, 0);
|
|
}
|
|
VECTOR_STORE(pg_true, 0, 0);
|
|
VECTOR_STORE(pg_true, 0, 1);
|
|
VECTOR_STORE(pg_true, 0, 2);
|
|
VECTOR_STORE(pg_true, 0, 3);
|
|
VECTOR_STORE(pg_true, 1, 0);
|
|
VECTOR_STORE(pg_true, 1, 1);
|
|
VECTOR_STORE(pg_true, 1, 2);
|
|
VECTOR_STORE(pg_true, 1, 3);
|
|
INCR_C_POINTER(0, v_size2);
|
|
INCR_C_POINTER(1, v_size2);
|
|
INCR_C_POINTER(2, v_size2);
|
|
INCR_C_POINTER(3, v_size2);
|
|
}
|
|
for (; i < v_m1; i += v_size) {
|
|
|
|
CREATE_A_POINTER(0, 0);
|
|
UPDATE_A_POINTER(v_size);
|
|
|
|
BLASLONG k = 0;
|
|
DECLARE_RESULT_VECTOR(0, 0);
|
|
DECLARE_RESULT_VECTOR(0, 1);
|
|
DECLARE_RESULT_VECTOR(0, 2);
|
|
DECLARE_RESULT_VECTOR(0, 3);
|
|
|
|
for (; k < K; k++) {
|
|
|
|
QUADWORD_LOAD_B(0, 0);
|
|
VECTOR_LOAD_A(pg_true, 0, 0);
|
|
UPDATE_RESULT_VECTOR_QUADWORD(0, 0, 0, 0, 0);
|
|
UPDATE_RESULT_VECTOR_QUADWORD(0, 1, 0, 1, 0);
|
|
QUADWORD_LOAD_B(2, 0);
|
|
UPDATE_RESULT_VECTOR_QUADWORD(0, 2, 2, 0, 0);
|
|
UPDATE_RESULT_VECTOR_QUADWORD(0, 3, 2, 1, 0);
|
|
}
|
|
VECTOR_STORE(pg_true, 0, 0);
|
|
VECTOR_STORE(pg_true, 0, 1);
|
|
VECTOR_STORE(pg_true, 0, 2);
|
|
VECTOR_STORE(pg_true, 0, 3);
|
|
INCR_C_POINTER(0, v_size);
|
|
INCR_C_POINTER(1, v_size);
|
|
INCR_C_POINTER(2, v_size);
|
|
INCR_C_POINTER(3, v_size);
|
|
}
|
|
for (; i < M; i += v_size) {
|
|
const svbool_t pg_tail = svwhilelt_b64((uint64_t)i, (uint64_t)(M));
|
|
CREATE_A_POINTER(0, 0);
|
|
UPDATE_A_POINTER(0);
|
|
|
|
BLASLONG k = 0;
|
|
DECLARE_RESULT_VECTOR(0, 0);
|
|
DECLARE_RESULT_VECTOR(0, 1);
|
|
DECLARE_RESULT_VECTOR(0, 2);
|
|
DECLARE_RESULT_VECTOR(0, 3);
|
|
|
|
for (; k < K; k++) {
|
|
|
|
QUADWORD_LOAD_B(0, 0);
|
|
VECTOR_LOAD_A(pg_tail, 0, 0);
|
|
UPDATE_RESULT_VECTOR_QUADWORD(0, 0, 0, 0, 0);
|
|
UPDATE_RESULT_VECTOR_QUADWORD(0, 1, 0, 1, 0);
|
|
QUADWORD_LOAD_B(2, 0);
|
|
UPDATE_RESULT_VECTOR_QUADWORD(0, 2, 2, 0, 0);
|
|
UPDATE_RESULT_VECTOR_QUADWORD(0, 3, 2, 1, 0);
|
|
}
|
|
VECTOR_STORE(pg_tail, 0, 0);
|
|
VECTOR_STORE(pg_tail, 0, 1);
|
|
VECTOR_STORE(pg_tail, 0, 2);
|
|
VECTOR_STORE(pg_tail, 0, 3);
|
|
INCR_C_POINTER(0, 0);
|
|
INCR_C_POINTER(1, 0);
|
|
INCR_C_POINTER(2, 0);
|
|
INCR_C_POINTER(3, 0);
|
|
}
|
|
|
|
UPDATE_B_POINTER(4);
|
|
RESET_A_POINTER();
|
|
UPDATE_C_POINTER(4);
|
|
}
|
|
for (; j < n2; j += 2) {
|
|
|
|
CREATE_C_POINTER(0, 0);
|
|
CREATE_C_POINTER(1, 1);
|
|
CREATE_B_POINTER(0, 0);
|
|
CREATE_B_POINTER(1, 1);
|
|
|
|
BLASLONG i = 0;
|
|
for (; i < v_m2; i += v_size2) {
|
|
|
|
CREATE_A_POINTER(0, 0);
|
|
CREATE_A_POINTER(1, v_size);
|
|
UPDATE_A_POINTER(v_size2);
|
|
|
|
BLASLONG k = 0;
|
|
DECLARE_RESULT_VECTOR(0, 0);
|
|
DECLARE_RESULT_VECTOR(0, 1);
|
|
DECLARE_RESULT_VECTOR(1, 0);
|
|
DECLARE_RESULT_VECTOR(1, 1);
|
|
|
|
for (; k < K; k++) {
|
|
|
|
QUADWORD_LOAD_B(0, 0);
|
|
VECTOR_LOAD_A(pg_true, 0, 0);
|
|
UPDATE_RESULT_VECTOR_QUADWORD(0, 0, 0, 0, 0);
|
|
UPDATE_RESULT_VECTOR_QUADWORD(0, 1, 0, 1, 0);
|
|
VECTOR_LOAD_A(pg_true, 1, 0);
|
|
UPDATE_RESULT_VECTOR_QUADWORD(1, 0, 0, 0, 0);
|
|
UPDATE_RESULT_VECTOR_QUADWORD(1, 1, 0, 1, 0);
|
|
}
|
|
VECTOR_STORE(pg_true, 0, 0);
|
|
VECTOR_STORE(pg_true, 0, 1);
|
|
VECTOR_STORE(pg_true, 1, 0);
|
|
VECTOR_STORE(pg_true, 1, 1);
|
|
INCR_C_POINTER(0, v_size2);
|
|
INCR_C_POINTER(1, v_size2);
|
|
}
|
|
for (; i < v_m1; i += v_size) {
|
|
|
|
CREATE_A_POINTER(0, 0);
|
|
UPDATE_A_POINTER(v_size);
|
|
|
|
BLASLONG k = 0;
|
|
DECLARE_RESULT_VECTOR(0, 0);
|
|
DECLARE_RESULT_VECTOR(0, 1);
|
|
|
|
for (; k < K; k++) {
|
|
|
|
QUADWORD_LOAD_B(0, 0);
|
|
VECTOR_LOAD_A(pg_true, 0, 0);
|
|
UPDATE_RESULT_VECTOR_QUADWORD(0, 0, 0, 0, 0);
|
|
UPDATE_RESULT_VECTOR_QUADWORD(0, 1, 0, 1, 0);
|
|
}
|
|
VECTOR_STORE(pg_true, 0, 0);
|
|
VECTOR_STORE(pg_true, 0, 1);
|
|
INCR_C_POINTER(0, v_size);
|
|
INCR_C_POINTER(1, v_size);
|
|
}
|
|
for (; i < M; i += v_size) {
|
|
const svbool_t pg_tail = svwhilelt_b64((uint64_t)i, (uint64_t)(M));
|
|
CREATE_A_POINTER(0, 0);
|
|
UPDATE_A_POINTER(0);
|
|
|
|
BLASLONG k = 0;
|
|
DECLARE_RESULT_VECTOR(0, 0);
|
|
DECLARE_RESULT_VECTOR(0, 1);
|
|
|
|
for (; k < K; k++) {
|
|
|
|
QUADWORD_LOAD_B(0, 0);
|
|
VECTOR_LOAD_A(pg_tail, 0, 0);
|
|
UPDATE_RESULT_VECTOR_QUADWORD(0, 0, 0, 0, 0);
|
|
UPDATE_RESULT_VECTOR_QUADWORD(0, 1, 0, 1, 0);
|
|
}
|
|
VECTOR_STORE(pg_tail, 0, 0);
|
|
VECTOR_STORE(pg_tail, 0, 1);
|
|
INCR_C_POINTER(0, 0);
|
|
INCR_C_POINTER(1, 0);
|
|
}
|
|
|
|
UPDATE_B_POINTER(2);
|
|
RESET_A_POINTER();
|
|
UPDATE_C_POINTER(2);
|
|
}
|
|
for (; j < N; j++) {
|
|
|
|
CREATE_C_POINTER(0, 0);
|
|
CREATE_B_POINTER(0, 0);
|
|
|
|
BLASLONG i = 0;
|
|
for (; i < v_m2; i += v_size2) {
|
|
|
|
CREATE_A_POINTER(0, 0);
|
|
CREATE_A_POINTER(1, v_size);
|
|
UPDATE_A_POINTER(v_size2);
|
|
|
|
BLASLONG k = 0;
|
|
DECLARE_RESULT_VECTOR(0, 0);
|
|
DECLARE_RESULT_VECTOR(1, 0);
|
|
|
|
for (; k < K; k++) {
|
|
|
|
BROADCAST_LOAD_B(0, 0);
|
|
VECTOR_LOAD_A(pg_true, 0, 0);
|
|
UPDATE_RESULT_VECTOR(pg_true, 0, 0, 0);
|
|
VECTOR_LOAD_A(pg_true, 1, 0);
|
|
UPDATE_RESULT_VECTOR(pg_true, 1, 0, 0);
|
|
}
|
|
VECTOR_STORE(pg_true, 0, 0);
|
|
VECTOR_STORE(pg_true, 1, 0);
|
|
INCR_C_POINTER(0, v_size2);
|
|
}
|
|
for (; i < v_m1; i += v_size) {
|
|
|
|
CREATE_A_POINTER(0, 0);
|
|
UPDATE_A_POINTER(v_size);
|
|
|
|
BLASLONG k = 0;
|
|
DECLARE_RESULT_VECTOR(0, 0);
|
|
|
|
for (; k < K; k++) {
|
|
|
|
BROADCAST_LOAD_B(0, 0);
|
|
VECTOR_LOAD_A(pg_true, 0, 0);
|
|
UPDATE_RESULT_VECTOR(pg_true, 0, 0, 0);
|
|
}
|
|
VECTOR_STORE(pg_true, 0, 0);
|
|
INCR_C_POINTER(0, v_size);
|
|
}
|
|
for (; i < M; i += v_size) {
|
|
const svbool_t pg_tail = svwhilelt_b64((uint64_t)i, (uint64_t)(M));
|
|
CREATE_A_POINTER(0, 0);
|
|
UPDATE_A_POINTER(0);
|
|
|
|
BLASLONG k = 0;
|
|
DECLARE_RESULT_VECTOR(0, 0);
|
|
|
|
for (; k < K; k++) {
|
|
|
|
BROADCAST_LOAD_B(0, 0);
|
|
VECTOR_LOAD_A(pg_tail, 0, 0);
|
|
UPDATE_RESULT_VECTOR(pg_tail, 0, 0, 0);
|
|
}
|
|
VECTOR_STORE(pg_tail, 0, 0);
|
|
INCR_C_POINTER(0, 0);
|
|
}
|
|
|
|
UPDATE_B_POINTER(1);
|
|
RESET_A_POINTER();
|
|
UPDATE_C_POINTER(1);
|
|
}
|
|
|
|
return 0;
|
|
} |