Optimize zscal function for POWER10
authorRajalakshmi Srinivasaraghavan <rajis@linux.ibm.com>
Wed, 10 Mar 2021 23:15:33 +0000 (17:15 -0600)
committerRajalakshmi Srinivasaraghavan <rajis@linux.ibm.com>
Wed, 10 Mar 2021 23:15:33 +0000 (17:15 -0600)
This patch makes use of new POWER10 vector pair instructions for
loads and stores.

kernel/power/zscal.c
kernel/power/zscal_microk_power10.c [new file with mode: 0644]

index 31b3682..0068138 100644 (file)
@@ -45,7 +45,7 @@ USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
 #endif
 #elif defined(POWER10)
 #if defined(DOUBLE)
-#include "zscal_microk_power8.c"
+#include "zscal_microk_power10.c"
 #else
 #include "cscal_microk_power10.c"
 #endif
diff --git a/kernel/power/zscal_microk_power10.c b/kernel/power/zscal_microk_power10.c
new file mode 100644 (file)
index 0000000..15b8323
--- /dev/null
@@ -0,0 +1,195 @@
+/***************************************************************************
+Copyright (c) 2021, The OpenBLAS Project
+All rights reserved.
+Redistribution and use in source and binary forms, with or without
+modification, are permitted provided that the following conditions are
+met:
+1. Redistributions of source code must retain the above copyright
+notice, this list of conditions and the following disclaimer.
+2. Redistributions in binary form must reproduce the above copyright
+notice, this list of conditions and the following disclaimer in
+the documentation and/or other materials provided with the
+distribution.
+3. Neither the name of the OpenBLAS project nor the names of
+its contributors may be used to endorse or promote products
+derived from this software without specific prior written permission.
+THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
+AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
+IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
+ARE DISCLAIMED. IN NO EVENT SHALL THE OPENBLAS PROJECT OR CONTRIBUTORS BE
+LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
+DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR
+SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
+CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY,
+OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE
+USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
+*****************************************************************************/
+
+#define HAVE_KERNEL_8 1
+
+static void zscal_kernel_8 (long n, double *x, double alpha_r, double alpha_i)
+{
+  __vector double t0;
+  __vector double t1;
+  __vector double t2;
+  __vector double t3;
+  __vector double t4;
+  __vector double t5;
+
+  __asm__
+    (
+       "dcbt           0, %2           \n\t"
+
+       "xsnegdp                33, %x10        \n\t"   // -alpha_i
+       XXSPLTD_S(32,%x9,0)     // alpha_r , alpha_r
+       XXMRGHD_S(33,%x10, 33)  // -alpha_i , alpha_i
+
+       "lxvp           40, 0(%2)       \n\t"
+       "lxvp           42, 32(%2)      \n\t"
+       "lxvp           44, 64(%2)      \n\t"
+       "lxvp           46, 96(%2)      \n\t"
+
+       "addic.         %1, %1, -8      \n\t"
+       "ble            two%=           \n\t"
+
+       ".align 5               \n"
+     "one%=:                           \n\t"
+
+       "xvmuldp                48, 40, 32      \n\t"   // x0_r * alpha_r, x0_i * alpha_r
+       "xvmuldp                49, 41, 32      \n\t"
+       "xvmuldp                50, 42, 32      \n\t"
+       "xvmuldp                51, 43, 32      \n\t"
+       "xvmuldp                34, 44, 32      \n\t"
+       "xvmuldp                35, 45, 32      \n\t"
+       "xvmuldp                36, 46, 32      \n\t"
+       "xvmuldp                37, 47, 32      \n\t"
+
+       XXSWAPD_S(38,40)
+       XXSWAPD_S(39,41)
+       XXSWAPD_S(%x3,42)
+       XXSWAPD_S(%x4,43)
+       XXSWAPD_S(%x5,44)
+       XXSWAPD_S(%x6,45)
+       XXSWAPD_S(%x7,46)
+       XXSWAPD_S(%x8,47)
+
+       "xvmuldp                38, 38, 33      \n\t"   // x0_i * -alpha_i, x0_r * alpha_i
+       "xvmuldp                39, 39, 33      \n\t"
+
+
+       "xvmuldp                %x3, %x3, 33    \n\t"
+       "xvmuldp                %x4, %x4, 33    \n\t"
+
+
+       "lxvp           40, 128(%2)     \n\t"
+       "lxvp           42, 160(%2)     \n\t"
+       "xvmuldp                %x5, %x5, 33    \n\t"
+       "xvmuldp                %x6, %x6, 33    \n\t"
+
+
+       "xvmuldp                %x7, %x7, 33    \n\t"
+       "xvmuldp                %x8, %x8, 33    \n\t"
+       "lxvp           44, 192(%2)     \n\t"
+       "lxvp           46, 224(%2)     \n\t"
+
+
+       "xvadddp                48, 48, 38      \n\t"
+       "xvadddp                49, 49, 39      \n\t"
+       "xvadddp                50, 50, %x3     \n\t"
+       "xvadddp                51, 51, %x4     \n\t"
+       "stxv           49, 0(%2)       \n\t"
+       "stxv           48, 16(%2)      \n\t"
+       "stxv           51, 32(%2)      \n\t"
+       "stxv           50, 48(%2)      \n\t"
+
+
+       "xvadddp                34, 34, %x5     \n\t"
+       "xvadddp                35, 35, %x6     \n\t"
+
+
+       "xvadddp                36, 36, %x7     \n\t"
+       "xvadddp                37, 37, %x8     \n\t"
+
+       "stxv           35, 64(%2)      \n\t"
+       "stxv           34, 80(%2)      \n\t"
+       "stxv           37, 96(%2)      \n\t"
+       "stxv           36, 112(%2)     \n\t"
+
+       "addi           %2, %2, 128     \n\t"
+
+       "addic.         %1, %1, -8      \n\t"
+       "bgt            one%=           \n"
+
+     "two%=:                           \n\t"
+
+       "xvmuldp                48, 40, 32      \n\t"   // x0_r * alpha_r, x0_i * alpha_r
+       "xvmuldp                49, 41, 32      \n\t"
+       "xvmuldp                50, 42, 32      \n\t"
+       "xvmuldp                51, 43, 32      \n\t"
+       "xvmuldp                34, 44, 32      \n\t"
+       "xvmuldp                35, 45, 32      \n\t"
+       "xvmuldp                36, 46, 32      \n\t"
+       "xvmuldp                37, 47, 32      \n\t"
+
+       XXSWAPD_S(38,40)
+       XXSWAPD_S(39,41)
+       XXSWAPD_S(%x3,42)
+       XXSWAPD_S(%x4,43)
+       XXSWAPD_S(%x5,44)
+       XXSWAPD_S(%x6,45)
+       XXSWAPD_S(%x7,46)
+       XXSWAPD_S(%x8,47)
+
+
+       "xvmuldp                38, 38, 33      \n\t"   // x0_i * -alpha_i, x0_r * alpha_i
+       "xvmuldp                39, 39, 33      \n\t"
+       "xvmuldp                %x3, %x3, 33    \n\t"
+       "xvmuldp                %x4, %x4, 33    \n\t"
+       "xvmuldp                %x5, %x5, 33    \n\t"
+       "xvmuldp                %x6, %x6, 33    \n\t"
+       "xvmuldp                %x7, %x7, 33    \n\t"
+       "xvmuldp                %x8, %x8, 33    \n\t"
+
+       "xvadddp                48, 48, 38      \n\t"
+       "xvadddp                49, 49, 39      \n\t"
+
+       "xvadddp                50, 50, %x3     \n\t"
+       "xvadddp                51, 51, %x4     \n\t"
+       "stxv           49, 0(%2)       \n\t"
+       "stxv           48, 16(%2)      \n\t"
+       "stxv           51, 32(%2)      \n\t"
+       "stxv           50, 48(%2)      \n\t"
+
+       "xvadddp                34, 34, %x5     \n\t"
+       "xvadddp                35, 35, %x6     \n\t"
+
+
+       "xvadddp                36, 36, %x7     \n\t"
+       "xvadddp                37, 37, %x8     \n\t"
+
+       "stxv           35, 64(%2)      \n\t"
+       "stxv           34, 80(%2)      \n\t"
+       "stxv           37, 96(%2)      \n\t"
+       "stxv           36, 112(%2)     \n\t"
+
+     "#n=%1 x=%0=%2 alpha=(%9,%10) \n"
+     :
+       "+m" (*x),
+       "+r" (n),       // 1
+       "+b" (x),       // 2
+       "=wa" (t0),     // 3
+       "=wa" (t1),     // 4
+       "=wa" (t2),     // 5
+       "=wa" (t3),     // 6
+       "=wa" (t4),     // 7
+       "=wa" (t5)      // 8
+     :
+       "d" (alpha_r),  // 9 
+       "d" (alpha_i)   // 10
+     :
+       "cr0",
+       "vs32","vs33","vs34","vs35","vs36","vs37","vs38","vs39",
+       "vs40","vs41","vs42","vs43","vs44","vs45","vs46","vs47",
+       "vs48","vs49","vs50","vs51"
+     );
+}