KVM: PPC: Add SR swapping code
authorAlexander Graf <agraf@suse.de>
Thu, 15 Apr 2010 22:11:34 +0000 (00:11 +0200)
committerAvi Kivity <avi@redhat.com>
Mon, 17 May 2010 09:18:17 +0000 (12:18 +0300)
Later in this series we will move the current segment switch code to
generic code and make that call hooks for the specific sub-archs (32
vs. 64 bit). This is the hook for 32 bits.

It enabled the entry and exit code to swap segment registers with
values from the shadow cpu structure.

Signed-off-by: Alexander Graf <agraf@suse.de>
Signed-off-by: Avi Kivity <avi@redhat.com>
arch/powerpc/kvm/book3s_32_sr.S [new file with mode: 0644]

diff --git a/arch/powerpc/kvm/book3s_32_sr.S b/arch/powerpc/kvm/book3s_32_sr.S
new file mode 100644 (file)
index 0000000..3608471
--- /dev/null
@@ -0,0 +1,143 @@
+/*
+ * This program is free software; you can redistribute it and/or modify
+ * it under the terms of the GNU General Public License, version 2, as
+ * published by the Free Software Foundation.
+ *
+ * This program is distributed in the hope that it will be useful,
+ * but WITHOUT ANY WARRANTY; without even the implied warranty of
+ * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
+ * GNU General Public License for more details.
+ *
+ * You should have received a copy of the GNU General Public License
+ * along with this program; if not, write to the Free Software
+ * Foundation, 51 Franklin Street, Fifth Floor, Boston, MA  02110-1301, USA.
+ *
+ * Copyright SUSE Linux Products GmbH 2009
+ *
+ * Authors: Alexander Graf <agraf@suse.de>
+ */
+
+/******************************************************************************
+ *                                                                            *
+ *                               Entry code                                   *
+ *                                                                            *
+ *****************************************************************************/
+
+.macro LOAD_GUEST_SEGMENTS
+
+       /* Required state:
+        *
+        * MSR = ~IR|DR
+        * R1 = host R1
+        * R2 = host R2
+        * R3 = shadow vcpu
+        * all other volatile GPRS = free
+        * SVCPU[CR]  = guest CR
+        * SVCPU[XER] = guest XER
+        * SVCPU[CTR] = guest CTR
+        * SVCPU[LR]  = guest LR
+        */
+
+#define XCHG_SR(n)     lwz     r9, (SVCPU_SR+(n*4))(r3);  \
+                       mtsr    n, r9
+
+       XCHG_SR(0)
+       XCHG_SR(1)
+       XCHG_SR(2)
+       XCHG_SR(3)
+       XCHG_SR(4)
+       XCHG_SR(5)
+       XCHG_SR(6)
+       XCHG_SR(7)
+       XCHG_SR(8)
+       XCHG_SR(9)
+       XCHG_SR(10)
+       XCHG_SR(11)
+       XCHG_SR(12)
+       XCHG_SR(13)
+       XCHG_SR(14)
+       XCHG_SR(15)
+
+       /* Clear BATs. */
+
+#define KVM_KILL_BAT(n, reg)           \
+        mtspr   SPRN_IBAT##n##U,reg;   \
+        mtspr   SPRN_IBAT##n##L,reg;   \
+        mtspr   SPRN_DBAT##n##U,reg;   \
+        mtspr   SPRN_DBAT##n##L,reg;   \
+
+        li     r9, 0
+       KVM_KILL_BAT(0, r9)
+       KVM_KILL_BAT(1, r9)
+       KVM_KILL_BAT(2, r9)
+       KVM_KILL_BAT(3, r9)
+
+.endm
+
+/******************************************************************************
+ *                                                                            *
+ *                               Exit code                                    *
+ *                                                                            *
+ *****************************************************************************/
+
+.macro LOAD_HOST_SEGMENTS
+
+       /* Register usage at this point:
+        *
+        * R1         = host R1
+        * R2         = host R2
+        * R12        = exit handler id
+        * R13        = shadow vcpu - SHADOW_VCPU_OFF
+        * SVCPU.*    = guest *
+        * SVCPU[CR]  = guest CR
+        * SVCPU[XER] = guest XER
+        * SVCPU[CTR] = guest CTR
+        * SVCPU[LR]  = guest LR
+        *
+        */
+
+       /* Restore BATs */
+
+       /* We only overwrite the upper part, so we only restoree
+          the upper part. */
+#define KVM_LOAD_BAT(n, reg, RA, RB)   \
+       lwz     RA,(n*16)+0(reg);       \
+       lwz     RB,(n*16)+4(reg);       \
+       mtspr   SPRN_IBAT##n##U,RA;     \
+       mtspr   SPRN_IBAT##n##L,RB;     \
+       lwz     RA,(n*16)+8(reg);       \
+       lwz     RB,(n*16)+12(reg);      \
+       mtspr   SPRN_DBAT##n##U,RA;     \
+       mtspr   SPRN_DBAT##n##L,RB;     \
+
+       lis     r9, BATS@ha
+       addi    r9, r9, BATS@l
+       tophys(r9, r9)
+       KVM_LOAD_BAT(0, r9, r10, r11)
+       KVM_LOAD_BAT(1, r9, r10, r11)
+       KVM_LOAD_BAT(2, r9, r10, r11)
+       KVM_LOAD_BAT(3, r9, r10, r11)
+
+       /* Restore Segment Registers */
+
+       /* 0xc - 0xf */
+
+        li      r0, 4
+        mtctr   r0
+       LOAD_REG_IMMEDIATE(r3, 0x20000000 | (0x111 * 0xc))
+        lis     r4, 0xc000
+3:      mtsrin  r3, r4
+        addi    r3, r3, 0x111     /* increment VSID */
+        addis   r4, r4, 0x1000    /* address of next segment */
+        bdnz    3b
+
+       /* 0x0 - 0xb */
+
+       /* 'current->mm' needs to be in r4 */
+       tophys(r4, r2)
+       lwz     r4, MM(r4)
+       tophys(r4, r4)
+       /* This only clobbers r0, r3, r4 and r5 */
+       bl      switch_mmu_context
+
+.endm