mirror of
https://github.com/torvalds/linux.git
synced 2024-11-17 09:31:50 +00:00
786f19daa8
Later in this series we will move the current segment switch code to generic code and make that call hooks for the specific sub-archs (32 vs. 64 bit). This is the hook for 32 bits. It enabled the entry and exit code to swap segment registers with values from the shadow cpu structure. Signed-off-by: Alexander Graf <agraf@suse.de> Signed-off-by: Avi Kivity <avi@redhat.com>
144 lines
3.7 KiB
ArmAsm
144 lines
3.7 KiB
ArmAsm
/*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License, version 2, as
|
|
* published by the Free Software Foundation.
|
|
*
|
|
* This program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program; if not, write to the Free Software
|
|
* Foundation, 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA.
|
|
*
|
|
* Copyright SUSE Linux Products GmbH 2009
|
|
*
|
|
* Authors: Alexander Graf <agraf@suse.de>
|
|
*/
|
|
|
|
/******************************************************************************
|
|
* *
|
|
* Entry code *
|
|
* *
|
|
*****************************************************************************/
|
|
|
|
.macro LOAD_GUEST_SEGMENTS
|
|
|
|
/* Required state:
|
|
*
|
|
* MSR = ~IR|DR
|
|
* R1 = host R1
|
|
* R2 = host R2
|
|
* R3 = shadow vcpu
|
|
* all other volatile GPRS = free
|
|
* SVCPU[CR] = guest CR
|
|
* SVCPU[XER] = guest XER
|
|
* SVCPU[CTR] = guest CTR
|
|
* SVCPU[LR] = guest LR
|
|
*/
|
|
|
|
#define XCHG_SR(n) lwz r9, (SVCPU_SR+(n*4))(r3); \
|
|
mtsr n, r9
|
|
|
|
XCHG_SR(0)
|
|
XCHG_SR(1)
|
|
XCHG_SR(2)
|
|
XCHG_SR(3)
|
|
XCHG_SR(4)
|
|
XCHG_SR(5)
|
|
XCHG_SR(6)
|
|
XCHG_SR(7)
|
|
XCHG_SR(8)
|
|
XCHG_SR(9)
|
|
XCHG_SR(10)
|
|
XCHG_SR(11)
|
|
XCHG_SR(12)
|
|
XCHG_SR(13)
|
|
XCHG_SR(14)
|
|
XCHG_SR(15)
|
|
|
|
/* Clear BATs. */
|
|
|
|
#define KVM_KILL_BAT(n, reg) \
|
|
mtspr SPRN_IBAT##n##U,reg; \
|
|
mtspr SPRN_IBAT##n##L,reg; \
|
|
mtspr SPRN_DBAT##n##U,reg; \
|
|
mtspr SPRN_DBAT##n##L,reg; \
|
|
|
|
li r9, 0
|
|
KVM_KILL_BAT(0, r9)
|
|
KVM_KILL_BAT(1, r9)
|
|
KVM_KILL_BAT(2, r9)
|
|
KVM_KILL_BAT(3, r9)
|
|
|
|
.endm
|
|
|
|
/******************************************************************************
|
|
* *
|
|
* Exit code *
|
|
* *
|
|
*****************************************************************************/
|
|
|
|
.macro LOAD_HOST_SEGMENTS
|
|
|
|
/* Register usage at this point:
|
|
*
|
|
* R1 = host R1
|
|
* R2 = host R2
|
|
* R12 = exit handler id
|
|
* R13 = shadow vcpu - SHADOW_VCPU_OFF
|
|
* SVCPU.* = guest *
|
|
* SVCPU[CR] = guest CR
|
|
* SVCPU[XER] = guest XER
|
|
* SVCPU[CTR] = guest CTR
|
|
* SVCPU[LR] = guest LR
|
|
*
|
|
*/
|
|
|
|
/* Restore BATs */
|
|
|
|
/* We only overwrite the upper part, so we only restoree
|
|
the upper part. */
|
|
#define KVM_LOAD_BAT(n, reg, RA, RB) \
|
|
lwz RA,(n*16)+0(reg); \
|
|
lwz RB,(n*16)+4(reg); \
|
|
mtspr SPRN_IBAT##n##U,RA; \
|
|
mtspr SPRN_IBAT##n##L,RB; \
|
|
lwz RA,(n*16)+8(reg); \
|
|
lwz RB,(n*16)+12(reg); \
|
|
mtspr SPRN_DBAT##n##U,RA; \
|
|
mtspr SPRN_DBAT##n##L,RB; \
|
|
|
|
lis r9, BATS@ha
|
|
addi r9, r9, BATS@l
|
|
tophys(r9, r9)
|
|
KVM_LOAD_BAT(0, r9, r10, r11)
|
|
KVM_LOAD_BAT(1, r9, r10, r11)
|
|
KVM_LOAD_BAT(2, r9, r10, r11)
|
|
KVM_LOAD_BAT(3, r9, r10, r11)
|
|
|
|
/* Restore Segment Registers */
|
|
|
|
/* 0xc - 0xf */
|
|
|
|
li r0, 4
|
|
mtctr r0
|
|
LOAD_REG_IMMEDIATE(r3, 0x20000000 | (0x111 * 0xc))
|
|
lis r4, 0xc000
|
|
3: mtsrin r3, r4
|
|
addi r3, r3, 0x111 /* increment VSID */
|
|
addis r4, r4, 0x1000 /* address of next segment */
|
|
bdnz 3b
|
|
|
|
/* 0x0 - 0xb */
|
|
|
|
/* 'current->mm' needs to be in r4 */
|
|
tophys(r4, r2)
|
|
lwz r4, MM(r4)
|
|
tophys(r4, r4)
|
|
/* This only clobbers r0, r3, r4 and r5 */
|
|
bl switch_mmu_context
|
|
|
|
.endm
|