mirror of
https://github.com/torvalds/linux.git
synced 2024-11-18 01:51:53 +00:00
8c99d34578
In the current kvm version, 'kvm_run' has been included in the 'kvm_vcpu' structure. For historical reasons, many kvm-related function parameters retain the 'kvm_run' and 'kvm_vcpu' parameters at the same time. This patch does a unified cleanup of these remaining redundant parameters. Signed-off-by: Tianjia Zhang <tianjia.zhang@linux.alibaba.com> Reviewed-by: Vitaly Kuznetsov <vkuznets@redhat.com> Reviewed-by: Paul Mackerras <paulus@ozlabs.org> Signed-off-by: Paul Mackerras <paulus@ozlabs.org>
453 lines
9.6 KiB
C
453 lines
9.6 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* Copyright (C) 2008-2011 Freescale Semiconductor, Inc. All rights reserved.
|
|
*
|
|
* Author: Yu Liu, <yu.liu@freescale.com>
|
|
*
|
|
* Description:
|
|
* This file is derived from arch/powerpc/kvm/44x_emulate.c,
|
|
* by Hollis Blanchard <hollisb@us.ibm.com>.
|
|
*/
|
|
|
|
#include <asm/kvm_ppc.h>
|
|
#include <asm/disassemble.h>
|
|
#include <asm/dbell.h>
|
|
#include <asm/reg_booke.h>
|
|
|
|
#include "booke.h"
|
|
#include "e500.h"
|
|
|
|
#define XOP_DCBTLS 166
|
|
#define XOP_MSGSND 206
|
|
#define XOP_MSGCLR 238
|
|
#define XOP_MFTMR 366
|
|
#define XOP_TLBIVAX 786
|
|
#define XOP_TLBSX 914
|
|
#define XOP_TLBRE 946
|
|
#define XOP_TLBWE 978
|
|
#define XOP_TLBILX 18
|
|
#define XOP_EHPRIV 270
|
|
|
|
#ifdef CONFIG_KVM_E500MC
|
|
static int dbell2prio(ulong param)
|
|
{
|
|
int msg = param & PPC_DBELL_TYPE_MASK;
|
|
int prio = -1;
|
|
|
|
switch (msg) {
|
|
case PPC_DBELL_TYPE(PPC_DBELL):
|
|
prio = BOOKE_IRQPRIO_DBELL;
|
|
break;
|
|
case PPC_DBELL_TYPE(PPC_DBELL_CRIT):
|
|
prio = BOOKE_IRQPRIO_DBELL_CRIT;
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
return prio;
|
|
}
|
|
|
|
static int kvmppc_e500_emul_msgclr(struct kvm_vcpu *vcpu, int rb)
|
|
{
|
|
ulong param = vcpu->arch.regs.gpr[rb];
|
|
int prio = dbell2prio(param);
|
|
|
|
if (prio < 0)
|
|
return EMULATE_FAIL;
|
|
|
|
clear_bit(prio, &vcpu->arch.pending_exceptions);
|
|
return EMULATE_DONE;
|
|
}
|
|
|
|
static int kvmppc_e500_emul_msgsnd(struct kvm_vcpu *vcpu, int rb)
|
|
{
|
|
ulong param = vcpu->arch.regs.gpr[rb];
|
|
int prio = dbell2prio(rb);
|
|
int pir = param & PPC_DBELL_PIR_MASK;
|
|
int i;
|
|
struct kvm_vcpu *cvcpu;
|
|
|
|
if (prio < 0)
|
|
return EMULATE_FAIL;
|
|
|
|
kvm_for_each_vcpu(i, cvcpu, vcpu->kvm) {
|
|
int cpir = cvcpu->arch.shared->pir;
|
|
if ((param & PPC_DBELL_MSG_BRDCAST) || (cpir == pir)) {
|
|
set_bit(prio, &cvcpu->arch.pending_exceptions);
|
|
kvm_vcpu_kick(cvcpu);
|
|
}
|
|
}
|
|
|
|
return EMULATE_DONE;
|
|
}
|
|
#endif
|
|
|
|
static int kvmppc_e500_emul_ehpriv(struct kvm_vcpu *vcpu,
|
|
unsigned int inst, int *advance)
|
|
{
|
|
int emulated = EMULATE_DONE;
|
|
|
|
switch (get_oc(inst)) {
|
|
case EHPRIV_OC_DEBUG:
|
|
vcpu->run->exit_reason = KVM_EXIT_DEBUG;
|
|
vcpu->run->debug.arch.address = vcpu->arch.regs.nip;
|
|
vcpu->run->debug.arch.status = 0;
|
|
kvmppc_account_exit(vcpu, DEBUG_EXITS);
|
|
emulated = EMULATE_EXIT_USER;
|
|
*advance = 0;
|
|
break;
|
|
default:
|
|
emulated = EMULATE_FAIL;
|
|
}
|
|
return emulated;
|
|
}
|
|
|
|
static int kvmppc_e500_emul_dcbtls(struct kvm_vcpu *vcpu)
|
|
{
|
|
struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
|
|
|
|
/* Always fail to lock the cache */
|
|
vcpu_e500->l1csr0 |= L1CSR0_CUL;
|
|
return EMULATE_DONE;
|
|
}
|
|
|
|
static int kvmppc_e500_emul_mftmr(struct kvm_vcpu *vcpu, unsigned int inst,
|
|
int rt)
|
|
{
|
|
/* Expose one thread per vcpu */
|
|
if (get_tmrn(inst) == TMRN_TMCFG0) {
|
|
kvmppc_set_gpr(vcpu, rt,
|
|
1 | (1 << TMRN_TMCFG0_NATHRD_SHIFT));
|
|
return EMULATE_DONE;
|
|
}
|
|
|
|
return EMULATE_FAIL;
|
|
}
|
|
|
|
int kvmppc_core_emulate_op_e500(struct kvm_vcpu *vcpu,
|
|
unsigned int inst, int *advance)
|
|
{
|
|
int emulated = EMULATE_DONE;
|
|
int ra = get_ra(inst);
|
|
int rb = get_rb(inst);
|
|
int rt = get_rt(inst);
|
|
gva_t ea;
|
|
|
|
switch (get_op(inst)) {
|
|
case 31:
|
|
switch (get_xop(inst)) {
|
|
|
|
case XOP_DCBTLS:
|
|
emulated = kvmppc_e500_emul_dcbtls(vcpu);
|
|
break;
|
|
|
|
#ifdef CONFIG_KVM_E500MC
|
|
case XOP_MSGSND:
|
|
emulated = kvmppc_e500_emul_msgsnd(vcpu, rb);
|
|
break;
|
|
|
|
case XOP_MSGCLR:
|
|
emulated = kvmppc_e500_emul_msgclr(vcpu, rb);
|
|
break;
|
|
#endif
|
|
|
|
case XOP_TLBRE:
|
|
emulated = kvmppc_e500_emul_tlbre(vcpu);
|
|
break;
|
|
|
|
case XOP_TLBWE:
|
|
emulated = kvmppc_e500_emul_tlbwe(vcpu);
|
|
break;
|
|
|
|
case XOP_TLBSX:
|
|
ea = kvmppc_get_ea_indexed(vcpu, ra, rb);
|
|
emulated = kvmppc_e500_emul_tlbsx(vcpu, ea);
|
|
break;
|
|
|
|
case XOP_TLBILX: {
|
|
int type = rt & 0x3;
|
|
ea = kvmppc_get_ea_indexed(vcpu, ra, rb);
|
|
emulated = kvmppc_e500_emul_tlbilx(vcpu, type, ea);
|
|
break;
|
|
}
|
|
|
|
case XOP_TLBIVAX:
|
|
ea = kvmppc_get_ea_indexed(vcpu, ra, rb);
|
|
emulated = kvmppc_e500_emul_tlbivax(vcpu, ea);
|
|
break;
|
|
|
|
case XOP_MFTMR:
|
|
emulated = kvmppc_e500_emul_mftmr(vcpu, inst, rt);
|
|
break;
|
|
|
|
case XOP_EHPRIV:
|
|
emulated = kvmppc_e500_emul_ehpriv(vcpu, inst, advance);
|
|
break;
|
|
|
|
default:
|
|
emulated = EMULATE_FAIL;
|
|
}
|
|
|
|
break;
|
|
|
|
default:
|
|
emulated = EMULATE_FAIL;
|
|
}
|
|
|
|
if (emulated == EMULATE_FAIL)
|
|
emulated = kvmppc_booke_emulate_op(vcpu, inst, advance);
|
|
|
|
return emulated;
|
|
}
|
|
|
|
int kvmppc_core_emulate_mtspr_e500(struct kvm_vcpu *vcpu, int sprn, ulong spr_val)
|
|
{
|
|
struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
|
|
int emulated = EMULATE_DONE;
|
|
|
|
switch (sprn) {
|
|
#ifndef CONFIG_KVM_BOOKE_HV
|
|
case SPRN_PID:
|
|
kvmppc_set_pid(vcpu, spr_val);
|
|
break;
|
|
case SPRN_PID1:
|
|
if (spr_val != 0)
|
|
return EMULATE_FAIL;
|
|
vcpu_e500->pid[1] = spr_val;
|
|
break;
|
|
case SPRN_PID2:
|
|
if (spr_val != 0)
|
|
return EMULATE_FAIL;
|
|
vcpu_e500->pid[2] = spr_val;
|
|
break;
|
|
case SPRN_MAS0:
|
|
vcpu->arch.shared->mas0 = spr_val;
|
|
break;
|
|
case SPRN_MAS1:
|
|
vcpu->arch.shared->mas1 = spr_val;
|
|
break;
|
|
case SPRN_MAS2:
|
|
vcpu->arch.shared->mas2 = spr_val;
|
|
break;
|
|
case SPRN_MAS3:
|
|
vcpu->arch.shared->mas7_3 &= ~(u64)0xffffffff;
|
|
vcpu->arch.shared->mas7_3 |= spr_val;
|
|
break;
|
|
case SPRN_MAS4:
|
|
vcpu->arch.shared->mas4 = spr_val;
|
|
break;
|
|
case SPRN_MAS6:
|
|
vcpu->arch.shared->mas6 = spr_val;
|
|
break;
|
|
case SPRN_MAS7:
|
|
vcpu->arch.shared->mas7_3 &= (u64)0xffffffff;
|
|
vcpu->arch.shared->mas7_3 |= (u64)spr_val << 32;
|
|
break;
|
|
#endif
|
|
case SPRN_L1CSR0:
|
|
vcpu_e500->l1csr0 = spr_val;
|
|
vcpu_e500->l1csr0 &= ~(L1CSR0_DCFI | L1CSR0_CLFC);
|
|
break;
|
|
case SPRN_L1CSR1:
|
|
vcpu_e500->l1csr1 = spr_val;
|
|
vcpu_e500->l1csr1 &= ~(L1CSR1_ICFI | L1CSR1_ICLFR);
|
|
break;
|
|
case SPRN_HID0:
|
|
vcpu_e500->hid0 = spr_val;
|
|
break;
|
|
case SPRN_HID1:
|
|
vcpu_e500->hid1 = spr_val;
|
|
break;
|
|
|
|
case SPRN_MMUCSR0:
|
|
emulated = kvmppc_e500_emul_mt_mmucsr0(vcpu_e500,
|
|
spr_val);
|
|
break;
|
|
|
|
case SPRN_PWRMGTCR0:
|
|
/*
|
|
* Guest relies on host power management configurations
|
|
* Treat the request as a general store
|
|
*/
|
|
vcpu->arch.pwrmgtcr0 = spr_val;
|
|
break;
|
|
|
|
case SPRN_BUCSR:
|
|
/*
|
|
* If we are here, it means that we have already flushed the
|
|
* branch predictor, so just return to guest.
|
|
*/
|
|
break;
|
|
|
|
/* extra exceptions */
|
|
#ifdef CONFIG_SPE_POSSIBLE
|
|
case SPRN_IVOR32:
|
|
vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_UNAVAIL] = spr_val;
|
|
break;
|
|
case SPRN_IVOR33:
|
|
vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_FP_DATA] = spr_val;
|
|
break;
|
|
case SPRN_IVOR34:
|
|
vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_FP_ROUND] = spr_val;
|
|
break;
|
|
#endif
|
|
#ifdef CONFIG_ALTIVEC
|
|
case SPRN_IVOR32:
|
|
vcpu->arch.ivor[BOOKE_IRQPRIO_ALTIVEC_UNAVAIL] = spr_val;
|
|
break;
|
|
case SPRN_IVOR33:
|
|
vcpu->arch.ivor[BOOKE_IRQPRIO_ALTIVEC_ASSIST] = spr_val;
|
|
break;
|
|
#endif
|
|
case SPRN_IVOR35:
|
|
vcpu->arch.ivor[BOOKE_IRQPRIO_PERFORMANCE_MONITOR] = spr_val;
|
|
break;
|
|
#ifdef CONFIG_KVM_BOOKE_HV
|
|
case SPRN_IVOR36:
|
|
vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL] = spr_val;
|
|
break;
|
|
case SPRN_IVOR37:
|
|
vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL_CRIT] = spr_val;
|
|
break;
|
|
#endif
|
|
default:
|
|
emulated = kvmppc_booke_emulate_mtspr(vcpu, sprn, spr_val);
|
|
}
|
|
|
|
return emulated;
|
|
}
|
|
|
|
int kvmppc_core_emulate_mfspr_e500(struct kvm_vcpu *vcpu, int sprn, ulong *spr_val)
|
|
{
|
|
struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
|
|
int emulated = EMULATE_DONE;
|
|
|
|
switch (sprn) {
|
|
#ifndef CONFIG_KVM_BOOKE_HV
|
|
case SPRN_PID:
|
|
*spr_val = vcpu_e500->pid[0];
|
|
break;
|
|
case SPRN_PID1:
|
|
*spr_val = vcpu_e500->pid[1];
|
|
break;
|
|
case SPRN_PID2:
|
|
*spr_val = vcpu_e500->pid[2];
|
|
break;
|
|
case SPRN_MAS0:
|
|
*spr_val = vcpu->arch.shared->mas0;
|
|
break;
|
|
case SPRN_MAS1:
|
|
*spr_val = vcpu->arch.shared->mas1;
|
|
break;
|
|
case SPRN_MAS2:
|
|
*spr_val = vcpu->arch.shared->mas2;
|
|
break;
|
|
case SPRN_MAS3:
|
|
*spr_val = (u32)vcpu->arch.shared->mas7_3;
|
|
break;
|
|
case SPRN_MAS4:
|
|
*spr_val = vcpu->arch.shared->mas4;
|
|
break;
|
|
case SPRN_MAS6:
|
|
*spr_val = vcpu->arch.shared->mas6;
|
|
break;
|
|
case SPRN_MAS7:
|
|
*spr_val = vcpu->arch.shared->mas7_3 >> 32;
|
|
break;
|
|
#endif
|
|
case SPRN_DECAR:
|
|
*spr_val = vcpu->arch.decar;
|
|
break;
|
|
case SPRN_TLB0CFG:
|
|
*spr_val = vcpu->arch.tlbcfg[0];
|
|
break;
|
|
case SPRN_TLB1CFG:
|
|
*spr_val = vcpu->arch.tlbcfg[1];
|
|
break;
|
|
case SPRN_TLB0PS:
|
|
if (!has_feature(vcpu, VCPU_FTR_MMU_V2))
|
|
return EMULATE_FAIL;
|
|
*spr_val = vcpu->arch.tlbps[0];
|
|
break;
|
|
case SPRN_TLB1PS:
|
|
if (!has_feature(vcpu, VCPU_FTR_MMU_V2))
|
|
return EMULATE_FAIL;
|
|
*spr_val = vcpu->arch.tlbps[1];
|
|
break;
|
|
case SPRN_L1CSR0:
|
|
*spr_val = vcpu_e500->l1csr0;
|
|
break;
|
|
case SPRN_L1CSR1:
|
|
*spr_val = vcpu_e500->l1csr1;
|
|
break;
|
|
case SPRN_HID0:
|
|
*spr_val = vcpu_e500->hid0;
|
|
break;
|
|
case SPRN_HID1:
|
|
*spr_val = vcpu_e500->hid1;
|
|
break;
|
|
case SPRN_SVR:
|
|
*spr_val = vcpu_e500->svr;
|
|
break;
|
|
|
|
case SPRN_MMUCSR0:
|
|
*spr_val = 0;
|
|
break;
|
|
|
|
case SPRN_MMUCFG:
|
|
*spr_val = vcpu->arch.mmucfg;
|
|
break;
|
|
case SPRN_EPTCFG:
|
|
if (!has_feature(vcpu, VCPU_FTR_MMU_V2))
|
|
return EMULATE_FAIL;
|
|
/*
|
|
* Legacy Linux guests access EPTCFG register even if the E.PT
|
|
* category is disabled in the VM. Give them a chance to live.
|
|
*/
|
|
*spr_val = vcpu->arch.eptcfg;
|
|
break;
|
|
|
|
case SPRN_PWRMGTCR0:
|
|
*spr_val = vcpu->arch.pwrmgtcr0;
|
|
break;
|
|
|
|
/* extra exceptions */
|
|
#ifdef CONFIG_SPE_POSSIBLE
|
|
case SPRN_IVOR32:
|
|
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_UNAVAIL];
|
|
break;
|
|
case SPRN_IVOR33:
|
|
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_FP_DATA];
|
|
break;
|
|
case SPRN_IVOR34:
|
|
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_FP_ROUND];
|
|
break;
|
|
#endif
|
|
#ifdef CONFIG_ALTIVEC
|
|
case SPRN_IVOR32:
|
|
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_ALTIVEC_UNAVAIL];
|
|
break;
|
|
case SPRN_IVOR33:
|
|
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_ALTIVEC_ASSIST];
|
|
break;
|
|
#endif
|
|
case SPRN_IVOR35:
|
|
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_PERFORMANCE_MONITOR];
|
|
break;
|
|
#ifdef CONFIG_KVM_BOOKE_HV
|
|
case SPRN_IVOR36:
|
|
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL];
|
|
break;
|
|
case SPRN_IVOR37:
|
|
*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL_CRIT];
|
|
break;
|
|
#endif
|
|
default:
|
|
emulated = kvmppc_booke_emulate_mfspr(vcpu, sprn, spr_val);
|
|
}
|
|
|
|
return emulated;
|
|
}
|
|
|