457 lines
		
	
	
		
			9.8 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			457 lines
		
	
	
		
			9.8 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /*
 | |
|  * Copyright (C) 2008-2011 Freescale Semiconductor, Inc. All rights reserved.
 | |
|  *
 | |
|  * Author: Yu Liu, <yu.liu@freescale.com>
 | |
|  *
 | |
|  * Description:
 | |
|  * This file is derived from arch/powerpc/kvm/44x_emulate.c,
 | |
|  * by Hollis Blanchard <hollisb@us.ibm.com>.
 | |
|  *
 | |
|  * This program is free software; you can redistribute it and/or modify
 | |
|  * it under the terms of the GNU General Public License, version 2, as
 | |
|  * published by the Free Software Foundation.
 | |
|  */
 | |
| 
 | |
| #include <asm/kvm_ppc.h>
 | |
| #include <asm/disassemble.h>
 | |
| #include <asm/dbell.h>
 | |
| #include <asm/reg_booke.h>
 | |
| 
 | |
| #include "booke.h"
 | |
| #include "e500.h"
 | |
| 
 | |
| #define XOP_DCBTLS  166
 | |
| #define XOP_MSGSND  206
 | |
| #define XOP_MSGCLR  238
 | |
| #define XOP_MFTMR   366
 | |
| #define XOP_TLBIVAX 786
 | |
| #define XOP_TLBSX   914
 | |
| #define XOP_TLBRE   946
 | |
| #define XOP_TLBWE   978
 | |
| #define XOP_TLBILX  18
 | |
| #define XOP_EHPRIV  270
 | |
| 
 | |
| #ifdef CONFIG_KVM_E500MC
 | |
| static int dbell2prio(ulong param)
 | |
| {
 | |
| 	int msg = param & PPC_DBELL_TYPE_MASK;
 | |
| 	int prio = -1;
 | |
| 
 | |
| 	switch (msg) {
 | |
| 	case PPC_DBELL_TYPE(PPC_DBELL):
 | |
| 		prio = BOOKE_IRQPRIO_DBELL;
 | |
| 		break;
 | |
| 	case PPC_DBELL_TYPE(PPC_DBELL_CRIT):
 | |
| 		prio = BOOKE_IRQPRIO_DBELL_CRIT;
 | |
| 		break;
 | |
| 	default:
 | |
| 		break;
 | |
| 	}
 | |
| 
 | |
| 	return prio;
 | |
| }
 | |
| 
 | |
| static int kvmppc_e500_emul_msgclr(struct kvm_vcpu *vcpu, int rb)
 | |
| {
 | |
| 	ulong param = vcpu->arch.regs.gpr[rb];
 | |
| 	int prio = dbell2prio(param);
 | |
| 
 | |
| 	if (prio < 0)
 | |
| 		return EMULATE_FAIL;
 | |
| 
 | |
| 	clear_bit(prio, &vcpu->arch.pending_exceptions);
 | |
| 	return EMULATE_DONE;
 | |
| }
 | |
| 
 | |
| static int kvmppc_e500_emul_msgsnd(struct kvm_vcpu *vcpu, int rb)
 | |
| {
 | |
| 	ulong param = vcpu->arch.regs.gpr[rb];
 | |
| 	int prio = dbell2prio(rb);
 | |
| 	int pir = param & PPC_DBELL_PIR_MASK;
 | |
| 	int i;
 | |
| 	struct kvm_vcpu *cvcpu;
 | |
| 
 | |
| 	if (prio < 0)
 | |
| 		return EMULATE_FAIL;
 | |
| 
 | |
| 	kvm_for_each_vcpu(i, cvcpu, vcpu->kvm) {
 | |
| 		int cpir = cvcpu->arch.shared->pir;
 | |
| 		if ((param & PPC_DBELL_MSG_BRDCAST) || (cpir == pir)) {
 | |
| 			set_bit(prio, &cvcpu->arch.pending_exceptions);
 | |
| 			kvm_vcpu_kick(cvcpu);
 | |
| 		}
 | |
| 	}
 | |
| 
 | |
| 	return EMULATE_DONE;
 | |
| }
 | |
| #endif
 | |
| 
 | |
| static int kvmppc_e500_emul_ehpriv(struct kvm_run *run, struct kvm_vcpu *vcpu,
 | |
| 				   unsigned int inst, int *advance)
 | |
| {
 | |
| 	int emulated = EMULATE_DONE;
 | |
| 
 | |
| 	switch (get_oc(inst)) {
 | |
| 	case EHPRIV_OC_DEBUG:
 | |
| 		run->exit_reason = KVM_EXIT_DEBUG;
 | |
| 		run->debug.arch.address = vcpu->arch.regs.nip;
 | |
| 		run->debug.arch.status = 0;
 | |
| 		kvmppc_account_exit(vcpu, DEBUG_EXITS);
 | |
| 		emulated = EMULATE_EXIT_USER;
 | |
| 		*advance = 0;
 | |
| 		break;
 | |
| 	default:
 | |
| 		emulated = EMULATE_FAIL;
 | |
| 	}
 | |
| 	return emulated;
 | |
| }
 | |
| 
 | |
| static int kvmppc_e500_emul_dcbtls(struct kvm_vcpu *vcpu)
 | |
| {
 | |
| 	struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
 | |
| 
 | |
| 	/* Always fail to lock the cache */
 | |
| 	vcpu_e500->l1csr0 |= L1CSR0_CUL;
 | |
| 	return EMULATE_DONE;
 | |
| }
 | |
| 
 | |
| static int kvmppc_e500_emul_mftmr(struct kvm_vcpu *vcpu, unsigned int inst,
 | |
| 				  int rt)
 | |
| {
 | |
| 	/* Expose one thread per vcpu */
 | |
| 	if (get_tmrn(inst) == TMRN_TMCFG0) {
 | |
| 		kvmppc_set_gpr(vcpu, rt,
 | |
| 			       1 | (1 << TMRN_TMCFG0_NATHRD_SHIFT));
 | |
| 		return EMULATE_DONE;
 | |
| 	}
 | |
| 
 | |
| 	return EMULATE_FAIL;
 | |
| }
 | |
| 
 | |
| int kvmppc_core_emulate_op_e500(struct kvm_run *run, struct kvm_vcpu *vcpu,
 | |
| 				unsigned int inst, int *advance)
 | |
| {
 | |
| 	int emulated = EMULATE_DONE;
 | |
| 	int ra = get_ra(inst);
 | |
| 	int rb = get_rb(inst);
 | |
| 	int rt = get_rt(inst);
 | |
| 	gva_t ea;
 | |
| 
 | |
| 	switch (get_op(inst)) {
 | |
| 	case 31:
 | |
| 		switch (get_xop(inst)) {
 | |
| 
 | |
| 		case XOP_DCBTLS:
 | |
| 			emulated = kvmppc_e500_emul_dcbtls(vcpu);
 | |
| 			break;
 | |
| 
 | |
| #ifdef CONFIG_KVM_E500MC
 | |
| 		case XOP_MSGSND:
 | |
| 			emulated = kvmppc_e500_emul_msgsnd(vcpu, rb);
 | |
| 			break;
 | |
| 
 | |
| 		case XOP_MSGCLR:
 | |
| 			emulated = kvmppc_e500_emul_msgclr(vcpu, rb);
 | |
| 			break;
 | |
| #endif
 | |
| 
 | |
| 		case XOP_TLBRE:
 | |
| 			emulated = kvmppc_e500_emul_tlbre(vcpu);
 | |
| 			break;
 | |
| 
 | |
| 		case XOP_TLBWE:
 | |
| 			emulated = kvmppc_e500_emul_tlbwe(vcpu);
 | |
| 			break;
 | |
| 
 | |
| 		case XOP_TLBSX:
 | |
| 			ea = kvmppc_get_ea_indexed(vcpu, ra, rb);
 | |
| 			emulated = kvmppc_e500_emul_tlbsx(vcpu, ea);
 | |
| 			break;
 | |
| 
 | |
| 		case XOP_TLBILX: {
 | |
| 			int type = rt & 0x3;
 | |
| 			ea = kvmppc_get_ea_indexed(vcpu, ra, rb);
 | |
| 			emulated = kvmppc_e500_emul_tlbilx(vcpu, type, ea);
 | |
| 			break;
 | |
| 		}
 | |
| 
 | |
| 		case XOP_TLBIVAX:
 | |
| 			ea = kvmppc_get_ea_indexed(vcpu, ra, rb);
 | |
| 			emulated = kvmppc_e500_emul_tlbivax(vcpu, ea);
 | |
| 			break;
 | |
| 
 | |
| 		case XOP_MFTMR:
 | |
| 			emulated = kvmppc_e500_emul_mftmr(vcpu, inst, rt);
 | |
| 			break;
 | |
| 
 | |
| 		case XOP_EHPRIV:
 | |
| 			emulated = kvmppc_e500_emul_ehpriv(run, vcpu, inst,
 | |
| 							   advance);
 | |
| 			break;
 | |
| 
 | |
| 		default:
 | |
| 			emulated = EMULATE_FAIL;
 | |
| 		}
 | |
| 
 | |
| 		break;
 | |
| 
 | |
| 	default:
 | |
| 		emulated = EMULATE_FAIL;
 | |
| 	}
 | |
| 
 | |
| 	if (emulated == EMULATE_FAIL)
 | |
| 		emulated = kvmppc_booke_emulate_op(run, vcpu, inst, advance);
 | |
| 
 | |
| 	return emulated;
 | |
| }
 | |
| 
 | |
| int kvmppc_core_emulate_mtspr_e500(struct kvm_vcpu *vcpu, int sprn, ulong spr_val)
 | |
| {
 | |
| 	struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
 | |
| 	int emulated = EMULATE_DONE;
 | |
| 
 | |
| 	switch (sprn) {
 | |
| #ifndef CONFIG_KVM_BOOKE_HV
 | |
| 	case SPRN_PID:
 | |
| 		kvmppc_set_pid(vcpu, spr_val);
 | |
| 		break;
 | |
| 	case SPRN_PID1:
 | |
| 		if (spr_val != 0)
 | |
| 			return EMULATE_FAIL;
 | |
| 		vcpu_e500->pid[1] = spr_val;
 | |
| 		break;
 | |
| 	case SPRN_PID2:
 | |
| 		if (spr_val != 0)
 | |
| 			return EMULATE_FAIL;
 | |
| 		vcpu_e500->pid[2] = spr_val;
 | |
| 		break;
 | |
| 	case SPRN_MAS0:
 | |
| 		vcpu->arch.shared->mas0 = spr_val;
 | |
| 		break;
 | |
| 	case SPRN_MAS1:
 | |
| 		vcpu->arch.shared->mas1 = spr_val;
 | |
| 		break;
 | |
| 	case SPRN_MAS2:
 | |
| 		vcpu->arch.shared->mas2 = spr_val;
 | |
| 		break;
 | |
| 	case SPRN_MAS3:
 | |
| 		vcpu->arch.shared->mas7_3 &= ~(u64)0xffffffff;
 | |
| 		vcpu->arch.shared->mas7_3 |= spr_val;
 | |
| 		break;
 | |
| 	case SPRN_MAS4:
 | |
| 		vcpu->arch.shared->mas4 = spr_val;
 | |
| 		break;
 | |
| 	case SPRN_MAS6:
 | |
| 		vcpu->arch.shared->mas6 = spr_val;
 | |
| 		break;
 | |
| 	case SPRN_MAS7:
 | |
| 		vcpu->arch.shared->mas7_3 &= (u64)0xffffffff;
 | |
| 		vcpu->arch.shared->mas7_3 |= (u64)spr_val << 32;
 | |
| 		break;
 | |
| #endif
 | |
| 	case SPRN_L1CSR0:
 | |
| 		vcpu_e500->l1csr0 = spr_val;
 | |
| 		vcpu_e500->l1csr0 &= ~(L1CSR0_DCFI | L1CSR0_CLFC);
 | |
| 		break;
 | |
| 	case SPRN_L1CSR1:
 | |
| 		vcpu_e500->l1csr1 = spr_val;
 | |
| 		vcpu_e500->l1csr1 &= ~(L1CSR1_ICFI | L1CSR1_ICLFR);
 | |
| 		break;
 | |
| 	case SPRN_HID0:
 | |
| 		vcpu_e500->hid0 = spr_val;
 | |
| 		break;
 | |
| 	case SPRN_HID1:
 | |
| 		vcpu_e500->hid1 = spr_val;
 | |
| 		break;
 | |
| 
 | |
| 	case SPRN_MMUCSR0:
 | |
| 		emulated = kvmppc_e500_emul_mt_mmucsr0(vcpu_e500,
 | |
| 				spr_val);
 | |
| 		break;
 | |
| 
 | |
| 	case SPRN_PWRMGTCR0:
 | |
| 		/*
 | |
| 		 * Guest relies on host power management configurations
 | |
| 		 * Treat the request as a general store
 | |
| 		 */
 | |
| 		vcpu->arch.pwrmgtcr0 = spr_val;
 | |
| 		break;
 | |
| 
 | |
| 	case SPRN_BUCSR:
 | |
| 		/*
 | |
| 		 * If we are here, it means that we have already flushed the
 | |
| 		 * branch predictor, so just return to guest.
 | |
| 		 */
 | |
| 		break;
 | |
| 
 | |
| 	/* extra exceptions */
 | |
| #ifdef CONFIG_SPE_POSSIBLE
 | |
| 	case SPRN_IVOR32:
 | |
| 		vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_UNAVAIL] = spr_val;
 | |
| 		break;
 | |
| 	case SPRN_IVOR33:
 | |
| 		vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_FP_DATA] = spr_val;
 | |
| 		break;
 | |
| 	case SPRN_IVOR34:
 | |
| 		vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_FP_ROUND] = spr_val;
 | |
| 		break;
 | |
| #endif
 | |
| #ifdef CONFIG_ALTIVEC
 | |
| 	case SPRN_IVOR32:
 | |
| 		vcpu->arch.ivor[BOOKE_IRQPRIO_ALTIVEC_UNAVAIL] = spr_val;
 | |
| 		break;
 | |
| 	case SPRN_IVOR33:
 | |
| 		vcpu->arch.ivor[BOOKE_IRQPRIO_ALTIVEC_ASSIST] = spr_val;
 | |
| 		break;
 | |
| #endif
 | |
| 	case SPRN_IVOR35:
 | |
| 		vcpu->arch.ivor[BOOKE_IRQPRIO_PERFORMANCE_MONITOR] = spr_val;
 | |
| 		break;
 | |
| #ifdef CONFIG_KVM_BOOKE_HV
 | |
| 	case SPRN_IVOR36:
 | |
| 		vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL] = spr_val;
 | |
| 		break;
 | |
| 	case SPRN_IVOR37:
 | |
| 		vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL_CRIT] = spr_val;
 | |
| 		break;
 | |
| #endif
 | |
| 	default:
 | |
| 		emulated = kvmppc_booke_emulate_mtspr(vcpu, sprn, spr_val);
 | |
| 	}
 | |
| 
 | |
| 	return emulated;
 | |
| }
 | |
| 
 | |
| int kvmppc_core_emulate_mfspr_e500(struct kvm_vcpu *vcpu, int sprn, ulong *spr_val)
 | |
| {
 | |
| 	struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
 | |
| 	int emulated = EMULATE_DONE;
 | |
| 
 | |
| 	switch (sprn) {
 | |
| #ifndef CONFIG_KVM_BOOKE_HV
 | |
| 	case SPRN_PID:
 | |
| 		*spr_val = vcpu_e500->pid[0];
 | |
| 		break;
 | |
| 	case SPRN_PID1:
 | |
| 		*spr_val = vcpu_e500->pid[1];
 | |
| 		break;
 | |
| 	case SPRN_PID2:
 | |
| 		*spr_val = vcpu_e500->pid[2];
 | |
| 		break;
 | |
| 	case SPRN_MAS0:
 | |
| 		*spr_val = vcpu->arch.shared->mas0;
 | |
| 		break;
 | |
| 	case SPRN_MAS1:
 | |
| 		*spr_val = vcpu->arch.shared->mas1;
 | |
| 		break;
 | |
| 	case SPRN_MAS2:
 | |
| 		*spr_val = vcpu->arch.shared->mas2;
 | |
| 		break;
 | |
| 	case SPRN_MAS3:
 | |
| 		*spr_val = (u32)vcpu->arch.shared->mas7_3;
 | |
| 		break;
 | |
| 	case SPRN_MAS4:
 | |
| 		*spr_val = vcpu->arch.shared->mas4;
 | |
| 		break;
 | |
| 	case SPRN_MAS6:
 | |
| 		*spr_val = vcpu->arch.shared->mas6;
 | |
| 		break;
 | |
| 	case SPRN_MAS7:
 | |
| 		*spr_val = vcpu->arch.shared->mas7_3 >> 32;
 | |
| 		break;
 | |
| #endif
 | |
| 	case SPRN_DECAR:
 | |
| 		*spr_val = vcpu->arch.decar;
 | |
| 		break;
 | |
| 	case SPRN_TLB0CFG:
 | |
| 		*spr_val = vcpu->arch.tlbcfg[0];
 | |
| 		break;
 | |
| 	case SPRN_TLB1CFG:
 | |
| 		*spr_val = vcpu->arch.tlbcfg[1];
 | |
| 		break;
 | |
| 	case SPRN_TLB0PS:
 | |
| 		if (!has_feature(vcpu, VCPU_FTR_MMU_V2))
 | |
| 			return EMULATE_FAIL;
 | |
| 		*spr_val = vcpu->arch.tlbps[0];
 | |
| 		break;
 | |
| 	case SPRN_TLB1PS:
 | |
| 		if (!has_feature(vcpu, VCPU_FTR_MMU_V2))
 | |
| 			return EMULATE_FAIL;
 | |
| 		*spr_val = vcpu->arch.tlbps[1];
 | |
| 		break;
 | |
| 	case SPRN_L1CSR0:
 | |
| 		*spr_val = vcpu_e500->l1csr0;
 | |
| 		break;
 | |
| 	case SPRN_L1CSR1:
 | |
| 		*spr_val = vcpu_e500->l1csr1;
 | |
| 		break;
 | |
| 	case SPRN_HID0:
 | |
| 		*spr_val = vcpu_e500->hid0;
 | |
| 		break;
 | |
| 	case SPRN_HID1:
 | |
| 		*spr_val = vcpu_e500->hid1;
 | |
| 		break;
 | |
| 	case SPRN_SVR:
 | |
| 		*spr_val = vcpu_e500->svr;
 | |
| 		break;
 | |
| 
 | |
| 	case SPRN_MMUCSR0:
 | |
| 		*spr_val = 0;
 | |
| 		break;
 | |
| 
 | |
| 	case SPRN_MMUCFG:
 | |
| 		*spr_val = vcpu->arch.mmucfg;
 | |
| 		break;
 | |
| 	case SPRN_EPTCFG:
 | |
| 		if (!has_feature(vcpu, VCPU_FTR_MMU_V2))
 | |
| 			return EMULATE_FAIL;
 | |
| 		/*
 | |
| 		 * Legacy Linux guests access EPTCFG register even if the E.PT
 | |
| 		 * category is disabled in the VM. Give them a chance to live.
 | |
| 		 */
 | |
| 		*spr_val = vcpu->arch.eptcfg;
 | |
| 		break;
 | |
| 
 | |
| 	case SPRN_PWRMGTCR0:
 | |
| 		*spr_val = vcpu->arch.pwrmgtcr0;
 | |
| 		break;
 | |
| 
 | |
| 	/* extra exceptions */
 | |
| #ifdef CONFIG_SPE_POSSIBLE
 | |
| 	case SPRN_IVOR32:
 | |
| 		*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_UNAVAIL];
 | |
| 		break;
 | |
| 	case SPRN_IVOR33:
 | |
| 		*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_FP_DATA];
 | |
| 		break;
 | |
| 	case SPRN_IVOR34:
 | |
| 		*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_FP_ROUND];
 | |
| 		break;
 | |
| #endif
 | |
| #ifdef CONFIG_ALTIVEC
 | |
| 	case SPRN_IVOR32:
 | |
| 		*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_ALTIVEC_UNAVAIL];
 | |
| 		break;
 | |
| 	case SPRN_IVOR33:
 | |
| 		*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_ALTIVEC_ASSIST];
 | |
| 		break;
 | |
| #endif
 | |
| 	case SPRN_IVOR35:
 | |
| 		*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_PERFORMANCE_MONITOR];
 | |
| 		break;
 | |
| #ifdef CONFIG_KVM_BOOKE_HV
 | |
| 	case SPRN_IVOR36:
 | |
| 		*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL];
 | |
| 		break;
 | |
| 	case SPRN_IVOR37:
 | |
| 		*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL_CRIT];
 | |
| 		break;
 | |
| #endif
 | |
| 	default:
 | |
| 		emulated = kvmppc_booke_emulate_mfspr(vcpu, sprn, spr_val);
 | |
| 	}
 | |
| 
 | |
| 	return emulated;
 | |
| }
 | |
| 
 | 
