cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

e500_emulate.c (9851B)


      1// SPDX-License-Identifier: GPL-2.0-only
      2/*
      3 * Copyright (C) 2008-2011 Freescale Semiconductor, Inc. All rights reserved.
      4 *
      5 * Author: Yu Liu, <yu.liu@freescale.com>
      6 *
      7 * Description:
      8 * This file is derived from arch/powerpc/kvm/44x_emulate.c,
      9 * by Hollis Blanchard <hollisb@us.ibm.com>.
     10 */
     11
     12#include <asm/kvm_ppc.h>
     13#include <asm/disassemble.h>
     14#include <asm/dbell.h>
     15#include <asm/reg_booke.h>
     16
     17#include "booke.h"
     18#include "e500.h"
     19
     20#define XOP_DCBTLS  166
     21#define XOP_MSGSND  206
     22#define XOP_MSGCLR  238
     23#define XOP_MFTMR   366
     24#define XOP_TLBIVAX 786
     25#define XOP_TLBSX   914
     26#define XOP_TLBRE   946
     27#define XOP_TLBWE   978
     28#define XOP_TLBILX  18
     29#define XOP_EHPRIV  270
     30
     31#ifdef CONFIG_KVM_E500MC
     32static int dbell2prio(ulong param)
     33{
     34	int msg = param & PPC_DBELL_TYPE_MASK;
     35	int prio = -1;
     36
     37	switch (msg) {
     38	case PPC_DBELL_TYPE(PPC_DBELL):
     39		prio = BOOKE_IRQPRIO_DBELL;
     40		break;
     41	case PPC_DBELL_TYPE(PPC_DBELL_CRIT):
     42		prio = BOOKE_IRQPRIO_DBELL_CRIT;
     43		break;
     44	default:
     45		break;
     46	}
     47
     48	return prio;
     49}
     50
     51static int kvmppc_e500_emul_msgclr(struct kvm_vcpu *vcpu, int rb)
     52{
     53	ulong param = vcpu->arch.regs.gpr[rb];
     54	int prio = dbell2prio(param);
     55
     56	if (prio < 0)
     57		return EMULATE_FAIL;
     58
     59	clear_bit(prio, &vcpu->arch.pending_exceptions);
     60	return EMULATE_DONE;
     61}
     62
     63static int kvmppc_e500_emul_msgsnd(struct kvm_vcpu *vcpu, int rb)
     64{
     65	ulong param = vcpu->arch.regs.gpr[rb];
     66	int prio = dbell2prio(rb);
     67	int pir = param & PPC_DBELL_PIR_MASK;
     68	unsigned long i;
     69	struct kvm_vcpu *cvcpu;
     70
     71	if (prio < 0)
     72		return EMULATE_FAIL;
     73
     74	kvm_for_each_vcpu(i, cvcpu, vcpu->kvm) {
     75		int cpir = cvcpu->arch.shared->pir;
     76		if ((param & PPC_DBELL_MSG_BRDCAST) || (cpir == pir)) {
     77			set_bit(prio, &cvcpu->arch.pending_exceptions);
     78			kvm_vcpu_kick(cvcpu);
     79		}
     80	}
     81
     82	return EMULATE_DONE;
     83}
     84#endif
     85
     86static int kvmppc_e500_emul_ehpriv(struct kvm_vcpu *vcpu,
     87				   unsigned int inst, int *advance)
     88{
     89	int emulated = EMULATE_DONE;
     90
     91	switch (get_oc(inst)) {
     92	case EHPRIV_OC_DEBUG:
     93		vcpu->run->exit_reason = KVM_EXIT_DEBUG;
     94		vcpu->run->debug.arch.address = vcpu->arch.regs.nip;
     95		vcpu->run->debug.arch.status = 0;
     96		kvmppc_account_exit(vcpu, DEBUG_EXITS);
     97		emulated = EMULATE_EXIT_USER;
     98		*advance = 0;
     99		break;
    100	default:
    101		emulated = EMULATE_FAIL;
    102	}
    103	return emulated;
    104}
    105
    106static int kvmppc_e500_emul_dcbtls(struct kvm_vcpu *vcpu)
    107{
    108	struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
    109
    110	/* Always fail to lock the cache */
    111	vcpu_e500->l1csr0 |= L1CSR0_CUL;
    112	return EMULATE_DONE;
    113}
    114
    115static int kvmppc_e500_emul_mftmr(struct kvm_vcpu *vcpu, unsigned int inst,
    116				  int rt)
    117{
    118	/* Expose one thread per vcpu */
    119	if (get_tmrn(inst) == TMRN_TMCFG0) {
    120		kvmppc_set_gpr(vcpu, rt,
    121			       1 | (1 << TMRN_TMCFG0_NATHRD_SHIFT));
    122		return EMULATE_DONE;
    123	}
    124
    125	return EMULATE_FAIL;
    126}
    127
    128int kvmppc_core_emulate_op_e500(struct kvm_vcpu *vcpu,
    129				unsigned int inst, int *advance)
    130{
    131	int emulated = EMULATE_DONE;
    132	int ra = get_ra(inst);
    133	int rb = get_rb(inst);
    134	int rt = get_rt(inst);
    135	gva_t ea;
    136
    137	switch (get_op(inst)) {
    138	case 31:
    139		switch (get_xop(inst)) {
    140
    141		case XOP_DCBTLS:
    142			emulated = kvmppc_e500_emul_dcbtls(vcpu);
    143			break;
    144
    145#ifdef CONFIG_KVM_E500MC
    146		case XOP_MSGSND:
    147			emulated = kvmppc_e500_emul_msgsnd(vcpu, rb);
    148			break;
    149
    150		case XOP_MSGCLR:
    151			emulated = kvmppc_e500_emul_msgclr(vcpu, rb);
    152			break;
    153#endif
    154
    155		case XOP_TLBRE:
    156			emulated = kvmppc_e500_emul_tlbre(vcpu);
    157			break;
    158
    159		case XOP_TLBWE:
    160			emulated = kvmppc_e500_emul_tlbwe(vcpu);
    161			break;
    162
    163		case XOP_TLBSX:
    164			ea = kvmppc_get_ea_indexed(vcpu, ra, rb);
    165			emulated = kvmppc_e500_emul_tlbsx(vcpu, ea);
    166			break;
    167
    168		case XOP_TLBILX: {
    169			int type = rt & 0x3;
    170			ea = kvmppc_get_ea_indexed(vcpu, ra, rb);
    171			emulated = kvmppc_e500_emul_tlbilx(vcpu, type, ea);
    172			break;
    173		}
    174
    175		case XOP_TLBIVAX:
    176			ea = kvmppc_get_ea_indexed(vcpu, ra, rb);
    177			emulated = kvmppc_e500_emul_tlbivax(vcpu, ea);
    178			break;
    179
    180		case XOP_MFTMR:
    181			emulated = kvmppc_e500_emul_mftmr(vcpu, inst, rt);
    182			break;
    183
    184		case XOP_EHPRIV:
    185			emulated = kvmppc_e500_emul_ehpriv(vcpu, inst, advance);
    186			break;
    187
    188		default:
    189			emulated = EMULATE_FAIL;
    190		}
    191
    192		break;
    193
    194	default:
    195		emulated = EMULATE_FAIL;
    196	}
    197
    198	if (emulated == EMULATE_FAIL)
    199		emulated = kvmppc_booke_emulate_op(vcpu, inst, advance);
    200
    201	return emulated;
    202}
    203
    204int kvmppc_core_emulate_mtspr_e500(struct kvm_vcpu *vcpu, int sprn, ulong spr_val)
    205{
    206	struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
    207	int emulated = EMULATE_DONE;
    208
    209	switch (sprn) {
    210#ifndef CONFIG_KVM_BOOKE_HV
    211	case SPRN_PID:
    212		kvmppc_set_pid(vcpu, spr_val);
    213		break;
    214	case SPRN_PID1:
    215		if (spr_val != 0)
    216			return EMULATE_FAIL;
    217		vcpu_e500->pid[1] = spr_val;
    218		break;
    219	case SPRN_PID2:
    220		if (spr_val != 0)
    221			return EMULATE_FAIL;
    222		vcpu_e500->pid[2] = spr_val;
    223		break;
    224	case SPRN_MAS0:
    225		vcpu->arch.shared->mas0 = spr_val;
    226		break;
    227	case SPRN_MAS1:
    228		vcpu->arch.shared->mas1 = spr_val;
    229		break;
    230	case SPRN_MAS2:
    231		vcpu->arch.shared->mas2 = spr_val;
    232		break;
    233	case SPRN_MAS3:
    234		vcpu->arch.shared->mas7_3 &= ~(u64)0xffffffff;
    235		vcpu->arch.shared->mas7_3 |= spr_val;
    236		break;
    237	case SPRN_MAS4:
    238		vcpu->arch.shared->mas4 = spr_val;
    239		break;
    240	case SPRN_MAS6:
    241		vcpu->arch.shared->mas6 = spr_val;
    242		break;
    243	case SPRN_MAS7:
    244		vcpu->arch.shared->mas7_3 &= (u64)0xffffffff;
    245		vcpu->arch.shared->mas7_3 |= (u64)spr_val << 32;
    246		break;
    247#endif
    248	case SPRN_L1CSR0:
    249		vcpu_e500->l1csr0 = spr_val;
    250		vcpu_e500->l1csr0 &= ~(L1CSR0_DCFI | L1CSR0_CLFC);
    251		break;
    252	case SPRN_L1CSR1:
    253		vcpu_e500->l1csr1 = spr_val;
    254		vcpu_e500->l1csr1 &= ~(L1CSR1_ICFI | L1CSR1_ICLFR);
    255		break;
    256	case SPRN_HID0:
    257		vcpu_e500->hid0 = spr_val;
    258		break;
    259	case SPRN_HID1:
    260		vcpu_e500->hid1 = spr_val;
    261		break;
    262
    263	case SPRN_MMUCSR0:
    264		emulated = kvmppc_e500_emul_mt_mmucsr0(vcpu_e500,
    265				spr_val);
    266		break;
    267
    268	case SPRN_PWRMGTCR0:
    269		/*
    270		 * Guest relies on host power management configurations
    271		 * Treat the request as a general store
    272		 */
    273		vcpu->arch.pwrmgtcr0 = spr_val;
    274		break;
    275
    276	case SPRN_BUCSR:
    277		/*
    278		 * If we are here, it means that we have already flushed the
    279		 * branch predictor, so just return to guest.
    280		 */
    281		break;
    282
    283	/* extra exceptions */
    284#ifdef CONFIG_SPE_POSSIBLE
    285	case SPRN_IVOR32:
    286		vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_UNAVAIL] = spr_val;
    287		break;
    288	case SPRN_IVOR33:
    289		vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_FP_DATA] = spr_val;
    290		break;
    291	case SPRN_IVOR34:
    292		vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_FP_ROUND] = spr_val;
    293		break;
    294#endif
    295#ifdef CONFIG_ALTIVEC
    296	case SPRN_IVOR32:
    297		vcpu->arch.ivor[BOOKE_IRQPRIO_ALTIVEC_UNAVAIL] = spr_val;
    298		break;
    299	case SPRN_IVOR33:
    300		vcpu->arch.ivor[BOOKE_IRQPRIO_ALTIVEC_ASSIST] = spr_val;
    301		break;
    302#endif
    303	case SPRN_IVOR35:
    304		vcpu->arch.ivor[BOOKE_IRQPRIO_PERFORMANCE_MONITOR] = spr_val;
    305		break;
    306#ifdef CONFIG_KVM_BOOKE_HV
    307	case SPRN_IVOR36:
    308		vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL] = spr_val;
    309		break;
    310	case SPRN_IVOR37:
    311		vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL_CRIT] = spr_val;
    312		break;
    313#endif
    314	default:
    315		emulated = kvmppc_booke_emulate_mtspr(vcpu, sprn, spr_val);
    316	}
    317
    318	return emulated;
    319}
    320
    321int kvmppc_core_emulate_mfspr_e500(struct kvm_vcpu *vcpu, int sprn, ulong *spr_val)
    322{
    323	struct kvmppc_vcpu_e500 *vcpu_e500 = to_e500(vcpu);
    324	int emulated = EMULATE_DONE;
    325
    326	switch (sprn) {
    327#ifndef CONFIG_KVM_BOOKE_HV
    328	case SPRN_PID:
    329		*spr_val = vcpu_e500->pid[0];
    330		break;
    331	case SPRN_PID1:
    332		*spr_val = vcpu_e500->pid[1];
    333		break;
    334	case SPRN_PID2:
    335		*spr_val = vcpu_e500->pid[2];
    336		break;
    337	case SPRN_MAS0:
    338		*spr_val = vcpu->arch.shared->mas0;
    339		break;
    340	case SPRN_MAS1:
    341		*spr_val = vcpu->arch.shared->mas1;
    342		break;
    343	case SPRN_MAS2:
    344		*spr_val = vcpu->arch.shared->mas2;
    345		break;
    346	case SPRN_MAS3:
    347		*spr_val = (u32)vcpu->arch.shared->mas7_3;
    348		break;
    349	case SPRN_MAS4:
    350		*spr_val = vcpu->arch.shared->mas4;
    351		break;
    352	case SPRN_MAS6:
    353		*spr_val = vcpu->arch.shared->mas6;
    354		break;
    355	case SPRN_MAS7:
    356		*spr_val = vcpu->arch.shared->mas7_3 >> 32;
    357		break;
    358#endif
    359	case SPRN_DECAR:
    360		*spr_val = vcpu->arch.decar;
    361		break;
    362	case SPRN_TLB0CFG:
    363		*spr_val = vcpu->arch.tlbcfg[0];
    364		break;
    365	case SPRN_TLB1CFG:
    366		*spr_val = vcpu->arch.tlbcfg[1];
    367		break;
    368	case SPRN_TLB0PS:
    369		if (!has_feature(vcpu, VCPU_FTR_MMU_V2))
    370			return EMULATE_FAIL;
    371		*spr_val = vcpu->arch.tlbps[0];
    372		break;
    373	case SPRN_TLB1PS:
    374		if (!has_feature(vcpu, VCPU_FTR_MMU_V2))
    375			return EMULATE_FAIL;
    376		*spr_val = vcpu->arch.tlbps[1];
    377		break;
    378	case SPRN_L1CSR0:
    379		*spr_val = vcpu_e500->l1csr0;
    380		break;
    381	case SPRN_L1CSR1:
    382		*spr_val = vcpu_e500->l1csr1;
    383		break;
    384	case SPRN_HID0:
    385		*spr_val = vcpu_e500->hid0;
    386		break;
    387	case SPRN_HID1:
    388		*spr_val = vcpu_e500->hid1;
    389		break;
    390	case SPRN_SVR:
    391		*spr_val = vcpu_e500->svr;
    392		break;
    393
    394	case SPRN_MMUCSR0:
    395		*spr_val = 0;
    396		break;
    397
    398	case SPRN_MMUCFG:
    399		*spr_val = vcpu->arch.mmucfg;
    400		break;
    401	case SPRN_EPTCFG:
    402		if (!has_feature(vcpu, VCPU_FTR_MMU_V2))
    403			return EMULATE_FAIL;
    404		/*
    405		 * Legacy Linux guests access EPTCFG register even if the E.PT
    406		 * category is disabled in the VM. Give them a chance to live.
    407		 */
    408		*spr_val = vcpu->arch.eptcfg;
    409		break;
    410
    411	case SPRN_PWRMGTCR0:
    412		*spr_val = vcpu->arch.pwrmgtcr0;
    413		break;
    414
    415	/* extra exceptions */
    416#ifdef CONFIG_SPE_POSSIBLE
    417	case SPRN_IVOR32:
    418		*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_UNAVAIL];
    419		break;
    420	case SPRN_IVOR33:
    421		*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_FP_DATA];
    422		break;
    423	case SPRN_IVOR34:
    424		*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_SPE_FP_ROUND];
    425		break;
    426#endif
    427#ifdef CONFIG_ALTIVEC
    428	case SPRN_IVOR32:
    429		*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_ALTIVEC_UNAVAIL];
    430		break;
    431	case SPRN_IVOR33:
    432		*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_ALTIVEC_ASSIST];
    433		break;
    434#endif
    435	case SPRN_IVOR35:
    436		*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_PERFORMANCE_MONITOR];
    437		break;
    438#ifdef CONFIG_KVM_BOOKE_HV
    439	case SPRN_IVOR36:
    440		*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL];
    441		break;
    442	case SPRN_IVOR37:
    443		*spr_val = vcpu->arch.ivor[BOOKE_IRQPRIO_DBELL_CRIT];
    444		break;
    445#endif
    446	default:
    447		emulated = kvmppc_booke_emulate_mfspr(vcpu, sprn, spr_val);
    448	}
    449
    450	return emulated;
    451}
    452