cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

hv_spinlock.c (2272B)


      1// SPDX-License-Identifier: GPL-2.0
      2
      3/*
      4 * Hyper-V specific spinlock code.
      5 *
      6 * Copyright (C) 2018, Intel, Inc.
      7 *
      8 * Author : Yi Sun <yi.y.sun@intel.com>
      9 */
     10
     11#define pr_fmt(fmt) "Hyper-V: " fmt
     12
     13#include <linux/spinlock.h>
     14
     15#include <asm/mshyperv.h>
     16#include <asm/paravirt.h>
     17#include <asm/apic.h>
     18
     19static bool __initdata hv_pvspin = true;
     20
     21static void hv_qlock_kick(int cpu)
     22{
     23	apic->send_IPI(cpu, X86_PLATFORM_IPI_VECTOR);
     24}
     25
     26static void hv_qlock_wait(u8 *byte, u8 val)
     27{
     28	unsigned long flags;
     29
     30	if (in_nmi())
     31		return;
     32
     33	/*
     34	 * Reading HV_X64_MSR_GUEST_IDLE MSR tells the hypervisor that the
     35	 * vCPU can be put into 'idle' state. This 'idle' state is
     36	 * terminated by an IPI, usually from hv_qlock_kick(), even if
     37	 * interrupts are disabled on the vCPU.
     38	 *
     39	 * To prevent a race against the unlock path it is required to
     40	 * disable interrupts before accessing the HV_X64_MSR_GUEST_IDLE
     41	 * MSR. Otherwise, if the IPI from hv_qlock_kick() arrives between
     42	 * the lock value check and the rdmsrl() then the vCPU might be put
     43	 * into 'idle' state by the hypervisor and kept in that state for
     44	 * an unspecified amount of time.
     45	 */
     46	local_irq_save(flags);
     47	/*
     48	 * Only issue the rdmsrl() when the lock state has not changed.
     49	 */
     50	if (READ_ONCE(*byte) == val) {
     51		unsigned long msr_val;
     52
     53		rdmsrl(HV_X64_MSR_GUEST_IDLE, msr_val);
     54
     55		(void)msr_val;
     56	}
     57	local_irq_restore(flags);
     58}
     59
     60/*
     61 * Hyper-V does not support this so far.
     62 */
     63__visible bool hv_vcpu_is_preempted(int vcpu)
     64{
     65	return false;
     66}
     67PV_CALLEE_SAVE_REGS_THUNK(hv_vcpu_is_preempted);
     68
     69void __init hv_init_spinlocks(void)
     70{
     71	if (!hv_pvspin || !apic ||
     72	    !(ms_hyperv.hints & HV_X64_CLUSTER_IPI_RECOMMENDED) ||
     73	    !(ms_hyperv.features & HV_MSR_GUEST_IDLE_AVAILABLE)) {
     74		pr_info("PV spinlocks disabled\n");
     75		return;
     76	}
     77	pr_info("PV spinlocks enabled\n");
     78
     79	__pv_init_lock_hash();
     80	pv_ops.lock.queued_spin_lock_slowpath = __pv_queued_spin_lock_slowpath;
     81	pv_ops.lock.queued_spin_unlock = PV_CALLEE_SAVE(__pv_queued_spin_unlock);
     82	pv_ops.lock.wait = hv_qlock_wait;
     83	pv_ops.lock.kick = hv_qlock_kick;
     84	pv_ops.lock.vcpu_is_preempted = PV_CALLEE_SAVE(hv_vcpu_is_preempted);
     85}
     86
     87static __init int hv_parse_nopvspin(char *arg)
     88{
     89	hv_pvspin = false;
     90	return 0;
     91}
     92early_param("hv_nopvspin", hv_parse_nopvspin);