cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

processor.h (5622B)


      1/* SPDX-License-Identifier: GPL-2.0 */
      2/*
      3 * AArch64 processor specific defines
      4 *
      5 * Copyright (C) 2018, Red Hat, Inc.
      6 */
      7#ifndef SELFTEST_KVM_PROCESSOR_H
      8#define SELFTEST_KVM_PROCESSOR_H
      9
     10#include "kvm_util.h"
     11#include <linux/stringify.h>
     12#include <linux/types.h>
     13#include <asm/sysreg.h>
     14
     15
     16#define ARM64_CORE_REG(x) (KVM_REG_ARM64 | KVM_REG_SIZE_U64 | \
     17			   KVM_REG_ARM_CORE | KVM_REG_ARM_CORE_REG(x))
     18
     19/*
     20 * KVM_ARM64_SYS_REG(sys_reg_id): Helper macro to convert
     21 * SYS_* register definitions in asm/sysreg.h to use in KVM
     22 * calls such as get_reg() and set_reg().
     23 */
     24#define KVM_ARM64_SYS_REG(sys_reg_id)			\
     25	ARM64_SYS_REG(sys_reg_Op0(sys_reg_id),		\
     26			sys_reg_Op1(sys_reg_id),	\
     27			sys_reg_CRn(sys_reg_id),	\
     28			sys_reg_CRm(sys_reg_id),	\
     29			sys_reg_Op2(sys_reg_id))
     30
     31/*
     32 * Default MAIR
     33 *                  index   attribute
     34 * DEVICE_nGnRnE      0     0000:0000
     35 * DEVICE_nGnRE       1     0000:0100
     36 * DEVICE_GRE         2     0000:1100
     37 * NORMAL_NC          3     0100:0100
     38 * NORMAL             4     1111:1111
     39 * NORMAL_WT          5     1011:1011
     40 */
     41#define DEFAULT_MAIR_EL1 ((0x00ul << (0 * 8)) | \
     42			  (0x04ul << (1 * 8)) | \
     43			  (0x0cul << (2 * 8)) | \
     44			  (0x44ul << (3 * 8)) | \
     45			  (0xfful << (4 * 8)) | \
     46			  (0xbbul << (5 * 8)))
     47
     48#define MPIDR_HWID_BITMASK (0xff00fffffful)
     49
     50static inline void get_reg(struct kvm_vm *vm, uint32_t vcpuid, uint64_t id, uint64_t *addr)
     51{
     52	struct kvm_one_reg reg;
     53	reg.id = id;
     54	reg.addr = (uint64_t)addr;
     55	vcpu_ioctl(vm, vcpuid, KVM_GET_ONE_REG, &reg);
     56}
     57
     58static inline void set_reg(struct kvm_vm *vm, uint32_t vcpuid, uint64_t id, uint64_t val)
     59{
     60	struct kvm_one_reg reg;
     61	reg.id = id;
     62	reg.addr = (uint64_t)&val;
     63	vcpu_ioctl(vm, vcpuid, KVM_SET_ONE_REG, &reg);
     64}
     65
     66void aarch64_vcpu_setup(struct kvm_vm *vm, uint32_t vcpuid, struct kvm_vcpu_init *init);
     67void aarch64_vcpu_add_default(struct kvm_vm *vm, uint32_t vcpuid,
     68			      struct kvm_vcpu_init *init, void *guest_code);
     69
     70struct ex_regs {
     71	u64 regs[31];
     72	u64 sp;
     73	u64 pc;
     74	u64 pstate;
     75};
     76
     77#define VECTOR_NUM	16
     78
     79enum {
     80	VECTOR_SYNC_CURRENT_SP0,
     81	VECTOR_IRQ_CURRENT_SP0,
     82	VECTOR_FIQ_CURRENT_SP0,
     83	VECTOR_ERROR_CURRENT_SP0,
     84
     85	VECTOR_SYNC_CURRENT,
     86	VECTOR_IRQ_CURRENT,
     87	VECTOR_FIQ_CURRENT,
     88	VECTOR_ERROR_CURRENT,
     89
     90	VECTOR_SYNC_LOWER_64,
     91	VECTOR_IRQ_LOWER_64,
     92	VECTOR_FIQ_LOWER_64,
     93	VECTOR_ERROR_LOWER_64,
     94
     95	VECTOR_SYNC_LOWER_32,
     96	VECTOR_IRQ_LOWER_32,
     97	VECTOR_FIQ_LOWER_32,
     98	VECTOR_ERROR_LOWER_32,
     99};
    100
    101#define VECTOR_IS_SYNC(v) ((v) == VECTOR_SYNC_CURRENT_SP0 || \
    102			   (v) == VECTOR_SYNC_CURRENT     || \
    103			   (v) == VECTOR_SYNC_LOWER_64    || \
    104			   (v) == VECTOR_SYNC_LOWER_32)
    105
    106#define ESR_EC_NUM		64
    107#define ESR_EC_SHIFT		26
    108#define ESR_EC_MASK		(ESR_EC_NUM - 1)
    109
    110#define ESR_EC_SVC64		0x15
    111#define ESR_EC_HW_BP_CURRENT	0x31
    112#define ESR_EC_SSTEP_CURRENT	0x33
    113#define ESR_EC_WP_CURRENT	0x35
    114#define ESR_EC_BRK_INS		0x3c
    115
    116void aarch64_get_supported_page_sizes(uint32_t ipa,
    117				      bool *ps4k, bool *ps16k, bool *ps64k);
    118
    119void vm_init_descriptor_tables(struct kvm_vm *vm);
    120void vcpu_init_descriptor_tables(struct kvm_vm *vm, uint32_t vcpuid);
    121
    122typedef void(*handler_fn)(struct ex_regs *);
    123void vm_install_exception_handler(struct kvm_vm *vm,
    124		int vector, handler_fn handler);
    125void vm_install_sync_handler(struct kvm_vm *vm,
    126		int vector, int ec, handler_fn handler);
    127
    128static inline void cpu_relax(void)
    129{
    130	asm volatile("yield" ::: "memory");
    131}
    132
    133#define isb()		asm volatile("isb" : : : "memory")
    134#define dsb(opt)	asm volatile("dsb " #opt : : : "memory")
    135#define dmb(opt)	asm volatile("dmb " #opt : : : "memory")
    136
    137#define dma_wmb()	dmb(oshst)
    138#define __iowmb()	dma_wmb()
    139
    140#define dma_rmb()	dmb(oshld)
    141
    142#define __iormb(v)							\
    143({									\
    144	unsigned long tmp;						\
    145									\
    146	dma_rmb();							\
    147									\
    148	/*								\
    149	 * Courtesy of arch/arm64/include/asm/io.h:			\
    150	 * Create a dummy control dependency from the IO read to any	\
    151	 * later instructions. This ensures that a subsequent call	\
    152	 * to udelay() will be ordered due to the ISB in __delay().	\
    153	 */								\
    154	asm volatile("eor	%0, %1, %1\n"				\
    155		     "cbnz	%0, ."					\
    156		     : "=r" (tmp) : "r" ((unsigned long)(v))		\
    157		     : "memory");					\
    158})
    159
    160static __always_inline void __raw_writel(u32 val, volatile void *addr)
    161{
    162	asm volatile("str %w0, [%1]" : : "rZ" (val), "r" (addr));
    163}
    164
    165static __always_inline u32 __raw_readl(const volatile void *addr)
    166{
    167	u32 val;
    168	asm volatile("ldr %w0, [%1]" : "=r" (val) : "r" (addr));
    169	return val;
    170}
    171
    172#define writel_relaxed(v,c)	((void)__raw_writel((__force u32)cpu_to_le32(v),(c)))
    173#define readl_relaxed(c)	({ u32 __r = le32_to_cpu((__force __le32)__raw_readl(c)); __r; })
    174
    175#define writel(v,c)		({ __iowmb(); writel_relaxed((v),(c));})
    176#define readl(c)		({ u32 __v = readl_relaxed(c); __iormb(__v); __v; })
    177
    178static inline void local_irq_enable(void)
    179{
    180	asm volatile("msr daifclr, #3" : : : "memory");
    181}
    182
    183static inline void local_irq_disable(void)
    184{
    185	asm volatile("msr daifset, #3" : : : "memory");
    186}
    187
    188/**
    189 * struct arm_smccc_res - Result from SMC/HVC call
    190 * @a0-a3 result values from registers 0 to 3
    191 */
    192struct arm_smccc_res {
    193	unsigned long a0;
    194	unsigned long a1;
    195	unsigned long a2;
    196	unsigned long a3;
    197};
    198
    199/**
    200 * smccc_hvc - Invoke a SMCCC function using the hvc conduit
    201 * @function_id: the SMCCC function to be called
    202 * @arg0-arg6: SMCCC function arguments, corresponding to registers x1-x7
    203 * @res: pointer to write the return values from registers x0-x3
    204 *
    205 */
    206void smccc_hvc(uint32_t function_id, uint64_t arg0, uint64_t arg1,
    207	       uint64_t arg2, uint64_t arg3, uint64_t arg4, uint64_t arg5,
    208	       uint64_t arg6, struct arm_smccc_res *res);
    209
    210#endif /* SELFTEST_KVM_PROCESSOR_H */