cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

cfi.c (8433B)


      1// SPDX-License-Identifier: GPL-2.0
      2/*
      3 * Clang Control Flow Integrity (CFI) error and slowpath handling.
      4 *
      5 * Copyright (C) 2021 Google LLC
      6 */
      7
      8#include <linux/hardirq.h>
      9#include <linux/kallsyms.h>
     10#include <linux/module.h>
     11#include <linux/mutex.h>
     12#include <linux/printk.h>
     13#include <linux/ratelimit.h>
     14#include <linux/rcupdate.h>
     15#include <linux/vmalloc.h>
     16#include <asm/cacheflush.h>
     17#include <asm/set_memory.h>
     18
     19/* Compiler-defined handler names */
     20#ifdef CONFIG_CFI_PERMISSIVE
     21#define cfi_failure_handler	__ubsan_handle_cfi_check_fail
     22#else
     23#define cfi_failure_handler	__ubsan_handle_cfi_check_fail_abort
     24#endif
     25
     26static inline void handle_cfi_failure(void *ptr)
     27{
     28	if (IS_ENABLED(CONFIG_CFI_PERMISSIVE))
     29		WARN_RATELIMIT(1, "CFI failure (target: %pS):\n", ptr);
     30	else
     31		panic("CFI failure (target: %pS)\n", ptr);
     32}
     33
     34#ifdef CONFIG_MODULES
     35#ifdef CONFIG_CFI_CLANG_SHADOW
     36/*
     37 * Index type. A 16-bit index can address at most (2^16)-2 pages (taking
     38 * into account SHADOW_INVALID), i.e. ~256M with 4k pages.
     39 */
     40typedef u16 shadow_t;
     41#define SHADOW_INVALID		((shadow_t)~0UL)
     42
     43struct cfi_shadow {
     44	/* Page index for the beginning of the shadow */
     45	unsigned long base;
     46	/* An array of __cfi_check locations (as indices to the shadow) */
     47	shadow_t shadow[1];
     48} __packed;
     49
     50/*
     51 * The shadow covers ~128M from the beginning of the module region. If
     52 * the region is larger, we fall back to __module_address for the rest.
     53 */
     54#define __SHADOW_RANGE		(_UL(SZ_128M) >> PAGE_SHIFT)
     55
     56/* The in-memory size of struct cfi_shadow, always at least one page */
     57#define __SHADOW_PAGES		((__SHADOW_RANGE * sizeof(shadow_t)) >> PAGE_SHIFT)
     58#define SHADOW_PAGES		max(1UL, __SHADOW_PAGES)
     59#define SHADOW_SIZE		(SHADOW_PAGES << PAGE_SHIFT)
     60
     61/* The actual size of the shadow array, minus metadata */
     62#define SHADOW_ARR_SIZE		(SHADOW_SIZE - offsetof(struct cfi_shadow, shadow))
     63#define SHADOW_ARR_SLOTS	(SHADOW_ARR_SIZE / sizeof(shadow_t))
     64
     65static DEFINE_MUTEX(shadow_update_lock);
     66static struct cfi_shadow __rcu *cfi_shadow __read_mostly;
     67
     68/* Returns the index in the shadow for the given address */
     69static inline int ptr_to_shadow(const struct cfi_shadow *s, unsigned long ptr)
     70{
     71	unsigned long index;
     72	unsigned long page = ptr >> PAGE_SHIFT;
     73
     74	if (unlikely(page < s->base))
     75		return -1; /* Outside of module area */
     76
     77	index = page - s->base;
     78
     79	if (index >= SHADOW_ARR_SLOTS)
     80		return -1; /* Cannot be addressed with shadow */
     81
     82	return (int)index;
     83}
     84
     85/* Returns the page address for an index in the shadow */
     86static inline unsigned long shadow_to_ptr(const struct cfi_shadow *s,
     87	int index)
     88{
     89	if (unlikely(index < 0 || index >= SHADOW_ARR_SLOTS))
     90		return 0;
     91
     92	return (s->base + index) << PAGE_SHIFT;
     93}
     94
     95/* Returns the __cfi_check function address for the given shadow location */
     96static inline unsigned long shadow_to_check_fn(const struct cfi_shadow *s,
     97	int index)
     98{
     99	if (unlikely(index < 0 || index >= SHADOW_ARR_SLOTS))
    100		return 0;
    101
    102	if (unlikely(s->shadow[index] == SHADOW_INVALID))
    103		return 0;
    104
    105	/* __cfi_check is always page aligned */
    106	return (s->base + s->shadow[index]) << PAGE_SHIFT;
    107}
    108
    109static void prepare_next_shadow(const struct cfi_shadow __rcu *prev,
    110		struct cfi_shadow *next)
    111{
    112	int i, index, check;
    113
    114	/* Mark everything invalid */
    115	memset(next->shadow, 0xFF, SHADOW_ARR_SIZE);
    116
    117	if (!prev)
    118		return; /* No previous shadow */
    119
    120	/* If the base address didn't change, an update is not needed */
    121	if (prev->base == next->base) {
    122		memcpy(next->shadow, prev->shadow, SHADOW_ARR_SIZE);
    123		return;
    124	}
    125
    126	/* Convert the previous shadow to the new address range */
    127	for (i = 0; i < SHADOW_ARR_SLOTS; ++i) {
    128		if (prev->shadow[i] == SHADOW_INVALID)
    129			continue;
    130
    131		index = ptr_to_shadow(next, shadow_to_ptr(prev, i));
    132		if (index < 0)
    133			continue;
    134
    135		check = ptr_to_shadow(next,
    136				shadow_to_check_fn(prev, prev->shadow[i]));
    137		if (check < 0)
    138			continue;
    139
    140		next->shadow[index] = (shadow_t)check;
    141	}
    142}
    143
    144static void add_module_to_shadow(struct cfi_shadow *s, struct module *mod,
    145			unsigned long min_addr, unsigned long max_addr)
    146{
    147	int check_index;
    148	unsigned long check = (unsigned long)mod->cfi_check;
    149	unsigned long ptr;
    150
    151	if (unlikely(!PAGE_ALIGNED(check))) {
    152		pr_warn("cfi: not using shadow for module %s\n", mod->name);
    153		return;
    154	}
    155
    156	check_index = ptr_to_shadow(s, check);
    157	if (check_index < 0)
    158		return; /* Module not addressable with shadow */
    159
    160	/* For each page, store the check function index in the shadow */
    161	for (ptr = min_addr; ptr <= max_addr; ptr += PAGE_SIZE) {
    162		int index = ptr_to_shadow(s, ptr);
    163
    164		if (index >= 0) {
    165			/* Each page must only contain one module */
    166			WARN_ON_ONCE(s->shadow[index] != SHADOW_INVALID);
    167			s->shadow[index] = (shadow_t)check_index;
    168		}
    169	}
    170}
    171
    172static void remove_module_from_shadow(struct cfi_shadow *s, struct module *mod,
    173		unsigned long min_addr, unsigned long max_addr)
    174{
    175	unsigned long ptr;
    176
    177	for (ptr = min_addr; ptr <= max_addr; ptr += PAGE_SIZE) {
    178		int index = ptr_to_shadow(s, ptr);
    179
    180		if (index >= 0)
    181			s->shadow[index] = SHADOW_INVALID;
    182	}
    183}
    184
    185typedef void (*update_shadow_fn)(struct cfi_shadow *, struct module *,
    186			unsigned long min_addr, unsigned long max_addr);
    187
    188static void update_shadow(struct module *mod, unsigned long base_addr,
    189		update_shadow_fn fn)
    190{
    191	struct cfi_shadow *prev;
    192	struct cfi_shadow *next;
    193	unsigned long min_addr, max_addr;
    194
    195	next = vmalloc(SHADOW_SIZE);
    196
    197	mutex_lock(&shadow_update_lock);
    198	prev = rcu_dereference_protected(cfi_shadow,
    199					 mutex_is_locked(&shadow_update_lock));
    200
    201	if (next) {
    202		next->base = base_addr >> PAGE_SHIFT;
    203		prepare_next_shadow(prev, next);
    204
    205		min_addr = (unsigned long)mod->core_layout.base;
    206		max_addr = min_addr + mod->core_layout.text_size;
    207		fn(next, mod, min_addr & PAGE_MASK, max_addr & PAGE_MASK);
    208
    209		set_memory_ro((unsigned long)next, SHADOW_PAGES);
    210	}
    211
    212	rcu_assign_pointer(cfi_shadow, next);
    213	mutex_unlock(&shadow_update_lock);
    214	synchronize_rcu();
    215
    216	if (prev) {
    217		set_memory_rw((unsigned long)prev, SHADOW_PAGES);
    218		vfree(prev);
    219	}
    220}
    221
    222void cfi_module_add(struct module *mod, unsigned long base_addr)
    223{
    224	update_shadow(mod, base_addr, add_module_to_shadow);
    225}
    226
    227void cfi_module_remove(struct module *mod, unsigned long base_addr)
    228{
    229	update_shadow(mod, base_addr, remove_module_from_shadow);
    230}
    231
    232static inline cfi_check_fn ptr_to_check_fn(const struct cfi_shadow __rcu *s,
    233	unsigned long ptr)
    234{
    235	int index;
    236
    237	if (unlikely(!s))
    238		return NULL; /* No shadow available */
    239
    240	index = ptr_to_shadow(s, ptr);
    241	if (index < 0)
    242		return NULL; /* Cannot be addressed with shadow */
    243
    244	return (cfi_check_fn)shadow_to_check_fn(s, index);
    245}
    246
    247static inline cfi_check_fn find_shadow_check_fn(unsigned long ptr)
    248{
    249	cfi_check_fn fn;
    250
    251	rcu_read_lock_sched_notrace();
    252	fn = ptr_to_check_fn(rcu_dereference_sched(cfi_shadow), ptr);
    253	rcu_read_unlock_sched_notrace();
    254
    255	return fn;
    256}
    257
    258#else /* !CONFIG_CFI_CLANG_SHADOW */
    259
    260static inline cfi_check_fn find_shadow_check_fn(unsigned long ptr)
    261{
    262	return NULL;
    263}
    264
    265#endif /* CONFIG_CFI_CLANG_SHADOW */
    266
    267static inline cfi_check_fn find_module_check_fn(unsigned long ptr)
    268{
    269	cfi_check_fn fn = NULL;
    270	struct module *mod;
    271
    272	rcu_read_lock_sched_notrace();
    273	mod = __module_address(ptr);
    274	if (mod)
    275		fn = mod->cfi_check;
    276	rcu_read_unlock_sched_notrace();
    277
    278	return fn;
    279}
    280
    281static inline cfi_check_fn find_check_fn(unsigned long ptr)
    282{
    283	cfi_check_fn fn = NULL;
    284	unsigned long flags;
    285	bool rcu_idle;
    286
    287	if (is_kernel_text(ptr))
    288		return __cfi_check;
    289
    290	/*
    291	 * Indirect call checks can happen when RCU is not watching. Both
    292	 * the shadow and __module_address use RCU, so we need to wake it
    293	 * up if necessary.
    294	 */
    295	rcu_idle = !rcu_is_watching();
    296	if (rcu_idle) {
    297		local_irq_save(flags);
    298		rcu_irq_enter();
    299	}
    300
    301	if (IS_ENABLED(CONFIG_CFI_CLANG_SHADOW))
    302		fn = find_shadow_check_fn(ptr);
    303	if (!fn)
    304		fn = find_module_check_fn(ptr);
    305
    306	if (rcu_idle) {
    307		rcu_irq_exit();
    308		local_irq_restore(flags);
    309	}
    310
    311	return fn;
    312}
    313
    314void __cfi_slowpath_diag(uint64_t id, void *ptr, void *diag)
    315{
    316	cfi_check_fn fn = find_check_fn((unsigned long)ptr);
    317
    318	if (likely(fn))
    319		fn(id, ptr, diag);
    320	else /* Don't allow unchecked modules */
    321		handle_cfi_failure(ptr);
    322}
    323EXPORT_SYMBOL(__cfi_slowpath_diag);
    324
    325#else /* !CONFIG_MODULES */
    326
    327void __cfi_slowpath_diag(uint64_t id, void *ptr, void *diag)
    328{
    329	handle_cfi_failure(ptr); /* No modules */
    330}
    331EXPORT_SYMBOL(__cfi_slowpath_diag);
    332
    333#endif /* CONFIG_MODULES */
    334
    335void cfi_failure_handler(void *data, void *ptr, void *vtable)
    336{
    337	handle_cfi_failure(ptr);
    338}
    339EXPORT_SYMBOL(cfi_failure_handler);