cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

fault.c (4373B)


      1// SPDX-License-Identifier: GPL-2.0-or-later
      2/*
      3 * Low-level SPU handling
      4 *
      5 * (C) Copyright IBM Deutschland Entwicklung GmbH 2005
      6 *
      7 * Author: Arnd Bergmann <arndb@de.ibm.com>
      8 */
      9#include <linux/sched/signal.h>
     10#include <linux/mm.h>
     11
     12#include <asm/spu.h>
     13#include <asm/spu_csa.h>
     14
     15#include "spufs.h"
     16
     17/**
     18 * Handle an SPE event, depending on context SPU_CREATE_EVENTS_ENABLED flag.
     19 *
     20 * If the context was created with events, we just set the return event.
     21 * Otherwise, send an appropriate signal to the process.
     22 */
     23static void spufs_handle_event(struct spu_context *ctx,
     24				unsigned long ea, int type)
     25{
     26	if (ctx->flags & SPU_CREATE_EVENTS_ENABLED) {
     27		ctx->event_return |= type;
     28		wake_up_all(&ctx->stop_wq);
     29		return;
     30	}
     31
     32	switch (type) {
     33	case SPE_EVENT_INVALID_DMA:
     34		force_sig_fault(SIGBUS, BUS_OBJERR, NULL);
     35		break;
     36	case SPE_EVENT_SPE_DATA_STORAGE:
     37		ctx->ops->restart_dma(ctx);
     38		force_sig_fault(SIGSEGV, SEGV_ACCERR, (void __user *)ea);
     39		break;
     40	case SPE_EVENT_DMA_ALIGNMENT:
     41		/* DAR isn't set for an alignment fault :( */
     42		force_sig_fault(SIGBUS, BUS_ADRALN, NULL);
     43		break;
     44	case SPE_EVENT_SPE_ERROR:
     45		force_sig_fault(
     46			SIGILL, ILL_ILLOPC,
     47			(void __user *)(unsigned long)
     48			ctx->ops->npc_read(ctx) - 4);
     49		break;
     50	}
     51}
     52
     53int spufs_handle_class0(struct spu_context *ctx)
     54{
     55	unsigned long stat = ctx->csa.class_0_pending & CLASS0_INTR_MASK;
     56
     57	if (likely(!stat))
     58		return 0;
     59
     60	if (stat & CLASS0_DMA_ALIGNMENT_INTR)
     61		spufs_handle_event(ctx, ctx->csa.class_0_dar,
     62			SPE_EVENT_DMA_ALIGNMENT);
     63
     64	if (stat & CLASS0_INVALID_DMA_COMMAND_INTR)
     65		spufs_handle_event(ctx, ctx->csa.class_0_dar,
     66			SPE_EVENT_INVALID_DMA);
     67
     68	if (stat & CLASS0_SPU_ERROR_INTR)
     69		spufs_handle_event(ctx, ctx->csa.class_0_dar,
     70			SPE_EVENT_SPE_ERROR);
     71
     72	ctx->csa.class_0_pending = 0;
     73
     74	return -EIO;
     75}
     76
     77/*
     78 * bottom half handler for page faults, we can't do this from
     79 * interrupt context, since we might need to sleep.
     80 * we also need to give up the mutex so we can get scheduled
     81 * out while waiting for the backing store.
     82 *
     83 * TODO: try calling hash_page from the interrupt handler first
     84 *       in order to speed up the easy case.
     85 */
     86int spufs_handle_class1(struct spu_context *ctx)
     87{
     88	u64 ea, dsisr, access;
     89	unsigned long flags;
     90	vm_fault_t flt = 0;
     91	int ret;
     92
     93	/*
     94	 * dar and dsisr get passed from the registers
     95	 * to the spu_context, to this function, but not
     96	 * back to the spu if it gets scheduled again.
     97	 *
     98	 * if we don't handle the fault for a saved context
     99	 * in time, we can still expect to get the same fault
    100	 * the immediately after the context restore.
    101	 */
    102	ea = ctx->csa.class_1_dar;
    103	dsisr = ctx->csa.class_1_dsisr;
    104
    105	if (!(dsisr & (MFC_DSISR_PTE_NOT_FOUND | MFC_DSISR_ACCESS_DENIED)))
    106		return 0;
    107
    108	spuctx_switch_state(ctx, SPU_UTIL_IOWAIT);
    109
    110	pr_debug("ctx %p: ea %016llx, dsisr %016llx state %d\n", ctx, ea,
    111		dsisr, ctx->state);
    112
    113	ctx->stats.hash_flt++;
    114	if (ctx->state == SPU_STATE_RUNNABLE)
    115		ctx->spu->stats.hash_flt++;
    116
    117	/* we must not hold the lock when entering copro_handle_mm_fault */
    118	spu_release(ctx);
    119
    120	access = (_PAGE_PRESENT | _PAGE_READ);
    121	access |= (dsisr & MFC_DSISR_ACCESS_PUT) ? _PAGE_WRITE : 0UL;
    122	local_irq_save(flags);
    123	ret = hash_page(ea, access, 0x300, dsisr);
    124	local_irq_restore(flags);
    125
    126	/* hashing failed, so try the actual fault handler */
    127	if (ret)
    128		ret = copro_handle_mm_fault(current->mm, ea, dsisr, &flt);
    129
    130	/*
    131	 * This is nasty: we need the state_mutex for all the bookkeeping even
    132	 * if the syscall was interrupted by a signal. ewww.
    133	 */
    134	mutex_lock(&ctx->state_mutex);
    135
    136	/*
    137	 * Clear dsisr under ctxt lock after handling the fault, so that
    138	 * time slicing will not preempt the context while the page fault
    139	 * handler is running. Context switch code removes mappings.
    140	 */
    141	ctx->csa.class_1_dar = ctx->csa.class_1_dsisr = 0;
    142
    143	/*
    144	 * If we handled the fault successfully and are in runnable
    145	 * state, restart the DMA.
    146	 * In case of unhandled error report the problem to user space.
    147	 */
    148	if (!ret) {
    149		if (flt & VM_FAULT_MAJOR)
    150			ctx->stats.maj_flt++;
    151		else
    152			ctx->stats.min_flt++;
    153		if (ctx->state == SPU_STATE_RUNNABLE) {
    154			if (flt & VM_FAULT_MAJOR)
    155				ctx->spu->stats.maj_flt++;
    156			else
    157				ctx->spu->stats.min_flt++;
    158		}
    159
    160		if (ctx->spu)
    161			ctx->ops->restart_dma(ctx);
    162	} else
    163		spufs_handle_event(ctx, ea, SPE_EVENT_SPE_DATA_STORAGE);
    164
    165	spuctx_switch_state(ctx, SPU_UTIL_SYSTEM);
    166	return ret;
    167}