cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

bpf-helper.h (7593B)


      1/* SPDX-License-Identifier: GPL-2.0 */
      2/*
      3 * Example wrapper around BPF macros.
      4 *
      5 * Copyright (c) 2012 The Chromium OS Authors <chromium-os-dev@chromium.org>
      6 * Author: Will Drewry <wad@chromium.org>
      7 *
      8 * The code may be used by anyone for any purpose,
      9 * and can serve as a starting point for developing
     10 * applications using prctl(PR_SET_SECCOMP, 2, ...).
     11 *
     12 * No guarantees are provided with respect to the correctness
     13 * or functionality of this code.
     14 */
     15#ifndef __BPF_HELPER_H__
     16#define __BPF_HELPER_H__
     17
     18#include <asm/bitsperlong.h>	/* for __BITS_PER_LONG */
     19#include <endian.h>
     20#include <linux/filter.h>
     21#include <linux/seccomp.h>	/* for seccomp_data */
     22#include <linux/types.h>
     23#include <linux/unistd.h>
     24#include <stddef.h>
     25
     26#define BPF_LABELS_MAX 256
     27struct bpf_labels {
     28	int count;
     29	struct __bpf_label {
     30		const char *label;
     31		__u32 location;
     32	} labels[BPF_LABELS_MAX];
     33};
     34
     35int bpf_resolve_jumps(struct bpf_labels *labels,
     36		      struct sock_filter *filter, size_t count);
     37__u32 seccomp_bpf_label(struct bpf_labels *labels, const char *label);
     38void seccomp_bpf_print(struct sock_filter *filter, size_t count);
     39
     40#define JUMP_JT 0xff
     41#define JUMP_JF 0xff
     42#define LABEL_JT 0xfe
     43#define LABEL_JF 0xfe
     44
     45#define ALLOW \
     46	BPF_STMT(BPF_RET+BPF_K, SECCOMP_RET_ALLOW)
     47#define DENY \
     48	BPF_STMT(BPF_RET+BPF_K, SECCOMP_RET_KILL)
     49#define JUMP(labels, label) \
     50	BPF_JUMP(BPF_JMP+BPF_JA, FIND_LABEL((labels), (label)), \
     51		 JUMP_JT, JUMP_JF)
     52#define LABEL(labels, label) \
     53	BPF_JUMP(BPF_JMP+BPF_JA, FIND_LABEL((labels), (label)), \
     54		 LABEL_JT, LABEL_JF)
     55#define SYSCALL(nr, jt) \
     56	BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (nr), 0, 1), \
     57	jt
     58
     59/* Lame, but just an example */
     60#define FIND_LABEL(labels, label) seccomp_bpf_label((labels), #label)
     61
     62#define EXPAND(...) __VA_ARGS__
     63
     64/* Ensure that we load the logically correct offset. */
     65#if __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__
     66#define LO_ARG(idx) offsetof(struct seccomp_data, args[(idx)])
     67#elif __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
     68#define LO_ARG(idx) offsetof(struct seccomp_data, args[(idx)]) + sizeof(__u32)
     69#else
     70#error "Unknown endianness"
     71#endif
     72
     73/* Map all width-sensitive operations */
     74#if __BITS_PER_LONG == 32
     75
     76#define JEQ(x, jt) JEQ32(x, EXPAND(jt))
     77#define JNE(x, jt) JNE32(x, EXPAND(jt))
     78#define JGT(x, jt) JGT32(x, EXPAND(jt))
     79#define JLT(x, jt) JLT32(x, EXPAND(jt))
     80#define JGE(x, jt) JGE32(x, EXPAND(jt))
     81#define JLE(x, jt) JLE32(x, EXPAND(jt))
     82#define JA(x, jt) JA32(x, EXPAND(jt))
     83#define ARG(i) ARG_32(i)
     84
     85#elif __BITS_PER_LONG == 64
     86
     87/* Ensure that we load the logically correct offset. */
     88#if __BYTE_ORDER__ == __ORDER_LITTLE_ENDIAN__
     89#define ENDIAN(_lo, _hi) _lo, _hi
     90#define HI_ARG(idx) offsetof(struct seccomp_data, args[(idx)]) + sizeof(__u32)
     91#elif __BYTE_ORDER__ == __ORDER_BIG_ENDIAN__
     92#define ENDIAN(_lo, _hi) _hi, _lo
     93#define HI_ARG(idx) offsetof(struct seccomp_data, args[(idx)])
     94#endif
     95
     96union arg64 {
     97	struct {
     98		__u32 ENDIAN(lo32, hi32);
     99	};
    100	__u64 u64;
    101};
    102
    103#define JEQ(x, jt) \
    104	JEQ64(((union arg64){.u64 = (x)}).lo32, \
    105	      ((union arg64){.u64 = (x)}).hi32, \
    106	      EXPAND(jt))
    107#define JGT(x, jt) \
    108	JGT64(((union arg64){.u64 = (x)}).lo32, \
    109	      ((union arg64){.u64 = (x)}).hi32, \
    110	      EXPAND(jt))
    111#define JGE(x, jt) \
    112	JGE64(((union arg64){.u64 = (x)}).lo32, \
    113	      ((union arg64){.u64 = (x)}).hi32, \
    114	      EXPAND(jt))
    115#define JNE(x, jt) \
    116	JNE64(((union arg64){.u64 = (x)}).lo32, \
    117	      ((union arg64){.u64 = (x)}).hi32, \
    118	      EXPAND(jt))
    119#define JLT(x, jt) \
    120	JLT64(((union arg64){.u64 = (x)}).lo32, \
    121	      ((union arg64){.u64 = (x)}).hi32, \
    122	      EXPAND(jt))
    123#define JLE(x, jt) \
    124	JLE64(((union arg64){.u64 = (x)}).lo32, \
    125	      ((union arg64){.u64 = (x)}).hi32, \
    126	      EXPAND(jt))
    127
    128#define JA(x, jt) \
    129	JA64(((union arg64){.u64 = (x)}).lo32, \
    130	       ((union arg64){.u64 = (x)}).hi32, \
    131	       EXPAND(jt))
    132#define ARG(i) ARG_64(i)
    133
    134#else
    135#error __BITS_PER_LONG value unusable.
    136#endif
    137
    138/* Loads the arg into A */
    139#define ARG_32(idx) \
    140	BPF_STMT(BPF_LD+BPF_W+BPF_ABS, LO_ARG(idx))
    141
    142/* Loads lo into M[0] and hi into M[1] and A */
    143#define ARG_64(idx) \
    144	BPF_STMT(BPF_LD+BPF_W+BPF_ABS, LO_ARG(idx)), \
    145	BPF_STMT(BPF_ST, 0), /* lo -> M[0] */ \
    146	BPF_STMT(BPF_LD+BPF_W+BPF_ABS, HI_ARG(idx)), \
    147	BPF_STMT(BPF_ST, 1) /* hi -> M[1] */
    148
    149#define JEQ32(value, jt) \
    150	BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (value), 0, 1), \
    151	jt
    152
    153#define JNE32(value, jt) \
    154	BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (value), 1, 0), \
    155	jt
    156
    157#define JA32(value, jt) \
    158	BPF_JUMP(BPF_JMP+BPF_JSET+BPF_K, (value), 0, 1), \
    159	jt
    160
    161#define JGE32(value, jt) \
    162	BPF_JUMP(BPF_JMP+BPF_JGE+BPF_K, (value), 0, 1), \
    163	jt
    164
    165#define JGT32(value, jt) \
    166	BPF_JUMP(BPF_JMP+BPF_JGT+BPF_K, (value), 0, 1), \
    167	jt
    168
    169#define JLE32(value, jt) \
    170	BPF_JUMP(BPF_JMP+BPF_JGT+BPF_K, (value), 1, 0), \
    171	jt
    172
    173#define JLT32(value, jt) \
    174	BPF_JUMP(BPF_JMP+BPF_JGE+BPF_K, (value), 1, 0), \
    175	jt
    176
    177/*
    178 * All the JXX64 checks assume lo is saved in M[0] and hi is saved in both
    179 * A and M[1]. This invariant is kept by restoring A if necessary.
    180 */
    181#define JEQ64(lo, hi, jt) \
    182	/* if (hi != arg.hi) goto NOMATCH; */ \
    183	BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (hi), 0, 5), \
    184	BPF_STMT(BPF_LD+BPF_MEM, 0), /* swap in lo */ \
    185	/* if (lo != arg.lo) goto NOMATCH; */ \
    186	BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (lo), 0, 2), \
    187	BPF_STMT(BPF_LD+BPF_MEM, 1), \
    188	jt, \
    189	BPF_STMT(BPF_LD+BPF_MEM, 1)
    190
    191#define JNE64(lo, hi, jt) \
    192	/* if (hi != arg.hi) goto MATCH; */ \
    193	BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (hi), 0, 3), \
    194	BPF_STMT(BPF_LD+BPF_MEM, 0), \
    195	/* if (lo != arg.lo) goto MATCH; */ \
    196	BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (lo), 2, 0), \
    197	BPF_STMT(BPF_LD+BPF_MEM, 1), \
    198	jt, \
    199	BPF_STMT(BPF_LD+BPF_MEM, 1)
    200
    201#define JA64(lo, hi, jt) \
    202	/* if (hi & arg.hi) goto MATCH; */ \
    203	BPF_JUMP(BPF_JMP+BPF_JSET+BPF_K, (hi), 3, 0), \
    204	BPF_STMT(BPF_LD+BPF_MEM, 0), \
    205	/* if (lo & arg.lo) goto MATCH; */ \
    206	BPF_JUMP(BPF_JMP+BPF_JSET+BPF_K, (lo), 0, 2), \
    207	BPF_STMT(BPF_LD+BPF_MEM, 1), \
    208	jt, \
    209	BPF_STMT(BPF_LD+BPF_MEM, 1)
    210
    211#define JGE64(lo, hi, jt) \
    212	/* if (hi > arg.hi) goto MATCH; */ \
    213	BPF_JUMP(BPF_JMP+BPF_JGT+BPF_K, (hi), 4, 0), \
    214	/* if (hi != arg.hi) goto NOMATCH; */ \
    215	BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (hi), 0, 5), \
    216	BPF_STMT(BPF_LD+BPF_MEM, 0), \
    217	/* if (lo >= arg.lo) goto MATCH; */ \
    218	BPF_JUMP(BPF_JMP+BPF_JGE+BPF_K, (lo), 0, 2), \
    219	BPF_STMT(BPF_LD+BPF_MEM, 1), \
    220	jt, \
    221	BPF_STMT(BPF_LD+BPF_MEM, 1)
    222
    223#define JGT64(lo, hi, jt) \
    224	/* if (hi > arg.hi) goto MATCH; */ \
    225	BPF_JUMP(BPF_JMP+BPF_JGT+BPF_K, (hi), 4, 0), \
    226	/* if (hi != arg.hi) goto NOMATCH; */ \
    227	BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (hi), 0, 5), \
    228	BPF_STMT(BPF_LD+BPF_MEM, 0), \
    229	/* if (lo > arg.lo) goto MATCH; */ \
    230	BPF_JUMP(BPF_JMP+BPF_JGT+BPF_K, (lo), 0, 2), \
    231	BPF_STMT(BPF_LD+BPF_MEM, 1), \
    232	jt, \
    233	BPF_STMT(BPF_LD+BPF_MEM, 1)
    234
    235#define JLE64(lo, hi, jt) \
    236	/* if (hi < arg.hi) goto MATCH; */ \
    237	BPF_JUMP(BPF_JMP+BPF_JGE+BPF_K, (hi), 0, 4), \
    238	/* if (hi != arg.hi) goto NOMATCH; */ \
    239	BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (hi), 0, 5), \
    240	BPF_STMT(BPF_LD+BPF_MEM, 0), \
    241	/* if (lo <= arg.lo) goto MATCH; */ \
    242	BPF_JUMP(BPF_JMP+BPF_JGT+BPF_K, (lo), 2, 0), \
    243	BPF_STMT(BPF_LD+BPF_MEM, 1), \
    244	jt, \
    245	BPF_STMT(BPF_LD+BPF_MEM, 1)
    246
    247#define JLT64(lo, hi, jt) \
    248	/* if (hi < arg.hi) goto MATCH; */ \
    249	BPF_JUMP(BPF_JMP+BPF_JGE+BPF_K, (hi), 0, 4), \
    250	/* if (hi != arg.hi) goto NOMATCH; */ \
    251	BPF_JUMP(BPF_JMP+BPF_JEQ+BPF_K, (hi), 0, 5), \
    252	BPF_STMT(BPF_LD+BPF_MEM, 0), \
    253	/* if (lo < arg.lo) goto MATCH; */ \
    254	BPF_JUMP(BPF_JMP+BPF_JGE+BPF_K, (lo), 2, 0), \
    255	BPF_STMT(BPF_LD+BPF_MEM, 1), \
    256	jt, \
    257	BPF_STMT(BPF_LD+BPF_MEM, 1)
    258
    259#define LOAD_SYSCALL_NR \
    260	BPF_STMT(BPF_LD+BPF_W+BPF_ABS, \
    261		 offsetof(struct seccomp_data, nr))
    262
    263#endif  /* __BPF_HELPER_H__ */