cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

xsk.h (9806B)


      1/* SPDX-License-Identifier: (LGPL-2.1 OR BSD-2-Clause) */
      2
      3/*
      4 * AF_XDP user-space access library.
      5 *
      6 * Copyright (c) 2018 - 2019 Intel Corporation.
      7 * Copyright (c) 2019 Facebook
      8 *
      9 * Author(s): Magnus Karlsson <magnus.karlsson@intel.com>
     10 */
     11
     12#ifndef __LIBBPF_XSK_H
     13#define __LIBBPF_XSK_H
     14
     15#include <stdio.h>
     16#include <stdint.h>
     17#include <stdbool.h>
     18#include <linux/if_xdp.h>
     19
     20#include "libbpf.h"
     21
     22#ifdef __cplusplus
     23extern "C" {
     24#endif
     25
     26/* This whole API has been deprecated and moved to libxdp that can be found at
     27 * https://github.com/xdp-project/xdp-tools. The APIs are exactly the same so
     28 * it should just be linking with libxdp instead of libbpf for this set of
     29 * functionality. If not, please submit a bug report on the aforementioned page.
     30 */
     31
     32/* Load-Acquire Store-Release barriers used by the XDP socket
     33 * library. The following macros should *NOT* be considered part of
     34 * the xsk.h API, and is subject to change anytime.
     35 *
     36 * LIBRARY INTERNAL
     37 */
     38
     39#define __XSK_READ_ONCE(x) (*(volatile typeof(x) *)&x)
     40#define __XSK_WRITE_ONCE(x, v) (*(volatile typeof(x) *)&x) = (v)
     41
     42#if defined(__i386__) || defined(__x86_64__)
     43# define libbpf_smp_store_release(p, v)					\
     44	do {								\
     45		asm volatile("" : : : "memory");			\
     46		__XSK_WRITE_ONCE(*p, v);				\
     47	} while (0)
     48# define libbpf_smp_load_acquire(p)					\
     49	({								\
     50		typeof(*p) ___p1 = __XSK_READ_ONCE(*p);			\
     51		asm volatile("" : : : "memory");			\
     52		___p1;							\
     53	})
     54#elif defined(__aarch64__)
     55# define libbpf_smp_store_release(p, v)					\
     56		asm volatile ("stlr %w1, %0" : "=Q" (*p) : "r" (v) : "memory")
     57# define libbpf_smp_load_acquire(p)					\
     58	({								\
     59		typeof(*p) ___p1;					\
     60		asm volatile ("ldar %w0, %1"				\
     61			      : "=r" (___p1) : "Q" (*p) : "memory");	\
     62		___p1;							\
     63	})
     64#elif defined(__riscv)
     65# define libbpf_smp_store_release(p, v)					\
     66	do {								\
     67		asm volatile ("fence rw,w" : : : "memory");		\
     68		__XSK_WRITE_ONCE(*p, v);				\
     69	} while (0)
     70# define libbpf_smp_load_acquire(p)					\
     71	({								\
     72		typeof(*p) ___p1 = __XSK_READ_ONCE(*p);			\
     73		asm volatile ("fence r,rw" : : : "memory");		\
     74		___p1;							\
     75	})
     76#endif
     77
     78#ifndef libbpf_smp_store_release
     79#define libbpf_smp_store_release(p, v)					\
     80	do {								\
     81		__sync_synchronize();					\
     82		__XSK_WRITE_ONCE(*p, v);				\
     83	} while (0)
     84#endif
     85
     86#ifndef libbpf_smp_load_acquire
     87#define libbpf_smp_load_acquire(p)					\
     88	({								\
     89		typeof(*p) ___p1 = __XSK_READ_ONCE(*p);			\
     90		__sync_synchronize();					\
     91		___p1;							\
     92	})
     93#endif
     94
     95/* LIBRARY INTERNAL -- END */
     96
     97/* Do not access these members directly. Use the functions below. */
     98#define DEFINE_XSK_RING(name) \
     99struct name { \
    100	__u32 cached_prod; \
    101	__u32 cached_cons; \
    102	__u32 mask; \
    103	__u32 size; \
    104	__u32 *producer; \
    105	__u32 *consumer; \
    106	void *ring; \
    107	__u32 *flags; \
    108}
    109
    110DEFINE_XSK_RING(xsk_ring_prod);
    111DEFINE_XSK_RING(xsk_ring_cons);
    112
    113/* For a detailed explanation on the memory barriers associated with the
    114 * ring, please take a look at net/xdp/xsk_queue.h.
    115 */
    116
    117struct xsk_umem;
    118struct xsk_socket;
    119
    120static inline __u64 *xsk_ring_prod__fill_addr(struct xsk_ring_prod *fill,
    121					      __u32 idx)
    122{
    123	__u64 *addrs = (__u64 *)fill->ring;
    124
    125	return &addrs[idx & fill->mask];
    126}
    127
    128static inline const __u64 *
    129xsk_ring_cons__comp_addr(const struct xsk_ring_cons *comp, __u32 idx)
    130{
    131	const __u64 *addrs = (const __u64 *)comp->ring;
    132
    133	return &addrs[idx & comp->mask];
    134}
    135
    136static inline struct xdp_desc *xsk_ring_prod__tx_desc(struct xsk_ring_prod *tx,
    137						      __u32 idx)
    138{
    139	struct xdp_desc *descs = (struct xdp_desc *)tx->ring;
    140
    141	return &descs[idx & tx->mask];
    142}
    143
    144static inline const struct xdp_desc *
    145xsk_ring_cons__rx_desc(const struct xsk_ring_cons *rx, __u32 idx)
    146{
    147	const struct xdp_desc *descs = (const struct xdp_desc *)rx->ring;
    148
    149	return &descs[idx & rx->mask];
    150}
    151
    152static inline int xsk_ring_prod__needs_wakeup(const struct xsk_ring_prod *r)
    153{
    154	return *r->flags & XDP_RING_NEED_WAKEUP;
    155}
    156
    157static inline __u32 xsk_prod_nb_free(struct xsk_ring_prod *r, __u32 nb)
    158{
    159	__u32 free_entries = r->cached_cons - r->cached_prod;
    160
    161	if (free_entries >= nb)
    162		return free_entries;
    163
    164	/* Refresh the local tail pointer.
    165	 * cached_cons is r->size bigger than the real consumer pointer so
    166	 * that this addition can be avoided in the more frequently
    167	 * executed code that computs free_entries in the beginning of
    168	 * this function. Without this optimization it whould have been
    169	 * free_entries = r->cached_prod - r->cached_cons + r->size.
    170	 */
    171	r->cached_cons = libbpf_smp_load_acquire(r->consumer);
    172	r->cached_cons += r->size;
    173
    174	return r->cached_cons - r->cached_prod;
    175}
    176
    177static inline __u32 xsk_cons_nb_avail(struct xsk_ring_cons *r, __u32 nb)
    178{
    179	__u32 entries = r->cached_prod - r->cached_cons;
    180
    181	if (entries == 0) {
    182		r->cached_prod = libbpf_smp_load_acquire(r->producer);
    183		entries = r->cached_prod - r->cached_cons;
    184	}
    185
    186	return (entries > nb) ? nb : entries;
    187}
    188
    189static inline __u32 xsk_ring_prod__reserve(struct xsk_ring_prod *prod, __u32 nb, __u32 *idx)
    190{
    191	if (xsk_prod_nb_free(prod, nb) < nb)
    192		return 0;
    193
    194	*idx = prod->cached_prod;
    195	prod->cached_prod += nb;
    196
    197	return nb;
    198}
    199
    200static inline void xsk_ring_prod__submit(struct xsk_ring_prod *prod, __u32 nb)
    201{
    202	/* Make sure everything has been written to the ring before indicating
    203	 * this to the kernel by writing the producer pointer.
    204	 */
    205	libbpf_smp_store_release(prod->producer, *prod->producer + nb);
    206}
    207
    208static inline __u32 xsk_ring_cons__peek(struct xsk_ring_cons *cons, __u32 nb, __u32 *idx)
    209{
    210	__u32 entries = xsk_cons_nb_avail(cons, nb);
    211
    212	if (entries > 0) {
    213		*idx = cons->cached_cons;
    214		cons->cached_cons += entries;
    215	}
    216
    217	return entries;
    218}
    219
    220static inline void xsk_ring_cons__cancel(struct xsk_ring_cons *cons, __u32 nb)
    221{
    222	cons->cached_cons -= nb;
    223}
    224
    225static inline void xsk_ring_cons__release(struct xsk_ring_cons *cons, __u32 nb)
    226{
    227	/* Make sure data has been read before indicating we are done
    228	 * with the entries by updating the consumer pointer.
    229	 */
    230	libbpf_smp_store_release(cons->consumer, *cons->consumer + nb);
    231
    232}
    233
    234static inline void *xsk_umem__get_data(void *umem_area, __u64 addr)
    235{
    236	return &((char *)umem_area)[addr];
    237}
    238
    239static inline __u64 xsk_umem__extract_addr(__u64 addr)
    240{
    241	return addr & XSK_UNALIGNED_BUF_ADDR_MASK;
    242}
    243
    244static inline __u64 xsk_umem__extract_offset(__u64 addr)
    245{
    246	return addr >> XSK_UNALIGNED_BUF_OFFSET_SHIFT;
    247}
    248
    249static inline __u64 xsk_umem__add_offset_to_addr(__u64 addr)
    250{
    251	return xsk_umem__extract_addr(addr) + xsk_umem__extract_offset(addr);
    252}
    253
    254LIBBPF_API LIBBPF_DEPRECATED_SINCE(0, 7, "AF_XDP support deprecated and moved to libxdp")
    255int xsk_umem__fd(const struct xsk_umem *umem);
    256LIBBPF_API LIBBPF_DEPRECATED_SINCE(0, 7, "AF_XDP support deprecated and moved to libxdp")
    257int xsk_socket__fd(const struct xsk_socket *xsk);
    258
    259#define XSK_RING_CONS__DEFAULT_NUM_DESCS      2048
    260#define XSK_RING_PROD__DEFAULT_NUM_DESCS      2048
    261#define XSK_UMEM__DEFAULT_FRAME_SHIFT    12 /* 4096 bytes */
    262#define XSK_UMEM__DEFAULT_FRAME_SIZE     (1 << XSK_UMEM__DEFAULT_FRAME_SHIFT)
    263#define XSK_UMEM__DEFAULT_FRAME_HEADROOM 0
    264#define XSK_UMEM__DEFAULT_FLAGS 0
    265
    266struct xsk_umem_config {
    267	__u32 fill_size;
    268	__u32 comp_size;
    269	__u32 frame_size;
    270	__u32 frame_headroom;
    271	__u32 flags;
    272};
    273
    274LIBBPF_API LIBBPF_DEPRECATED_SINCE(0, 7, "AF_XDP support deprecated and moved to libxdp")
    275int xsk_setup_xdp_prog(int ifindex, int *xsks_map_fd);
    276LIBBPF_API LIBBPF_DEPRECATED_SINCE(0, 7, "AF_XDP support deprecated and moved to libxdp")
    277int xsk_socket__update_xskmap(struct xsk_socket *xsk, int xsks_map_fd);
    278
    279/* Flags for the libbpf_flags field. */
    280#define XSK_LIBBPF_FLAGS__INHIBIT_PROG_LOAD (1 << 0)
    281
    282struct xsk_socket_config {
    283	__u32 rx_size;
    284	__u32 tx_size;
    285	__u32 libbpf_flags;
    286	__u32 xdp_flags;
    287	__u16 bind_flags;
    288};
    289
    290/* Set config to NULL to get the default configuration. */
    291LIBBPF_API LIBBPF_DEPRECATED_SINCE(0, 7, "AF_XDP support deprecated and moved to libxdp")
    292int xsk_umem__create(struct xsk_umem **umem,
    293		     void *umem_area, __u64 size,
    294		     struct xsk_ring_prod *fill,
    295		     struct xsk_ring_cons *comp,
    296		     const struct xsk_umem_config *config);
    297LIBBPF_API LIBBPF_DEPRECATED_SINCE(0, 7, "AF_XDP support deprecated and moved to libxdp")
    298int xsk_umem__create_v0_0_2(struct xsk_umem **umem,
    299			    void *umem_area, __u64 size,
    300			    struct xsk_ring_prod *fill,
    301			    struct xsk_ring_cons *comp,
    302			    const struct xsk_umem_config *config);
    303LIBBPF_API LIBBPF_DEPRECATED_SINCE(0, 7, "AF_XDP support deprecated and moved to libxdp")
    304int xsk_umem__create_v0_0_4(struct xsk_umem **umem,
    305			    void *umem_area, __u64 size,
    306			    struct xsk_ring_prod *fill,
    307			    struct xsk_ring_cons *comp,
    308			    const struct xsk_umem_config *config);
    309LIBBPF_API LIBBPF_DEPRECATED_SINCE(0, 7, "AF_XDP support deprecated and moved to libxdp")
    310int xsk_socket__create(struct xsk_socket **xsk,
    311		       const char *ifname, __u32 queue_id,
    312		       struct xsk_umem *umem,
    313		       struct xsk_ring_cons *rx,
    314		       struct xsk_ring_prod *tx,
    315		       const struct xsk_socket_config *config);
    316LIBBPF_API LIBBPF_DEPRECATED_SINCE(0, 7, "AF_XDP support deprecated and moved to libxdp")
    317int xsk_socket__create_shared(struct xsk_socket **xsk_ptr,
    318			      const char *ifname,
    319			      __u32 queue_id, struct xsk_umem *umem,
    320			      struct xsk_ring_cons *rx,
    321			      struct xsk_ring_prod *tx,
    322			      struct xsk_ring_prod *fill,
    323			      struct xsk_ring_cons *comp,
    324			      const struct xsk_socket_config *config);
    325
    326/* Returns 0 for success and -EBUSY if the umem is still in use. */
    327LIBBPF_API LIBBPF_DEPRECATED_SINCE(0, 7, "AF_XDP support deprecated and moved to libxdp")
    328int xsk_umem__delete(struct xsk_umem *umem);
    329LIBBPF_API LIBBPF_DEPRECATED_SINCE(0, 7, "AF_XDP support deprecated and moved to libxdp")
    330void xsk_socket__delete(struct xsk_socket *xsk);
    331
    332#ifdef __cplusplus
    333} /* extern "C" */
    334#endif
    335
    336#endif /* __LIBBPF_XSK_H */