cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

bpf_tcp_helpers.h (6711B)


      1/* SPDX-License-Identifier: GPL-2.0 */
      2#ifndef __BPF_TCP_HELPERS_H
      3#define __BPF_TCP_HELPERS_H
      4
      5#include <stdbool.h>
      6#include <linux/types.h>
      7#include <bpf/bpf_helpers.h>
      8#include <bpf/bpf_core_read.h>
      9#include <bpf/bpf_tracing.h>
     10
     11#define BPF_STRUCT_OPS(name, args...) \
     12SEC("struct_ops/"#name) \
     13BPF_PROG(name, args)
     14
     15#ifndef SOL_TCP
     16#define SOL_TCP 6
     17#endif
     18
     19#ifndef TCP_CA_NAME_MAX
     20#define TCP_CA_NAME_MAX	16
     21#endif
     22
     23#define tcp_jiffies32 ((__u32)bpf_jiffies64())
     24
     25struct sock_common {
     26	unsigned char	skc_state;
     27	__u16		skc_num;
     28} __attribute__((preserve_access_index));
     29
     30enum sk_pacing {
     31	SK_PACING_NONE		= 0,
     32	SK_PACING_NEEDED	= 1,
     33	SK_PACING_FQ		= 2,
     34};
     35
     36struct sock {
     37	struct sock_common	__sk_common;
     38#define sk_state		__sk_common.skc_state
     39	unsigned long		sk_pacing_rate;
     40	__u32			sk_pacing_status; /* see enum sk_pacing */
     41} __attribute__((preserve_access_index));
     42
     43struct inet_sock {
     44	struct sock		sk;
     45} __attribute__((preserve_access_index));
     46
     47struct inet_connection_sock {
     48	struct inet_sock	  icsk_inet;
     49	__u8			  icsk_ca_state:6,
     50				  icsk_ca_setsockopt:1,
     51				  icsk_ca_dst_locked:1;
     52	struct {
     53		__u8		  pending;
     54	} icsk_ack;
     55	__u64			  icsk_ca_priv[104 / sizeof(__u64)];
     56} __attribute__((preserve_access_index));
     57
     58struct request_sock {
     59	struct sock_common		__req_common;
     60} __attribute__((preserve_access_index));
     61
     62struct tcp_sock {
     63	struct inet_connection_sock	inet_conn;
     64
     65	__u32	rcv_nxt;
     66	__u32	snd_nxt;
     67	__u32	snd_una;
     68	__u32	window_clamp;
     69	__u8	ecn_flags;
     70	__u32	delivered;
     71	__u32	delivered_ce;
     72	__u32	snd_cwnd;
     73	__u32	snd_cwnd_cnt;
     74	__u32	snd_cwnd_clamp;
     75	__u32	snd_ssthresh;
     76	__u8	syn_data:1,	/* SYN includes data */
     77		syn_fastopen:1,	/* SYN includes Fast Open option */
     78		syn_fastopen_exp:1,/* SYN includes Fast Open exp. option */
     79		syn_fastopen_ch:1, /* Active TFO re-enabling probe */
     80		syn_data_acked:1,/* data in SYN is acked by SYN-ACK */
     81		save_syn:1,	/* Save headers of SYN packet */
     82		is_cwnd_limited:1,/* forward progress limited by snd_cwnd? */
     83		syn_smc:1;	/* SYN includes SMC */
     84	__u32	max_packets_out;
     85	__u32	lsndtime;
     86	__u32	prior_cwnd;
     87	__u64	tcp_mstamp;	/* most recent packet received/sent */
     88	bool	is_mptcp;
     89} __attribute__((preserve_access_index));
     90
     91static __always_inline struct inet_connection_sock *inet_csk(const struct sock *sk)
     92{
     93	return (struct inet_connection_sock *)sk;
     94}
     95
     96static __always_inline void *inet_csk_ca(const struct sock *sk)
     97{
     98	return (void *)inet_csk(sk)->icsk_ca_priv;
     99}
    100
    101static __always_inline struct tcp_sock *tcp_sk(const struct sock *sk)
    102{
    103	return (struct tcp_sock *)sk;
    104}
    105
    106static __always_inline bool before(__u32 seq1, __u32 seq2)
    107{
    108	return (__s32)(seq1-seq2) < 0;
    109}
    110#define after(seq2, seq1) 	before(seq1, seq2)
    111
    112#define	TCP_ECN_OK		1
    113#define	TCP_ECN_QUEUE_CWR	2
    114#define	TCP_ECN_DEMAND_CWR	4
    115#define	TCP_ECN_SEEN		8
    116
    117enum inet_csk_ack_state_t {
    118	ICSK_ACK_SCHED	= 1,
    119	ICSK_ACK_TIMER  = 2,
    120	ICSK_ACK_PUSHED = 4,
    121	ICSK_ACK_PUSHED2 = 8,
    122	ICSK_ACK_NOW = 16	/* Send the next ACK immediately (once) */
    123};
    124
    125enum tcp_ca_event {
    126	CA_EVENT_TX_START = 0,
    127	CA_EVENT_CWND_RESTART = 1,
    128	CA_EVENT_COMPLETE_CWR = 2,
    129	CA_EVENT_LOSS = 3,
    130	CA_EVENT_ECN_NO_CE = 4,
    131	CA_EVENT_ECN_IS_CE = 5,
    132};
    133
    134struct ack_sample {
    135	__u32 pkts_acked;
    136	__s32 rtt_us;
    137	__u32 in_flight;
    138} __attribute__((preserve_access_index));
    139
    140struct rate_sample {
    141	__u64  prior_mstamp; /* starting timestamp for interval */
    142	__u32  prior_delivered;	/* tp->delivered at "prior_mstamp" */
    143	__s32  delivered;		/* number of packets delivered over interval */
    144	long interval_us;	/* time for tp->delivered to incr "delivered" */
    145	__u32 snd_interval_us;	/* snd interval for delivered packets */
    146	__u32 rcv_interval_us;	/* rcv interval for delivered packets */
    147	long rtt_us;		/* RTT of last (S)ACKed packet (or -1) */
    148	int  losses;		/* number of packets marked lost upon ACK */
    149	__u32  acked_sacked;	/* number of packets newly (S)ACKed upon ACK */
    150	__u32  prior_in_flight;	/* in flight before this ACK */
    151	bool is_app_limited;	/* is sample from packet with bubble in pipe? */
    152	bool is_retrans;	/* is sample from retransmission? */
    153	bool is_ack_delayed;	/* is this (likely) a delayed ACK? */
    154} __attribute__((preserve_access_index));
    155
    156#define TCP_CA_NAME_MAX		16
    157#define TCP_CONG_NEEDS_ECN	0x2
    158
    159struct tcp_congestion_ops {
    160	char name[TCP_CA_NAME_MAX];
    161	__u32 flags;
    162
    163	/* initialize private data (optional) */
    164	void (*init)(struct sock *sk);
    165	/* cleanup private data  (optional) */
    166	void (*release)(struct sock *sk);
    167
    168	/* return slow start threshold (required) */
    169	__u32 (*ssthresh)(struct sock *sk);
    170	/* do new cwnd calculation (required) */
    171	void (*cong_avoid)(struct sock *sk, __u32 ack, __u32 acked);
    172	/* call before changing ca_state (optional) */
    173	void (*set_state)(struct sock *sk, __u8 new_state);
    174	/* call when cwnd event occurs (optional) */
    175	void (*cwnd_event)(struct sock *sk, enum tcp_ca_event ev);
    176	/* call when ack arrives (optional) */
    177	void (*in_ack_event)(struct sock *sk, __u32 flags);
    178	/* new value of cwnd after loss (required) */
    179	__u32  (*undo_cwnd)(struct sock *sk);
    180	/* hook for packet ack accounting (optional) */
    181	void (*pkts_acked)(struct sock *sk, const struct ack_sample *sample);
    182	/* override sysctl_tcp_min_tso_segs */
    183	__u32 (*min_tso_segs)(struct sock *sk);
    184	/* returns the multiplier used in tcp_sndbuf_expand (optional) */
    185	__u32 (*sndbuf_expand)(struct sock *sk);
    186	/* call when packets are delivered to update cwnd and pacing rate,
    187	 * after all the ca_state processing. (optional)
    188	 */
    189	void (*cong_control)(struct sock *sk, const struct rate_sample *rs);
    190	void *owner;
    191};
    192
    193#define min(a, b) ((a) < (b) ? (a) : (b))
    194#define max(a, b) ((a) > (b) ? (a) : (b))
    195#define min_not_zero(x, y) ({			\
    196	typeof(x) __x = (x);			\
    197	typeof(y) __y = (y);			\
    198	__x == 0 ? __y : ((__y == 0) ? __x : min(__x, __y)); })
    199
    200static __always_inline bool tcp_in_slow_start(const struct tcp_sock *tp)
    201{
    202	return tp->snd_cwnd < tp->snd_ssthresh;
    203}
    204
    205static __always_inline bool tcp_is_cwnd_limited(const struct sock *sk)
    206{
    207	const struct tcp_sock *tp = tcp_sk(sk);
    208
    209	/* If in slow start, ensure cwnd grows to twice what was ACKed. */
    210	if (tcp_in_slow_start(tp))
    211		return tp->snd_cwnd < 2 * tp->max_packets_out;
    212
    213	return !!BPF_CORE_READ_BITFIELD(tp, is_cwnd_limited);
    214}
    215
    216static __always_inline bool tcp_cc_eq(const char *a, const char *b)
    217{
    218	int i;
    219
    220	for (i = 0; i < TCP_CA_NAME_MAX; i++) {
    221		if (a[i] != b[i])
    222			return false;
    223		if (!a[i])
    224			break;
    225	}
    226
    227	return true;
    228}
    229
    230extern __u32 tcp_slow_start(struct tcp_sock *tp, __u32 acked) __ksym;
    231extern void tcp_cong_avoid_ai(struct tcp_sock *tp, __u32 w, __u32 acked) __ksym;
    232
    233struct mptcp_sock {
    234	struct inet_connection_sock	sk;
    235
    236	__u32		token;
    237	struct sock	*first;
    238	char		ca_name[TCP_CA_NAME_MAX];
    239} __attribute__((preserve_access_index));
    240
    241#endif