cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

amt.c (88203B)


      1// SPDX-License-Identifier: GPL-2.0-or-later
      2/* Copyright (c) 2021 Taehee Yoo <ap420073@gmail.com> */
      3
      4#define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
      5
      6#include <linux/module.h>
      7#include <linux/skbuff.h>
      8#include <linux/udp.h>
      9#include <linux/jhash.h>
     10#include <linux/if_tunnel.h>
     11#include <linux/net.h>
     12#include <linux/igmp.h>
     13#include <linux/workqueue.h>
     14#include <net/sch_generic.h>
     15#include <net/net_namespace.h>
     16#include <net/ip.h>
     17#include <net/udp.h>
     18#include <net/udp_tunnel.h>
     19#include <net/icmp.h>
     20#include <net/mld.h>
     21#include <net/amt.h>
     22#include <uapi/linux/amt.h>
     23#include <linux/security.h>
     24#include <net/gro_cells.h>
     25#include <net/ipv6.h>
     26#include <net/if_inet6.h>
     27#include <net/ndisc.h>
     28#include <net/addrconf.h>
     29#include <net/ip6_route.h>
     30#include <net/inet_common.h>
     31#include <net/ip6_checksum.h>
     32
     33static struct workqueue_struct *amt_wq;
     34
     35static HLIST_HEAD(source_gc_list);
     36/* Lock for source_gc_list */
     37static spinlock_t source_gc_lock;
     38static struct delayed_work source_gc_wq;
     39static char *status_str[] = {
     40	"AMT_STATUS_INIT",
     41	"AMT_STATUS_SENT_DISCOVERY",
     42	"AMT_STATUS_RECEIVED_DISCOVERY",
     43	"AMT_STATUS_SENT_ADVERTISEMENT",
     44	"AMT_STATUS_RECEIVED_ADVERTISEMENT",
     45	"AMT_STATUS_SENT_REQUEST",
     46	"AMT_STATUS_RECEIVED_REQUEST",
     47	"AMT_STATUS_SENT_QUERY",
     48	"AMT_STATUS_RECEIVED_QUERY",
     49	"AMT_STATUS_SENT_UPDATE",
     50	"AMT_STATUS_RECEIVED_UPDATE",
     51};
     52
     53static char *type_str[] = {
     54	"", /* Type 0 is not defined */
     55	"AMT_MSG_DISCOVERY",
     56	"AMT_MSG_ADVERTISEMENT",
     57	"AMT_MSG_REQUEST",
     58	"AMT_MSG_MEMBERSHIP_QUERY",
     59	"AMT_MSG_MEMBERSHIP_UPDATE",
     60	"AMT_MSG_MULTICAST_DATA",
     61	"AMT_MSG_TEARDOWN",
     62};
     63
     64static char *action_str[] = {
     65	"AMT_ACT_GMI",
     66	"AMT_ACT_GMI_ZERO",
     67	"AMT_ACT_GT",
     68	"AMT_ACT_STATUS_FWD_NEW",
     69	"AMT_ACT_STATUS_D_FWD_NEW",
     70	"AMT_ACT_STATUS_NONE_NEW",
     71};
     72
     73static struct igmpv3_grec igmpv3_zero_grec;
     74
     75#if IS_ENABLED(CONFIG_IPV6)
     76#define MLD2_ALL_NODE_INIT { { { 0xff, 0x02, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0x01 } } }
     77static struct in6_addr mld2_all_node = MLD2_ALL_NODE_INIT;
     78static struct mld2_grec mldv2_zero_grec;
     79#endif
     80
     81static struct amt_skb_cb *amt_skb_cb(struct sk_buff *skb)
     82{
     83	BUILD_BUG_ON(sizeof(struct amt_skb_cb) + sizeof(struct qdisc_skb_cb) >
     84		     sizeof_field(struct sk_buff, cb));
     85
     86	return (struct amt_skb_cb *)((void *)skb->cb +
     87		sizeof(struct qdisc_skb_cb));
     88}
     89
     90static void __amt_source_gc_work(void)
     91{
     92	struct amt_source_node *snode;
     93	struct hlist_head gc_list;
     94	struct hlist_node *t;
     95
     96	spin_lock_bh(&source_gc_lock);
     97	hlist_move_list(&source_gc_list, &gc_list);
     98	spin_unlock_bh(&source_gc_lock);
     99
    100	hlist_for_each_entry_safe(snode, t, &gc_list, node) {
    101		hlist_del_rcu(&snode->node);
    102		kfree_rcu(snode, rcu);
    103	}
    104}
    105
    106static void amt_source_gc_work(struct work_struct *work)
    107{
    108	__amt_source_gc_work();
    109
    110	spin_lock_bh(&source_gc_lock);
    111	mod_delayed_work(amt_wq, &source_gc_wq,
    112			 msecs_to_jiffies(AMT_GC_INTERVAL));
    113	spin_unlock_bh(&source_gc_lock);
    114}
    115
    116static bool amt_addr_equal(union amt_addr *a, union amt_addr *b)
    117{
    118	return !memcmp(a, b, sizeof(union amt_addr));
    119}
    120
    121static u32 amt_source_hash(struct amt_tunnel_list *tunnel, union amt_addr *src)
    122{
    123	u32 hash = jhash(src, sizeof(*src), tunnel->amt->hash_seed);
    124
    125	return reciprocal_scale(hash, tunnel->amt->hash_buckets);
    126}
    127
    128static bool amt_status_filter(struct amt_source_node *snode,
    129			      enum amt_filter filter)
    130{
    131	bool rc = false;
    132
    133	switch (filter) {
    134	case AMT_FILTER_FWD:
    135		if (snode->status == AMT_SOURCE_STATUS_FWD &&
    136		    snode->flags == AMT_SOURCE_OLD)
    137			rc = true;
    138		break;
    139	case AMT_FILTER_D_FWD:
    140		if (snode->status == AMT_SOURCE_STATUS_D_FWD &&
    141		    snode->flags == AMT_SOURCE_OLD)
    142			rc = true;
    143		break;
    144	case AMT_FILTER_FWD_NEW:
    145		if (snode->status == AMT_SOURCE_STATUS_FWD &&
    146		    snode->flags == AMT_SOURCE_NEW)
    147			rc = true;
    148		break;
    149	case AMT_FILTER_D_FWD_NEW:
    150		if (snode->status == AMT_SOURCE_STATUS_D_FWD &&
    151		    snode->flags == AMT_SOURCE_NEW)
    152			rc = true;
    153		break;
    154	case AMT_FILTER_ALL:
    155		rc = true;
    156		break;
    157	case AMT_FILTER_NONE_NEW:
    158		if (snode->status == AMT_SOURCE_STATUS_NONE &&
    159		    snode->flags == AMT_SOURCE_NEW)
    160			rc = true;
    161		break;
    162	case AMT_FILTER_BOTH:
    163		if ((snode->status == AMT_SOURCE_STATUS_D_FWD ||
    164		     snode->status == AMT_SOURCE_STATUS_FWD) &&
    165		    snode->flags == AMT_SOURCE_OLD)
    166			rc = true;
    167		break;
    168	case AMT_FILTER_BOTH_NEW:
    169		if ((snode->status == AMT_SOURCE_STATUS_D_FWD ||
    170		     snode->status == AMT_SOURCE_STATUS_FWD) &&
    171		    snode->flags == AMT_SOURCE_NEW)
    172			rc = true;
    173		break;
    174	default:
    175		WARN_ON_ONCE(1);
    176		break;
    177	}
    178
    179	return rc;
    180}
    181
    182static struct amt_source_node *amt_lookup_src(struct amt_tunnel_list *tunnel,
    183					      struct amt_group_node *gnode,
    184					      enum amt_filter filter,
    185					      union amt_addr *src)
    186{
    187	u32 hash = amt_source_hash(tunnel, src);
    188	struct amt_source_node *snode;
    189
    190	hlist_for_each_entry_rcu(snode, &gnode->sources[hash], node)
    191		if (amt_status_filter(snode, filter) &&
    192		    amt_addr_equal(&snode->source_addr, src))
    193			return snode;
    194
    195	return NULL;
    196}
    197
    198static u32 amt_group_hash(struct amt_tunnel_list *tunnel, union amt_addr *group)
    199{
    200	u32 hash = jhash(group, sizeof(*group), tunnel->amt->hash_seed);
    201
    202	return reciprocal_scale(hash, tunnel->amt->hash_buckets);
    203}
    204
    205static struct amt_group_node *amt_lookup_group(struct amt_tunnel_list *tunnel,
    206					       union amt_addr *group,
    207					       union amt_addr *host,
    208					       bool v6)
    209{
    210	u32 hash = amt_group_hash(tunnel, group);
    211	struct amt_group_node *gnode;
    212
    213	hlist_for_each_entry_rcu(gnode, &tunnel->groups[hash], node) {
    214		if (amt_addr_equal(&gnode->group_addr, group) &&
    215		    amt_addr_equal(&gnode->host_addr, host) &&
    216		    gnode->v6 == v6)
    217			return gnode;
    218	}
    219
    220	return NULL;
    221}
    222
    223static void amt_destroy_source(struct amt_source_node *snode)
    224{
    225	struct amt_group_node *gnode = snode->gnode;
    226	struct amt_tunnel_list *tunnel;
    227
    228	tunnel = gnode->tunnel_list;
    229
    230	if (!gnode->v6) {
    231		netdev_dbg(snode->gnode->amt->dev,
    232			   "Delete source %pI4 from %pI4\n",
    233			   &snode->source_addr.ip4,
    234			   &gnode->group_addr.ip4);
    235#if IS_ENABLED(CONFIG_IPV6)
    236	} else {
    237		netdev_dbg(snode->gnode->amt->dev,
    238			   "Delete source %pI6 from %pI6\n",
    239			   &snode->source_addr.ip6,
    240			   &gnode->group_addr.ip6);
    241#endif
    242	}
    243
    244	cancel_delayed_work(&snode->source_timer);
    245	hlist_del_init_rcu(&snode->node);
    246	tunnel->nr_sources--;
    247	gnode->nr_sources--;
    248	spin_lock_bh(&source_gc_lock);
    249	hlist_add_head_rcu(&snode->node, &source_gc_list);
    250	spin_unlock_bh(&source_gc_lock);
    251}
    252
    253static void amt_del_group(struct amt_dev *amt, struct amt_group_node *gnode)
    254{
    255	struct amt_source_node *snode;
    256	struct hlist_node *t;
    257	int i;
    258
    259	if (cancel_delayed_work(&gnode->group_timer))
    260		dev_put(amt->dev);
    261	hlist_del_rcu(&gnode->node);
    262	gnode->tunnel_list->nr_groups--;
    263
    264	if (!gnode->v6)
    265		netdev_dbg(amt->dev, "Leave group %pI4\n",
    266			   &gnode->group_addr.ip4);
    267#if IS_ENABLED(CONFIG_IPV6)
    268	else
    269		netdev_dbg(amt->dev, "Leave group %pI6\n",
    270			   &gnode->group_addr.ip6);
    271#endif
    272	for (i = 0; i < amt->hash_buckets; i++)
    273		hlist_for_each_entry_safe(snode, t, &gnode->sources[i], node)
    274			amt_destroy_source(snode);
    275
    276	/* tunnel->lock was acquired outside of amt_del_group()
    277	 * But rcu_read_lock() was acquired too so It's safe.
    278	 */
    279	kfree_rcu(gnode, rcu);
    280}
    281
    282/* If a source timer expires with a router filter-mode for the group of
    283 * INCLUDE, the router concludes that traffic from this particular
    284 * source is no longer desired on the attached network, and deletes the
    285 * associated source record.
    286 */
    287static void amt_source_work(struct work_struct *work)
    288{
    289	struct amt_source_node *snode = container_of(to_delayed_work(work),
    290						     struct amt_source_node,
    291						     source_timer);
    292	struct amt_group_node *gnode = snode->gnode;
    293	struct amt_dev *amt = gnode->amt;
    294	struct amt_tunnel_list *tunnel;
    295
    296	tunnel = gnode->tunnel_list;
    297	spin_lock_bh(&tunnel->lock);
    298	rcu_read_lock();
    299	if (gnode->filter_mode == MCAST_INCLUDE) {
    300		amt_destroy_source(snode);
    301		if (!gnode->nr_sources)
    302			amt_del_group(amt, gnode);
    303	} else {
    304		/* When a router filter-mode for a group is EXCLUDE,
    305		 * source records are only deleted when the group timer expires
    306		 */
    307		snode->status = AMT_SOURCE_STATUS_D_FWD;
    308	}
    309	rcu_read_unlock();
    310	spin_unlock_bh(&tunnel->lock);
    311}
    312
    313static void amt_act_src(struct amt_tunnel_list *tunnel,
    314			struct amt_group_node *gnode,
    315			struct amt_source_node *snode,
    316			enum amt_act act)
    317{
    318	struct amt_dev *amt = tunnel->amt;
    319
    320	switch (act) {
    321	case AMT_ACT_GMI:
    322		mod_delayed_work(amt_wq, &snode->source_timer,
    323				 msecs_to_jiffies(amt_gmi(amt)));
    324		break;
    325	case AMT_ACT_GMI_ZERO:
    326		cancel_delayed_work(&snode->source_timer);
    327		break;
    328	case AMT_ACT_GT:
    329		mod_delayed_work(amt_wq, &snode->source_timer,
    330				 gnode->group_timer.timer.expires);
    331		break;
    332	case AMT_ACT_STATUS_FWD_NEW:
    333		snode->status = AMT_SOURCE_STATUS_FWD;
    334		snode->flags = AMT_SOURCE_NEW;
    335		break;
    336	case AMT_ACT_STATUS_D_FWD_NEW:
    337		snode->status = AMT_SOURCE_STATUS_D_FWD;
    338		snode->flags = AMT_SOURCE_NEW;
    339		break;
    340	case AMT_ACT_STATUS_NONE_NEW:
    341		cancel_delayed_work(&snode->source_timer);
    342		snode->status = AMT_SOURCE_STATUS_NONE;
    343		snode->flags = AMT_SOURCE_NEW;
    344		break;
    345	default:
    346		WARN_ON_ONCE(1);
    347		return;
    348	}
    349
    350	if (!gnode->v6)
    351		netdev_dbg(amt->dev, "Source %pI4 from %pI4 Acted %s\n",
    352			   &snode->source_addr.ip4,
    353			   &gnode->group_addr.ip4,
    354			   action_str[act]);
    355#if IS_ENABLED(CONFIG_IPV6)
    356	else
    357		netdev_dbg(amt->dev, "Source %pI6 from %pI6 Acted %s\n",
    358			   &snode->source_addr.ip6,
    359			   &gnode->group_addr.ip6,
    360			   action_str[act]);
    361#endif
    362}
    363
    364static struct amt_source_node *amt_alloc_snode(struct amt_group_node *gnode,
    365					       union amt_addr *src)
    366{
    367	struct amt_source_node *snode;
    368
    369	snode = kzalloc(sizeof(*snode), GFP_ATOMIC);
    370	if (!snode)
    371		return NULL;
    372
    373	memcpy(&snode->source_addr, src, sizeof(union amt_addr));
    374	snode->gnode = gnode;
    375	snode->status = AMT_SOURCE_STATUS_NONE;
    376	snode->flags = AMT_SOURCE_NEW;
    377	INIT_HLIST_NODE(&snode->node);
    378	INIT_DELAYED_WORK(&snode->source_timer, amt_source_work);
    379
    380	return snode;
    381}
    382
    383/* RFC 3810 - 7.2.2.  Definition of Filter Timers
    384 *
    385 *  Router Mode          Filter Timer         Actions/Comments
    386 *  -----------       -----------------       ----------------
    387 *
    388 *    INCLUDE             Not Used            All listeners in
    389 *                                            INCLUDE mode.
    390 *
    391 *    EXCLUDE             Timer > 0           At least one listener
    392 *                                            in EXCLUDE mode.
    393 *
    394 *    EXCLUDE             Timer == 0          No more listeners in
    395 *                                            EXCLUDE mode for the
    396 *                                            multicast address.
    397 *                                            If the Requested List
    398 *                                            is empty, delete
    399 *                                            Multicast Address
    400 *                                            Record.  If not, switch
    401 *                                            to INCLUDE filter mode;
    402 *                                            the sources in the
    403 *                                            Requested List are
    404 *                                            moved to the Include
    405 *                                            List, and the Exclude
    406 *                                            List is deleted.
    407 */
    408static void amt_group_work(struct work_struct *work)
    409{
    410	struct amt_group_node *gnode = container_of(to_delayed_work(work),
    411						    struct amt_group_node,
    412						    group_timer);
    413	struct amt_tunnel_list *tunnel = gnode->tunnel_list;
    414	struct amt_dev *amt = gnode->amt;
    415	struct amt_source_node *snode;
    416	bool delete_group = true;
    417	struct hlist_node *t;
    418	int i, buckets;
    419
    420	buckets = amt->hash_buckets;
    421
    422	spin_lock_bh(&tunnel->lock);
    423	if (gnode->filter_mode == MCAST_INCLUDE) {
    424		/* Not Used */
    425		spin_unlock_bh(&tunnel->lock);
    426		goto out;
    427	}
    428
    429	rcu_read_lock();
    430	for (i = 0; i < buckets; i++) {
    431		hlist_for_each_entry_safe(snode, t,
    432					  &gnode->sources[i], node) {
    433			if (!delayed_work_pending(&snode->source_timer) ||
    434			    snode->status == AMT_SOURCE_STATUS_D_FWD) {
    435				amt_destroy_source(snode);
    436			} else {
    437				delete_group = false;
    438				snode->status = AMT_SOURCE_STATUS_FWD;
    439			}
    440		}
    441	}
    442	if (delete_group)
    443		amt_del_group(amt, gnode);
    444	else
    445		gnode->filter_mode = MCAST_INCLUDE;
    446	rcu_read_unlock();
    447	spin_unlock_bh(&tunnel->lock);
    448out:
    449	dev_put(amt->dev);
    450}
    451
    452/* Non-existant group is created as INCLUDE {empty}:
    453 *
    454 * RFC 3376 - 5.1. Action on Change of Interface State
    455 *
    456 * If no interface state existed for that multicast address before
    457 * the change (i.e., the change consisted of creating a new
    458 * per-interface record), or if no state exists after the change
    459 * (i.e., the change consisted of deleting a per-interface record),
    460 * then the "non-existent" state is considered to have a filter mode
    461 * of INCLUDE and an empty source list.
    462 */
    463static struct amt_group_node *amt_add_group(struct amt_dev *amt,
    464					    struct amt_tunnel_list *tunnel,
    465					    union amt_addr *group,
    466					    union amt_addr *host,
    467					    bool v6)
    468{
    469	struct amt_group_node *gnode;
    470	u32 hash;
    471	int i;
    472
    473	if (tunnel->nr_groups >= amt->max_groups)
    474		return ERR_PTR(-ENOSPC);
    475
    476	gnode = kzalloc(sizeof(*gnode) +
    477			(sizeof(struct hlist_head) * amt->hash_buckets),
    478			GFP_ATOMIC);
    479	if (unlikely(!gnode))
    480		return ERR_PTR(-ENOMEM);
    481
    482	gnode->amt = amt;
    483	gnode->group_addr = *group;
    484	gnode->host_addr = *host;
    485	gnode->v6 = v6;
    486	gnode->tunnel_list = tunnel;
    487	gnode->filter_mode = MCAST_INCLUDE;
    488	INIT_HLIST_NODE(&gnode->node);
    489	INIT_DELAYED_WORK(&gnode->group_timer, amt_group_work);
    490	for (i = 0; i < amt->hash_buckets; i++)
    491		INIT_HLIST_HEAD(&gnode->sources[i]);
    492
    493	hash = amt_group_hash(tunnel, group);
    494	hlist_add_head_rcu(&gnode->node, &tunnel->groups[hash]);
    495	tunnel->nr_groups++;
    496
    497	if (!gnode->v6)
    498		netdev_dbg(amt->dev, "Join group %pI4\n",
    499			   &gnode->group_addr.ip4);
    500#if IS_ENABLED(CONFIG_IPV6)
    501	else
    502		netdev_dbg(amt->dev, "Join group %pI6\n",
    503			   &gnode->group_addr.ip6);
    504#endif
    505
    506	return gnode;
    507}
    508
    509static struct sk_buff *amt_build_igmp_gq(struct amt_dev *amt)
    510{
    511	u8 ra[AMT_IPHDR_OPTS] = { IPOPT_RA, 4, 0, 0 };
    512	int hlen = LL_RESERVED_SPACE(amt->dev);
    513	int tlen = amt->dev->needed_tailroom;
    514	struct igmpv3_query *ihv3;
    515	void *csum_start = NULL;
    516	__sum16 *csum = NULL;
    517	struct sk_buff *skb;
    518	struct ethhdr *eth;
    519	struct iphdr *iph;
    520	unsigned int len;
    521	int offset;
    522
    523	len = hlen + tlen + sizeof(*iph) + AMT_IPHDR_OPTS + sizeof(*ihv3);
    524	skb = netdev_alloc_skb_ip_align(amt->dev, len);
    525	if (!skb)
    526		return NULL;
    527
    528	skb_reserve(skb, hlen);
    529	skb_push(skb, sizeof(*eth));
    530	skb->protocol = htons(ETH_P_IP);
    531	skb_reset_mac_header(skb);
    532	skb->priority = TC_PRIO_CONTROL;
    533	skb_put(skb, sizeof(*iph));
    534	skb_put_data(skb, ra, sizeof(ra));
    535	skb_put(skb, sizeof(*ihv3));
    536	skb_pull(skb, sizeof(*eth));
    537	skb_reset_network_header(skb);
    538
    539	iph		= ip_hdr(skb);
    540	iph->version	= 4;
    541	iph->ihl	= (sizeof(struct iphdr) + AMT_IPHDR_OPTS) >> 2;
    542	iph->tos	= AMT_TOS;
    543	iph->tot_len	= htons(sizeof(*iph) + AMT_IPHDR_OPTS + sizeof(*ihv3));
    544	iph->frag_off	= htons(IP_DF);
    545	iph->ttl	= 1;
    546	iph->id		= 0;
    547	iph->protocol	= IPPROTO_IGMP;
    548	iph->daddr	= htonl(INADDR_ALLHOSTS_GROUP);
    549	iph->saddr	= htonl(INADDR_ANY);
    550	ip_send_check(iph);
    551
    552	eth = eth_hdr(skb);
    553	ether_addr_copy(eth->h_source, amt->dev->dev_addr);
    554	ip_eth_mc_map(htonl(INADDR_ALLHOSTS_GROUP), eth->h_dest);
    555	eth->h_proto = htons(ETH_P_IP);
    556
    557	ihv3		= skb_pull(skb, sizeof(*iph) + AMT_IPHDR_OPTS);
    558	skb_reset_transport_header(skb);
    559	ihv3->type	= IGMP_HOST_MEMBERSHIP_QUERY;
    560	ihv3->code	= 1;
    561	ihv3->group	= 0;
    562	ihv3->qqic	= amt->qi;
    563	ihv3->nsrcs	= 0;
    564	ihv3->resv	= 0;
    565	ihv3->suppress	= false;
    566	ihv3->qrv	= amt->net->ipv4.sysctl_igmp_qrv;
    567	ihv3->csum	= 0;
    568	csum		= &ihv3->csum;
    569	csum_start	= (void *)ihv3;
    570	*csum		= ip_compute_csum(csum_start, sizeof(*ihv3));
    571	offset		= skb_transport_offset(skb);
    572	skb->csum	= skb_checksum(skb, offset, skb->len - offset, 0);
    573	skb->ip_summed	= CHECKSUM_NONE;
    574
    575	skb_push(skb, sizeof(*eth) + sizeof(*iph) + AMT_IPHDR_OPTS);
    576
    577	return skb;
    578}
    579
    580static void __amt_update_gw_status(struct amt_dev *amt, enum amt_status status,
    581				   bool validate)
    582{
    583	if (validate && amt->status >= status)
    584		return;
    585	netdev_dbg(amt->dev, "Update GW status %s -> %s",
    586		   status_str[amt->status], status_str[status]);
    587	amt->status = status;
    588}
    589
    590static void __amt_update_relay_status(struct amt_tunnel_list *tunnel,
    591				      enum amt_status status,
    592				      bool validate)
    593{
    594	if (validate && tunnel->status >= status)
    595		return;
    596	netdev_dbg(tunnel->amt->dev,
    597		   "Update Tunnel(IP = %pI4, PORT = %u) status %s -> %s",
    598		   &tunnel->ip4, ntohs(tunnel->source_port),
    599		   status_str[tunnel->status], status_str[status]);
    600	tunnel->status = status;
    601}
    602
    603static void amt_update_gw_status(struct amt_dev *amt, enum amt_status status,
    604				 bool validate)
    605{
    606	spin_lock_bh(&amt->lock);
    607	__amt_update_gw_status(amt, status, validate);
    608	spin_unlock_bh(&amt->lock);
    609}
    610
    611static void amt_update_relay_status(struct amt_tunnel_list *tunnel,
    612				    enum amt_status status, bool validate)
    613{
    614	spin_lock_bh(&tunnel->lock);
    615	__amt_update_relay_status(tunnel, status, validate);
    616	spin_unlock_bh(&tunnel->lock);
    617}
    618
    619static void amt_send_discovery(struct amt_dev *amt)
    620{
    621	struct amt_header_discovery *amtd;
    622	int hlen, tlen, offset;
    623	struct socket *sock;
    624	struct udphdr *udph;
    625	struct sk_buff *skb;
    626	struct iphdr *iph;
    627	struct rtable *rt;
    628	struct flowi4 fl4;
    629	u32 len;
    630	int err;
    631
    632	rcu_read_lock();
    633	sock = rcu_dereference(amt->sock);
    634	if (!sock)
    635		goto out;
    636
    637	if (!netif_running(amt->stream_dev) || !netif_running(amt->dev))
    638		goto out;
    639
    640	rt = ip_route_output_ports(amt->net, &fl4, sock->sk,
    641				   amt->discovery_ip, amt->local_ip,
    642				   amt->gw_port, amt->relay_port,
    643				   IPPROTO_UDP, 0,
    644				   amt->stream_dev->ifindex);
    645	if (IS_ERR(rt)) {
    646		amt->dev->stats.tx_errors++;
    647		goto out;
    648	}
    649
    650	hlen = LL_RESERVED_SPACE(amt->dev);
    651	tlen = amt->dev->needed_tailroom;
    652	len = hlen + tlen + sizeof(*iph) + sizeof(*udph) + sizeof(*amtd);
    653	skb = netdev_alloc_skb_ip_align(amt->dev, len);
    654	if (!skb) {
    655		ip_rt_put(rt);
    656		amt->dev->stats.tx_errors++;
    657		goto out;
    658	}
    659
    660	skb->priority = TC_PRIO_CONTROL;
    661	skb_dst_set(skb, &rt->dst);
    662
    663	len = sizeof(*iph) + sizeof(*udph) + sizeof(*amtd);
    664	skb_reset_network_header(skb);
    665	skb_put(skb, len);
    666	amtd = skb_pull(skb, sizeof(*iph) + sizeof(*udph));
    667	amtd->version	= 0;
    668	amtd->type	= AMT_MSG_DISCOVERY;
    669	amtd->reserved	= 0;
    670	amtd->nonce	= amt->nonce;
    671	skb_push(skb, sizeof(*udph));
    672	skb_reset_transport_header(skb);
    673	udph		= udp_hdr(skb);
    674	udph->source	= amt->gw_port;
    675	udph->dest	= amt->relay_port;
    676	udph->len	= htons(sizeof(*udph) + sizeof(*amtd));
    677	udph->check	= 0;
    678	offset = skb_transport_offset(skb);
    679	skb->csum = skb_checksum(skb, offset, skb->len - offset, 0);
    680	udph->check = csum_tcpudp_magic(amt->local_ip, amt->discovery_ip,
    681					sizeof(*udph) + sizeof(*amtd),
    682					IPPROTO_UDP, skb->csum);
    683
    684	skb_push(skb, sizeof(*iph));
    685	iph		= ip_hdr(skb);
    686	iph->version	= 4;
    687	iph->ihl	= (sizeof(struct iphdr)) >> 2;
    688	iph->tos	= AMT_TOS;
    689	iph->frag_off	= 0;
    690	iph->ttl	= ip4_dst_hoplimit(&rt->dst);
    691	iph->daddr	= amt->discovery_ip;
    692	iph->saddr	= amt->local_ip;
    693	iph->protocol	= IPPROTO_UDP;
    694	iph->tot_len	= htons(len);
    695
    696	skb->ip_summed = CHECKSUM_NONE;
    697	ip_select_ident(amt->net, skb, NULL);
    698	ip_send_check(iph);
    699	err = ip_local_out(amt->net, sock->sk, skb);
    700	if (unlikely(net_xmit_eval(err)))
    701		amt->dev->stats.tx_errors++;
    702
    703	spin_lock_bh(&amt->lock);
    704	__amt_update_gw_status(amt, AMT_STATUS_SENT_DISCOVERY, true);
    705	spin_unlock_bh(&amt->lock);
    706out:
    707	rcu_read_unlock();
    708}
    709
    710static void amt_send_request(struct amt_dev *amt, bool v6)
    711{
    712	struct amt_header_request *amtrh;
    713	int hlen, tlen, offset;
    714	struct socket *sock;
    715	struct udphdr *udph;
    716	struct sk_buff *skb;
    717	struct iphdr *iph;
    718	struct rtable *rt;
    719	struct flowi4 fl4;
    720	u32 len;
    721	int err;
    722
    723	rcu_read_lock();
    724	sock = rcu_dereference(amt->sock);
    725	if (!sock)
    726		goto out;
    727
    728	if (!netif_running(amt->stream_dev) || !netif_running(amt->dev))
    729		goto out;
    730
    731	rt = ip_route_output_ports(amt->net, &fl4, sock->sk,
    732				   amt->remote_ip, amt->local_ip,
    733				   amt->gw_port, amt->relay_port,
    734				   IPPROTO_UDP, 0,
    735				   amt->stream_dev->ifindex);
    736	if (IS_ERR(rt)) {
    737		amt->dev->stats.tx_errors++;
    738		goto out;
    739	}
    740
    741	hlen = LL_RESERVED_SPACE(amt->dev);
    742	tlen = amt->dev->needed_tailroom;
    743	len = hlen + tlen + sizeof(*iph) + sizeof(*udph) + sizeof(*amtrh);
    744	skb = netdev_alloc_skb_ip_align(amt->dev, len);
    745	if (!skb) {
    746		ip_rt_put(rt);
    747		amt->dev->stats.tx_errors++;
    748		goto out;
    749	}
    750
    751	skb->priority = TC_PRIO_CONTROL;
    752	skb_dst_set(skb, &rt->dst);
    753
    754	len = sizeof(*iph) + sizeof(*udph) + sizeof(*amtrh);
    755	skb_reset_network_header(skb);
    756	skb_put(skb, len);
    757	amtrh = skb_pull(skb, sizeof(*iph) + sizeof(*udph));
    758	amtrh->version	 = 0;
    759	amtrh->type	 = AMT_MSG_REQUEST;
    760	amtrh->reserved1 = 0;
    761	amtrh->p	 = v6;
    762	amtrh->reserved2 = 0;
    763	amtrh->nonce	 = amt->nonce;
    764	skb_push(skb, sizeof(*udph));
    765	skb_reset_transport_header(skb);
    766	udph		= udp_hdr(skb);
    767	udph->source	= amt->gw_port;
    768	udph->dest	= amt->relay_port;
    769	udph->len	= htons(sizeof(*amtrh) + sizeof(*udph));
    770	udph->check	= 0;
    771	offset = skb_transport_offset(skb);
    772	skb->csum = skb_checksum(skb, offset, skb->len - offset, 0);
    773	udph->check = csum_tcpudp_magic(amt->local_ip, amt->remote_ip,
    774					sizeof(*udph) + sizeof(*amtrh),
    775					IPPROTO_UDP, skb->csum);
    776
    777	skb_push(skb, sizeof(*iph));
    778	iph		= ip_hdr(skb);
    779	iph->version	= 4;
    780	iph->ihl	= (sizeof(struct iphdr)) >> 2;
    781	iph->tos	= AMT_TOS;
    782	iph->frag_off	= 0;
    783	iph->ttl	= ip4_dst_hoplimit(&rt->dst);
    784	iph->daddr	= amt->remote_ip;
    785	iph->saddr	= amt->local_ip;
    786	iph->protocol	= IPPROTO_UDP;
    787	iph->tot_len	= htons(len);
    788
    789	skb->ip_summed = CHECKSUM_NONE;
    790	ip_select_ident(amt->net, skb, NULL);
    791	ip_send_check(iph);
    792	err = ip_local_out(amt->net, sock->sk, skb);
    793	if (unlikely(net_xmit_eval(err)))
    794		amt->dev->stats.tx_errors++;
    795
    796out:
    797	rcu_read_unlock();
    798}
    799
    800static void amt_send_igmp_gq(struct amt_dev *amt,
    801			     struct amt_tunnel_list *tunnel)
    802{
    803	struct sk_buff *skb;
    804
    805	skb = amt_build_igmp_gq(amt);
    806	if (!skb)
    807		return;
    808
    809	amt_skb_cb(skb)->tunnel = tunnel;
    810	dev_queue_xmit(skb);
    811}
    812
    813#if IS_ENABLED(CONFIG_IPV6)
    814static struct sk_buff *amt_build_mld_gq(struct amt_dev *amt)
    815{
    816	u8 ra[AMT_IP6HDR_OPTS] = { IPPROTO_ICMPV6, 0, IPV6_TLV_ROUTERALERT,
    817				   2, 0, 0, IPV6_TLV_PAD1, IPV6_TLV_PAD1 };
    818	int hlen = LL_RESERVED_SPACE(amt->dev);
    819	int tlen = amt->dev->needed_tailroom;
    820	struct mld2_query *mld2q;
    821	void *csum_start = NULL;
    822	struct ipv6hdr *ip6h;
    823	struct sk_buff *skb;
    824	struct ethhdr *eth;
    825	u32 len;
    826
    827	len = hlen + tlen + sizeof(*ip6h) + sizeof(ra) + sizeof(*mld2q);
    828	skb = netdev_alloc_skb_ip_align(amt->dev, len);
    829	if (!skb)
    830		return NULL;
    831
    832	skb_reserve(skb, hlen);
    833	skb_push(skb, sizeof(*eth));
    834	skb_reset_mac_header(skb);
    835	eth = eth_hdr(skb);
    836	skb->priority = TC_PRIO_CONTROL;
    837	skb->protocol = htons(ETH_P_IPV6);
    838	skb_put_zero(skb, sizeof(*ip6h));
    839	skb_put_data(skb, ra, sizeof(ra));
    840	skb_put_zero(skb, sizeof(*mld2q));
    841	skb_pull(skb, sizeof(*eth));
    842	skb_reset_network_header(skb);
    843	ip6h			= ipv6_hdr(skb);
    844	ip6h->payload_len	= htons(sizeof(ra) + sizeof(*mld2q));
    845	ip6h->nexthdr		= NEXTHDR_HOP;
    846	ip6h->hop_limit		= 1;
    847	ip6h->daddr		= mld2_all_node;
    848	ip6_flow_hdr(ip6h, 0, 0);
    849
    850	if (ipv6_dev_get_saddr(amt->net, amt->dev, &ip6h->daddr, 0,
    851			       &ip6h->saddr)) {
    852		amt->dev->stats.tx_errors++;
    853		kfree_skb(skb);
    854		return NULL;
    855	}
    856
    857	eth->h_proto = htons(ETH_P_IPV6);
    858	ether_addr_copy(eth->h_source, amt->dev->dev_addr);
    859	ipv6_eth_mc_map(&mld2_all_node, eth->h_dest);
    860
    861	skb_pull(skb, sizeof(*ip6h) + sizeof(ra));
    862	skb_reset_transport_header(skb);
    863	mld2q			= (struct mld2_query *)icmp6_hdr(skb);
    864	mld2q->mld2q_mrc	= htons(1);
    865	mld2q->mld2q_type	= ICMPV6_MGM_QUERY;
    866	mld2q->mld2q_code	= 0;
    867	mld2q->mld2q_cksum	= 0;
    868	mld2q->mld2q_resv1	= 0;
    869	mld2q->mld2q_resv2	= 0;
    870	mld2q->mld2q_suppress	= 0;
    871	mld2q->mld2q_qrv	= amt->qrv;
    872	mld2q->mld2q_nsrcs	= 0;
    873	mld2q->mld2q_qqic	= amt->qi;
    874	csum_start		= (void *)mld2q;
    875	mld2q->mld2q_cksum = csum_ipv6_magic(&ip6h->saddr, &ip6h->daddr,
    876					     sizeof(*mld2q),
    877					     IPPROTO_ICMPV6,
    878					     csum_partial(csum_start,
    879							  sizeof(*mld2q), 0));
    880
    881	skb->ip_summed = CHECKSUM_NONE;
    882	skb_push(skb, sizeof(*eth) + sizeof(*ip6h) + sizeof(ra));
    883	return skb;
    884}
    885
    886static void amt_send_mld_gq(struct amt_dev *amt, struct amt_tunnel_list *tunnel)
    887{
    888	struct sk_buff *skb;
    889
    890	skb = amt_build_mld_gq(amt);
    891	if (!skb)
    892		return;
    893
    894	amt_skb_cb(skb)->tunnel = tunnel;
    895	dev_queue_xmit(skb);
    896}
    897#else
    898static void amt_send_mld_gq(struct amt_dev *amt, struct amt_tunnel_list *tunnel)
    899{
    900}
    901#endif
    902
    903static void amt_secret_work(struct work_struct *work)
    904{
    905	struct amt_dev *amt = container_of(to_delayed_work(work),
    906					   struct amt_dev,
    907					   secret_wq);
    908
    909	spin_lock_bh(&amt->lock);
    910	get_random_bytes(&amt->key, sizeof(siphash_key_t));
    911	spin_unlock_bh(&amt->lock);
    912	mod_delayed_work(amt_wq, &amt->secret_wq,
    913			 msecs_to_jiffies(AMT_SECRET_TIMEOUT));
    914}
    915
    916static void amt_discovery_work(struct work_struct *work)
    917{
    918	struct amt_dev *amt = container_of(to_delayed_work(work),
    919					   struct amt_dev,
    920					   discovery_wq);
    921
    922	spin_lock_bh(&amt->lock);
    923	if (amt->status > AMT_STATUS_SENT_DISCOVERY)
    924		goto out;
    925	get_random_bytes(&amt->nonce, sizeof(__be32));
    926	spin_unlock_bh(&amt->lock);
    927
    928	amt_send_discovery(amt);
    929	spin_lock_bh(&amt->lock);
    930out:
    931	mod_delayed_work(amt_wq, &amt->discovery_wq,
    932			 msecs_to_jiffies(AMT_DISCOVERY_TIMEOUT));
    933	spin_unlock_bh(&amt->lock);
    934}
    935
    936static void amt_req_work(struct work_struct *work)
    937{
    938	struct amt_dev *amt = container_of(to_delayed_work(work),
    939					   struct amt_dev,
    940					   req_wq);
    941	u32 exp;
    942
    943	spin_lock_bh(&amt->lock);
    944	if (amt->status < AMT_STATUS_RECEIVED_ADVERTISEMENT)
    945		goto out;
    946
    947	if (amt->req_cnt > AMT_MAX_REQ_COUNT) {
    948		netdev_dbg(amt->dev, "Gateway is not ready");
    949		amt->qi = AMT_INIT_REQ_TIMEOUT;
    950		amt->ready4 = false;
    951		amt->ready6 = false;
    952		amt->remote_ip = 0;
    953		__amt_update_gw_status(amt, AMT_STATUS_INIT, false);
    954		amt->req_cnt = 0;
    955		goto out;
    956	}
    957	spin_unlock_bh(&amt->lock);
    958
    959	amt_send_request(amt, false);
    960	amt_send_request(amt, true);
    961	spin_lock_bh(&amt->lock);
    962	__amt_update_gw_status(amt, AMT_STATUS_SENT_REQUEST, true);
    963	amt->req_cnt++;
    964out:
    965	exp = min_t(u32, (1 * (1 << amt->req_cnt)), AMT_MAX_REQ_TIMEOUT);
    966	mod_delayed_work(amt_wq, &amt->req_wq, msecs_to_jiffies(exp * 1000));
    967	spin_unlock_bh(&amt->lock);
    968}
    969
    970static bool amt_send_membership_update(struct amt_dev *amt,
    971				       struct sk_buff *skb,
    972				       bool v6)
    973{
    974	struct amt_header_membership_update *amtmu;
    975	struct socket *sock;
    976	struct iphdr *iph;
    977	struct flowi4 fl4;
    978	struct rtable *rt;
    979	int err;
    980
    981	sock = rcu_dereference_bh(amt->sock);
    982	if (!sock)
    983		return true;
    984
    985	err = skb_cow_head(skb, LL_RESERVED_SPACE(amt->dev) + sizeof(*amtmu) +
    986			   sizeof(*iph) + sizeof(struct udphdr));
    987	if (err)
    988		return true;
    989
    990	skb_reset_inner_headers(skb);
    991	memset(&fl4, 0, sizeof(struct flowi4));
    992	fl4.flowi4_oif         = amt->stream_dev->ifindex;
    993	fl4.daddr              = amt->remote_ip;
    994	fl4.saddr              = amt->local_ip;
    995	fl4.flowi4_tos         = AMT_TOS;
    996	fl4.flowi4_proto       = IPPROTO_UDP;
    997	rt = ip_route_output_key(amt->net, &fl4);
    998	if (IS_ERR(rt)) {
    999		netdev_dbg(amt->dev, "no route to %pI4\n", &amt->remote_ip);
   1000		return true;
   1001	}
   1002
   1003	amtmu			= skb_push(skb, sizeof(*amtmu));
   1004	amtmu->version		= 0;
   1005	amtmu->type		= AMT_MSG_MEMBERSHIP_UPDATE;
   1006	amtmu->reserved		= 0;
   1007	amtmu->nonce		= amt->nonce;
   1008	amtmu->response_mac	= amt->mac;
   1009
   1010	if (!v6)
   1011		skb_set_inner_protocol(skb, htons(ETH_P_IP));
   1012	else
   1013		skb_set_inner_protocol(skb, htons(ETH_P_IPV6));
   1014	udp_tunnel_xmit_skb(rt, sock->sk, skb,
   1015			    fl4.saddr,
   1016			    fl4.daddr,
   1017			    AMT_TOS,
   1018			    ip4_dst_hoplimit(&rt->dst),
   1019			    0,
   1020			    amt->gw_port,
   1021			    amt->relay_port,
   1022			    false,
   1023			    false);
   1024	amt_update_gw_status(amt, AMT_STATUS_SENT_UPDATE, true);
   1025	return false;
   1026}
   1027
   1028static void amt_send_multicast_data(struct amt_dev *amt,
   1029				    const struct sk_buff *oskb,
   1030				    struct amt_tunnel_list *tunnel,
   1031				    bool v6)
   1032{
   1033	struct amt_header_mcast_data *amtmd;
   1034	struct socket *sock;
   1035	struct sk_buff *skb;
   1036	struct iphdr *iph;
   1037	struct flowi4 fl4;
   1038	struct rtable *rt;
   1039
   1040	sock = rcu_dereference_bh(amt->sock);
   1041	if (!sock)
   1042		return;
   1043
   1044	skb = skb_copy_expand(oskb, sizeof(*amtmd) + sizeof(*iph) +
   1045			      sizeof(struct udphdr), 0, GFP_ATOMIC);
   1046	if (!skb)
   1047		return;
   1048
   1049	skb_reset_inner_headers(skb);
   1050	memset(&fl4, 0, sizeof(struct flowi4));
   1051	fl4.flowi4_oif         = amt->stream_dev->ifindex;
   1052	fl4.daddr              = tunnel->ip4;
   1053	fl4.saddr              = amt->local_ip;
   1054	fl4.flowi4_proto       = IPPROTO_UDP;
   1055	rt = ip_route_output_key(amt->net, &fl4);
   1056	if (IS_ERR(rt)) {
   1057		netdev_dbg(amt->dev, "no route to %pI4\n", &tunnel->ip4);
   1058		kfree_skb(skb);
   1059		return;
   1060	}
   1061
   1062	amtmd = skb_push(skb, sizeof(*amtmd));
   1063	amtmd->version = 0;
   1064	amtmd->reserved = 0;
   1065	amtmd->type = AMT_MSG_MULTICAST_DATA;
   1066
   1067	if (!v6)
   1068		skb_set_inner_protocol(skb, htons(ETH_P_IP));
   1069	else
   1070		skb_set_inner_protocol(skb, htons(ETH_P_IPV6));
   1071	udp_tunnel_xmit_skb(rt, sock->sk, skb,
   1072			    fl4.saddr,
   1073			    fl4.daddr,
   1074			    AMT_TOS,
   1075			    ip4_dst_hoplimit(&rt->dst),
   1076			    0,
   1077			    amt->relay_port,
   1078			    tunnel->source_port,
   1079			    false,
   1080			    false);
   1081}
   1082
   1083static bool amt_send_membership_query(struct amt_dev *amt,
   1084				      struct sk_buff *skb,
   1085				      struct amt_tunnel_list *tunnel,
   1086				      bool v6)
   1087{
   1088	struct amt_header_membership_query *amtmq;
   1089	struct socket *sock;
   1090	struct rtable *rt;
   1091	struct flowi4 fl4;
   1092	int err;
   1093
   1094	sock = rcu_dereference_bh(amt->sock);
   1095	if (!sock)
   1096		return true;
   1097
   1098	err = skb_cow_head(skb, LL_RESERVED_SPACE(amt->dev) + sizeof(*amtmq) +
   1099			   sizeof(struct iphdr) + sizeof(struct udphdr));
   1100	if (err)
   1101		return true;
   1102
   1103	skb_reset_inner_headers(skb);
   1104	memset(&fl4, 0, sizeof(struct flowi4));
   1105	fl4.flowi4_oif         = amt->stream_dev->ifindex;
   1106	fl4.daddr              = tunnel->ip4;
   1107	fl4.saddr              = amt->local_ip;
   1108	fl4.flowi4_tos         = AMT_TOS;
   1109	fl4.flowi4_proto       = IPPROTO_UDP;
   1110	rt = ip_route_output_key(amt->net, &fl4);
   1111	if (IS_ERR(rt)) {
   1112		netdev_dbg(amt->dev, "no route to %pI4\n", &tunnel->ip4);
   1113		return true;
   1114	}
   1115
   1116	amtmq		= skb_push(skb, sizeof(*amtmq));
   1117	amtmq->version	= 0;
   1118	amtmq->type	= AMT_MSG_MEMBERSHIP_QUERY;
   1119	amtmq->reserved = 0;
   1120	amtmq->l	= 0;
   1121	amtmq->g	= 0;
   1122	amtmq->nonce	= tunnel->nonce;
   1123	amtmq->response_mac = tunnel->mac;
   1124
   1125	if (!v6)
   1126		skb_set_inner_protocol(skb, htons(ETH_P_IP));
   1127	else
   1128		skb_set_inner_protocol(skb, htons(ETH_P_IPV6));
   1129	udp_tunnel_xmit_skb(rt, sock->sk, skb,
   1130			    fl4.saddr,
   1131			    fl4.daddr,
   1132			    AMT_TOS,
   1133			    ip4_dst_hoplimit(&rt->dst),
   1134			    0,
   1135			    amt->relay_port,
   1136			    tunnel->source_port,
   1137			    false,
   1138			    false);
   1139	amt_update_relay_status(tunnel, AMT_STATUS_SENT_QUERY, true);
   1140	return false;
   1141}
   1142
   1143static netdev_tx_t amt_dev_xmit(struct sk_buff *skb, struct net_device *dev)
   1144{
   1145	struct amt_dev *amt = netdev_priv(dev);
   1146	struct amt_tunnel_list *tunnel;
   1147	struct amt_group_node *gnode;
   1148	union amt_addr group = {0,};
   1149#if IS_ENABLED(CONFIG_IPV6)
   1150	struct ipv6hdr *ip6h;
   1151	struct mld_msg *mld;
   1152#endif
   1153	bool report = false;
   1154	struct igmphdr *ih;
   1155	bool query = false;
   1156	struct iphdr *iph;
   1157	bool data = false;
   1158	bool v6 = false;
   1159	u32 hash;
   1160
   1161	iph = ip_hdr(skb);
   1162	if (iph->version == 4) {
   1163		if (!ipv4_is_multicast(iph->daddr))
   1164			goto free;
   1165
   1166		if (!ip_mc_check_igmp(skb)) {
   1167			ih = igmp_hdr(skb);
   1168			switch (ih->type) {
   1169			case IGMPV3_HOST_MEMBERSHIP_REPORT:
   1170			case IGMP_HOST_MEMBERSHIP_REPORT:
   1171				report = true;
   1172				break;
   1173			case IGMP_HOST_MEMBERSHIP_QUERY:
   1174				query = true;
   1175				break;
   1176			default:
   1177				goto free;
   1178			}
   1179		} else {
   1180			data = true;
   1181		}
   1182		v6 = false;
   1183		group.ip4 = iph->daddr;
   1184#if IS_ENABLED(CONFIG_IPV6)
   1185	} else if (iph->version == 6) {
   1186		ip6h = ipv6_hdr(skb);
   1187		if (!ipv6_addr_is_multicast(&ip6h->daddr))
   1188			goto free;
   1189
   1190		if (!ipv6_mc_check_mld(skb)) {
   1191			mld = (struct mld_msg *)skb_transport_header(skb);
   1192			switch (mld->mld_type) {
   1193			case ICMPV6_MGM_REPORT:
   1194			case ICMPV6_MLD2_REPORT:
   1195				report = true;
   1196				break;
   1197			case ICMPV6_MGM_QUERY:
   1198				query = true;
   1199				break;
   1200			default:
   1201				goto free;
   1202			}
   1203		} else {
   1204			data = true;
   1205		}
   1206		v6 = true;
   1207		group.ip6 = ip6h->daddr;
   1208#endif
   1209	} else {
   1210		dev->stats.tx_errors++;
   1211		goto free;
   1212	}
   1213
   1214	if (!pskb_may_pull(skb, sizeof(struct ethhdr)))
   1215		goto free;
   1216
   1217	skb_pull(skb, sizeof(struct ethhdr));
   1218
   1219	if (amt->mode == AMT_MODE_GATEWAY) {
   1220		/* Gateway only passes IGMP/MLD packets */
   1221		if (!report)
   1222			goto free;
   1223		if ((!v6 && !amt->ready4) || (v6 && !amt->ready6))
   1224			goto free;
   1225		if (amt_send_membership_update(amt, skb,  v6))
   1226			goto free;
   1227		goto unlock;
   1228	} else if (amt->mode == AMT_MODE_RELAY) {
   1229		if (query) {
   1230			tunnel = amt_skb_cb(skb)->tunnel;
   1231			if (!tunnel) {
   1232				WARN_ON(1);
   1233				goto free;
   1234			}
   1235
   1236			/* Do not forward unexpected query */
   1237			if (amt_send_membership_query(amt, skb, tunnel, v6))
   1238				goto free;
   1239			goto unlock;
   1240		}
   1241
   1242		if (!data)
   1243			goto free;
   1244		list_for_each_entry_rcu(tunnel, &amt->tunnel_list, list) {
   1245			hash = amt_group_hash(tunnel, &group);
   1246			hlist_for_each_entry_rcu(gnode, &tunnel->groups[hash],
   1247						 node) {
   1248				if (!v6) {
   1249					if (gnode->group_addr.ip4 == iph->daddr)
   1250						goto found;
   1251#if IS_ENABLED(CONFIG_IPV6)
   1252				} else {
   1253					if (ipv6_addr_equal(&gnode->group_addr.ip6,
   1254							    &ip6h->daddr))
   1255						goto found;
   1256#endif
   1257				}
   1258			}
   1259			continue;
   1260found:
   1261			amt_send_multicast_data(amt, skb, tunnel, v6);
   1262		}
   1263	}
   1264
   1265	dev_kfree_skb(skb);
   1266	return NETDEV_TX_OK;
   1267free:
   1268	dev_kfree_skb(skb);
   1269unlock:
   1270	dev->stats.tx_dropped++;
   1271	return NETDEV_TX_OK;
   1272}
   1273
   1274static int amt_parse_type(struct sk_buff *skb)
   1275{
   1276	struct amt_header *amth;
   1277
   1278	if (!pskb_may_pull(skb, sizeof(struct udphdr) +
   1279			   sizeof(struct amt_header)))
   1280		return -1;
   1281
   1282	amth = (struct amt_header *)(udp_hdr(skb) + 1);
   1283
   1284	if (amth->version != 0)
   1285		return -1;
   1286
   1287	if (amth->type >= __AMT_MSG_MAX || !amth->type)
   1288		return -1;
   1289	return amth->type;
   1290}
   1291
   1292static void amt_clear_groups(struct amt_tunnel_list *tunnel)
   1293{
   1294	struct amt_dev *amt = tunnel->amt;
   1295	struct amt_group_node *gnode;
   1296	struct hlist_node *t;
   1297	int i;
   1298
   1299	spin_lock_bh(&tunnel->lock);
   1300	rcu_read_lock();
   1301	for (i = 0; i < amt->hash_buckets; i++)
   1302		hlist_for_each_entry_safe(gnode, t, &tunnel->groups[i], node)
   1303			amt_del_group(amt, gnode);
   1304	rcu_read_unlock();
   1305	spin_unlock_bh(&tunnel->lock);
   1306}
   1307
   1308static void amt_tunnel_expire(struct work_struct *work)
   1309{
   1310	struct amt_tunnel_list *tunnel = container_of(to_delayed_work(work),
   1311						      struct amt_tunnel_list,
   1312						      gc_wq);
   1313	struct amt_dev *amt = tunnel->amt;
   1314
   1315	spin_lock_bh(&amt->lock);
   1316	rcu_read_lock();
   1317	list_del_rcu(&tunnel->list);
   1318	amt->nr_tunnels--;
   1319	amt_clear_groups(tunnel);
   1320	rcu_read_unlock();
   1321	spin_unlock_bh(&amt->lock);
   1322	kfree_rcu(tunnel, rcu);
   1323}
   1324
   1325static void amt_cleanup_srcs(struct amt_dev *amt,
   1326			     struct amt_tunnel_list *tunnel,
   1327			     struct amt_group_node *gnode)
   1328{
   1329	struct amt_source_node *snode;
   1330	struct hlist_node *t;
   1331	int i;
   1332
   1333	/* Delete old sources */
   1334	for (i = 0; i < amt->hash_buckets; i++) {
   1335		hlist_for_each_entry_safe(snode, t, &gnode->sources[i], node) {
   1336			if (snode->flags == AMT_SOURCE_OLD)
   1337				amt_destroy_source(snode);
   1338		}
   1339	}
   1340
   1341	/* switch from new to old */
   1342	for (i = 0; i < amt->hash_buckets; i++)  {
   1343		hlist_for_each_entry_rcu(snode, &gnode->sources[i], node) {
   1344			snode->flags = AMT_SOURCE_OLD;
   1345			if (!gnode->v6)
   1346				netdev_dbg(snode->gnode->amt->dev,
   1347					   "Add source as OLD %pI4 from %pI4\n",
   1348					   &snode->source_addr.ip4,
   1349					   &gnode->group_addr.ip4);
   1350#if IS_ENABLED(CONFIG_IPV6)
   1351			else
   1352				netdev_dbg(snode->gnode->amt->dev,
   1353					   "Add source as OLD %pI6 from %pI6\n",
   1354					   &snode->source_addr.ip6,
   1355					   &gnode->group_addr.ip6);
   1356#endif
   1357		}
   1358	}
   1359}
   1360
   1361static void amt_add_srcs(struct amt_dev *amt, struct amt_tunnel_list *tunnel,
   1362			 struct amt_group_node *gnode, void *grec,
   1363			 bool v6)
   1364{
   1365	struct igmpv3_grec *igmp_grec;
   1366	struct amt_source_node *snode;
   1367#if IS_ENABLED(CONFIG_IPV6)
   1368	struct mld2_grec *mld_grec;
   1369#endif
   1370	union amt_addr src = {0,};
   1371	u16 nsrcs;
   1372	u32 hash;
   1373	int i;
   1374
   1375	if (!v6) {
   1376		igmp_grec = (struct igmpv3_grec *)grec;
   1377		nsrcs = ntohs(igmp_grec->grec_nsrcs);
   1378	} else {
   1379#if IS_ENABLED(CONFIG_IPV6)
   1380		mld_grec = (struct mld2_grec *)grec;
   1381		nsrcs = ntohs(mld_grec->grec_nsrcs);
   1382#else
   1383	return;
   1384#endif
   1385	}
   1386	for (i = 0; i < nsrcs; i++) {
   1387		if (tunnel->nr_sources >= amt->max_sources)
   1388			return;
   1389		if (!v6)
   1390			src.ip4 = igmp_grec->grec_src[i];
   1391#if IS_ENABLED(CONFIG_IPV6)
   1392		else
   1393			memcpy(&src.ip6, &mld_grec->grec_src[i],
   1394			       sizeof(struct in6_addr));
   1395#endif
   1396		if (amt_lookup_src(tunnel, gnode, AMT_FILTER_ALL, &src))
   1397			continue;
   1398
   1399		snode = amt_alloc_snode(gnode, &src);
   1400		if (snode) {
   1401			hash = amt_source_hash(tunnel, &snode->source_addr);
   1402			hlist_add_head_rcu(&snode->node, &gnode->sources[hash]);
   1403			tunnel->nr_sources++;
   1404			gnode->nr_sources++;
   1405
   1406			if (!gnode->v6)
   1407				netdev_dbg(snode->gnode->amt->dev,
   1408					   "Add source as NEW %pI4 from %pI4\n",
   1409					   &snode->source_addr.ip4,
   1410					   &gnode->group_addr.ip4);
   1411#if IS_ENABLED(CONFIG_IPV6)
   1412			else
   1413				netdev_dbg(snode->gnode->amt->dev,
   1414					   "Add source as NEW %pI6 from %pI6\n",
   1415					   &snode->source_addr.ip6,
   1416					   &gnode->group_addr.ip6);
   1417#endif
   1418		}
   1419	}
   1420}
   1421
   1422/* Router State   Report Rec'd New Router State
   1423 * ------------   ------------ ----------------
   1424 * EXCLUDE (X,Y)  IS_IN (A)    EXCLUDE (X+A,Y-A)
   1425 *
   1426 * -----------+-----------+-----------+
   1427 *            |    OLD    |    NEW    |
   1428 * -----------+-----------+-----------+
   1429 *    FWD     |     X     |    X+A    |
   1430 * -----------+-----------+-----------+
   1431 *    D_FWD   |     Y     |    Y-A    |
   1432 * -----------+-----------+-----------+
   1433 *    NONE    |           |     A     |
   1434 * -----------+-----------+-----------+
   1435 *
   1436 * a) Received sources are NONE/NEW
   1437 * b) All NONE will be deleted by amt_cleanup_srcs().
   1438 * c) All OLD will be deleted by amt_cleanup_srcs().
   1439 * d) After delete, NEW source will be switched to OLD.
   1440 */
   1441static void amt_lookup_act_srcs(struct amt_tunnel_list *tunnel,
   1442				struct amt_group_node *gnode,
   1443				void *grec,
   1444				enum amt_ops ops,
   1445				enum amt_filter filter,
   1446				enum amt_act act,
   1447				bool v6)
   1448{
   1449	struct amt_dev *amt = tunnel->amt;
   1450	struct amt_source_node *snode;
   1451	struct igmpv3_grec *igmp_grec;
   1452#if IS_ENABLED(CONFIG_IPV6)
   1453	struct mld2_grec *mld_grec;
   1454#endif
   1455	union amt_addr src = {0,};
   1456	struct hlist_node *t;
   1457	u16 nsrcs;
   1458	int i, j;
   1459
   1460	if (!v6) {
   1461		igmp_grec = (struct igmpv3_grec *)grec;
   1462		nsrcs = ntohs(igmp_grec->grec_nsrcs);
   1463	} else {
   1464#if IS_ENABLED(CONFIG_IPV6)
   1465		mld_grec = (struct mld2_grec *)grec;
   1466		nsrcs = ntohs(mld_grec->grec_nsrcs);
   1467#else
   1468	return;
   1469#endif
   1470	}
   1471
   1472	memset(&src, 0, sizeof(union amt_addr));
   1473	switch (ops) {
   1474	case AMT_OPS_INT:
   1475		/* A*B */
   1476		for (i = 0; i < nsrcs; i++) {
   1477			if (!v6)
   1478				src.ip4 = igmp_grec->grec_src[i];
   1479#if IS_ENABLED(CONFIG_IPV6)
   1480			else
   1481				memcpy(&src.ip6, &mld_grec->grec_src[i],
   1482				       sizeof(struct in6_addr));
   1483#endif
   1484			snode = amt_lookup_src(tunnel, gnode, filter, &src);
   1485			if (!snode)
   1486				continue;
   1487			amt_act_src(tunnel, gnode, snode, act);
   1488		}
   1489		break;
   1490	case AMT_OPS_UNI:
   1491		/* A+B */
   1492		for (i = 0; i < amt->hash_buckets; i++) {
   1493			hlist_for_each_entry_safe(snode, t, &gnode->sources[i],
   1494						  node) {
   1495				if (amt_status_filter(snode, filter))
   1496					amt_act_src(tunnel, gnode, snode, act);
   1497			}
   1498		}
   1499		for (i = 0; i < nsrcs; i++) {
   1500			if (!v6)
   1501				src.ip4 = igmp_grec->grec_src[i];
   1502#if IS_ENABLED(CONFIG_IPV6)
   1503			else
   1504				memcpy(&src.ip6, &mld_grec->grec_src[i],
   1505				       sizeof(struct in6_addr));
   1506#endif
   1507			snode = amt_lookup_src(tunnel, gnode, filter, &src);
   1508			if (!snode)
   1509				continue;
   1510			amt_act_src(tunnel, gnode, snode, act);
   1511		}
   1512		break;
   1513	case AMT_OPS_SUB:
   1514		/* A-B */
   1515		for (i = 0; i < amt->hash_buckets; i++) {
   1516			hlist_for_each_entry_safe(snode, t, &gnode->sources[i],
   1517						  node) {
   1518				if (!amt_status_filter(snode, filter))
   1519					continue;
   1520				for (j = 0; j < nsrcs; j++) {
   1521					if (!v6)
   1522						src.ip4 = igmp_grec->grec_src[j];
   1523#if IS_ENABLED(CONFIG_IPV6)
   1524					else
   1525						memcpy(&src.ip6,
   1526						       &mld_grec->grec_src[j],
   1527						       sizeof(struct in6_addr));
   1528#endif
   1529					if (amt_addr_equal(&snode->source_addr,
   1530							   &src))
   1531						goto out_sub;
   1532				}
   1533				amt_act_src(tunnel, gnode, snode, act);
   1534				continue;
   1535out_sub:;
   1536			}
   1537		}
   1538		break;
   1539	case AMT_OPS_SUB_REV:
   1540		/* B-A */
   1541		for (i = 0; i < nsrcs; i++) {
   1542			if (!v6)
   1543				src.ip4 = igmp_grec->grec_src[i];
   1544#if IS_ENABLED(CONFIG_IPV6)
   1545			else
   1546				memcpy(&src.ip6, &mld_grec->grec_src[i],
   1547				       sizeof(struct in6_addr));
   1548#endif
   1549			snode = amt_lookup_src(tunnel, gnode, AMT_FILTER_ALL,
   1550					       &src);
   1551			if (!snode) {
   1552				snode = amt_lookup_src(tunnel, gnode,
   1553						       filter, &src);
   1554				if (snode)
   1555					amt_act_src(tunnel, gnode, snode, act);
   1556			}
   1557		}
   1558		break;
   1559	default:
   1560		netdev_dbg(amt->dev, "Invalid type\n");
   1561		return;
   1562	}
   1563}
   1564
   1565static void amt_mcast_is_in_handler(struct amt_dev *amt,
   1566				    struct amt_tunnel_list *tunnel,
   1567				    struct amt_group_node *gnode,
   1568				    void *grec, void *zero_grec, bool v6)
   1569{
   1570	if (gnode->filter_mode == MCAST_INCLUDE) {
   1571/* Router State   Report Rec'd New Router State        Actions
   1572 * ------------   ------------ ----------------        -------
   1573 * INCLUDE (A)    IS_IN (B)    INCLUDE (A+B)           (B)=GMI
   1574 */
   1575		/* Update IS_IN (B) as FWD/NEW */
   1576		amt_lookup_act_srcs(tunnel, gnode, grec, AMT_OPS_UNI,
   1577				    AMT_FILTER_NONE_NEW,
   1578				    AMT_ACT_STATUS_FWD_NEW,
   1579				    v6);
   1580		/* Update INCLUDE (A) as NEW */
   1581		amt_lookup_act_srcs(tunnel, gnode, grec, AMT_OPS_UNI,
   1582				    AMT_FILTER_FWD,
   1583				    AMT_ACT_STATUS_FWD_NEW,
   1584				    v6);
   1585		/* (B)=GMI */
   1586		amt_lookup_act_srcs(tunnel, gnode, grec, AMT_OPS_INT,
   1587				    AMT_FILTER_FWD_NEW,
   1588				    AMT_ACT_GMI,
   1589				    v6);
   1590	} else {
   1591/* State        Actions
   1592 * ------------   ------------ ----------------        -------
   1593 * EXCLUDE (X,Y)  IS_IN (A)    EXCLUDE (X+A,Y-A)       (A)=GMI
   1594 */
   1595		/* Update (A) in (X, Y) as NONE/NEW */
   1596		amt_lookup_act_srcs(tunnel, gnode, grec, AMT_OPS_INT,
   1597				    AMT_FILTER_BOTH,
   1598				    AMT_ACT_STATUS_NONE_NEW,
   1599				    v6);
   1600		/* Update FWD/OLD as FWD/NEW */
   1601		amt_lookup_act_srcs(tunnel, gnode, zero_grec, AMT_OPS_UNI,
   1602				    AMT_FILTER_FWD,
   1603				    AMT_ACT_STATUS_FWD_NEW,
   1604				    v6);
   1605		/* Update IS_IN (A) as FWD/NEW */
   1606		amt_lookup_act_srcs(tunnel, gnode, grec, AMT_OPS_INT,
   1607				    AMT_FILTER_NONE_NEW,
   1608				    AMT_ACT_STATUS_FWD_NEW,
   1609				    v6);
   1610		/* Update EXCLUDE (, Y-A) as D_FWD_NEW */
   1611		amt_lookup_act_srcs(tunnel, gnode, grec, AMT_OPS_SUB,
   1612				    AMT_FILTER_D_FWD,
   1613				    AMT_ACT_STATUS_D_FWD_NEW,
   1614				    v6);
   1615	}
   1616}
   1617
   1618static void amt_mcast_is_ex_handler(struct amt_dev *amt,
   1619				    struct amt_tunnel_list *tunnel,
   1620				    struct amt_group_node *gnode,
   1621				    void *grec, void *zero_grec, bool v6)
   1622{
   1623	if (gnode->filter_mode == MCAST_INCLUDE) {
   1624/* Router State   Report Rec'd  New Router State         Actions
   1625 * ------------   ------------  ----------------         -------
   1626 * INCLUDE (A)    IS_EX (B)     EXCLUDE (A*B,B-A)        (B-A)=0
   1627 *                                                       Delete (A-B)
   1628 *                                                       Group Timer=GMI
   1629 */
   1630		/* EXCLUDE(A*B, ) */
   1631		amt_lookup_act_srcs(tunnel, gnode, grec, AMT_OPS_INT,
   1632				    AMT_FILTER_FWD,
   1633				    AMT_ACT_STATUS_FWD_NEW,
   1634				    v6);
   1635		/* EXCLUDE(, B-A) */
   1636		amt_lookup_act_srcs(tunnel, gnode, grec, AMT_OPS_SUB_REV,
   1637				    AMT_FILTER_FWD,
   1638				    AMT_ACT_STATUS_D_FWD_NEW,
   1639				    v6);
   1640		/* (B-A)=0 */
   1641		amt_lookup_act_srcs(tunnel, gnode, zero_grec, AMT_OPS_UNI,
   1642				    AMT_FILTER_D_FWD_NEW,
   1643				    AMT_ACT_GMI_ZERO,
   1644				    v6);
   1645		/* Group Timer=GMI */
   1646		if (!mod_delayed_work(amt_wq, &gnode->group_timer,
   1647				      msecs_to_jiffies(amt_gmi(amt))))
   1648			dev_hold(amt->dev);
   1649		gnode->filter_mode = MCAST_EXCLUDE;
   1650		/* Delete (A-B) will be worked by amt_cleanup_srcs(). */
   1651	} else {
   1652/* Router State   Report Rec'd  New Router State	Actions
   1653 * ------------   ------------  ----------------	-------
   1654 * EXCLUDE (X,Y)  IS_EX (A)     EXCLUDE (A-Y,Y*A)	(A-X-Y)=GMI
   1655 *							Delete (X-A)
   1656 *							Delete (Y-A)
   1657 *							Group Timer=GMI
   1658 */
   1659		/* EXCLUDE (A-Y, ) */
   1660		amt_lookup_act_srcs(tunnel, gnode, grec, AMT_OPS_SUB_REV,
   1661				    AMT_FILTER_D_FWD,
   1662				    AMT_ACT_STATUS_FWD_NEW,
   1663				    v6);
   1664		/* EXCLUDE (, Y*A ) */
   1665		amt_lookup_act_srcs(tunnel, gnode, grec, AMT_OPS_INT,
   1666				    AMT_FILTER_D_FWD,
   1667				    AMT_ACT_STATUS_D_FWD_NEW,
   1668				    v6);
   1669		/* (A-X-Y)=GMI */
   1670		amt_lookup_act_srcs(tunnel, gnode, grec, AMT_OPS_SUB_REV,
   1671				    AMT_FILTER_BOTH_NEW,
   1672				    AMT_ACT_GMI,
   1673				    v6);
   1674		/* Group Timer=GMI */
   1675		if (!mod_delayed_work(amt_wq, &gnode->group_timer,
   1676				      msecs_to_jiffies(amt_gmi(amt))))
   1677			dev_hold(amt->dev);
   1678		/* Delete (X-A), (Y-A) will be worked by amt_cleanup_srcs(). */
   1679	}
   1680}
   1681
   1682static void amt_mcast_to_in_handler(struct amt_dev *amt,
   1683				    struct amt_tunnel_list *tunnel,
   1684				    struct amt_group_node *gnode,
   1685				    void *grec, void *zero_grec, bool v6)
   1686{
   1687	if (gnode->filter_mode == MCAST_INCLUDE) {
   1688/* Router State   Report Rec'd New Router State        Actions
   1689 * ------------   ------------ ----------------        -------
   1690 * INCLUDE (A)    TO_IN (B)    INCLUDE (A+B)           (B)=GMI
   1691 *						       Send Q(G,A-B)
   1692 */
   1693		/* Update TO_IN (B) sources as FWD/NEW */
   1694		amt_lookup_act_srcs(tunnel, gnode, grec, AMT_OPS_UNI,
   1695				    AMT_FILTER_NONE_NEW,
   1696				    AMT_ACT_STATUS_FWD_NEW,
   1697				    v6);
   1698		/* Update INCLUDE (A) sources as NEW */
   1699		amt_lookup_act_srcs(tunnel, gnode, grec, AMT_OPS_UNI,
   1700				    AMT_FILTER_FWD,
   1701				    AMT_ACT_STATUS_FWD_NEW,
   1702				    v6);
   1703		/* (B)=GMI */
   1704		amt_lookup_act_srcs(tunnel, gnode, grec, AMT_OPS_INT,
   1705				    AMT_FILTER_FWD_NEW,
   1706				    AMT_ACT_GMI,
   1707				    v6);
   1708	} else {
   1709/* Router State   Report Rec'd New Router State        Actions
   1710 * ------------   ------------ ----------------        -------
   1711 * EXCLUDE (X,Y)  TO_IN (A)    EXCLUDE (X+A,Y-A)       (A)=GMI
   1712 *						       Send Q(G,X-A)
   1713 *						       Send Q(G)
   1714 */
   1715		/* Update TO_IN (A) sources as FWD/NEW */
   1716		amt_lookup_act_srcs(tunnel, gnode, grec, AMT_OPS_UNI,
   1717				    AMT_FILTER_NONE_NEW,
   1718				    AMT_ACT_STATUS_FWD_NEW,
   1719				    v6);
   1720		/* Update EXCLUDE(X,) sources as FWD/NEW */
   1721		amt_lookup_act_srcs(tunnel, gnode, grec, AMT_OPS_UNI,
   1722				    AMT_FILTER_FWD,
   1723				    AMT_ACT_STATUS_FWD_NEW,
   1724				    v6);
   1725		/* EXCLUDE (, Y-A)
   1726		 * (A) are already switched to FWD_NEW.
   1727		 * So, D_FWD/OLD -> D_FWD/NEW is okay.
   1728		 */
   1729		amt_lookup_act_srcs(tunnel, gnode, zero_grec, AMT_OPS_UNI,
   1730				    AMT_FILTER_D_FWD,
   1731				    AMT_ACT_STATUS_D_FWD_NEW,
   1732				    v6);
   1733		/* (A)=GMI
   1734		 * Only FWD_NEW will have (A) sources.
   1735		 */
   1736		amt_lookup_act_srcs(tunnel, gnode, grec, AMT_OPS_INT,
   1737				    AMT_FILTER_FWD_NEW,
   1738				    AMT_ACT_GMI,
   1739				    v6);
   1740	}
   1741}
   1742
   1743static void amt_mcast_to_ex_handler(struct amt_dev *amt,
   1744				    struct amt_tunnel_list *tunnel,
   1745				    struct amt_group_node *gnode,
   1746				    void *grec, void *zero_grec, bool v6)
   1747{
   1748	if (gnode->filter_mode == MCAST_INCLUDE) {
   1749/* Router State   Report Rec'd New Router State        Actions
   1750 * ------------   ------------ ----------------        -------
   1751 * INCLUDE (A)    TO_EX (B)    EXCLUDE (A*B,B-A)       (B-A)=0
   1752 *						       Delete (A-B)
   1753 *						       Send Q(G,A*B)
   1754 *						       Group Timer=GMI
   1755 */
   1756		/* EXCLUDE (A*B, ) */
   1757		amt_lookup_act_srcs(tunnel, gnode, grec, AMT_OPS_INT,
   1758				    AMT_FILTER_FWD,
   1759				    AMT_ACT_STATUS_FWD_NEW,
   1760				    v6);
   1761		/* EXCLUDE (, B-A) */
   1762		amt_lookup_act_srcs(tunnel, gnode, grec, AMT_OPS_SUB_REV,
   1763				    AMT_FILTER_FWD,
   1764				    AMT_ACT_STATUS_D_FWD_NEW,
   1765				    v6);
   1766		/* (B-A)=0 */
   1767		amt_lookup_act_srcs(tunnel, gnode, zero_grec, AMT_OPS_UNI,
   1768				    AMT_FILTER_D_FWD_NEW,
   1769				    AMT_ACT_GMI_ZERO,
   1770				    v6);
   1771		/* Group Timer=GMI */
   1772		if (!mod_delayed_work(amt_wq, &gnode->group_timer,
   1773				      msecs_to_jiffies(amt_gmi(amt))))
   1774			dev_hold(amt->dev);
   1775		gnode->filter_mode = MCAST_EXCLUDE;
   1776		/* Delete (A-B) will be worked by amt_cleanup_srcs(). */
   1777	} else {
   1778/* Router State   Report Rec'd New Router State        Actions
   1779 * ------------   ------------ ----------------        -------
   1780 * EXCLUDE (X,Y)  TO_EX (A)    EXCLUDE (A-Y,Y*A)       (A-X-Y)=Group Timer
   1781 *						       Delete (X-A)
   1782 *						       Delete (Y-A)
   1783 *						       Send Q(G,A-Y)
   1784 *						       Group Timer=GMI
   1785 */
   1786		/* Update (A-X-Y) as NONE/OLD */
   1787		amt_lookup_act_srcs(tunnel, gnode, grec, AMT_OPS_SUB_REV,
   1788				    AMT_FILTER_BOTH,
   1789				    AMT_ACT_GT,
   1790				    v6);
   1791		/* EXCLUDE (A-Y, ) */
   1792		amt_lookup_act_srcs(tunnel, gnode, grec, AMT_OPS_SUB_REV,
   1793				    AMT_FILTER_D_FWD,
   1794				    AMT_ACT_STATUS_FWD_NEW,
   1795				    v6);
   1796		/* EXCLUDE (, Y*A) */
   1797		amt_lookup_act_srcs(tunnel, gnode, grec, AMT_OPS_INT,
   1798				    AMT_FILTER_D_FWD,
   1799				    AMT_ACT_STATUS_D_FWD_NEW,
   1800				    v6);
   1801		/* Group Timer=GMI */
   1802		if (!mod_delayed_work(amt_wq, &gnode->group_timer,
   1803				      msecs_to_jiffies(amt_gmi(amt))))
   1804			dev_hold(amt->dev);
   1805		/* Delete (X-A), (Y-A) will be worked by amt_cleanup_srcs(). */
   1806	}
   1807}
   1808
   1809static void amt_mcast_allow_handler(struct amt_dev *amt,
   1810				    struct amt_tunnel_list *tunnel,
   1811				    struct amt_group_node *gnode,
   1812				    void *grec, void *zero_grec, bool v6)
   1813{
   1814	if (gnode->filter_mode == MCAST_INCLUDE) {
   1815/* Router State   Report Rec'd New Router State        Actions
   1816 * ------------   ------------ ----------------        -------
   1817 * INCLUDE (A)    ALLOW (B)    INCLUDE (A+B)	       (B)=GMI
   1818 */
   1819		/* INCLUDE (A+B) */
   1820		amt_lookup_act_srcs(tunnel, gnode, grec, AMT_OPS_UNI,
   1821				    AMT_FILTER_FWD,
   1822				    AMT_ACT_STATUS_FWD_NEW,
   1823				    v6);
   1824		/* (B)=GMI */
   1825		amt_lookup_act_srcs(tunnel, gnode, grec, AMT_OPS_INT,
   1826				    AMT_FILTER_FWD_NEW,
   1827				    AMT_ACT_GMI,
   1828				    v6);
   1829	} else {
   1830/* Router State   Report Rec'd New Router State        Actions
   1831 * ------------   ------------ ----------------        -------
   1832 * EXCLUDE (X,Y)  ALLOW (A)    EXCLUDE (X+A,Y-A)       (A)=GMI
   1833 */
   1834		/* EXCLUDE (X+A, ) */
   1835		amt_lookup_act_srcs(tunnel, gnode, grec, AMT_OPS_UNI,
   1836				    AMT_FILTER_FWD,
   1837				    AMT_ACT_STATUS_FWD_NEW,
   1838				    v6);
   1839		/* EXCLUDE (, Y-A) */
   1840		amt_lookup_act_srcs(tunnel, gnode, grec, AMT_OPS_SUB,
   1841				    AMT_FILTER_D_FWD,
   1842				    AMT_ACT_STATUS_D_FWD_NEW,
   1843				    v6);
   1844		/* (A)=GMI
   1845		 * All (A) source are now FWD/NEW status.
   1846		 */
   1847		amt_lookup_act_srcs(tunnel, gnode, grec, AMT_OPS_INT,
   1848				    AMT_FILTER_FWD_NEW,
   1849				    AMT_ACT_GMI,
   1850				    v6);
   1851	}
   1852}
   1853
   1854static void amt_mcast_block_handler(struct amt_dev *amt,
   1855				    struct amt_tunnel_list *tunnel,
   1856				    struct amt_group_node *gnode,
   1857				    void *grec, void *zero_grec, bool v6)
   1858{
   1859	if (gnode->filter_mode == MCAST_INCLUDE) {
   1860/* Router State   Report Rec'd New Router State        Actions
   1861 * ------------   ------------ ----------------        -------
   1862 * INCLUDE (A)    BLOCK (B)    INCLUDE (A)             Send Q(G,A*B)
   1863 */
   1864		/* INCLUDE (A) */
   1865		amt_lookup_act_srcs(tunnel, gnode, zero_grec, AMT_OPS_UNI,
   1866				    AMT_FILTER_FWD,
   1867				    AMT_ACT_STATUS_FWD_NEW,
   1868				    v6);
   1869	} else {
   1870/* Router State   Report Rec'd New Router State        Actions
   1871 * ------------   ------------ ----------------        -------
   1872 * EXCLUDE (X,Y)  BLOCK (A)    EXCLUDE (X+(A-Y),Y)     (A-X-Y)=Group Timer
   1873 *						       Send Q(G,A-Y)
   1874 */
   1875		/* (A-X-Y)=Group Timer */
   1876		amt_lookup_act_srcs(tunnel, gnode, grec, AMT_OPS_SUB_REV,
   1877				    AMT_FILTER_BOTH,
   1878				    AMT_ACT_GT,
   1879				    v6);
   1880		/* EXCLUDE (X, ) */
   1881		amt_lookup_act_srcs(tunnel, gnode, grec, AMT_OPS_UNI,
   1882				    AMT_FILTER_FWD,
   1883				    AMT_ACT_STATUS_FWD_NEW,
   1884				    v6);
   1885		/* EXCLUDE (X+(A-Y) */
   1886		amt_lookup_act_srcs(tunnel, gnode, grec, AMT_OPS_SUB_REV,
   1887				    AMT_FILTER_D_FWD,
   1888				    AMT_ACT_STATUS_FWD_NEW,
   1889				    v6);
   1890		/* EXCLUDE (, Y) */
   1891		amt_lookup_act_srcs(tunnel, gnode, grec, AMT_OPS_UNI,
   1892				    AMT_FILTER_D_FWD,
   1893				    AMT_ACT_STATUS_D_FWD_NEW,
   1894				    v6);
   1895	}
   1896}
   1897
   1898/* RFC 3376
   1899 * 7.3.2. In the Presence of Older Version Group Members
   1900 *
   1901 * When Group Compatibility Mode is IGMPv2, a router internally
   1902 * translates the following IGMPv2 messages for that group to their
   1903 * IGMPv3 equivalents:
   1904 *
   1905 * IGMPv2 Message                IGMPv3 Equivalent
   1906 * --------------                -----------------
   1907 * Report                        IS_EX( {} )
   1908 * Leave                         TO_IN( {} )
   1909 */
   1910static void amt_igmpv2_report_handler(struct amt_dev *amt, struct sk_buff *skb,
   1911				      struct amt_tunnel_list *tunnel)
   1912{
   1913	struct igmphdr *ih = igmp_hdr(skb);
   1914	struct iphdr *iph = ip_hdr(skb);
   1915	struct amt_group_node *gnode;
   1916	union amt_addr group, host;
   1917
   1918	memset(&group, 0, sizeof(union amt_addr));
   1919	group.ip4 = ih->group;
   1920	memset(&host, 0, sizeof(union amt_addr));
   1921	host.ip4 = iph->saddr;
   1922
   1923	gnode = amt_lookup_group(tunnel, &group, &host, false);
   1924	if (!gnode) {
   1925		gnode = amt_add_group(amt, tunnel, &group, &host, false);
   1926		if (!IS_ERR(gnode)) {
   1927			gnode->filter_mode = MCAST_EXCLUDE;
   1928			if (!mod_delayed_work(amt_wq, &gnode->group_timer,
   1929					      msecs_to_jiffies(amt_gmi(amt))))
   1930				dev_hold(amt->dev);
   1931		}
   1932	}
   1933}
   1934
   1935/* RFC 3376
   1936 * 7.3.2. In the Presence of Older Version Group Members
   1937 *
   1938 * When Group Compatibility Mode is IGMPv2, a router internally
   1939 * translates the following IGMPv2 messages for that group to their
   1940 * IGMPv3 equivalents:
   1941 *
   1942 * IGMPv2 Message                IGMPv3 Equivalent
   1943 * --------------                -----------------
   1944 * Report                        IS_EX( {} )
   1945 * Leave                         TO_IN( {} )
   1946 */
   1947static void amt_igmpv2_leave_handler(struct amt_dev *amt, struct sk_buff *skb,
   1948				     struct amt_tunnel_list *tunnel)
   1949{
   1950	struct igmphdr *ih = igmp_hdr(skb);
   1951	struct iphdr *iph = ip_hdr(skb);
   1952	struct amt_group_node *gnode;
   1953	union amt_addr group, host;
   1954
   1955	memset(&group, 0, sizeof(union amt_addr));
   1956	group.ip4 = ih->group;
   1957	memset(&host, 0, sizeof(union amt_addr));
   1958	host.ip4 = iph->saddr;
   1959
   1960	gnode = amt_lookup_group(tunnel, &group, &host, false);
   1961	if (gnode)
   1962		amt_del_group(amt, gnode);
   1963}
   1964
   1965static void amt_igmpv3_report_handler(struct amt_dev *amt, struct sk_buff *skb,
   1966				      struct amt_tunnel_list *tunnel)
   1967{
   1968	struct igmpv3_report *ihrv3 = igmpv3_report_hdr(skb);
   1969	int len = skb_transport_offset(skb) + sizeof(*ihrv3);
   1970	void *zero_grec = (void *)&igmpv3_zero_grec;
   1971	struct iphdr *iph = ip_hdr(skb);
   1972	struct amt_group_node *gnode;
   1973	union amt_addr group, host;
   1974	struct igmpv3_grec *grec;
   1975	u16 nsrcs;
   1976	int i;
   1977
   1978	for (i = 0; i < ntohs(ihrv3->ngrec); i++) {
   1979		len += sizeof(*grec);
   1980		if (!ip_mc_may_pull(skb, len))
   1981			break;
   1982
   1983		grec = (void *)(skb->data + len - sizeof(*grec));
   1984		nsrcs = ntohs(grec->grec_nsrcs);
   1985
   1986		len += nsrcs * sizeof(__be32);
   1987		if (!ip_mc_may_pull(skb, len))
   1988			break;
   1989
   1990		memset(&group, 0, sizeof(union amt_addr));
   1991		group.ip4 = grec->grec_mca;
   1992		memset(&host, 0, sizeof(union amt_addr));
   1993		host.ip4 = iph->saddr;
   1994		gnode = amt_lookup_group(tunnel, &group, &host, false);
   1995		if (!gnode) {
   1996			gnode = amt_add_group(amt, tunnel, &group, &host,
   1997					      false);
   1998			if (IS_ERR(gnode))
   1999				continue;
   2000		}
   2001
   2002		amt_add_srcs(amt, tunnel, gnode, grec, false);
   2003		switch (grec->grec_type) {
   2004		case IGMPV3_MODE_IS_INCLUDE:
   2005			amt_mcast_is_in_handler(amt, tunnel, gnode, grec,
   2006						zero_grec, false);
   2007			break;
   2008		case IGMPV3_MODE_IS_EXCLUDE:
   2009			amt_mcast_is_ex_handler(amt, tunnel, gnode, grec,
   2010						zero_grec, false);
   2011			break;
   2012		case IGMPV3_CHANGE_TO_INCLUDE:
   2013			amt_mcast_to_in_handler(amt, tunnel, gnode, grec,
   2014						zero_grec, false);
   2015			break;
   2016		case IGMPV3_CHANGE_TO_EXCLUDE:
   2017			amt_mcast_to_ex_handler(amt, tunnel, gnode, grec,
   2018						zero_grec, false);
   2019			break;
   2020		case IGMPV3_ALLOW_NEW_SOURCES:
   2021			amt_mcast_allow_handler(amt, tunnel, gnode, grec,
   2022						zero_grec, false);
   2023			break;
   2024		case IGMPV3_BLOCK_OLD_SOURCES:
   2025			amt_mcast_block_handler(amt, tunnel, gnode, grec,
   2026						zero_grec, false);
   2027			break;
   2028		default:
   2029			break;
   2030		}
   2031		amt_cleanup_srcs(amt, tunnel, gnode);
   2032	}
   2033}
   2034
   2035/* caller held tunnel->lock */
   2036static void amt_igmp_report_handler(struct amt_dev *amt, struct sk_buff *skb,
   2037				    struct amt_tunnel_list *tunnel)
   2038{
   2039	struct igmphdr *ih = igmp_hdr(skb);
   2040
   2041	switch (ih->type) {
   2042	case IGMPV3_HOST_MEMBERSHIP_REPORT:
   2043		amt_igmpv3_report_handler(amt, skb, tunnel);
   2044		break;
   2045	case IGMPV2_HOST_MEMBERSHIP_REPORT:
   2046		amt_igmpv2_report_handler(amt, skb, tunnel);
   2047		break;
   2048	case IGMP_HOST_LEAVE_MESSAGE:
   2049		amt_igmpv2_leave_handler(amt, skb, tunnel);
   2050		break;
   2051	default:
   2052		break;
   2053	}
   2054}
   2055
   2056#if IS_ENABLED(CONFIG_IPV6)
   2057/* RFC 3810
   2058 * 8.3.2. In the Presence of MLDv1 Multicast Address Listeners
   2059 *
   2060 * When Multicast Address Compatibility Mode is MLDv2, a router acts
   2061 * using the MLDv2 protocol for that multicast address.  When Multicast
   2062 * Address Compatibility Mode is MLDv1, a router internally translates
   2063 * the following MLDv1 messages for that multicast address to their
   2064 * MLDv2 equivalents:
   2065 *
   2066 * MLDv1 Message                 MLDv2 Equivalent
   2067 * --------------                -----------------
   2068 * Report                        IS_EX( {} )
   2069 * Done                          TO_IN( {} )
   2070 */
   2071static void amt_mldv1_report_handler(struct amt_dev *amt, struct sk_buff *skb,
   2072				     struct amt_tunnel_list *tunnel)
   2073{
   2074	struct mld_msg *mld = (struct mld_msg *)icmp6_hdr(skb);
   2075	struct ipv6hdr *ip6h = ipv6_hdr(skb);
   2076	struct amt_group_node *gnode;
   2077	union amt_addr group, host;
   2078
   2079	memcpy(&group.ip6, &mld->mld_mca, sizeof(struct in6_addr));
   2080	memcpy(&host.ip6, &ip6h->saddr, sizeof(struct in6_addr));
   2081
   2082	gnode = amt_lookup_group(tunnel, &group, &host, true);
   2083	if (!gnode) {
   2084		gnode = amt_add_group(amt, tunnel, &group, &host, true);
   2085		if (!IS_ERR(gnode)) {
   2086			gnode->filter_mode = MCAST_EXCLUDE;
   2087			if (!mod_delayed_work(amt_wq, &gnode->group_timer,
   2088					      msecs_to_jiffies(amt_gmi(amt))))
   2089				dev_hold(amt->dev);
   2090		}
   2091	}
   2092}
   2093
   2094/* RFC 3810
   2095 * 8.3.2. In the Presence of MLDv1 Multicast Address Listeners
   2096 *
   2097 * When Multicast Address Compatibility Mode is MLDv2, a router acts
   2098 * using the MLDv2 protocol for that multicast address.  When Multicast
   2099 * Address Compatibility Mode is MLDv1, a router internally translates
   2100 * the following MLDv1 messages for that multicast address to their
   2101 * MLDv2 equivalents:
   2102 *
   2103 * MLDv1 Message                 MLDv2 Equivalent
   2104 * --------------                -----------------
   2105 * Report                        IS_EX( {} )
   2106 * Done                          TO_IN( {} )
   2107 */
   2108static void amt_mldv1_leave_handler(struct amt_dev *amt, struct sk_buff *skb,
   2109				    struct amt_tunnel_list *tunnel)
   2110{
   2111	struct mld_msg *mld = (struct mld_msg *)icmp6_hdr(skb);
   2112	struct iphdr *iph = ip_hdr(skb);
   2113	struct amt_group_node *gnode;
   2114	union amt_addr group, host;
   2115
   2116	memcpy(&group.ip6, &mld->mld_mca, sizeof(struct in6_addr));
   2117	memset(&host, 0, sizeof(union amt_addr));
   2118	host.ip4 = iph->saddr;
   2119
   2120	gnode = amt_lookup_group(tunnel, &group, &host, true);
   2121	if (gnode) {
   2122		amt_del_group(amt, gnode);
   2123		return;
   2124	}
   2125}
   2126
   2127static void amt_mldv2_report_handler(struct amt_dev *amt, struct sk_buff *skb,
   2128				     struct amt_tunnel_list *tunnel)
   2129{
   2130	struct mld2_report *mld2r = (struct mld2_report *)icmp6_hdr(skb);
   2131	int len = skb_transport_offset(skb) + sizeof(*mld2r);
   2132	void *zero_grec = (void *)&mldv2_zero_grec;
   2133	struct ipv6hdr *ip6h = ipv6_hdr(skb);
   2134	struct amt_group_node *gnode;
   2135	union amt_addr group, host;
   2136	struct mld2_grec *grec;
   2137	u16 nsrcs;
   2138	int i;
   2139
   2140	for (i = 0; i < ntohs(mld2r->mld2r_ngrec); i++) {
   2141		len += sizeof(*grec);
   2142		if (!ipv6_mc_may_pull(skb, len))
   2143			break;
   2144
   2145		grec = (void *)(skb->data + len - sizeof(*grec));
   2146		nsrcs = ntohs(grec->grec_nsrcs);
   2147
   2148		len += nsrcs * sizeof(struct in6_addr);
   2149		if (!ipv6_mc_may_pull(skb, len))
   2150			break;
   2151
   2152		memset(&group, 0, sizeof(union amt_addr));
   2153		group.ip6 = grec->grec_mca;
   2154		memset(&host, 0, sizeof(union amt_addr));
   2155		host.ip6 = ip6h->saddr;
   2156		gnode = amt_lookup_group(tunnel, &group, &host, true);
   2157		if (!gnode) {
   2158			gnode = amt_add_group(amt, tunnel, &group, &host,
   2159					      ETH_P_IPV6);
   2160			if (IS_ERR(gnode))
   2161				continue;
   2162		}
   2163
   2164		amt_add_srcs(amt, tunnel, gnode, grec, true);
   2165		switch (grec->grec_type) {
   2166		case MLD2_MODE_IS_INCLUDE:
   2167			amt_mcast_is_in_handler(amt, tunnel, gnode, grec,
   2168						zero_grec, true);
   2169			break;
   2170		case MLD2_MODE_IS_EXCLUDE:
   2171			amt_mcast_is_ex_handler(amt, tunnel, gnode, grec,
   2172						zero_grec, true);
   2173			break;
   2174		case MLD2_CHANGE_TO_INCLUDE:
   2175			amt_mcast_to_in_handler(amt, tunnel, gnode, grec,
   2176						zero_grec, true);
   2177			break;
   2178		case MLD2_CHANGE_TO_EXCLUDE:
   2179			amt_mcast_to_ex_handler(amt, tunnel, gnode, grec,
   2180						zero_grec, true);
   2181			break;
   2182		case MLD2_ALLOW_NEW_SOURCES:
   2183			amt_mcast_allow_handler(amt, tunnel, gnode, grec,
   2184						zero_grec, true);
   2185			break;
   2186		case MLD2_BLOCK_OLD_SOURCES:
   2187			amt_mcast_block_handler(amt, tunnel, gnode, grec,
   2188						zero_grec, true);
   2189			break;
   2190		default:
   2191			break;
   2192		}
   2193		amt_cleanup_srcs(amt, tunnel, gnode);
   2194	}
   2195}
   2196
   2197/* caller held tunnel->lock */
   2198static void amt_mld_report_handler(struct amt_dev *amt, struct sk_buff *skb,
   2199				   struct amt_tunnel_list *tunnel)
   2200{
   2201	struct mld_msg *mld = (struct mld_msg *)icmp6_hdr(skb);
   2202
   2203	switch (mld->mld_type) {
   2204	case ICMPV6_MGM_REPORT:
   2205		amt_mldv1_report_handler(amt, skb, tunnel);
   2206		break;
   2207	case ICMPV6_MLD2_REPORT:
   2208		amt_mldv2_report_handler(amt, skb, tunnel);
   2209		break;
   2210	case ICMPV6_MGM_REDUCTION:
   2211		amt_mldv1_leave_handler(amt, skb, tunnel);
   2212		break;
   2213	default:
   2214		break;
   2215	}
   2216}
   2217#endif
   2218
   2219static bool amt_advertisement_handler(struct amt_dev *amt, struct sk_buff *skb)
   2220{
   2221	struct amt_header_advertisement *amta;
   2222	int hdr_size;
   2223
   2224	hdr_size = sizeof(*amta) + sizeof(struct udphdr);
   2225	if (!pskb_may_pull(skb, hdr_size))
   2226		return true;
   2227
   2228	amta = (struct amt_header_advertisement *)(udp_hdr(skb) + 1);
   2229	if (!amta->ip4)
   2230		return true;
   2231
   2232	if (amta->reserved || amta->version)
   2233		return true;
   2234
   2235	if (ipv4_is_loopback(amta->ip4) || ipv4_is_multicast(amta->ip4) ||
   2236	    ipv4_is_zeronet(amta->ip4))
   2237		return true;
   2238
   2239	amt->remote_ip = amta->ip4;
   2240	netdev_dbg(amt->dev, "advertised remote ip = %pI4\n", &amt->remote_ip);
   2241	mod_delayed_work(amt_wq, &amt->req_wq, 0);
   2242
   2243	amt_update_gw_status(amt, AMT_STATUS_RECEIVED_ADVERTISEMENT, true);
   2244	return false;
   2245}
   2246
   2247static bool amt_multicast_data_handler(struct amt_dev *amt, struct sk_buff *skb)
   2248{
   2249	struct amt_header_mcast_data *amtmd;
   2250	int hdr_size, len, err;
   2251	struct ethhdr *eth;
   2252	struct iphdr *iph;
   2253
   2254	hdr_size = sizeof(*amtmd) + sizeof(struct udphdr);
   2255	if (!pskb_may_pull(skb, hdr_size))
   2256		return true;
   2257
   2258	amtmd = (struct amt_header_mcast_data *)(udp_hdr(skb) + 1);
   2259	if (amtmd->reserved || amtmd->version)
   2260		return true;
   2261
   2262	if (iptunnel_pull_header(skb, hdr_size, htons(ETH_P_IP), false))
   2263		return true;
   2264
   2265	skb_reset_network_header(skb);
   2266	skb_push(skb, sizeof(*eth));
   2267	skb_reset_mac_header(skb);
   2268	skb_pull(skb, sizeof(*eth));
   2269	eth = eth_hdr(skb);
   2270
   2271	if (!pskb_may_pull(skb, sizeof(*iph)))
   2272		return true;
   2273	iph = ip_hdr(skb);
   2274
   2275	if (iph->version == 4) {
   2276		if (!ipv4_is_multicast(iph->daddr))
   2277			return true;
   2278		skb->protocol = htons(ETH_P_IP);
   2279		eth->h_proto = htons(ETH_P_IP);
   2280		ip_eth_mc_map(iph->daddr, eth->h_dest);
   2281#if IS_ENABLED(CONFIG_IPV6)
   2282	} else if (iph->version == 6) {
   2283		struct ipv6hdr *ip6h;
   2284
   2285		if (!pskb_may_pull(skb, sizeof(*ip6h)))
   2286			return true;
   2287
   2288		ip6h = ipv6_hdr(skb);
   2289		if (!ipv6_addr_is_multicast(&ip6h->daddr))
   2290			return true;
   2291		skb->protocol = htons(ETH_P_IPV6);
   2292		eth->h_proto = htons(ETH_P_IPV6);
   2293		ipv6_eth_mc_map(&ip6h->daddr, eth->h_dest);
   2294#endif
   2295	} else {
   2296		return true;
   2297	}
   2298
   2299	skb->pkt_type = PACKET_MULTICAST;
   2300	skb->ip_summed = CHECKSUM_NONE;
   2301	len = skb->len;
   2302	err = gro_cells_receive(&amt->gro_cells, skb);
   2303	if (likely(err == NET_RX_SUCCESS))
   2304		dev_sw_netstats_rx_add(amt->dev, len);
   2305	else
   2306		amt->dev->stats.rx_dropped++;
   2307
   2308	return false;
   2309}
   2310
   2311static bool amt_membership_query_handler(struct amt_dev *amt,
   2312					 struct sk_buff *skb)
   2313{
   2314	struct amt_header_membership_query *amtmq;
   2315	struct igmpv3_query *ihv3;
   2316	struct ethhdr *eth, *oeth;
   2317	struct iphdr *iph;
   2318	int hdr_size, len;
   2319
   2320	hdr_size = sizeof(*amtmq) + sizeof(struct udphdr);
   2321	if (!pskb_may_pull(skb, hdr_size))
   2322		return true;
   2323
   2324	amtmq = (struct amt_header_membership_query *)(udp_hdr(skb) + 1);
   2325	if (amtmq->reserved || amtmq->version)
   2326		return true;
   2327
   2328	hdr_size -= sizeof(*eth);
   2329	if (iptunnel_pull_header(skb, hdr_size, htons(ETH_P_TEB), false))
   2330		return true;
   2331
   2332	oeth = eth_hdr(skb);
   2333	skb_reset_mac_header(skb);
   2334	skb_pull(skb, sizeof(*eth));
   2335	skb_reset_network_header(skb);
   2336	eth = eth_hdr(skb);
   2337	if (!pskb_may_pull(skb, sizeof(*iph)))
   2338		return true;
   2339
   2340	iph = ip_hdr(skb);
   2341	if (iph->version == 4) {
   2342		if (!pskb_may_pull(skb, sizeof(*iph) + AMT_IPHDR_OPTS +
   2343				   sizeof(*ihv3)))
   2344			return true;
   2345
   2346		if (!ipv4_is_multicast(iph->daddr))
   2347			return true;
   2348
   2349		ihv3 = skb_pull(skb, sizeof(*iph) + AMT_IPHDR_OPTS);
   2350		skb_reset_transport_header(skb);
   2351		skb_push(skb, sizeof(*iph) + AMT_IPHDR_OPTS);
   2352		spin_lock_bh(&amt->lock);
   2353		amt->ready4 = true;
   2354		amt->mac = amtmq->response_mac;
   2355		amt->req_cnt = 0;
   2356		amt->qi = ihv3->qqic;
   2357		spin_unlock_bh(&amt->lock);
   2358		skb->protocol = htons(ETH_P_IP);
   2359		eth->h_proto = htons(ETH_P_IP);
   2360		ip_eth_mc_map(iph->daddr, eth->h_dest);
   2361#if IS_ENABLED(CONFIG_IPV6)
   2362	} else if (iph->version == 6) {
   2363		struct mld2_query *mld2q;
   2364		struct ipv6hdr *ip6h;
   2365
   2366		if (!pskb_may_pull(skb, sizeof(*ip6h) + AMT_IP6HDR_OPTS +
   2367				   sizeof(*mld2q)))
   2368			return true;
   2369
   2370		ip6h = ipv6_hdr(skb);
   2371		if (!ipv6_addr_is_multicast(&ip6h->daddr))
   2372			return true;
   2373
   2374		mld2q = skb_pull(skb, sizeof(*ip6h) + AMT_IP6HDR_OPTS);
   2375		skb_reset_transport_header(skb);
   2376		skb_push(skb, sizeof(*ip6h) + AMT_IP6HDR_OPTS);
   2377		spin_lock_bh(&amt->lock);
   2378		amt->ready6 = true;
   2379		amt->mac = amtmq->response_mac;
   2380		amt->req_cnt = 0;
   2381		amt->qi = mld2q->mld2q_qqic;
   2382		spin_unlock_bh(&amt->lock);
   2383		skb->protocol = htons(ETH_P_IPV6);
   2384		eth->h_proto = htons(ETH_P_IPV6);
   2385		ipv6_eth_mc_map(&ip6h->daddr, eth->h_dest);
   2386#endif
   2387	} else {
   2388		return true;
   2389	}
   2390
   2391	ether_addr_copy(eth->h_source, oeth->h_source);
   2392	skb->pkt_type = PACKET_MULTICAST;
   2393	skb->ip_summed = CHECKSUM_NONE;
   2394	len = skb->len;
   2395	if (__netif_rx(skb) == NET_RX_SUCCESS) {
   2396		amt_update_gw_status(amt, AMT_STATUS_RECEIVED_QUERY, true);
   2397		dev_sw_netstats_rx_add(amt->dev, len);
   2398	} else {
   2399		amt->dev->stats.rx_dropped++;
   2400	}
   2401
   2402	return false;
   2403}
   2404
   2405static bool amt_update_handler(struct amt_dev *amt, struct sk_buff *skb)
   2406{
   2407	struct amt_header_membership_update *amtmu;
   2408	struct amt_tunnel_list *tunnel;
   2409	struct ethhdr *eth;
   2410	struct iphdr *iph;
   2411	int len, hdr_size;
   2412
   2413	iph = ip_hdr(skb);
   2414
   2415	hdr_size = sizeof(*amtmu) + sizeof(struct udphdr);
   2416	if (!pskb_may_pull(skb, hdr_size))
   2417		return true;
   2418
   2419	amtmu = (struct amt_header_membership_update *)(udp_hdr(skb) + 1);
   2420	if (amtmu->reserved || amtmu->version)
   2421		return true;
   2422
   2423	if (iptunnel_pull_header(skb, hdr_size, skb->protocol, false))
   2424		return true;
   2425
   2426	skb_reset_network_header(skb);
   2427
   2428	list_for_each_entry_rcu(tunnel, &amt->tunnel_list, list) {
   2429		if (tunnel->ip4 == iph->saddr) {
   2430			if ((amtmu->nonce == tunnel->nonce &&
   2431			     amtmu->response_mac == tunnel->mac)) {
   2432				mod_delayed_work(amt_wq, &tunnel->gc_wq,
   2433						 msecs_to_jiffies(amt_gmi(amt))
   2434								  * 3);
   2435				goto report;
   2436			} else {
   2437				netdev_dbg(amt->dev, "Invalid MAC\n");
   2438				return true;
   2439			}
   2440		}
   2441	}
   2442
   2443	return true;
   2444
   2445report:
   2446	if (!pskb_may_pull(skb, sizeof(*iph)))
   2447		return true;
   2448
   2449	iph = ip_hdr(skb);
   2450	if (iph->version == 4) {
   2451		if (ip_mc_check_igmp(skb)) {
   2452			netdev_dbg(amt->dev, "Invalid IGMP\n");
   2453			return true;
   2454		}
   2455
   2456		spin_lock_bh(&tunnel->lock);
   2457		amt_igmp_report_handler(amt, skb, tunnel);
   2458		spin_unlock_bh(&tunnel->lock);
   2459
   2460		skb_push(skb, sizeof(struct ethhdr));
   2461		skb_reset_mac_header(skb);
   2462		eth = eth_hdr(skb);
   2463		skb->protocol = htons(ETH_P_IP);
   2464		eth->h_proto = htons(ETH_P_IP);
   2465		ip_eth_mc_map(iph->daddr, eth->h_dest);
   2466#if IS_ENABLED(CONFIG_IPV6)
   2467	} else if (iph->version == 6) {
   2468		struct ipv6hdr *ip6h = ipv6_hdr(skb);
   2469
   2470		if (ipv6_mc_check_mld(skb)) {
   2471			netdev_dbg(amt->dev, "Invalid MLD\n");
   2472			return true;
   2473		}
   2474
   2475		spin_lock_bh(&tunnel->lock);
   2476		amt_mld_report_handler(amt, skb, tunnel);
   2477		spin_unlock_bh(&tunnel->lock);
   2478
   2479		skb_push(skb, sizeof(struct ethhdr));
   2480		skb_reset_mac_header(skb);
   2481		eth = eth_hdr(skb);
   2482		skb->protocol = htons(ETH_P_IPV6);
   2483		eth->h_proto = htons(ETH_P_IPV6);
   2484		ipv6_eth_mc_map(&ip6h->daddr, eth->h_dest);
   2485#endif
   2486	} else {
   2487		netdev_dbg(amt->dev, "Unsupported Protocol\n");
   2488		return true;
   2489	}
   2490
   2491	skb_pull(skb, sizeof(struct ethhdr));
   2492	skb->pkt_type = PACKET_MULTICAST;
   2493	skb->ip_summed = CHECKSUM_NONE;
   2494	len = skb->len;
   2495	if (__netif_rx(skb) == NET_RX_SUCCESS) {
   2496		amt_update_relay_status(tunnel, AMT_STATUS_RECEIVED_UPDATE,
   2497					true);
   2498		dev_sw_netstats_rx_add(amt->dev, len);
   2499	} else {
   2500		amt->dev->stats.rx_dropped++;
   2501	}
   2502
   2503	return false;
   2504}
   2505
   2506static void amt_send_advertisement(struct amt_dev *amt, __be32 nonce,
   2507				   __be32 daddr, __be16 dport)
   2508{
   2509	struct amt_header_advertisement *amta;
   2510	int hlen, tlen, offset;
   2511	struct socket *sock;
   2512	struct udphdr *udph;
   2513	struct sk_buff *skb;
   2514	struct iphdr *iph;
   2515	struct rtable *rt;
   2516	struct flowi4 fl4;
   2517	u32 len;
   2518	int err;
   2519
   2520	rcu_read_lock();
   2521	sock = rcu_dereference(amt->sock);
   2522	if (!sock)
   2523		goto out;
   2524
   2525	if (!netif_running(amt->stream_dev) || !netif_running(amt->dev))
   2526		goto out;
   2527
   2528	rt = ip_route_output_ports(amt->net, &fl4, sock->sk,
   2529				   daddr, amt->local_ip,
   2530				   dport, amt->relay_port,
   2531				   IPPROTO_UDP, 0,
   2532				   amt->stream_dev->ifindex);
   2533	if (IS_ERR(rt)) {
   2534		amt->dev->stats.tx_errors++;
   2535		goto out;
   2536	}
   2537
   2538	hlen = LL_RESERVED_SPACE(amt->dev);
   2539	tlen = amt->dev->needed_tailroom;
   2540	len = hlen + tlen + sizeof(*iph) + sizeof(*udph) + sizeof(*amta);
   2541	skb = netdev_alloc_skb_ip_align(amt->dev, len);
   2542	if (!skb) {
   2543		ip_rt_put(rt);
   2544		amt->dev->stats.tx_errors++;
   2545		goto out;
   2546	}
   2547
   2548	skb->priority = TC_PRIO_CONTROL;
   2549	skb_dst_set(skb, &rt->dst);
   2550
   2551	len = sizeof(*iph) + sizeof(*udph) + sizeof(*amta);
   2552	skb_reset_network_header(skb);
   2553	skb_put(skb, len);
   2554	amta = skb_pull(skb, sizeof(*iph) + sizeof(*udph));
   2555	amta->version	= 0;
   2556	amta->type	= AMT_MSG_ADVERTISEMENT;
   2557	amta->reserved	= 0;
   2558	amta->nonce	= nonce;
   2559	amta->ip4	= amt->local_ip;
   2560	skb_push(skb, sizeof(*udph));
   2561	skb_reset_transport_header(skb);
   2562	udph		= udp_hdr(skb);
   2563	udph->source	= amt->relay_port;
   2564	udph->dest	= dport;
   2565	udph->len	= htons(sizeof(*amta) + sizeof(*udph));
   2566	udph->check	= 0;
   2567	offset = skb_transport_offset(skb);
   2568	skb->csum = skb_checksum(skb, offset, skb->len - offset, 0);
   2569	udph->check = csum_tcpudp_magic(amt->local_ip, daddr,
   2570					sizeof(*udph) + sizeof(*amta),
   2571					IPPROTO_UDP, skb->csum);
   2572
   2573	skb_push(skb, sizeof(*iph));
   2574	iph		= ip_hdr(skb);
   2575	iph->version	= 4;
   2576	iph->ihl	= (sizeof(struct iphdr)) >> 2;
   2577	iph->tos	= AMT_TOS;
   2578	iph->frag_off	= 0;
   2579	iph->ttl	= ip4_dst_hoplimit(&rt->dst);
   2580	iph->daddr	= daddr;
   2581	iph->saddr	= amt->local_ip;
   2582	iph->protocol	= IPPROTO_UDP;
   2583	iph->tot_len	= htons(len);
   2584
   2585	skb->ip_summed = CHECKSUM_NONE;
   2586	ip_select_ident(amt->net, skb, NULL);
   2587	ip_send_check(iph);
   2588	err = ip_local_out(amt->net, sock->sk, skb);
   2589	if (unlikely(net_xmit_eval(err)))
   2590		amt->dev->stats.tx_errors++;
   2591
   2592out:
   2593	rcu_read_unlock();
   2594}
   2595
   2596static bool amt_discovery_handler(struct amt_dev *amt, struct sk_buff *skb)
   2597{
   2598	struct amt_header_discovery *amtd;
   2599	struct udphdr *udph;
   2600	struct iphdr *iph;
   2601
   2602	if (!pskb_may_pull(skb, sizeof(*udph) + sizeof(*amtd)))
   2603		return true;
   2604
   2605	iph = ip_hdr(skb);
   2606	udph = udp_hdr(skb);
   2607	amtd = (struct amt_header_discovery *)(udp_hdr(skb) + 1);
   2608
   2609	if (amtd->reserved || amtd->version)
   2610		return true;
   2611
   2612	amt_send_advertisement(amt, amtd->nonce, iph->saddr, udph->source);
   2613
   2614	return false;
   2615}
   2616
   2617static bool amt_request_handler(struct amt_dev *amt, struct sk_buff *skb)
   2618{
   2619	struct amt_header_request *amtrh;
   2620	struct amt_tunnel_list *tunnel;
   2621	unsigned long long key;
   2622	struct udphdr *udph;
   2623	struct iphdr *iph;
   2624	u64 mac;
   2625	int i;
   2626
   2627	if (!pskb_may_pull(skb, sizeof(*udph) + sizeof(*amtrh)))
   2628		return true;
   2629
   2630	iph = ip_hdr(skb);
   2631	udph = udp_hdr(skb);
   2632	amtrh = (struct amt_header_request *)(udp_hdr(skb) + 1);
   2633
   2634	if (amtrh->reserved1 || amtrh->reserved2 || amtrh->version)
   2635		return true;
   2636
   2637	list_for_each_entry_rcu(tunnel, &amt->tunnel_list, list)
   2638		if (tunnel->ip4 == iph->saddr)
   2639			goto send;
   2640
   2641	if (amt->nr_tunnels >= amt->max_tunnels) {
   2642		icmp_ndo_send(skb, ICMP_DEST_UNREACH, ICMP_HOST_UNREACH, 0);
   2643		return true;
   2644	}
   2645
   2646	tunnel = kzalloc(sizeof(*tunnel) +
   2647			 (sizeof(struct hlist_head) * amt->hash_buckets),
   2648			 GFP_ATOMIC);
   2649	if (!tunnel)
   2650		return true;
   2651
   2652	tunnel->source_port = udph->source;
   2653	tunnel->ip4 = iph->saddr;
   2654
   2655	memcpy(&key, &tunnel->key, sizeof(unsigned long long));
   2656	tunnel->amt = amt;
   2657	spin_lock_init(&tunnel->lock);
   2658	for (i = 0; i < amt->hash_buckets; i++)
   2659		INIT_HLIST_HEAD(&tunnel->groups[i]);
   2660
   2661	INIT_DELAYED_WORK(&tunnel->gc_wq, amt_tunnel_expire);
   2662
   2663	spin_lock_bh(&amt->lock);
   2664	list_add_tail_rcu(&tunnel->list, &amt->tunnel_list);
   2665	tunnel->key = amt->key;
   2666	amt_update_relay_status(tunnel, AMT_STATUS_RECEIVED_REQUEST, true);
   2667	amt->nr_tunnels++;
   2668	mod_delayed_work(amt_wq, &tunnel->gc_wq,
   2669			 msecs_to_jiffies(amt_gmi(amt)));
   2670	spin_unlock_bh(&amt->lock);
   2671
   2672send:
   2673	tunnel->nonce = amtrh->nonce;
   2674	mac = siphash_3u32((__force u32)tunnel->ip4,
   2675			   (__force u32)tunnel->source_port,
   2676			   (__force u32)tunnel->nonce,
   2677			   &tunnel->key);
   2678	tunnel->mac = mac >> 16;
   2679
   2680	if (!netif_running(amt->dev) || !netif_running(amt->stream_dev))
   2681		return true;
   2682
   2683	if (!amtrh->p)
   2684		amt_send_igmp_gq(amt, tunnel);
   2685	else
   2686		amt_send_mld_gq(amt, tunnel);
   2687
   2688	return false;
   2689}
   2690
   2691static int amt_rcv(struct sock *sk, struct sk_buff *skb)
   2692{
   2693	struct amt_dev *amt;
   2694	struct iphdr *iph;
   2695	int type;
   2696	bool err;
   2697
   2698	rcu_read_lock_bh();
   2699	amt = rcu_dereference_sk_user_data(sk);
   2700	if (!amt) {
   2701		err = true;
   2702		kfree_skb(skb);
   2703		goto out;
   2704	}
   2705
   2706	skb->dev = amt->dev;
   2707	iph = ip_hdr(skb);
   2708	type = amt_parse_type(skb);
   2709	if (type == -1) {
   2710		err = true;
   2711		goto drop;
   2712	}
   2713
   2714	if (amt->mode == AMT_MODE_GATEWAY) {
   2715		switch (type) {
   2716		case AMT_MSG_ADVERTISEMENT:
   2717			if (iph->saddr != amt->discovery_ip) {
   2718				netdev_dbg(amt->dev, "Invalid Relay IP\n");
   2719				err = true;
   2720				goto drop;
   2721			}
   2722			err = amt_advertisement_handler(amt, skb);
   2723			break;
   2724		case AMT_MSG_MULTICAST_DATA:
   2725			if (iph->saddr != amt->remote_ip) {
   2726				netdev_dbg(amt->dev, "Invalid Relay IP\n");
   2727				err = true;
   2728				goto drop;
   2729			}
   2730			err = amt_multicast_data_handler(amt, skb);
   2731			if (err)
   2732				goto drop;
   2733			else
   2734				goto out;
   2735		case AMT_MSG_MEMBERSHIP_QUERY:
   2736			if (iph->saddr != amt->remote_ip) {
   2737				netdev_dbg(amt->dev, "Invalid Relay IP\n");
   2738				err = true;
   2739				goto drop;
   2740			}
   2741			err = amt_membership_query_handler(amt, skb);
   2742			if (err)
   2743				goto drop;
   2744			else
   2745				goto out;
   2746		default:
   2747			err = true;
   2748			netdev_dbg(amt->dev, "Invalid type of Gateway\n");
   2749			break;
   2750		}
   2751	} else {
   2752		switch (type) {
   2753		case AMT_MSG_DISCOVERY:
   2754			err = amt_discovery_handler(amt, skb);
   2755			break;
   2756		case AMT_MSG_REQUEST:
   2757			err = amt_request_handler(amt, skb);
   2758			break;
   2759		case AMT_MSG_MEMBERSHIP_UPDATE:
   2760			err = amt_update_handler(amt, skb);
   2761			if (err)
   2762				goto drop;
   2763			else
   2764				goto out;
   2765		default:
   2766			err = true;
   2767			netdev_dbg(amt->dev, "Invalid type of relay\n");
   2768			break;
   2769		}
   2770	}
   2771drop:
   2772	if (err) {
   2773		amt->dev->stats.rx_dropped++;
   2774		kfree_skb(skb);
   2775	} else {
   2776		consume_skb(skb);
   2777	}
   2778out:
   2779	rcu_read_unlock_bh();
   2780	return 0;
   2781}
   2782
   2783static int amt_err_lookup(struct sock *sk, struct sk_buff *skb)
   2784{
   2785	struct amt_dev *amt;
   2786	int type;
   2787
   2788	rcu_read_lock_bh();
   2789	amt = rcu_dereference_sk_user_data(sk);
   2790	if (!amt)
   2791		goto out;
   2792
   2793	if (amt->mode != AMT_MODE_GATEWAY)
   2794		goto drop;
   2795
   2796	type = amt_parse_type(skb);
   2797	if (type == -1)
   2798		goto drop;
   2799
   2800	netdev_dbg(amt->dev, "Received IGMP Unreachable of %s\n",
   2801		   type_str[type]);
   2802	switch (type) {
   2803	case AMT_MSG_DISCOVERY:
   2804		break;
   2805	case AMT_MSG_REQUEST:
   2806	case AMT_MSG_MEMBERSHIP_UPDATE:
   2807		if (amt->status >= AMT_STATUS_RECEIVED_ADVERTISEMENT)
   2808			mod_delayed_work(amt_wq, &amt->req_wq, 0);
   2809		break;
   2810	default:
   2811		goto drop;
   2812	}
   2813out:
   2814	rcu_read_unlock_bh();
   2815	return 0;
   2816drop:
   2817	rcu_read_unlock_bh();
   2818	amt->dev->stats.rx_dropped++;
   2819	return 0;
   2820}
   2821
   2822static struct socket *amt_create_sock(struct net *net, __be16 port)
   2823{
   2824	struct udp_port_cfg udp_conf;
   2825	struct socket *sock;
   2826	int err;
   2827
   2828	memset(&udp_conf, 0, sizeof(udp_conf));
   2829	udp_conf.family = AF_INET;
   2830	udp_conf.local_ip.s_addr = htonl(INADDR_ANY);
   2831
   2832	udp_conf.local_udp_port = port;
   2833
   2834	err = udp_sock_create(net, &udp_conf, &sock);
   2835	if (err < 0)
   2836		return ERR_PTR(err);
   2837
   2838	return sock;
   2839}
   2840
   2841static int amt_socket_create(struct amt_dev *amt)
   2842{
   2843	struct udp_tunnel_sock_cfg tunnel_cfg;
   2844	struct socket *sock;
   2845
   2846	sock = amt_create_sock(amt->net, amt->relay_port);
   2847	if (IS_ERR(sock))
   2848		return PTR_ERR(sock);
   2849
   2850	/* Mark socket as an encapsulation socket */
   2851	memset(&tunnel_cfg, 0, sizeof(tunnel_cfg));
   2852	tunnel_cfg.sk_user_data = amt;
   2853	tunnel_cfg.encap_type = 1;
   2854	tunnel_cfg.encap_rcv = amt_rcv;
   2855	tunnel_cfg.encap_err_lookup = amt_err_lookup;
   2856	tunnel_cfg.encap_destroy = NULL;
   2857	setup_udp_tunnel_sock(amt->net, sock, &tunnel_cfg);
   2858
   2859	rcu_assign_pointer(amt->sock, sock);
   2860	return 0;
   2861}
   2862
   2863static int amt_dev_open(struct net_device *dev)
   2864{
   2865	struct amt_dev *amt = netdev_priv(dev);
   2866	int err;
   2867
   2868	amt->ready4 = false;
   2869	amt->ready6 = false;
   2870
   2871	err = amt_socket_create(amt);
   2872	if (err)
   2873		return err;
   2874
   2875	amt->req_cnt = 0;
   2876	amt->remote_ip = 0;
   2877	get_random_bytes(&amt->key, sizeof(siphash_key_t));
   2878
   2879	amt->status = AMT_STATUS_INIT;
   2880	if (amt->mode == AMT_MODE_GATEWAY) {
   2881		mod_delayed_work(amt_wq, &amt->discovery_wq, 0);
   2882		mod_delayed_work(amt_wq, &amt->req_wq, 0);
   2883	} else if (amt->mode == AMT_MODE_RELAY) {
   2884		mod_delayed_work(amt_wq, &amt->secret_wq,
   2885				 msecs_to_jiffies(AMT_SECRET_TIMEOUT));
   2886	}
   2887	return err;
   2888}
   2889
   2890static int amt_dev_stop(struct net_device *dev)
   2891{
   2892	struct amt_dev *amt = netdev_priv(dev);
   2893	struct amt_tunnel_list *tunnel, *tmp;
   2894	struct socket *sock;
   2895
   2896	cancel_delayed_work_sync(&amt->req_wq);
   2897	cancel_delayed_work_sync(&amt->discovery_wq);
   2898	cancel_delayed_work_sync(&amt->secret_wq);
   2899
   2900	/* shutdown */
   2901	sock = rtnl_dereference(amt->sock);
   2902	RCU_INIT_POINTER(amt->sock, NULL);
   2903	synchronize_net();
   2904	if (sock)
   2905		udp_tunnel_sock_release(sock);
   2906
   2907	amt->ready4 = false;
   2908	amt->ready6 = false;
   2909	amt->req_cnt = 0;
   2910	amt->remote_ip = 0;
   2911
   2912	list_for_each_entry_safe(tunnel, tmp, &amt->tunnel_list, list) {
   2913		list_del_rcu(&tunnel->list);
   2914		amt->nr_tunnels--;
   2915		cancel_delayed_work_sync(&tunnel->gc_wq);
   2916		amt_clear_groups(tunnel);
   2917		kfree_rcu(tunnel, rcu);
   2918	}
   2919
   2920	return 0;
   2921}
   2922
   2923static const struct device_type amt_type = {
   2924	.name = "amt",
   2925};
   2926
   2927static int amt_dev_init(struct net_device *dev)
   2928{
   2929	struct amt_dev *amt = netdev_priv(dev);
   2930	int err;
   2931
   2932	amt->dev = dev;
   2933	dev->tstats = netdev_alloc_pcpu_stats(struct pcpu_sw_netstats);
   2934	if (!dev->tstats)
   2935		return -ENOMEM;
   2936
   2937	err = gro_cells_init(&amt->gro_cells, dev);
   2938	if (err) {
   2939		free_percpu(dev->tstats);
   2940		return err;
   2941	}
   2942
   2943	return 0;
   2944}
   2945
   2946static void amt_dev_uninit(struct net_device *dev)
   2947{
   2948	struct amt_dev *amt = netdev_priv(dev);
   2949
   2950	gro_cells_destroy(&amt->gro_cells);
   2951	free_percpu(dev->tstats);
   2952}
   2953
   2954static const struct net_device_ops amt_netdev_ops = {
   2955	.ndo_init               = amt_dev_init,
   2956	.ndo_uninit             = amt_dev_uninit,
   2957	.ndo_open		= amt_dev_open,
   2958	.ndo_stop		= amt_dev_stop,
   2959	.ndo_start_xmit         = amt_dev_xmit,
   2960	.ndo_get_stats64        = dev_get_tstats64,
   2961};
   2962
   2963static void amt_link_setup(struct net_device *dev)
   2964{
   2965	dev->netdev_ops         = &amt_netdev_ops;
   2966	dev->needs_free_netdev  = true;
   2967	SET_NETDEV_DEVTYPE(dev, &amt_type);
   2968	dev->min_mtu		= ETH_MIN_MTU;
   2969	dev->max_mtu		= ETH_MAX_MTU;
   2970	dev->type		= ARPHRD_NONE;
   2971	dev->flags		= IFF_POINTOPOINT | IFF_NOARP | IFF_MULTICAST;
   2972	dev->hard_header_len	= 0;
   2973	dev->addr_len		= 0;
   2974	dev->priv_flags		|= IFF_NO_QUEUE;
   2975	dev->features		|= NETIF_F_LLTX;
   2976	dev->features		|= NETIF_F_GSO_SOFTWARE;
   2977	dev->features		|= NETIF_F_NETNS_LOCAL;
   2978	dev->hw_features	|= NETIF_F_SG | NETIF_F_HW_CSUM;
   2979	dev->hw_features	|= NETIF_F_FRAGLIST | NETIF_F_RXCSUM;
   2980	dev->hw_features	|= NETIF_F_GSO_SOFTWARE;
   2981	eth_hw_addr_random(dev);
   2982	eth_zero_addr(dev->broadcast);
   2983	ether_setup(dev);
   2984}
   2985
   2986static const struct nla_policy amt_policy[IFLA_AMT_MAX + 1] = {
   2987	[IFLA_AMT_MODE]		= { .type = NLA_U32 },
   2988	[IFLA_AMT_RELAY_PORT]	= { .type = NLA_U16 },
   2989	[IFLA_AMT_GATEWAY_PORT]	= { .type = NLA_U16 },
   2990	[IFLA_AMT_LINK]		= { .type = NLA_U32 },
   2991	[IFLA_AMT_LOCAL_IP]	= { .len = sizeof_field(struct iphdr, daddr) },
   2992	[IFLA_AMT_REMOTE_IP]	= { .len = sizeof_field(struct iphdr, daddr) },
   2993	[IFLA_AMT_DISCOVERY_IP]	= { .len = sizeof_field(struct iphdr, daddr) },
   2994	[IFLA_AMT_MAX_TUNNELS]	= { .type = NLA_U32 },
   2995};
   2996
   2997static int amt_validate(struct nlattr *tb[], struct nlattr *data[],
   2998			struct netlink_ext_ack *extack)
   2999{
   3000	if (!data)
   3001		return -EINVAL;
   3002
   3003	if (!data[IFLA_AMT_LINK]) {
   3004		NL_SET_ERR_MSG_ATTR(extack, data[IFLA_AMT_LINK],
   3005				    "Link attribute is required");
   3006		return -EINVAL;
   3007	}
   3008
   3009	if (!data[IFLA_AMT_MODE]) {
   3010		NL_SET_ERR_MSG_ATTR(extack, data[IFLA_AMT_MODE],
   3011				    "Mode attribute is required");
   3012		return -EINVAL;
   3013	}
   3014
   3015	if (nla_get_u32(data[IFLA_AMT_MODE]) > AMT_MODE_MAX) {
   3016		NL_SET_ERR_MSG_ATTR(extack, data[IFLA_AMT_MODE],
   3017				    "Mode attribute is not valid");
   3018		return -EINVAL;
   3019	}
   3020
   3021	if (!data[IFLA_AMT_LOCAL_IP]) {
   3022		NL_SET_ERR_MSG_ATTR(extack, data[IFLA_AMT_DISCOVERY_IP],
   3023				    "Local attribute is required");
   3024		return -EINVAL;
   3025	}
   3026
   3027	if (!data[IFLA_AMT_DISCOVERY_IP] &&
   3028	    nla_get_u32(data[IFLA_AMT_MODE]) == AMT_MODE_GATEWAY) {
   3029		NL_SET_ERR_MSG_ATTR(extack, data[IFLA_AMT_LOCAL_IP],
   3030				    "Discovery attribute is required");
   3031		return -EINVAL;
   3032	}
   3033
   3034	return 0;
   3035}
   3036
   3037static int amt_newlink(struct net *net, struct net_device *dev,
   3038		       struct nlattr *tb[], struct nlattr *data[],
   3039		       struct netlink_ext_ack *extack)
   3040{
   3041	struct amt_dev *amt = netdev_priv(dev);
   3042	int err = -EINVAL;
   3043
   3044	amt->net = net;
   3045	amt->mode = nla_get_u32(data[IFLA_AMT_MODE]);
   3046
   3047	if (data[IFLA_AMT_MAX_TUNNELS] &&
   3048	    nla_get_u32(data[IFLA_AMT_MAX_TUNNELS]))
   3049		amt->max_tunnels = nla_get_u32(data[IFLA_AMT_MAX_TUNNELS]);
   3050	else
   3051		amt->max_tunnels = AMT_MAX_TUNNELS;
   3052
   3053	spin_lock_init(&amt->lock);
   3054	amt->max_groups = AMT_MAX_GROUP;
   3055	amt->max_sources = AMT_MAX_SOURCE;
   3056	amt->hash_buckets = AMT_HSIZE;
   3057	amt->nr_tunnels = 0;
   3058	get_random_bytes(&amt->hash_seed, sizeof(amt->hash_seed));
   3059	amt->stream_dev = dev_get_by_index(net,
   3060					   nla_get_u32(data[IFLA_AMT_LINK]));
   3061	if (!amt->stream_dev) {
   3062		NL_SET_ERR_MSG_ATTR(extack, tb[IFLA_AMT_LINK],
   3063				    "Can't find stream device");
   3064		return -ENODEV;
   3065	}
   3066
   3067	if (amt->stream_dev->type != ARPHRD_ETHER) {
   3068		NL_SET_ERR_MSG_ATTR(extack, tb[IFLA_AMT_LINK],
   3069				    "Invalid stream device type");
   3070		goto err;
   3071	}
   3072
   3073	amt->local_ip = nla_get_in_addr(data[IFLA_AMT_LOCAL_IP]);
   3074	if (ipv4_is_loopback(amt->local_ip) ||
   3075	    ipv4_is_zeronet(amt->local_ip) ||
   3076	    ipv4_is_multicast(amt->local_ip)) {
   3077		NL_SET_ERR_MSG_ATTR(extack, tb[IFLA_AMT_LOCAL_IP],
   3078				    "Invalid Local address");
   3079		goto err;
   3080	}
   3081
   3082	if (data[IFLA_AMT_RELAY_PORT])
   3083		amt->relay_port = nla_get_be16(data[IFLA_AMT_RELAY_PORT]);
   3084	else
   3085		amt->relay_port = htons(IANA_AMT_UDP_PORT);
   3086
   3087	if (data[IFLA_AMT_GATEWAY_PORT])
   3088		amt->gw_port = nla_get_be16(data[IFLA_AMT_GATEWAY_PORT]);
   3089	else
   3090		amt->gw_port = htons(IANA_AMT_UDP_PORT);
   3091
   3092	if (!amt->relay_port) {
   3093		NL_SET_ERR_MSG_ATTR(extack, tb[IFLA_AMT_DISCOVERY_IP],
   3094				    "relay port must not be 0");
   3095		goto err;
   3096	}
   3097	if (amt->mode == AMT_MODE_RELAY) {
   3098		amt->qrv = amt->net->ipv4.sysctl_igmp_qrv;
   3099		amt->qri = 10;
   3100		dev->needed_headroom = amt->stream_dev->needed_headroom +
   3101				       AMT_RELAY_HLEN;
   3102		dev->mtu = amt->stream_dev->mtu - AMT_RELAY_HLEN;
   3103		dev->max_mtu = dev->mtu;
   3104		dev->min_mtu = ETH_MIN_MTU + AMT_RELAY_HLEN;
   3105	} else {
   3106		if (!data[IFLA_AMT_DISCOVERY_IP]) {
   3107			NL_SET_ERR_MSG_ATTR(extack, tb[IFLA_AMT_DISCOVERY_IP],
   3108					    "discovery must be set in gateway mode");
   3109			goto err;
   3110		}
   3111		if (!amt->gw_port) {
   3112			NL_SET_ERR_MSG_ATTR(extack, tb[IFLA_AMT_DISCOVERY_IP],
   3113					    "gateway port must not be 0");
   3114			goto err;
   3115		}
   3116		amt->remote_ip = 0;
   3117		amt->discovery_ip = nla_get_in_addr(data[IFLA_AMT_DISCOVERY_IP]);
   3118		if (ipv4_is_loopback(amt->discovery_ip) ||
   3119		    ipv4_is_zeronet(amt->discovery_ip) ||
   3120		    ipv4_is_multicast(amt->discovery_ip)) {
   3121			NL_SET_ERR_MSG_ATTR(extack, tb[IFLA_AMT_DISCOVERY_IP],
   3122					    "discovery must be unicast");
   3123			goto err;
   3124		}
   3125
   3126		dev->needed_headroom = amt->stream_dev->needed_headroom +
   3127				       AMT_GW_HLEN;
   3128		dev->mtu = amt->stream_dev->mtu - AMT_GW_HLEN;
   3129		dev->max_mtu = dev->mtu;
   3130		dev->min_mtu = ETH_MIN_MTU + AMT_GW_HLEN;
   3131	}
   3132	amt->qi = AMT_INIT_QUERY_INTERVAL;
   3133
   3134	err = register_netdevice(dev);
   3135	if (err < 0) {
   3136		netdev_dbg(dev, "failed to register new netdev %d\n", err);
   3137		goto err;
   3138	}
   3139
   3140	err = netdev_upper_dev_link(amt->stream_dev, dev, extack);
   3141	if (err < 0) {
   3142		unregister_netdevice(dev);
   3143		goto err;
   3144	}
   3145
   3146	INIT_DELAYED_WORK(&amt->discovery_wq, amt_discovery_work);
   3147	INIT_DELAYED_WORK(&amt->req_wq, amt_req_work);
   3148	INIT_DELAYED_WORK(&amt->secret_wq, amt_secret_work);
   3149	INIT_LIST_HEAD(&amt->tunnel_list);
   3150
   3151	return 0;
   3152err:
   3153	dev_put(amt->stream_dev);
   3154	return err;
   3155}
   3156
   3157static void amt_dellink(struct net_device *dev, struct list_head *head)
   3158{
   3159	struct amt_dev *amt = netdev_priv(dev);
   3160
   3161	unregister_netdevice_queue(dev, head);
   3162	netdev_upper_dev_unlink(amt->stream_dev, dev);
   3163	dev_put(amt->stream_dev);
   3164}
   3165
   3166static size_t amt_get_size(const struct net_device *dev)
   3167{
   3168	return nla_total_size(sizeof(__u32)) + /* IFLA_AMT_MODE */
   3169	       nla_total_size(sizeof(__u16)) + /* IFLA_AMT_RELAY_PORT */
   3170	       nla_total_size(sizeof(__u16)) + /* IFLA_AMT_GATEWAY_PORT */
   3171	       nla_total_size(sizeof(__u32)) + /* IFLA_AMT_LINK */
   3172	       nla_total_size(sizeof(__u32)) + /* IFLA_MAX_TUNNELS */
   3173	       nla_total_size(sizeof(struct iphdr)) + /* IFLA_AMT_DISCOVERY_IP */
   3174	       nla_total_size(sizeof(struct iphdr)) + /* IFLA_AMT_REMOTE_IP */
   3175	       nla_total_size(sizeof(struct iphdr)); /* IFLA_AMT_LOCAL_IP */
   3176}
   3177
   3178static int amt_fill_info(struct sk_buff *skb, const struct net_device *dev)
   3179{
   3180	struct amt_dev *amt = netdev_priv(dev);
   3181
   3182	if (nla_put_u32(skb, IFLA_AMT_MODE, amt->mode))
   3183		goto nla_put_failure;
   3184	if (nla_put_be16(skb, IFLA_AMT_RELAY_PORT, amt->relay_port))
   3185		goto nla_put_failure;
   3186	if (nla_put_be16(skb, IFLA_AMT_GATEWAY_PORT, amt->gw_port))
   3187		goto nla_put_failure;
   3188	if (nla_put_u32(skb, IFLA_AMT_LINK, amt->stream_dev->ifindex))
   3189		goto nla_put_failure;
   3190	if (nla_put_in_addr(skb, IFLA_AMT_LOCAL_IP, amt->local_ip))
   3191		goto nla_put_failure;
   3192	if (nla_put_in_addr(skb, IFLA_AMT_DISCOVERY_IP, amt->discovery_ip))
   3193		goto nla_put_failure;
   3194	if (amt->remote_ip)
   3195		if (nla_put_in_addr(skb, IFLA_AMT_REMOTE_IP, amt->remote_ip))
   3196			goto nla_put_failure;
   3197	if (nla_put_u32(skb, IFLA_AMT_MAX_TUNNELS, amt->max_tunnels))
   3198		goto nla_put_failure;
   3199
   3200	return 0;
   3201
   3202nla_put_failure:
   3203	return -EMSGSIZE;
   3204}
   3205
   3206static struct rtnl_link_ops amt_link_ops __read_mostly = {
   3207	.kind		= "amt",
   3208	.maxtype	= IFLA_AMT_MAX,
   3209	.policy		= amt_policy,
   3210	.priv_size	= sizeof(struct amt_dev),
   3211	.setup		= amt_link_setup,
   3212	.validate	= amt_validate,
   3213	.newlink	= amt_newlink,
   3214	.dellink	= amt_dellink,
   3215	.get_size       = amt_get_size,
   3216	.fill_info      = amt_fill_info,
   3217};
   3218
   3219static struct net_device *amt_lookup_upper_dev(struct net_device *dev)
   3220{
   3221	struct net_device *upper_dev;
   3222	struct amt_dev *amt;
   3223
   3224	for_each_netdev(dev_net(dev), upper_dev) {
   3225		if (netif_is_amt(upper_dev)) {
   3226			amt = netdev_priv(upper_dev);
   3227			if (amt->stream_dev == dev)
   3228				return upper_dev;
   3229		}
   3230	}
   3231
   3232	return NULL;
   3233}
   3234
   3235static int amt_device_event(struct notifier_block *unused,
   3236			    unsigned long event, void *ptr)
   3237{
   3238	struct net_device *dev = netdev_notifier_info_to_dev(ptr);
   3239	struct net_device *upper_dev;
   3240	struct amt_dev *amt;
   3241	LIST_HEAD(list);
   3242	int new_mtu;
   3243
   3244	upper_dev = amt_lookup_upper_dev(dev);
   3245	if (!upper_dev)
   3246		return NOTIFY_DONE;
   3247	amt = netdev_priv(upper_dev);
   3248
   3249	switch (event) {
   3250	case NETDEV_UNREGISTER:
   3251		amt_dellink(amt->dev, &list);
   3252		unregister_netdevice_many(&list);
   3253		break;
   3254	case NETDEV_CHANGEMTU:
   3255		if (amt->mode == AMT_MODE_RELAY)
   3256			new_mtu = dev->mtu - AMT_RELAY_HLEN;
   3257		else
   3258			new_mtu = dev->mtu - AMT_GW_HLEN;
   3259
   3260		dev_set_mtu(amt->dev, new_mtu);
   3261		break;
   3262	}
   3263
   3264	return NOTIFY_DONE;
   3265}
   3266
   3267static struct notifier_block amt_notifier_block __read_mostly = {
   3268	.notifier_call = amt_device_event,
   3269};
   3270
   3271static int __init amt_init(void)
   3272{
   3273	int err;
   3274
   3275	err = register_netdevice_notifier(&amt_notifier_block);
   3276	if (err < 0)
   3277		goto err;
   3278
   3279	err = rtnl_link_register(&amt_link_ops);
   3280	if (err < 0)
   3281		goto unregister_notifier;
   3282
   3283	amt_wq = alloc_workqueue("amt", WQ_UNBOUND, 1);
   3284	if (!amt_wq) {
   3285		err = -ENOMEM;
   3286		goto rtnl_unregister;
   3287	}
   3288
   3289	spin_lock_init(&source_gc_lock);
   3290	spin_lock_bh(&source_gc_lock);
   3291	INIT_DELAYED_WORK(&source_gc_wq, amt_source_gc_work);
   3292	mod_delayed_work(amt_wq, &source_gc_wq,
   3293			 msecs_to_jiffies(AMT_GC_INTERVAL));
   3294	spin_unlock_bh(&source_gc_lock);
   3295
   3296	return 0;
   3297
   3298rtnl_unregister:
   3299	rtnl_link_unregister(&amt_link_ops);
   3300unregister_notifier:
   3301	unregister_netdevice_notifier(&amt_notifier_block);
   3302err:
   3303	pr_err("error loading AMT module loaded\n");
   3304	return err;
   3305}
   3306late_initcall(amt_init);
   3307
   3308static void __exit amt_fini(void)
   3309{
   3310	rtnl_link_unregister(&amt_link_ops);
   3311	unregister_netdevice_notifier(&amt_notifier_block);
   3312	cancel_delayed_work_sync(&source_gc_wq);
   3313	__amt_source_gc_work();
   3314	destroy_workqueue(amt_wq);
   3315}
   3316module_exit(amt_fini);
   3317
   3318MODULE_LICENSE("GPL");
   3319MODULE_AUTHOR("Taehee Yoo <ap420073@gmail.com>");
   3320MODULE_ALIAS_RTNL_LINK("amt");