cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

dsa_priv.h (19431B)


      1/* SPDX-License-Identifier: GPL-2.0-or-later */
      2/*
      3 * net/dsa/dsa_priv.h - Hardware switch handling
      4 * Copyright (c) 2008-2009 Marvell Semiconductor
      5 */
      6
      7#ifndef __DSA_PRIV_H
      8#define __DSA_PRIV_H
      9
     10#include <linux/if_bridge.h>
     11#include <linux/if_vlan.h>
     12#include <linux/phy.h>
     13#include <linux/netdevice.h>
     14#include <linux/netpoll.h>
     15#include <net/dsa.h>
     16#include <net/gro_cells.h>
     17
     18#define DSA_MAX_NUM_OFFLOADING_BRIDGES		BITS_PER_LONG
     19
     20enum {
     21	DSA_NOTIFIER_AGEING_TIME,
     22	DSA_NOTIFIER_BRIDGE_JOIN,
     23	DSA_NOTIFIER_BRIDGE_LEAVE,
     24	DSA_NOTIFIER_FDB_ADD,
     25	DSA_NOTIFIER_FDB_DEL,
     26	DSA_NOTIFIER_HOST_FDB_ADD,
     27	DSA_NOTIFIER_HOST_FDB_DEL,
     28	DSA_NOTIFIER_LAG_FDB_ADD,
     29	DSA_NOTIFIER_LAG_FDB_DEL,
     30	DSA_NOTIFIER_LAG_CHANGE,
     31	DSA_NOTIFIER_LAG_JOIN,
     32	DSA_NOTIFIER_LAG_LEAVE,
     33	DSA_NOTIFIER_MDB_ADD,
     34	DSA_NOTIFIER_MDB_DEL,
     35	DSA_NOTIFIER_HOST_MDB_ADD,
     36	DSA_NOTIFIER_HOST_MDB_DEL,
     37	DSA_NOTIFIER_VLAN_ADD,
     38	DSA_NOTIFIER_VLAN_DEL,
     39	DSA_NOTIFIER_HOST_VLAN_ADD,
     40	DSA_NOTIFIER_HOST_VLAN_DEL,
     41	DSA_NOTIFIER_MTU,
     42	DSA_NOTIFIER_TAG_PROTO,
     43	DSA_NOTIFIER_TAG_PROTO_CONNECT,
     44	DSA_NOTIFIER_TAG_PROTO_DISCONNECT,
     45	DSA_NOTIFIER_TAG_8021Q_VLAN_ADD,
     46	DSA_NOTIFIER_TAG_8021Q_VLAN_DEL,
     47	DSA_NOTIFIER_MASTER_STATE_CHANGE,
     48};
     49
     50/* DSA_NOTIFIER_AGEING_TIME */
     51struct dsa_notifier_ageing_time_info {
     52	unsigned int ageing_time;
     53};
     54
     55/* DSA_NOTIFIER_BRIDGE_* */
     56struct dsa_notifier_bridge_info {
     57	const struct dsa_port *dp;
     58	struct dsa_bridge bridge;
     59	bool tx_fwd_offload;
     60	struct netlink_ext_ack *extack;
     61};
     62
     63/* DSA_NOTIFIER_FDB_* */
     64struct dsa_notifier_fdb_info {
     65	const struct dsa_port *dp;
     66	const unsigned char *addr;
     67	u16 vid;
     68	struct dsa_db db;
     69};
     70
     71/* DSA_NOTIFIER_LAG_FDB_* */
     72struct dsa_notifier_lag_fdb_info {
     73	struct dsa_lag *lag;
     74	const unsigned char *addr;
     75	u16 vid;
     76	struct dsa_db db;
     77};
     78
     79/* DSA_NOTIFIER_MDB_* */
     80struct dsa_notifier_mdb_info {
     81	const struct dsa_port *dp;
     82	const struct switchdev_obj_port_mdb *mdb;
     83	struct dsa_db db;
     84};
     85
     86/* DSA_NOTIFIER_LAG_* */
     87struct dsa_notifier_lag_info {
     88	const struct dsa_port *dp;
     89	struct dsa_lag lag;
     90	struct netdev_lag_upper_info *info;
     91};
     92
     93/* DSA_NOTIFIER_VLAN_* */
     94struct dsa_notifier_vlan_info {
     95	const struct dsa_port *dp;
     96	const struct switchdev_obj_port_vlan *vlan;
     97	struct netlink_ext_ack *extack;
     98};
     99
    100/* DSA_NOTIFIER_MTU */
    101struct dsa_notifier_mtu_info {
    102	const struct dsa_port *dp;
    103	int mtu;
    104};
    105
    106/* DSA_NOTIFIER_TAG_PROTO_* */
    107struct dsa_notifier_tag_proto_info {
    108	const struct dsa_device_ops *tag_ops;
    109};
    110
    111/* DSA_NOTIFIER_TAG_8021Q_VLAN_* */
    112struct dsa_notifier_tag_8021q_vlan_info {
    113	const struct dsa_port *dp;
    114	u16 vid;
    115};
    116
    117/* DSA_NOTIFIER_MASTER_STATE_CHANGE */
    118struct dsa_notifier_master_state_info {
    119	const struct net_device *master;
    120	bool operational;
    121};
    122
    123struct dsa_switchdev_event_work {
    124	struct net_device *dev;
    125	struct net_device *orig_dev;
    126	struct work_struct work;
    127	unsigned long event;
    128	/* Specific for SWITCHDEV_FDB_ADD_TO_DEVICE and
    129	 * SWITCHDEV_FDB_DEL_TO_DEVICE
    130	 */
    131	unsigned char addr[ETH_ALEN];
    132	u16 vid;
    133	bool host_addr;
    134};
    135
    136enum dsa_standalone_event {
    137	DSA_UC_ADD,
    138	DSA_UC_DEL,
    139	DSA_MC_ADD,
    140	DSA_MC_DEL,
    141};
    142
    143struct dsa_standalone_event_work {
    144	struct work_struct work;
    145	struct net_device *dev;
    146	enum dsa_standalone_event event;
    147	unsigned char addr[ETH_ALEN];
    148	u16 vid;
    149};
    150
    151struct dsa_slave_priv {
    152	/* Copy of CPU port xmit for faster access in slave transmit hot path */
    153	struct sk_buff *	(*xmit)(struct sk_buff *skb,
    154					struct net_device *dev);
    155
    156	struct gro_cells	gcells;
    157
    158	/* DSA port data, such as switch, port index, etc. */
    159	struct dsa_port		*dp;
    160
    161#ifdef CONFIG_NET_POLL_CONTROLLER
    162	struct netpoll		*netpoll;
    163#endif
    164
    165	/* TC context */
    166	struct list_head	mall_tc_list;
    167};
    168
    169/* dsa.c */
    170const struct dsa_device_ops *dsa_tag_driver_get(int tag_protocol);
    171void dsa_tag_driver_put(const struct dsa_device_ops *ops);
    172const struct dsa_device_ops *dsa_find_tagger_by_name(const char *buf);
    173
    174bool dsa_db_equal(const struct dsa_db *a, const struct dsa_db *b);
    175
    176bool dsa_schedule_work(struct work_struct *work);
    177const char *dsa_tag_protocol_to_str(const struct dsa_device_ops *ops);
    178
    179static inline int dsa_tag_protocol_overhead(const struct dsa_device_ops *ops)
    180{
    181	return ops->needed_headroom + ops->needed_tailroom;
    182}
    183
    184/* master.c */
    185int dsa_master_setup(struct net_device *dev, struct dsa_port *cpu_dp);
    186void dsa_master_teardown(struct net_device *dev);
    187
    188static inline struct net_device *dsa_master_find_slave(struct net_device *dev,
    189						       int device, int port)
    190{
    191	struct dsa_port *cpu_dp = dev->dsa_ptr;
    192	struct dsa_switch_tree *dst = cpu_dp->dst;
    193	struct dsa_port *dp;
    194
    195	list_for_each_entry(dp, &dst->ports, list)
    196		if (dp->ds->index == device && dp->index == port &&
    197		    dp->type == DSA_PORT_TYPE_USER)
    198			return dp->slave;
    199
    200	return NULL;
    201}
    202
    203/* port.c */
    204void dsa_port_set_tag_protocol(struct dsa_port *cpu_dp,
    205			       const struct dsa_device_ops *tag_ops);
    206int dsa_port_set_state(struct dsa_port *dp, u8 state, bool do_fast_age);
    207int dsa_port_set_mst_state(struct dsa_port *dp,
    208			   const struct switchdev_mst_state *state,
    209			   struct netlink_ext_ack *extack);
    210int dsa_port_enable_rt(struct dsa_port *dp, struct phy_device *phy);
    211int dsa_port_enable(struct dsa_port *dp, struct phy_device *phy);
    212void dsa_port_disable_rt(struct dsa_port *dp);
    213void dsa_port_disable(struct dsa_port *dp);
    214int dsa_port_bridge_join(struct dsa_port *dp, struct net_device *br,
    215			 struct netlink_ext_ack *extack);
    216void dsa_port_pre_bridge_leave(struct dsa_port *dp, struct net_device *br);
    217void dsa_port_bridge_leave(struct dsa_port *dp, struct net_device *br);
    218int dsa_port_lag_change(struct dsa_port *dp,
    219			struct netdev_lag_lower_state_info *linfo);
    220int dsa_port_lag_join(struct dsa_port *dp, struct net_device *lag_dev,
    221		      struct netdev_lag_upper_info *uinfo,
    222		      struct netlink_ext_ack *extack);
    223void dsa_port_pre_lag_leave(struct dsa_port *dp, struct net_device *lag_dev);
    224void dsa_port_lag_leave(struct dsa_port *dp, struct net_device *lag_dev);
    225int dsa_port_vlan_filtering(struct dsa_port *dp, bool vlan_filtering,
    226			    struct netlink_ext_ack *extack);
    227bool dsa_port_skip_vlan_configuration(struct dsa_port *dp);
    228int dsa_port_ageing_time(struct dsa_port *dp, clock_t ageing_clock);
    229int dsa_port_mst_enable(struct dsa_port *dp, bool on,
    230			struct netlink_ext_ack *extack);
    231int dsa_port_vlan_msti(struct dsa_port *dp,
    232		       const struct switchdev_vlan_msti *msti);
    233int dsa_port_mtu_change(struct dsa_port *dp, int new_mtu);
    234int dsa_port_fdb_add(struct dsa_port *dp, const unsigned char *addr,
    235		     u16 vid);
    236int dsa_port_fdb_del(struct dsa_port *dp, const unsigned char *addr,
    237		     u16 vid);
    238int dsa_port_standalone_host_fdb_add(struct dsa_port *dp,
    239				     const unsigned char *addr, u16 vid);
    240int dsa_port_standalone_host_fdb_del(struct dsa_port *dp,
    241				     const unsigned char *addr, u16 vid);
    242int dsa_port_bridge_host_fdb_add(struct dsa_port *dp, const unsigned char *addr,
    243				 u16 vid);
    244int dsa_port_bridge_host_fdb_del(struct dsa_port *dp, const unsigned char *addr,
    245				 u16 vid);
    246int dsa_port_lag_fdb_add(struct dsa_port *dp, const unsigned char *addr,
    247			 u16 vid);
    248int dsa_port_lag_fdb_del(struct dsa_port *dp, const unsigned char *addr,
    249			 u16 vid);
    250int dsa_port_fdb_dump(struct dsa_port *dp, dsa_fdb_dump_cb_t *cb, void *data);
    251int dsa_port_mdb_add(const struct dsa_port *dp,
    252		     const struct switchdev_obj_port_mdb *mdb);
    253int dsa_port_mdb_del(const struct dsa_port *dp,
    254		     const struct switchdev_obj_port_mdb *mdb);
    255int dsa_port_standalone_host_mdb_add(const struct dsa_port *dp,
    256				     const struct switchdev_obj_port_mdb *mdb);
    257int dsa_port_standalone_host_mdb_del(const struct dsa_port *dp,
    258				     const struct switchdev_obj_port_mdb *mdb);
    259int dsa_port_bridge_host_mdb_add(const struct dsa_port *dp,
    260				 const struct switchdev_obj_port_mdb *mdb);
    261int dsa_port_bridge_host_mdb_del(const struct dsa_port *dp,
    262				 const struct switchdev_obj_port_mdb *mdb);
    263int dsa_port_pre_bridge_flags(const struct dsa_port *dp,
    264			      struct switchdev_brport_flags flags,
    265			      struct netlink_ext_ack *extack);
    266int dsa_port_bridge_flags(struct dsa_port *dp,
    267			  struct switchdev_brport_flags flags,
    268			  struct netlink_ext_ack *extack);
    269int dsa_port_vlan_add(struct dsa_port *dp,
    270		      const struct switchdev_obj_port_vlan *vlan,
    271		      struct netlink_ext_ack *extack);
    272int dsa_port_vlan_del(struct dsa_port *dp,
    273		      const struct switchdev_obj_port_vlan *vlan);
    274int dsa_port_host_vlan_add(struct dsa_port *dp,
    275			   const struct switchdev_obj_port_vlan *vlan,
    276			   struct netlink_ext_ack *extack);
    277int dsa_port_host_vlan_del(struct dsa_port *dp,
    278			   const struct switchdev_obj_port_vlan *vlan);
    279int dsa_port_mrp_add(const struct dsa_port *dp,
    280		     const struct switchdev_obj_mrp *mrp);
    281int dsa_port_mrp_del(const struct dsa_port *dp,
    282		     const struct switchdev_obj_mrp *mrp);
    283int dsa_port_mrp_add_ring_role(const struct dsa_port *dp,
    284			       const struct switchdev_obj_ring_role_mrp *mrp);
    285int dsa_port_mrp_del_ring_role(const struct dsa_port *dp,
    286			       const struct switchdev_obj_ring_role_mrp *mrp);
    287int dsa_port_phylink_create(struct dsa_port *dp);
    288int dsa_port_link_register_of(struct dsa_port *dp);
    289void dsa_port_link_unregister_of(struct dsa_port *dp);
    290int dsa_port_hsr_join(struct dsa_port *dp, struct net_device *hsr);
    291void dsa_port_hsr_leave(struct dsa_port *dp, struct net_device *hsr);
    292int dsa_port_tag_8021q_vlan_add(struct dsa_port *dp, u16 vid, bool broadcast);
    293void dsa_port_tag_8021q_vlan_del(struct dsa_port *dp, u16 vid, bool broadcast);
    294void dsa_port_set_host_flood(struct dsa_port *dp, bool uc, bool mc);
    295
    296/* slave.c */
    297extern const struct dsa_device_ops notag_netdev_ops;
    298extern struct notifier_block dsa_slave_switchdev_notifier;
    299extern struct notifier_block dsa_slave_switchdev_blocking_notifier;
    300
    301void dsa_slave_mii_bus_init(struct dsa_switch *ds);
    302int dsa_slave_create(struct dsa_port *dp);
    303void dsa_slave_destroy(struct net_device *slave_dev);
    304int dsa_slave_suspend(struct net_device *slave_dev);
    305int dsa_slave_resume(struct net_device *slave_dev);
    306int dsa_slave_register_notifier(void);
    307void dsa_slave_unregister_notifier(void);
    308void dsa_slave_setup_tagger(struct net_device *slave);
    309int dsa_slave_change_mtu(struct net_device *dev, int new_mtu);
    310int dsa_slave_manage_vlan_filtering(struct net_device *dev,
    311				    bool vlan_filtering);
    312
    313static inline struct dsa_port *dsa_slave_to_port(const struct net_device *dev)
    314{
    315	struct dsa_slave_priv *p = netdev_priv(dev);
    316
    317	return p->dp;
    318}
    319
    320static inline struct net_device *
    321dsa_slave_to_master(const struct net_device *dev)
    322{
    323	struct dsa_port *dp = dsa_slave_to_port(dev);
    324
    325	return dp->cpu_dp->master;
    326}
    327
    328/* If under a bridge with vlan_filtering=0, make sure to send pvid-tagged
    329 * frames as untagged, since the bridge will not untag them.
    330 */
    331static inline struct sk_buff *dsa_untag_bridge_pvid(struct sk_buff *skb)
    332{
    333	struct dsa_port *dp = dsa_slave_to_port(skb->dev);
    334	struct net_device *br = dsa_port_bridge_dev_get(dp);
    335	struct net_device *dev = skb->dev;
    336	struct net_device *upper_dev;
    337	u16 vid, pvid, proto;
    338	int err;
    339
    340	if (!br || br_vlan_enabled(br))
    341		return skb;
    342
    343	err = br_vlan_get_proto(br, &proto);
    344	if (err)
    345		return skb;
    346
    347	/* Move VLAN tag from data to hwaccel */
    348	if (!skb_vlan_tag_present(skb) && skb->protocol == htons(proto)) {
    349		skb = skb_vlan_untag(skb);
    350		if (!skb)
    351			return NULL;
    352	}
    353
    354	if (!skb_vlan_tag_present(skb))
    355		return skb;
    356
    357	vid = skb_vlan_tag_get_id(skb);
    358
    359	/* We already run under an RCU read-side critical section since
    360	 * we are called from netif_receive_skb_list_internal().
    361	 */
    362	err = br_vlan_get_pvid_rcu(dev, &pvid);
    363	if (err)
    364		return skb;
    365
    366	if (vid != pvid)
    367		return skb;
    368
    369	/* The sad part about attempting to untag from DSA is that we
    370	 * don't know, unless we check, if the skb will end up in
    371	 * the bridge's data path - br_allowed_ingress() - or not.
    372	 * For example, there might be an 8021q upper for the
    373	 * default_pvid of the bridge, which will steal VLAN-tagged traffic
    374	 * from the bridge's data path. This is a configuration that DSA
    375	 * supports because vlan_filtering is 0. In that case, we should
    376	 * definitely keep the tag, to make sure it keeps working.
    377	 */
    378	upper_dev = __vlan_find_dev_deep_rcu(br, htons(proto), vid);
    379	if (upper_dev)
    380		return skb;
    381
    382	__vlan_hwaccel_clear_tag(skb);
    383
    384	return skb;
    385}
    386
    387/* For switches without hardware support for DSA tagging to be able
    388 * to support termination through the bridge.
    389 */
    390static inline struct net_device *
    391dsa_find_designated_bridge_port_by_vid(struct net_device *master, u16 vid)
    392{
    393	struct dsa_port *cpu_dp = master->dsa_ptr;
    394	struct dsa_switch_tree *dst = cpu_dp->dst;
    395	struct bridge_vlan_info vinfo;
    396	struct net_device *slave;
    397	struct dsa_port *dp;
    398	int err;
    399
    400	list_for_each_entry(dp, &dst->ports, list) {
    401		if (dp->type != DSA_PORT_TYPE_USER)
    402			continue;
    403
    404		if (!dp->bridge)
    405			continue;
    406
    407		if (dp->stp_state != BR_STATE_LEARNING &&
    408		    dp->stp_state != BR_STATE_FORWARDING)
    409			continue;
    410
    411		/* Since the bridge might learn this packet, keep the CPU port
    412		 * affinity with the port that will be used for the reply on
    413		 * xmit.
    414		 */
    415		if (dp->cpu_dp != cpu_dp)
    416			continue;
    417
    418		slave = dp->slave;
    419
    420		err = br_vlan_get_info_rcu(slave, vid, &vinfo);
    421		if (err)
    422			continue;
    423
    424		return slave;
    425	}
    426
    427	return NULL;
    428}
    429
    430/* If the ingress port offloads the bridge, we mark the frame as autonomously
    431 * forwarded by hardware, so the software bridge doesn't forward in twice, back
    432 * to us, because we already did. However, if we're in fallback mode and we do
    433 * software bridging, we are not offloading it, therefore the dp->bridge
    434 * pointer is not populated, and flooding needs to be done by software (we are
    435 * effectively operating in standalone ports mode).
    436 */
    437static inline void dsa_default_offload_fwd_mark(struct sk_buff *skb)
    438{
    439	struct dsa_port *dp = dsa_slave_to_port(skb->dev);
    440
    441	skb->offload_fwd_mark = !!(dp->bridge);
    442}
    443
    444/* Helper for removing DSA header tags from packets in the RX path.
    445 * Must not be called before skb_pull(len).
    446 *                                                                 skb->data
    447 *                                                                         |
    448 *                                                                         v
    449 * |   |   |   |   |   |   |   |   |   |   |   |   |   |   |   |   |   |   |
    450 * +-----------------------+-----------------------+---------------+-------+
    451 * |    Destination MAC    |      Source MAC       |  DSA header   | EType |
    452 * +-----------------------+-----------------------+---------------+-------+
    453 *                                                 |               |
    454 * <----- len ----->                               <----- len ----->
    455 *                 |
    456 *       >>>>>>>   v
    457 *       >>>>>>>   |   |   |   |   |   |   |   |   |   |   |   |   |   |   |
    458 *       >>>>>>>   +-----------------------+-----------------------+-------+
    459 *       >>>>>>>   |    Destination MAC    |      Source MAC       | EType |
    460 *                 +-----------------------+-----------------------+-------+
    461 *                                                                         ^
    462 *                                                                         |
    463 *                                                                 skb->data
    464 */
    465static inline void dsa_strip_etype_header(struct sk_buff *skb, int len)
    466{
    467	memmove(skb->data - ETH_HLEN, skb->data - ETH_HLEN - len, 2 * ETH_ALEN);
    468}
    469
    470/* Helper for creating space for DSA header tags in TX path packets.
    471 * Must not be called before skb_push(len).
    472 *
    473 * Before:
    474 *
    475 *       <<<<<<<   |   |   |   |   |   |   |   |   |   |   |   |   |   |   |
    476 * ^     <<<<<<<   +-----------------------+-----------------------+-------+
    477 * |     <<<<<<<   |    Destination MAC    |      Source MAC       | EType |
    478 * |               +-----------------------+-----------------------+-------+
    479 * <----- len ----->
    480 * |
    481 * |
    482 * skb->data
    483 *
    484 * After:
    485 *
    486 * |   |   |   |   |   |   |   |   |   |   |   |   |   |   |   |   |   |   |
    487 * +-----------------------+-----------------------+---------------+-------+
    488 * |    Destination MAC    |      Source MAC       |  DSA header   | EType |
    489 * +-----------------------+-----------------------+---------------+-------+
    490 * ^                                               |               |
    491 * |                                               <----- len ----->
    492 * skb->data
    493 */
    494static inline void dsa_alloc_etype_header(struct sk_buff *skb, int len)
    495{
    496	memmove(skb->data, skb->data + len, 2 * ETH_ALEN);
    497}
    498
    499/* On RX, eth_type_trans() on the DSA master pulls ETH_HLEN bytes starting from
    500 * skb_mac_header(skb), which leaves skb->data pointing at the first byte after
    501 * what the DSA master perceives as the EtherType (the beginning of the L3
    502 * protocol). Since DSA EtherType header taggers treat the EtherType as part of
    503 * the DSA tag itself, and the EtherType is 2 bytes in length, the DSA header
    504 * is located 2 bytes behind skb->data. Note that EtherType in this context
    505 * means the first 2 bytes of the DSA header, not the encapsulated EtherType
    506 * that will become visible after the DSA header is stripped.
    507 */
    508static inline void *dsa_etype_header_pos_rx(struct sk_buff *skb)
    509{
    510	return skb->data - 2;
    511}
    512
    513/* On TX, skb->data points to skb_mac_header(skb), which means that EtherType
    514 * header taggers start exactly where the EtherType is (the EtherType is
    515 * treated as part of the DSA header).
    516 */
    517static inline void *dsa_etype_header_pos_tx(struct sk_buff *skb)
    518{
    519	return skb->data + 2 * ETH_ALEN;
    520}
    521
    522/* switch.c */
    523int dsa_switch_register_notifier(struct dsa_switch *ds);
    524void dsa_switch_unregister_notifier(struct dsa_switch *ds);
    525
    526static inline bool dsa_switch_supports_uc_filtering(struct dsa_switch *ds)
    527{
    528	return ds->ops->port_fdb_add && ds->ops->port_fdb_del &&
    529	       ds->fdb_isolation && !ds->vlan_filtering_is_global &&
    530	       !ds->needs_standalone_vlan_filtering;
    531}
    532
    533static inline bool dsa_switch_supports_mc_filtering(struct dsa_switch *ds)
    534{
    535	return ds->ops->port_mdb_add && ds->ops->port_mdb_del &&
    536	       ds->fdb_isolation && !ds->vlan_filtering_is_global &&
    537	       !ds->needs_standalone_vlan_filtering;
    538}
    539
    540/* dsa2.c */
    541void dsa_lag_map(struct dsa_switch_tree *dst, struct dsa_lag *lag);
    542void dsa_lag_unmap(struct dsa_switch_tree *dst, struct dsa_lag *lag);
    543struct dsa_lag *dsa_tree_lag_find(struct dsa_switch_tree *dst,
    544				  const struct net_device *lag_dev);
    545int dsa_tree_notify(struct dsa_switch_tree *dst, unsigned long e, void *v);
    546int dsa_broadcast(unsigned long e, void *v);
    547int dsa_tree_change_tag_proto(struct dsa_switch_tree *dst,
    548			      struct net_device *master,
    549			      const struct dsa_device_ops *tag_ops,
    550			      const struct dsa_device_ops *old_tag_ops);
    551void dsa_tree_master_admin_state_change(struct dsa_switch_tree *dst,
    552					struct net_device *master,
    553					bool up);
    554void dsa_tree_master_oper_state_change(struct dsa_switch_tree *dst,
    555				       struct net_device *master,
    556				       bool up);
    557unsigned int dsa_bridge_num_get(const struct net_device *bridge_dev, int max);
    558void dsa_bridge_num_put(const struct net_device *bridge_dev,
    559			unsigned int bridge_num);
    560struct dsa_bridge *dsa_tree_bridge_find(struct dsa_switch_tree *dst,
    561					const struct net_device *br);
    562
    563/* tag_8021q.c */
    564int dsa_switch_tag_8021q_vlan_add(struct dsa_switch *ds,
    565				  struct dsa_notifier_tag_8021q_vlan_info *info);
    566int dsa_switch_tag_8021q_vlan_del(struct dsa_switch *ds,
    567				  struct dsa_notifier_tag_8021q_vlan_info *info);
    568
    569extern struct list_head dsa_tree_list;
    570
    571#endif