cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

fm10k_iov.c (19221B)


      1// SPDX-License-Identifier: GPL-2.0
      2/* Copyright(c) 2013 - 2019 Intel Corporation. */
      3
      4#include "fm10k.h"
      5#include "fm10k_vf.h"
      6#include "fm10k_pf.h"
      7
      8static s32 fm10k_iov_msg_error(struct fm10k_hw *hw, u32 **results,
      9			       struct fm10k_mbx_info *mbx)
     10{
     11	struct fm10k_vf_info *vf_info = (struct fm10k_vf_info *)mbx;
     12	struct fm10k_intfc *interface = hw->back;
     13	struct pci_dev *pdev = interface->pdev;
     14
     15	dev_err(&pdev->dev, "Unknown message ID %u on VF %d\n",
     16		**results & FM10K_TLV_ID_MASK, vf_info->vf_idx);
     17
     18	return fm10k_tlv_msg_error(hw, results, mbx);
     19}
     20
     21/**
     22 *  fm10k_iov_msg_queue_mac_vlan - Message handler for MAC/VLAN request from VF
     23 *  @hw: Pointer to hardware structure
     24 *  @results: Pointer array to message, results[0] is pointer to message
     25 *  @mbx: Pointer to mailbox information structure
     26 *
     27 *  This function is a custom handler for MAC/VLAN requests from the VF. The
     28 *  assumption is that it is acceptable to directly hand off the message from
     29 *  the VF to the PF's switch manager. However, we use a MAC/VLAN message
     30 *  queue to avoid overloading the mailbox when a large number of requests
     31 *  come in.
     32 **/
     33static s32 fm10k_iov_msg_queue_mac_vlan(struct fm10k_hw *hw, u32 **results,
     34					struct fm10k_mbx_info *mbx)
     35{
     36	struct fm10k_vf_info *vf_info = (struct fm10k_vf_info *)mbx;
     37	struct fm10k_intfc *interface = hw->back;
     38	u8 mac[ETH_ALEN];
     39	u32 *result;
     40	int err = 0;
     41	bool set;
     42	u16 vlan;
     43	u32 vid;
     44
     45	/* we shouldn't be updating rules on a disabled interface */
     46	if (!FM10K_VF_FLAG_ENABLED(vf_info))
     47		err = FM10K_ERR_PARAM;
     48
     49	if (!err && !!results[FM10K_MAC_VLAN_MSG_VLAN]) {
     50		result = results[FM10K_MAC_VLAN_MSG_VLAN];
     51
     52		/* record VLAN id requested */
     53		err = fm10k_tlv_attr_get_u32(result, &vid);
     54		if (err)
     55			return err;
     56
     57		set = !(vid & FM10K_VLAN_CLEAR);
     58		vid &= ~FM10K_VLAN_CLEAR;
     59
     60		/* if the length field has been set, this is a multi-bit
     61		 * update request. For multi-bit requests, simply disallow
     62		 * them when the pf_vid has been set. In this case, the PF
     63		 * should have already cleared the VLAN_TABLE, and if we
     64		 * allowed them, it could allow a rogue VF to receive traffic
     65		 * on a VLAN it was not assigned. In the single-bit case, we
     66		 * need to modify requests for VLAN 0 to use the default PF or
     67		 * SW vid when assigned.
     68		 */
     69
     70		if (vid >> 16) {
     71			/* prevent multi-bit requests when PF has
     72			 * administratively set the VLAN for this VF
     73			 */
     74			if (vf_info->pf_vid)
     75				return FM10K_ERR_PARAM;
     76		} else {
     77			err = fm10k_iov_select_vid(vf_info, (u16)vid);
     78			if (err < 0)
     79				return err;
     80
     81			vid = err;
     82		}
     83
     84		/* update VSI info for VF in regards to VLAN table */
     85		err = hw->mac.ops.update_vlan(hw, vid, vf_info->vsi, set);
     86	}
     87
     88	if (!err && !!results[FM10K_MAC_VLAN_MSG_MAC]) {
     89		result = results[FM10K_MAC_VLAN_MSG_MAC];
     90
     91		/* record unicast MAC address requested */
     92		err = fm10k_tlv_attr_get_mac_vlan(result, mac, &vlan);
     93		if (err)
     94			return err;
     95
     96		/* block attempts to set MAC for a locked device */
     97		if (is_valid_ether_addr(vf_info->mac) &&
     98		    !ether_addr_equal(mac, vf_info->mac))
     99			return FM10K_ERR_PARAM;
    100
    101		set = !(vlan & FM10K_VLAN_CLEAR);
    102		vlan &= ~FM10K_VLAN_CLEAR;
    103
    104		err = fm10k_iov_select_vid(vf_info, vlan);
    105		if (err < 0)
    106			return err;
    107
    108		vlan = (u16)err;
    109
    110		/* Add this request to the MAC/VLAN queue */
    111		err = fm10k_queue_mac_request(interface, vf_info->glort,
    112					      mac, vlan, set);
    113	}
    114
    115	if (!err && !!results[FM10K_MAC_VLAN_MSG_MULTICAST]) {
    116		result = results[FM10K_MAC_VLAN_MSG_MULTICAST];
    117
    118		/* record multicast MAC address requested */
    119		err = fm10k_tlv_attr_get_mac_vlan(result, mac, &vlan);
    120		if (err)
    121			return err;
    122
    123		/* verify that the VF is allowed to request multicast */
    124		if (!(vf_info->vf_flags & FM10K_VF_FLAG_MULTI_ENABLED))
    125			return FM10K_ERR_PARAM;
    126
    127		set = !(vlan & FM10K_VLAN_CLEAR);
    128		vlan &= ~FM10K_VLAN_CLEAR;
    129
    130		err = fm10k_iov_select_vid(vf_info, vlan);
    131		if (err < 0)
    132			return err;
    133
    134		vlan = (u16)err;
    135
    136		/* Add this request to the MAC/VLAN queue */
    137		err = fm10k_queue_mac_request(interface, vf_info->glort,
    138					      mac, vlan, set);
    139	}
    140
    141	return err;
    142}
    143
    144static const struct fm10k_msg_data iov_mbx_data[] = {
    145	FM10K_TLV_MSG_TEST_HANDLER(fm10k_tlv_msg_test),
    146	FM10K_VF_MSG_MSIX_HANDLER(fm10k_iov_msg_msix_pf),
    147	FM10K_VF_MSG_MAC_VLAN_HANDLER(fm10k_iov_msg_queue_mac_vlan),
    148	FM10K_VF_MSG_LPORT_STATE_HANDLER(fm10k_iov_msg_lport_state_pf),
    149	FM10K_TLV_MSG_ERROR_HANDLER(fm10k_iov_msg_error),
    150};
    151
    152s32 fm10k_iov_event(struct fm10k_intfc *interface)
    153{
    154	struct fm10k_hw *hw = &interface->hw;
    155	struct fm10k_iov_data *iov_data;
    156	s64 vflre;
    157	int i;
    158
    159	/* if there is no iov_data then there is no mailbox to process */
    160	if (!READ_ONCE(interface->iov_data))
    161		return 0;
    162
    163	rcu_read_lock();
    164
    165	iov_data = interface->iov_data;
    166
    167	/* check again now that we are in the RCU block */
    168	if (!iov_data)
    169		goto read_unlock;
    170
    171	if (!(fm10k_read_reg(hw, FM10K_EICR) & FM10K_EICR_VFLR))
    172		goto read_unlock;
    173
    174	/* read VFLRE to determine if any VFs have been reset */
    175	vflre = fm10k_read_reg(hw, FM10K_PFVFLRE(1));
    176	vflre <<= 32;
    177	vflre |= fm10k_read_reg(hw, FM10K_PFVFLRE(0));
    178
    179	i = iov_data->num_vfs;
    180
    181	for (vflre <<= 64 - i; vflre && i--; vflre += vflre) {
    182		struct fm10k_vf_info *vf_info = &iov_data->vf_info[i];
    183
    184		if (vflre >= 0)
    185			continue;
    186
    187		hw->iov.ops.reset_resources(hw, vf_info);
    188		vf_info->mbx.ops.connect(hw, &vf_info->mbx);
    189	}
    190
    191read_unlock:
    192	rcu_read_unlock();
    193
    194	return 0;
    195}
    196
    197s32 fm10k_iov_mbx(struct fm10k_intfc *interface)
    198{
    199	struct fm10k_hw *hw = &interface->hw;
    200	struct fm10k_iov_data *iov_data;
    201	int i;
    202
    203	/* if there is no iov_data then there is no mailbox to process */
    204	if (!READ_ONCE(interface->iov_data))
    205		return 0;
    206
    207	rcu_read_lock();
    208
    209	iov_data = interface->iov_data;
    210
    211	/* check again now that we are in the RCU block */
    212	if (!iov_data)
    213		goto read_unlock;
    214
    215	/* lock the mailbox for transmit and receive */
    216	fm10k_mbx_lock(interface);
    217
    218	/* Most VF messages sent to the PF cause the PF to respond by
    219	 * requesting from the SM mailbox. This means that too many VF
    220	 * messages processed at once could cause a mailbox timeout on the PF.
    221	 * To prevent this, store a pointer to the next VF mbx to process. Use
    222	 * that as the start of the loop so that we don't starve whichever VF
    223	 * got ignored on the previous run.
    224	 */
    225process_mbx:
    226	for (i = iov_data->next_vf_mbx ? : iov_data->num_vfs; i--;) {
    227		struct fm10k_vf_info *vf_info = &iov_data->vf_info[i];
    228		struct fm10k_mbx_info *mbx = &vf_info->mbx;
    229		u16 glort = vf_info->glort;
    230
    231		/* process the SM mailbox first to drain outgoing messages */
    232		hw->mbx.ops.process(hw, &hw->mbx);
    233
    234		/* verify port mapping is valid, if not reset port */
    235		if (vf_info->vf_flags && !fm10k_glort_valid_pf(hw, glort)) {
    236			hw->iov.ops.reset_lport(hw, vf_info);
    237			fm10k_clear_macvlan_queue(interface, glort, false);
    238		}
    239
    240		/* reset VFs that have mailbox timed out */
    241		if (!mbx->timeout) {
    242			hw->iov.ops.reset_resources(hw, vf_info);
    243			mbx->ops.connect(hw, mbx);
    244		}
    245
    246		/* guarantee we have free space in the SM mailbox */
    247		if (hw->mbx.state == FM10K_STATE_OPEN &&
    248		    !hw->mbx.ops.tx_ready(&hw->mbx, FM10K_VFMBX_MSG_MTU)) {
    249			/* keep track of how many times this occurs */
    250			interface->hw_sm_mbx_full++;
    251
    252			/* make sure we try again momentarily */
    253			fm10k_service_event_schedule(interface);
    254
    255			break;
    256		}
    257
    258		/* cleanup mailbox and process received messages */
    259		mbx->ops.process(hw, mbx);
    260	}
    261
    262	/* if we stopped processing mailboxes early, update next_vf_mbx.
    263	 * Otherwise, reset next_vf_mbx, and restart loop so that we process
    264	 * the remaining mailboxes we skipped at the start.
    265	 */
    266	if (i >= 0) {
    267		iov_data->next_vf_mbx = i + 1;
    268	} else if (iov_data->next_vf_mbx) {
    269		iov_data->next_vf_mbx = 0;
    270		goto process_mbx;
    271	}
    272
    273	/* free the lock */
    274	fm10k_mbx_unlock(interface);
    275
    276read_unlock:
    277	rcu_read_unlock();
    278
    279	return 0;
    280}
    281
    282void fm10k_iov_suspend(struct pci_dev *pdev)
    283{
    284	struct fm10k_intfc *interface = pci_get_drvdata(pdev);
    285	struct fm10k_iov_data *iov_data = interface->iov_data;
    286	struct fm10k_hw *hw = &interface->hw;
    287	int num_vfs, i;
    288
    289	/* pull out num_vfs from iov_data */
    290	num_vfs = iov_data ? iov_data->num_vfs : 0;
    291
    292	/* shut down queue mapping for VFs */
    293	fm10k_write_reg(hw, FM10K_DGLORTMAP(fm10k_dglort_vf_rss),
    294			FM10K_DGLORTMAP_NONE);
    295
    296	/* Stop any active VFs and reset their resources */
    297	for (i = 0; i < num_vfs; i++) {
    298		struct fm10k_vf_info *vf_info = &iov_data->vf_info[i];
    299
    300		hw->iov.ops.reset_resources(hw, vf_info);
    301		hw->iov.ops.reset_lport(hw, vf_info);
    302		fm10k_clear_macvlan_queue(interface, vf_info->glort, false);
    303	}
    304}
    305
    306static void fm10k_mask_aer_comp_abort(struct pci_dev *pdev)
    307{
    308	u32 err_mask;
    309	int pos;
    310
    311	pos = pci_find_ext_capability(pdev, PCI_EXT_CAP_ID_ERR);
    312	if (!pos)
    313		return;
    314
    315	/* Mask the completion abort bit in the ERR_UNCOR_MASK register,
    316	 * preventing the device from reporting these errors to the upstream
    317	 * PCIe root device. This avoids bringing down platforms which upgrade
    318	 * non-fatal completer aborts into machine check exceptions. Completer
    319	 * aborts can occur whenever a VF reads a queue it doesn't own.
    320	 */
    321	pci_read_config_dword(pdev, pos + PCI_ERR_UNCOR_MASK, &err_mask);
    322	err_mask |= PCI_ERR_UNC_COMP_ABORT;
    323	pci_write_config_dword(pdev, pos + PCI_ERR_UNCOR_MASK, err_mask);
    324}
    325
    326int fm10k_iov_resume(struct pci_dev *pdev)
    327{
    328	struct fm10k_intfc *interface = pci_get_drvdata(pdev);
    329	struct fm10k_iov_data *iov_data = interface->iov_data;
    330	struct fm10k_dglort_cfg dglort = { 0 };
    331	struct fm10k_hw *hw = &interface->hw;
    332	int num_vfs, i;
    333
    334	/* pull out num_vfs from iov_data */
    335	num_vfs = iov_data ? iov_data->num_vfs : 0;
    336
    337	/* return error if iov_data is not already populated */
    338	if (!iov_data)
    339		return -ENOMEM;
    340
    341	/* Lower severity of completer abort error reporting as
    342	 * the VFs can trigger this any time they read a queue
    343	 * that they don't own.
    344	 */
    345	fm10k_mask_aer_comp_abort(pdev);
    346
    347	/* allocate hardware resources for the VFs */
    348	hw->iov.ops.assign_resources(hw, num_vfs, num_vfs);
    349
    350	/* configure DGLORT mapping for RSS */
    351	dglort.glort = hw->mac.dglort_map & FM10K_DGLORTMAP_NONE;
    352	dglort.idx = fm10k_dglort_vf_rss;
    353	dglort.inner_rss = 1;
    354	dglort.rss_l = fls(fm10k_queues_per_pool(hw) - 1);
    355	dglort.queue_b = fm10k_vf_queue_index(hw, 0);
    356	dglort.vsi_l = fls(hw->iov.total_vfs - 1);
    357	dglort.vsi_b = 1;
    358
    359	hw->mac.ops.configure_dglort_map(hw, &dglort);
    360
    361	/* assign resources to the device */
    362	for (i = 0; i < num_vfs; i++) {
    363		struct fm10k_vf_info *vf_info = &iov_data->vf_info[i];
    364
    365		/* allocate all but the last GLORT to the VFs */
    366		if (i == (~hw->mac.dglort_map >> FM10K_DGLORTMAP_MASK_SHIFT))
    367			break;
    368
    369		/* assign GLORT to VF, and restrict it to multicast */
    370		hw->iov.ops.set_lport(hw, vf_info, i,
    371				      FM10K_VF_FLAG_MULTI_CAPABLE);
    372
    373		/* mailbox is disconnected so we don't send a message */
    374		hw->iov.ops.assign_default_mac_vlan(hw, vf_info);
    375
    376		/* now we are ready so we can connect */
    377		vf_info->mbx.ops.connect(hw, &vf_info->mbx);
    378	}
    379
    380	return 0;
    381}
    382
    383s32 fm10k_iov_update_pvid(struct fm10k_intfc *interface, u16 glort, u16 pvid)
    384{
    385	struct fm10k_iov_data *iov_data = interface->iov_data;
    386	struct fm10k_hw *hw = &interface->hw;
    387	struct fm10k_vf_info *vf_info;
    388	u16 vf_idx = (glort - hw->mac.dglort_map) & FM10K_DGLORTMAP_NONE;
    389
    390	/* no IOV support, not our message to process */
    391	if (!iov_data)
    392		return FM10K_ERR_PARAM;
    393
    394	/* glort outside our range, not our message to process */
    395	if (vf_idx >= iov_data->num_vfs)
    396		return FM10K_ERR_PARAM;
    397
    398	/* determine if an update has occurred and if so notify the VF */
    399	vf_info = &iov_data->vf_info[vf_idx];
    400	if (vf_info->sw_vid != pvid) {
    401		vf_info->sw_vid = pvid;
    402		hw->iov.ops.assign_default_mac_vlan(hw, vf_info);
    403	}
    404
    405	return 0;
    406}
    407
    408static void fm10k_iov_free_data(struct pci_dev *pdev)
    409{
    410	struct fm10k_intfc *interface = pci_get_drvdata(pdev);
    411
    412	if (!interface->iov_data)
    413		return;
    414
    415	/* reclaim hardware resources */
    416	fm10k_iov_suspend(pdev);
    417
    418	/* drop iov_data from interface */
    419	kfree_rcu(interface->iov_data, rcu);
    420	interface->iov_data = NULL;
    421}
    422
    423static s32 fm10k_iov_alloc_data(struct pci_dev *pdev, int num_vfs)
    424{
    425	struct fm10k_intfc *interface = pci_get_drvdata(pdev);
    426	struct fm10k_iov_data *iov_data = interface->iov_data;
    427	struct fm10k_hw *hw = &interface->hw;
    428	size_t size;
    429	int i;
    430
    431	/* return error if iov_data is already populated */
    432	if (iov_data)
    433		return -EBUSY;
    434
    435	/* The PF should always be able to assign resources */
    436	if (!hw->iov.ops.assign_resources)
    437		return -ENODEV;
    438
    439	/* nothing to do if no VFs are requested */
    440	if (!num_vfs)
    441		return 0;
    442
    443	/* allocate memory for VF storage */
    444	size = offsetof(struct fm10k_iov_data, vf_info[num_vfs]);
    445	iov_data = kzalloc(size, GFP_KERNEL);
    446	if (!iov_data)
    447		return -ENOMEM;
    448
    449	/* record number of VFs */
    450	iov_data->num_vfs = num_vfs;
    451
    452	/* loop through vf_info structures initializing each entry */
    453	for (i = 0; i < num_vfs; i++) {
    454		struct fm10k_vf_info *vf_info = &iov_data->vf_info[i];
    455		int err;
    456
    457		/* Record VF VSI value */
    458		vf_info->vsi = i + 1;
    459		vf_info->vf_idx = i;
    460
    461		/* initialize mailbox memory */
    462		err = fm10k_pfvf_mbx_init(hw, &vf_info->mbx, iov_mbx_data, i);
    463		if (err) {
    464			dev_err(&pdev->dev,
    465				"Unable to initialize SR-IOV mailbox\n");
    466			kfree(iov_data);
    467			return err;
    468		}
    469	}
    470
    471	/* assign iov_data to interface */
    472	interface->iov_data = iov_data;
    473
    474	/* allocate hardware resources for the VFs */
    475	fm10k_iov_resume(pdev);
    476
    477	return 0;
    478}
    479
    480void fm10k_iov_disable(struct pci_dev *pdev)
    481{
    482	if (pci_num_vf(pdev) && pci_vfs_assigned(pdev))
    483		dev_err(&pdev->dev,
    484			"Cannot disable SR-IOV while VFs are assigned\n");
    485	else
    486		pci_disable_sriov(pdev);
    487
    488	fm10k_iov_free_data(pdev);
    489}
    490
    491int fm10k_iov_configure(struct pci_dev *pdev, int num_vfs)
    492{
    493	int current_vfs = pci_num_vf(pdev);
    494	int err = 0;
    495
    496	if (current_vfs && pci_vfs_assigned(pdev)) {
    497		dev_err(&pdev->dev,
    498			"Cannot modify SR-IOV while VFs are assigned\n");
    499		num_vfs = current_vfs;
    500	} else {
    501		pci_disable_sriov(pdev);
    502		fm10k_iov_free_data(pdev);
    503	}
    504
    505	/* allocate resources for the VFs */
    506	err = fm10k_iov_alloc_data(pdev, num_vfs);
    507	if (err)
    508		return err;
    509
    510	/* allocate VFs if not already allocated */
    511	if (num_vfs && num_vfs != current_vfs) {
    512		err = pci_enable_sriov(pdev, num_vfs);
    513		if (err) {
    514			dev_err(&pdev->dev,
    515				"Enable PCI SR-IOV failed: %d\n", err);
    516			return err;
    517		}
    518	}
    519
    520	return num_vfs;
    521}
    522
    523/**
    524 * fm10k_iov_update_stats - Update stats for all VFs
    525 * @interface: device private structure
    526 *
    527 * Updates the VF statistics for all enabled VFs. Expects to be called by
    528 * fm10k_update_stats and assumes that locking via the __FM10K_UPDATING_STATS
    529 * bit is already handled.
    530 */
    531void fm10k_iov_update_stats(struct fm10k_intfc *interface)
    532{
    533	struct fm10k_iov_data *iov_data = interface->iov_data;
    534	struct fm10k_hw *hw = &interface->hw;
    535	int i;
    536
    537	if (!iov_data)
    538		return;
    539
    540	for (i = 0; i < iov_data->num_vfs; i++)
    541		hw->iov.ops.update_stats(hw, iov_data->vf_info[i].stats, i);
    542}
    543
    544static inline void fm10k_reset_vf_info(struct fm10k_intfc *interface,
    545				       struct fm10k_vf_info *vf_info)
    546{
    547	struct fm10k_hw *hw = &interface->hw;
    548
    549	/* assigning the MAC address will send a mailbox message */
    550	fm10k_mbx_lock(interface);
    551
    552	/* disable LPORT for this VF which clears switch rules */
    553	hw->iov.ops.reset_lport(hw, vf_info);
    554
    555	fm10k_clear_macvlan_queue(interface, vf_info->glort, false);
    556
    557	/* assign new MAC+VLAN for this VF */
    558	hw->iov.ops.assign_default_mac_vlan(hw, vf_info);
    559
    560	/* re-enable the LPORT for this VF */
    561	hw->iov.ops.set_lport(hw, vf_info, vf_info->vf_idx,
    562			      FM10K_VF_FLAG_MULTI_CAPABLE);
    563
    564	fm10k_mbx_unlock(interface);
    565}
    566
    567int fm10k_ndo_set_vf_mac(struct net_device *netdev, int vf_idx, u8 *mac)
    568{
    569	struct fm10k_intfc *interface = netdev_priv(netdev);
    570	struct fm10k_iov_data *iov_data = interface->iov_data;
    571	struct fm10k_vf_info *vf_info;
    572
    573	/* verify SR-IOV is active and that vf idx is valid */
    574	if (!iov_data || vf_idx >= iov_data->num_vfs)
    575		return -EINVAL;
    576
    577	/* verify MAC addr is valid */
    578	if (!is_zero_ether_addr(mac) && !is_valid_ether_addr(mac))
    579		return -EINVAL;
    580
    581	/* record new MAC address */
    582	vf_info = &iov_data->vf_info[vf_idx];
    583	ether_addr_copy(vf_info->mac, mac);
    584
    585	fm10k_reset_vf_info(interface, vf_info);
    586
    587	return 0;
    588}
    589
    590int fm10k_ndo_set_vf_vlan(struct net_device *netdev, int vf_idx, u16 vid,
    591			  u8 qos, __be16 vlan_proto)
    592{
    593	struct fm10k_intfc *interface = netdev_priv(netdev);
    594	struct fm10k_iov_data *iov_data = interface->iov_data;
    595	struct fm10k_hw *hw = &interface->hw;
    596	struct fm10k_vf_info *vf_info;
    597
    598	/* verify SR-IOV is active and that vf idx is valid */
    599	if (!iov_data || vf_idx >= iov_data->num_vfs)
    600		return -EINVAL;
    601
    602	/* QOS is unsupported and VLAN IDs accepted range 0-4094 */
    603	if (qos || (vid > (VLAN_VID_MASK - 1)))
    604		return -EINVAL;
    605
    606	/* VF VLAN Protocol part to default is unsupported */
    607	if (vlan_proto != htons(ETH_P_8021Q))
    608		return -EPROTONOSUPPORT;
    609
    610	vf_info = &iov_data->vf_info[vf_idx];
    611
    612	/* exit if there is nothing to do */
    613	if (vf_info->pf_vid == vid)
    614		return 0;
    615
    616	/* record default VLAN ID for VF */
    617	vf_info->pf_vid = vid;
    618
    619	/* Clear the VLAN table for the VF */
    620	hw->mac.ops.update_vlan(hw, FM10K_VLAN_ALL, vf_info->vsi, false);
    621
    622	fm10k_reset_vf_info(interface, vf_info);
    623
    624	return 0;
    625}
    626
    627int fm10k_ndo_set_vf_bw(struct net_device *netdev, int vf_idx,
    628			int __always_unused min_rate, int max_rate)
    629{
    630	struct fm10k_intfc *interface = netdev_priv(netdev);
    631	struct fm10k_iov_data *iov_data = interface->iov_data;
    632	struct fm10k_hw *hw = &interface->hw;
    633
    634	/* verify SR-IOV is active and that vf idx is valid */
    635	if (!iov_data || vf_idx >= iov_data->num_vfs)
    636		return -EINVAL;
    637
    638	/* rate limit cannot be less than 10Mbs or greater than link speed */
    639	if (max_rate &&
    640	    (max_rate < FM10K_VF_TC_MIN || max_rate > FM10K_VF_TC_MAX))
    641		return -EINVAL;
    642
    643	/* store values */
    644	iov_data->vf_info[vf_idx].rate = max_rate;
    645
    646	/* update hardware configuration */
    647	hw->iov.ops.configure_tc(hw, vf_idx, max_rate);
    648
    649	return 0;
    650}
    651
    652int fm10k_ndo_get_vf_config(struct net_device *netdev,
    653			    int vf_idx, struct ifla_vf_info *ivi)
    654{
    655	struct fm10k_intfc *interface = netdev_priv(netdev);
    656	struct fm10k_iov_data *iov_data = interface->iov_data;
    657	struct fm10k_vf_info *vf_info;
    658
    659	/* verify SR-IOV is active and that vf idx is valid */
    660	if (!iov_data || vf_idx >= iov_data->num_vfs)
    661		return -EINVAL;
    662
    663	vf_info = &iov_data->vf_info[vf_idx];
    664
    665	ivi->vf = vf_idx;
    666	ivi->max_tx_rate = vf_info->rate;
    667	ivi->min_tx_rate = 0;
    668	ether_addr_copy(ivi->mac, vf_info->mac);
    669	ivi->vlan = vf_info->pf_vid;
    670	ivi->qos = 0;
    671
    672	return 0;
    673}
    674
    675int fm10k_ndo_get_vf_stats(struct net_device *netdev,
    676			   int vf_idx, struct ifla_vf_stats *stats)
    677{
    678	struct fm10k_intfc *interface = netdev_priv(netdev);
    679	struct fm10k_iov_data *iov_data = interface->iov_data;
    680	struct fm10k_hw *hw = &interface->hw;
    681	struct fm10k_hw_stats_q *hw_stats;
    682	u32 idx, qpp;
    683
    684	/* verify SR-IOV is active and that vf idx is valid */
    685	if (!iov_data || vf_idx >= iov_data->num_vfs)
    686		return -EINVAL;
    687
    688	qpp = fm10k_queues_per_pool(hw);
    689	hw_stats = iov_data->vf_info[vf_idx].stats;
    690
    691	for (idx = 0; idx < qpp; idx++) {
    692		stats->rx_packets += hw_stats[idx].rx_packets.count;
    693		stats->tx_packets += hw_stats[idx].tx_packets.count;
    694		stats->rx_bytes += hw_stats[idx].rx_bytes.count;
    695		stats->tx_bytes += hw_stats[idx].tx_bytes.count;
    696		stats->rx_dropped += hw_stats[idx].rx_drops.count;
    697	}
    698
    699	return 0;
    700}