cachepc-linux

Fork of AMDESE/linux with modifications for CachePC side-channel attack
git clone https://git.sinitax.com/sinitax/cachepc-linux
Log | Files | Refs | README | LICENSE | sfeed.txt

nfp_netvf_main.c (8821B)


      1// SPDX-License-Identifier: (GPL-2.0-only OR BSD-2-Clause)
      2/* Copyright (C) 2015-2018 Netronome Systems, Inc. */
      3
      4/*
      5 * nfp_netvf_main.c
      6 * Netronome virtual function network device driver: Main entry point
      7 * Author: Jason McMullan <jason.mcmullan@netronome.com>
      8 *         Rolf Neugebauer <rolf.neugebauer@netronome.com>
      9 */
     10
     11#include <linux/module.h>
     12#include <linux/kernel.h>
     13#include <linux/init.h>
     14#include <linux/etherdevice.h>
     15
     16#include "nfpcore/nfp_dev.h"
     17#include "nfp_net_ctrl.h"
     18#include "nfp_net.h"
     19#include "nfp_main.h"
     20
     21/**
     22 * struct nfp_net_vf - NFP VF-specific device structure
     23 * @nn:		NFP Net structure for this device
     24 * @irq_entries: Pre-allocated array of MSI-X entries
     25 * @q_bar:	Pointer to mapped QC memory (NULL if TX/RX mapped directly)
     26 * @ddir:	Per-device debugfs directory
     27 */
     28struct nfp_net_vf {
     29	struct nfp_net *nn;
     30
     31	struct msix_entry irq_entries[NFP_NET_NON_Q_VECTORS +
     32				      NFP_NET_MAX_TX_RINGS];
     33	u8 __iomem *q_bar;
     34
     35	struct dentry *ddir;
     36};
     37
     38static const char nfp_net_driver_name[] = "nfp_netvf";
     39
     40static const struct pci_device_id nfp_netvf_pci_device_ids[] = {
     41	{ PCI_VENDOR_ID_NETRONOME, PCI_DEVICE_ID_NFP3800_VF,
     42	  PCI_VENDOR_ID_NETRONOME, PCI_ANY_ID,
     43	  PCI_ANY_ID, 0, NFP_DEV_NFP3800_VF,
     44	},
     45	{ PCI_VENDOR_ID_NETRONOME, PCI_DEVICE_ID_NFP6000_VF,
     46	  PCI_VENDOR_ID_NETRONOME, PCI_ANY_ID,
     47	  PCI_ANY_ID, 0, NFP_DEV_NFP6000_VF,
     48	},
     49	{ PCI_VENDOR_ID_CORIGINE, PCI_DEVICE_ID_NFP3800_VF,
     50	  PCI_VENDOR_ID_CORIGINE, PCI_ANY_ID,
     51	  PCI_ANY_ID, 0, NFP_DEV_NFP3800_VF,
     52	},
     53	{ PCI_VENDOR_ID_CORIGINE, PCI_DEVICE_ID_NFP6000_VF,
     54	  PCI_VENDOR_ID_CORIGINE, PCI_ANY_ID,
     55	  PCI_ANY_ID, 0, NFP_DEV_NFP6000_VF,
     56	},
     57	{ 0, } /* Required last entry. */
     58};
     59MODULE_DEVICE_TABLE(pci, nfp_netvf_pci_device_ids);
     60
     61static void nfp_netvf_get_mac_addr(struct nfp_net *nn)
     62{
     63	u8 mac_addr[ETH_ALEN];
     64
     65	put_unaligned_be32(nn_readl(nn, NFP_NET_CFG_MACADDR + 0), &mac_addr[0]);
     66	put_unaligned_be16(nn_readw(nn, NFP_NET_CFG_MACADDR + 6), &mac_addr[4]);
     67
     68	if (!is_valid_ether_addr(mac_addr)) {
     69		eth_hw_addr_random(nn->dp.netdev);
     70		return;
     71	}
     72
     73	eth_hw_addr_set(nn->dp.netdev, mac_addr);
     74	ether_addr_copy(nn->dp.netdev->perm_addr, mac_addr);
     75}
     76
     77static int nfp_netvf_pci_probe(struct pci_dev *pdev,
     78			       const struct pci_device_id *pci_id)
     79{
     80	const struct nfp_dev_info *dev_info;
     81	struct nfp_net_fw_version fw_ver;
     82	int max_tx_rings, max_rx_rings;
     83	u32 tx_bar_off, rx_bar_off;
     84	u32 tx_bar_sz, rx_bar_sz;
     85	int tx_bar_no, rx_bar_no;
     86	struct nfp_net_vf *vf;
     87	unsigned int num_irqs;
     88	u8 __iomem *ctrl_bar;
     89	struct nfp_net *nn;
     90	u32 startq;
     91	int stride;
     92	int err;
     93
     94	dev_info = &nfp_dev_info[pci_id->driver_data];
     95
     96	vf = kzalloc(sizeof(*vf), GFP_KERNEL);
     97	if (!vf)
     98		return -ENOMEM;
     99	pci_set_drvdata(pdev, vf);
    100
    101	err = pci_enable_device_mem(pdev);
    102	if (err)
    103		goto err_free_vf;
    104
    105	err = pci_request_regions(pdev, nfp_net_driver_name);
    106	if (err) {
    107		dev_err(&pdev->dev, "Unable to allocate device memory.\n");
    108		goto err_pci_disable;
    109	}
    110
    111	pci_set_master(pdev);
    112
    113	err = dma_set_mask_and_coherent(&pdev->dev, dev_info->dma_mask);
    114	if (err)
    115		goto err_pci_regions;
    116
    117	/* Map the Control BAR.
    118	 *
    119	 * Irrespective of the advertised BAR size we only map the
    120	 * first NFP_NET_CFG_BAR_SZ of the BAR.  This keeps the code
    121	 * the identical for PF and VF drivers.
    122	 */
    123	ctrl_bar = ioremap(pci_resource_start(pdev, NFP_NET_CTRL_BAR),
    124				   NFP_NET_CFG_BAR_SZ);
    125	if (!ctrl_bar) {
    126		dev_err(&pdev->dev,
    127			"Failed to map resource %d\n", NFP_NET_CTRL_BAR);
    128		err = -EIO;
    129		goto err_pci_regions;
    130	}
    131
    132	nfp_net_get_fw_version(&fw_ver, ctrl_bar);
    133	if (fw_ver.extend & NFP_NET_CFG_VERSION_RESERVED_MASK ||
    134	    fw_ver.class != NFP_NET_CFG_VERSION_CLASS_GENERIC) {
    135		dev_err(&pdev->dev, "Unknown Firmware ABI %d.%d.%d.%d\n",
    136			fw_ver.extend, fw_ver.class,
    137			fw_ver.major, fw_ver.minor);
    138		err = -EINVAL;
    139		goto err_ctrl_unmap;
    140	}
    141
    142	/* Determine stride */
    143	if (nfp_net_fw_ver_eq(&fw_ver, 0, 0, 0, 1)) {
    144		stride = 2;
    145		tx_bar_no = NFP_NET_Q0_BAR;
    146		rx_bar_no = NFP_NET_Q1_BAR;
    147		dev_warn(&pdev->dev, "OBSOLETE Firmware detected - VF isolation not available\n");
    148	} else {
    149		switch (fw_ver.major) {
    150		case 1 ... 5:
    151			stride = 4;
    152			tx_bar_no = NFP_NET_Q0_BAR;
    153			rx_bar_no = tx_bar_no;
    154			break;
    155		default:
    156			dev_err(&pdev->dev, "Unsupported Firmware ABI %d.%d.%d.%d\n",
    157				fw_ver.extend, fw_ver.class,
    158				fw_ver.major, fw_ver.minor);
    159			err = -EINVAL;
    160			goto err_ctrl_unmap;
    161		}
    162	}
    163
    164	/* Find out how many rings are supported */
    165	max_tx_rings = readl(ctrl_bar + NFP_NET_CFG_MAX_TXRINGS);
    166	max_rx_rings = readl(ctrl_bar + NFP_NET_CFG_MAX_RXRINGS);
    167
    168	tx_bar_sz = NFP_QCP_QUEUE_ADDR_SZ * max_tx_rings * stride;
    169	rx_bar_sz = NFP_QCP_QUEUE_ADDR_SZ * max_rx_rings * stride;
    170
    171	/* Sanity checks */
    172	if (tx_bar_sz > pci_resource_len(pdev, tx_bar_no)) {
    173		dev_err(&pdev->dev,
    174			"TX BAR too small for number of TX rings. Adjusting\n");
    175		tx_bar_sz = pci_resource_len(pdev, tx_bar_no);
    176		max_tx_rings = (tx_bar_sz / NFP_QCP_QUEUE_ADDR_SZ) / 2;
    177	}
    178	if (rx_bar_sz > pci_resource_len(pdev, rx_bar_no)) {
    179		dev_err(&pdev->dev,
    180			"RX BAR too small for number of RX rings. Adjusting\n");
    181		rx_bar_sz = pci_resource_len(pdev, rx_bar_no);
    182		max_rx_rings = (rx_bar_sz / NFP_QCP_QUEUE_ADDR_SZ) / 2;
    183	}
    184
    185	startq = readl(ctrl_bar + NFP_NET_CFG_START_TXQ);
    186	tx_bar_off = nfp_qcp_queue_offset(dev_info, startq);
    187	startq = readl(ctrl_bar + NFP_NET_CFG_START_RXQ);
    188	rx_bar_off = nfp_qcp_queue_offset(dev_info, startq);
    189
    190	/* Allocate and initialise the netdev */
    191	nn = nfp_net_alloc(pdev, dev_info, ctrl_bar, true,
    192			   max_tx_rings, max_rx_rings);
    193	if (IS_ERR(nn)) {
    194		err = PTR_ERR(nn);
    195		goto err_ctrl_unmap;
    196	}
    197	vf->nn = nn;
    198
    199	nn->dp.is_vf = 1;
    200	nn->stride_tx = stride;
    201	nn->stride_rx = stride;
    202
    203	if (rx_bar_no == tx_bar_no) {
    204		u32 bar_off, bar_sz;
    205		resource_size_t map_addr;
    206
    207		/* Make a single overlapping BAR mapping */
    208		if (tx_bar_off < rx_bar_off)
    209			bar_off = tx_bar_off;
    210		else
    211			bar_off = rx_bar_off;
    212
    213		if ((tx_bar_off + tx_bar_sz) > (rx_bar_off + rx_bar_sz))
    214			bar_sz = (tx_bar_off + tx_bar_sz) - bar_off;
    215		else
    216			bar_sz = (rx_bar_off + rx_bar_sz) - bar_off;
    217
    218		map_addr = pci_resource_start(pdev, tx_bar_no) + bar_off;
    219		vf->q_bar = ioremap(map_addr, bar_sz);
    220		if (!vf->q_bar) {
    221			nn_err(nn, "Failed to map resource %d\n", tx_bar_no);
    222			err = -EIO;
    223			goto err_netdev_free;
    224		}
    225
    226		/* TX queues */
    227		nn->tx_bar = vf->q_bar + (tx_bar_off - bar_off);
    228		/* RX queues */
    229		nn->rx_bar = vf->q_bar + (rx_bar_off - bar_off);
    230	} else {
    231		resource_size_t map_addr;
    232
    233		/* TX queues */
    234		map_addr = pci_resource_start(pdev, tx_bar_no) + tx_bar_off;
    235		nn->tx_bar = ioremap(map_addr, tx_bar_sz);
    236		if (!nn->tx_bar) {
    237			nn_err(nn, "Failed to map resource %d\n", tx_bar_no);
    238			err = -EIO;
    239			goto err_netdev_free;
    240		}
    241
    242		/* RX queues */
    243		map_addr = pci_resource_start(pdev, rx_bar_no) + rx_bar_off;
    244		nn->rx_bar = ioremap(map_addr, rx_bar_sz);
    245		if (!nn->rx_bar) {
    246			nn_err(nn, "Failed to map resource %d\n", rx_bar_no);
    247			err = -EIO;
    248			goto err_unmap_tx;
    249		}
    250	}
    251
    252	nfp_netvf_get_mac_addr(nn);
    253
    254	num_irqs = nfp_net_irqs_alloc(pdev, vf->irq_entries,
    255				      NFP_NET_MIN_VNIC_IRQS,
    256				      NFP_NET_NON_Q_VECTORS +
    257				      nn->dp.num_r_vecs);
    258	if (!num_irqs) {
    259		nn_warn(nn, "Unable to allocate MSI-X Vectors. Exiting\n");
    260		err = -EIO;
    261		goto err_unmap_rx;
    262	}
    263	nfp_net_irqs_assign(nn, vf->irq_entries, num_irqs);
    264
    265	err = nfp_net_init(nn);
    266	if (err)
    267		goto err_irqs_disable;
    268
    269	nfp_net_info(nn);
    270	vf->ddir = nfp_net_debugfs_device_add(pdev);
    271	nfp_net_debugfs_vnic_add(nn, vf->ddir);
    272
    273	return 0;
    274
    275err_irqs_disable:
    276	nfp_net_irqs_disable(pdev);
    277err_unmap_rx:
    278	if (!vf->q_bar)
    279		iounmap(nn->rx_bar);
    280err_unmap_tx:
    281	if (!vf->q_bar)
    282		iounmap(nn->tx_bar);
    283	else
    284		iounmap(vf->q_bar);
    285err_netdev_free:
    286	nfp_net_free(nn);
    287err_ctrl_unmap:
    288	iounmap(ctrl_bar);
    289err_pci_regions:
    290	pci_release_regions(pdev);
    291err_pci_disable:
    292	pci_disable_device(pdev);
    293err_free_vf:
    294	pci_set_drvdata(pdev, NULL);
    295	kfree(vf);
    296	return err;
    297}
    298
    299static void nfp_netvf_pci_remove(struct pci_dev *pdev)
    300{
    301	struct nfp_net_vf *vf;
    302	struct nfp_net *nn;
    303
    304	vf = pci_get_drvdata(pdev);
    305	if (!vf)
    306		return;
    307
    308	nn = vf->nn;
    309
    310	/* Note, the order is slightly different from above as we need
    311	 * to keep the nn pointer around till we have freed everything.
    312	 */
    313	nfp_net_debugfs_dir_clean(&nn->debugfs_dir);
    314	nfp_net_debugfs_dir_clean(&vf->ddir);
    315
    316	nfp_net_clean(nn);
    317
    318	nfp_net_irqs_disable(pdev);
    319
    320	if (!vf->q_bar) {
    321		iounmap(nn->rx_bar);
    322		iounmap(nn->tx_bar);
    323	} else {
    324		iounmap(vf->q_bar);
    325	}
    326	iounmap(nn->dp.ctrl_bar);
    327
    328	nfp_net_free(nn);
    329
    330	pci_release_regions(pdev);
    331	pci_disable_device(pdev);
    332
    333	pci_set_drvdata(pdev, NULL);
    334	kfree(vf);
    335}
    336
    337struct pci_driver nfp_netvf_pci_driver = {
    338	.name        = nfp_net_driver_name,
    339	.id_table    = nfp_netvf_pci_device_ids,
    340	.probe       = nfp_netvf_pci_probe,
    341	.remove      = nfp_netvf_pci_remove,
    342	.shutdown    = nfp_netvf_pci_remove,
    343};