root/drivers/net/ethernet/netronome/nfp/nfp_netvf_main.c
// SPDX-License-Identifier: (GPL-2.0-only OR BSD-2-Clause)
/* Copyright (C) 2015-2018 Netronome Systems, Inc. */

/*
 * nfp_netvf_main.c
 * Netronome virtual function network device driver: Main entry point
 * Author: Jason McMullan <jason.mcmullan@netronome.com>
 *         Rolf Neugebauer <rolf.neugebauer@netronome.com>
 */

#include <linux/module.h>
#include <linux/kernel.h>
#include <linux/init.h>
#include <linux/etherdevice.h>

#include "nfpcore/nfp_dev.h"
#include "nfp_net_ctrl.h"
#include "nfp_net.h"
#include "nfp_main.h"

/**
 * struct nfp_net_vf - NFP VF-specific device structure
 * @nn:         NFP Net structure for this device
 * @irq_entries: Pre-allocated array of MSI-X entries
 * @q_bar:      Pointer to mapped QC memory (NULL if TX/RX mapped directly)
 * @ddir:       Per-device debugfs directory
 */
struct nfp_net_vf {
        struct nfp_net *nn;

        struct msix_entry irq_entries[NFP_NET_NON_Q_VECTORS +
                                      NFP_NET_MAX_TX_RINGS];
        u8 __iomem *q_bar;

        struct dentry *ddir;
};

static const char nfp_net_driver_name[] = "nfp_netvf";

static const struct pci_device_id nfp_netvf_pci_device_ids[] = {
        { PCI_VENDOR_ID_NETRONOME, PCI_DEVICE_ID_NFP3800_VF,
          PCI_VENDOR_ID_NETRONOME, PCI_ANY_ID,
          PCI_ANY_ID, 0, NFP_DEV_NFP3800_VF,
        },
        { PCI_VENDOR_ID_NETRONOME, PCI_DEVICE_ID_NFP6000_VF,
          PCI_VENDOR_ID_NETRONOME, PCI_ANY_ID,
          PCI_ANY_ID, 0, NFP_DEV_NFP6000_VF,
        },
        { PCI_VENDOR_ID_CORIGINE, PCI_DEVICE_ID_NFP3800_VF,
          PCI_VENDOR_ID_CORIGINE, PCI_ANY_ID,
          PCI_ANY_ID, 0, NFP_DEV_NFP3800_VF,
        },
        { PCI_VENDOR_ID_CORIGINE, PCI_DEVICE_ID_NFP6000_VF,
          PCI_VENDOR_ID_CORIGINE, PCI_ANY_ID,
          PCI_ANY_ID, 0, NFP_DEV_NFP6000_VF,
        },
        { 0, } /* Required last entry. */
};
MODULE_DEVICE_TABLE(pci, nfp_netvf_pci_device_ids);

static void nfp_netvf_get_mac_addr(struct nfp_net *nn)
{
        u8 mac_addr[ETH_ALEN];

        put_unaligned_be32(nn_readl(nn, NFP_NET_CFG_MACADDR + 0), &mac_addr[0]);
        put_unaligned_be16(nn_readw(nn, NFP_NET_CFG_MACADDR + 6), &mac_addr[4]);

        if (!is_valid_ether_addr(mac_addr)) {
                eth_hw_addr_random(nn->dp.netdev);
                return;
        }

        eth_hw_addr_set(nn->dp.netdev, mac_addr);
        ether_addr_copy(nn->dp.netdev->perm_addr, mac_addr);
}

static int nfp_netvf_pci_probe(struct pci_dev *pdev,
                               const struct pci_device_id *pci_id)
{
        const struct nfp_dev_info *dev_info;
        struct nfp_net_fw_version fw_ver;
        int max_tx_rings, max_rx_rings;
        u32 tx_bar_off, rx_bar_off;
        u32 tx_bar_sz, rx_bar_sz;
        int tx_bar_no, rx_bar_no;
        struct nfp_net_vf *vf;
        unsigned int num_irqs;
        u8 __iomem *ctrl_bar;
        struct nfp_net *nn;
        u32 startq;
        int stride;
        int err;

        dev_info = &nfp_dev_info[pci_id->driver_data];

        vf = kzalloc_obj(*vf);
        if (!vf)
                return -ENOMEM;
        pci_set_drvdata(pdev, vf);

        err = pci_enable_device_mem(pdev);
        if (err)
                goto err_free_vf;

        err = pci_request_regions(pdev, nfp_net_driver_name);
        if (err) {
                dev_err(&pdev->dev, "Unable to allocate device memory.\n");
                goto err_pci_disable;
        }

        pci_set_master(pdev);

        err = dma_set_mask_and_coherent(&pdev->dev, dev_info->dma_mask);
        if (err)
                goto err_pci_regions;

        /* Map the Control BAR.
         *
         * Irrespective of the advertised BAR size we only map the
         * first NFP_NET_CFG_BAR_SZ of the BAR.  This keeps the code
         * the identical for PF and VF drivers.
         */
        ctrl_bar = ioremap(pci_resource_start(pdev, NFP_NET_CTRL_BAR),
                                   NFP_NET_CFG_BAR_SZ);
        if (!ctrl_bar) {
                dev_err(&pdev->dev,
                        "Failed to map resource %d\n", NFP_NET_CTRL_BAR);
                err = -EIO;
                goto err_pci_regions;
        }

        nfp_net_get_fw_version(&fw_ver, ctrl_bar);
        if (fw_ver.extend & NFP_NET_CFG_VERSION_RESERVED_MASK ||
            fw_ver.class != NFP_NET_CFG_VERSION_CLASS_GENERIC) {
                dev_err(&pdev->dev, "Unknown Firmware ABI %d.%d.%d.%d\n",
                        fw_ver.extend, fw_ver.class,
                        fw_ver.major, fw_ver.minor);
                err = -EINVAL;
                goto err_ctrl_unmap;
        }

        /* Determine stride */
        if (nfp_net_fw_ver_eq(&fw_ver, 0, 0, 0, 1)) {
                stride = 2;
                tx_bar_no = NFP_NET_Q0_BAR;
                rx_bar_no = NFP_NET_Q1_BAR;
                dev_warn(&pdev->dev, "OBSOLETE Firmware detected - VF isolation not available\n");
        } else {
                switch (fw_ver.major) {
                case 1 ... 5:
                        stride = 4;
                        tx_bar_no = NFP_NET_Q0_BAR;
                        rx_bar_no = tx_bar_no;
                        break;
                default:
                        dev_err(&pdev->dev, "Unsupported Firmware ABI %d.%d.%d.%d\n",
                                fw_ver.extend, fw_ver.class,
                                fw_ver.major, fw_ver.minor);
                        err = -EINVAL;
                        goto err_ctrl_unmap;
                }
        }

        /* Find out how many rings are supported */
        max_tx_rings = readl(ctrl_bar + NFP_NET_CFG_MAX_TXRINGS);
        max_rx_rings = readl(ctrl_bar + NFP_NET_CFG_MAX_RXRINGS);

        tx_bar_sz = NFP_QCP_QUEUE_ADDR_SZ * max_tx_rings * stride;
        rx_bar_sz = NFP_QCP_QUEUE_ADDR_SZ * max_rx_rings * stride;

        /* Sanity checks */
        if (tx_bar_sz > pci_resource_len(pdev, tx_bar_no)) {
                dev_err(&pdev->dev,
                        "TX BAR too small for number of TX rings. Adjusting\n");
                tx_bar_sz = pci_resource_len(pdev, tx_bar_no);
                max_tx_rings = (tx_bar_sz / NFP_QCP_QUEUE_ADDR_SZ) / 2;
        }
        if (rx_bar_sz > pci_resource_len(pdev, rx_bar_no)) {
                dev_err(&pdev->dev,
                        "RX BAR too small for number of RX rings. Adjusting\n");
                rx_bar_sz = pci_resource_len(pdev, rx_bar_no);
                max_rx_rings = (rx_bar_sz / NFP_QCP_QUEUE_ADDR_SZ) / 2;
        }

        startq = readl(ctrl_bar + NFP_NET_CFG_START_TXQ);
        tx_bar_off = nfp_qcp_queue_offset(dev_info, startq);
        startq = readl(ctrl_bar + NFP_NET_CFG_START_RXQ);
        rx_bar_off = nfp_qcp_queue_offset(dev_info, startq);

        /* Allocate and initialise the netdev */
        nn = nfp_net_alloc(pdev, dev_info, ctrl_bar, true,
                           max_tx_rings, max_rx_rings);
        if (IS_ERR(nn)) {
                err = PTR_ERR(nn);
                goto err_ctrl_unmap;
        }
        vf->nn = nn;

        nn->dp.is_vf = 1;
        nn->stride_tx = stride;
        nn->stride_rx = stride;

        if (rx_bar_no == tx_bar_no) {
                u32 bar_off, bar_sz;
                resource_size_t map_addr;

                /* Make a single overlapping BAR mapping */
                if (tx_bar_off < rx_bar_off)
                        bar_off = tx_bar_off;
                else
                        bar_off = rx_bar_off;

                if ((tx_bar_off + tx_bar_sz) > (rx_bar_off + rx_bar_sz))
                        bar_sz = (tx_bar_off + tx_bar_sz) - bar_off;
                else
                        bar_sz = (rx_bar_off + rx_bar_sz) - bar_off;

                map_addr = pci_resource_start(pdev, tx_bar_no) + bar_off;
                vf->q_bar = ioremap(map_addr, bar_sz);
                if (!vf->q_bar) {
                        nn_err(nn, "Failed to map resource %d\n", tx_bar_no);
                        err = -EIO;
                        goto err_netdev_free;
                }

                /* TX queues */
                nn->tx_bar = vf->q_bar + (tx_bar_off - bar_off);
                /* RX queues */
                nn->rx_bar = vf->q_bar + (rx_bar_off - bar_off);
        } else {
                resource_size_t map_addr;

                /* TX queues */
                map_addr = pci_resource_start(pdev, tx_bar_no) + tx_bar_off;
                nn->tx_bar = ioremap(map_addr, tx_bar_sz);
                if (!nn->tx_bar) {
                        nn_err(nn, "Failed to map resource %d\n", tx_bar_no);
                        err = -EIO;
                        goto err_netdev_free;
                }

                /* RX queues */
                map_addr = pci_resource_start(pdev, rx_bar_no) + rx_bar_off;
                nn->rx_bar = ioremap(map_addr, rx_bar_sz);
                if (!nn->rx_bar) {
                        nn_err(nn, "Failed to map resource %d\n", rx_bar_no);
                        err = -EIO;
                        goto err_unmap_tx;
                }
        }

        nfp_netvf_get_mac_addr(nn);

        num_irqs = nfp_net_irqs_alloc(pdev, vf->irq_entries,
                                      NFP_NET_MIN_VNIC_IRQS,
                                      NFP_NET_NON_Q_VECTORS +
                                      nn->dp.num_r_vecs);
        if (!num_irqs) {
                nn_warn(nn, "Unable to allocate MSI-X Vectors. Exiting\n");
                err = -EIO;
                goto err_unmap_rx;
        }
        nfp_net_irqs_assign(nn, vf->irq_entries, num_irqs);

        err = nfp_net_init(nn);
        if (err)
                goto err_irqs_disable;

        nfp_net_info(nn);
        vf->ddir = nfp_net_debugfs_device_add(pdev);
        nfp_net_debugfs_vnic_add(nn, vf->ddir);

        return 0;

err_irqs_disable:
        nfp_net_irqs_disable(pdev);
err_unmap_rx:
        if (!vf->q_bar)
                iounmap(nn->rx_bar);
err_unmap_tx:
        if (!vf->q_bar)
                iounmap(nn->tx_bar);
        else
                iounmap(vf->q_bar);
err_netdev_free:
        nfp_net_free(nn);
err_ctrl_unmap:
        iounmap(ctrl_bar);
err_pci_regions:
        pci_release_regions(pdev);
err_pci_disable:
        pci_disable_device(pdev);
err_free_vf:
        pci_set_drvdata(pdev, NULL);
        kfree(vf);
        return err;
}

static void nfp_netvf_pci_remove(struct pci_dev *pdev)
{
        struct nfp_net_vf *vf;
        struct nfp_net *nn;

        vf = pci_get_drvdata(pdev);
        if (!vf)
                return;

        nn = vf->nn;

        /* Note, the order is slightly different from above as we need
         * to keep the nn pointer around till we have freed everything.
         */
        nfp_net_debugfs_dir_clean(&nn->debugfs_dir);
        nfp_net_debugfs_dir_clean(&vf->ddir);

        nfp_net_clean(nn);

        nfp_net_irqs_disable(pdev);

        if (!vf->q_bar) {
                iounmap(nn->rx_bar);
                iounmap(nn->tx_bar);
        } else {
                iounmap(vf->q_bar);
        }
        iounmap(nn->dp.ctrl_bar);

        nfp_net_free(nn);

        pci_release_regions(pdev);
        pci_disable_device(pdev);

        pci_set_drvdata(pdev, NULL);
        kfree(vf);
}

struct pci_driver nfp_netvf_pci_driver = {
        .name        = nfp_net_driver_name,
        .id_table    = nfp_netvf_pci_device_ids,
        .probe       = nfp_netvf_pci_probe,
        .remove      = nfp_netvf_pci_remove,
        .shutdown    = nfp_netvf_pci_remove,
};