This patch registers supported virtual function
and initialization/deinit routines for the same.

Signed-off-by: Aman Kumar <aman.ku...@vvdntech.in>
---
 drivers/net/qdma/meson.build      |   1 +
 drivers/net/qdma/qdma.h           |   9 +
 drivers/net/qdma/qdma_ethdev.c    |  22 +++
 drivers/net/qdma/qdma_vf_ethdev.c | 319 ++++++++++++++++++++++++++++++
 4 files changed, 351 insertions(+)
 create mode 100644 drivers/net/qdma/qdma_vf_ethdev.c

diff --git a/drivers/net/qdma/meson.build b/drivers/net/qdma/meson.build
index dd2478be6c..c453d556b6 100644
--- a/drivers/net/qdma/meson.build
+++ b/drivers/net/qdma/meson.build
@@ -28,6 +28,7 @@ sources = files(
         'qdma_mbox.c',
         'qdma_user.c',
         'qdma_rxtx.c',
+        'qdma_vf_ethdev.c',
         'qdma_access/eqdma_soft_access/eqdma_soft_access.c',
         'qdma_access/eqdma_soft_access/eqdma_soft_reg_dump.c',
         'qdma_access/qdma_s80_hard_access/qdma_s80_hard_access.c',
diff --git a/drivers/net/qdma/qdma.h b/drivers/net/qdma/qdma.h
index 20a1b72dd1..d9239f34a7 100644
--- a/drivers/net/qdma/qdma.h
+++ b/drivers/net/qdma/qdma.h
@@ -25,6 +25,7 @@
 
 #define QDMA_NUM_BARS          (6)
 #define DEFAULT_PF_CONFIG_BAR  (0)
+#define DEFAULT_VF_CONFIG_BAR  (0)
 #define BAR_ID_INVALID         (-1)
 
 #define QDMA_FUNC_ID_INVALID    0xFFFF
@@ -306,6 +307,10 @@ int qdma_pf_csr_read(struct rte_eth_dev *dev);
 int qdma_vf_csr_read(struct rte_eth_dev *dev);
 
 uint8_t qmda_get_desc_sz_idx(enum rte_pmd_qdma_bypass_desc_len);
+int qdma_vf_dev_rx_queue_start(struct rte_eth_dev *dev, uint16_t qid);
+int qdma_vf_dev_rx_queue_stop(struct rte_eth_dev *dev, uint16_t qid);
+int qdma_vf_dev_tx_queue_start(struct rte_eth_dev *dev, uint16_t qid);
+int qdma_vf_dev_tx_queue_stop(struct rte_eth_dev *dev, uint16_t qid);
 
 int qdma_init_rx_queue(struct qdma_rx_queue *rxq);
 void qdma_reset_tx_queue(struct qdma_tx_queue *txq);
@@ -342,5 +347,9 @@ struct rte_memzone *qdma_zone_reserve(struct rte_eth_dev 
*dev,
                                                socket_id, 0, QDMA_ALIGN);
 }
 
+bool is_qdma_supported(struct rte_eth_dev *dev);
+bool is_vf_device_supported(struct rte_eth_dev *dev);
+bool is_pf_device_supported(struct rte_eth_dev *dev);
+
 void qdma_check_errors(void *arg);
 #endif /* ifndef __QDMA_H__ */
diff --git a/drivers/net/qdma/qdma_ethdev.c b/drivers/net/qdma/qdma_ethdev.c
index 466a9e9284..a33d5efc5a 100644
--- a/drivers/net/qdma/qdma_ethdev.c
+++ b/drivers/net/qdma/qdma_ethdev.c
@@ -695,6 +695,28 @@ static struct rte_pci_driver rte_qdma_pmd = {
        .remove = eth_qdma_pci_remove,
 };
 
+bool
+is_pf_device_supported(struct rte_eth_dev *dev)
+{
+       if (strcmp(dev->device->driver->name, rte_qdma_pmd.driver.name))
+               return false;
+
+       return true;
+}
+
+bool is_qdma_supported(struct rte_eth_dev *dev)
+{
+       bool is_pf, is_vf;
+
+       is_pf = is_pf_device_supported(dev);
+       is_vf = is_vf_device_supported(dev);
+
+       if (!is_pf && !is_vf)
+               return false;
+
+       return true;
+}
+
 RTE_PMD_REGISTER_PCI(net_qdma, rte_qdma_pmd);
 RTE_PMD_REGISTER_PCI_TABLE(net_qdma, qdma_pci_id_tbl);
 RTE_LOG_REGISTER_DEFAULT(qdma_logtype_pmd, NOTICE);
diff --git a/drivers/net/qdma/qdma_vf_ethdev.c 
b/drivers/net/qdma/qdma_vf_ethdev.c
new file mode 100644
index 0000000000..ca3d21b688
--- /dev/null
+++ b/drivers/net/qdma/qdma_vf_ethdev.c
@@ -0,0 +1,319 @@
+/* SPDX-License-Identifier: BSD-3-Clause
+ * Copyright(c) 2017-2022 Xilinx, Inc. All rights reserved.
+ */
+
+#include <stdint.h>
+#include <stdbool.h>
+#include <sys/mman.h>
+#include <sys/fcntl.h>
+#include <rte_memzone.h>
+#include <rte_string_fns.h>
+#include <ethdev_pci.h>
+#include <rte_malloc.h>
+#include <rte_dev.h>
+#include <rte_pci.h>
+#include <rte_ether.h>
+#include <rte_ethdev.h>
+#include <rte_cycles.h>
+#include <rte_alarm.h>
+#include <unistd.h>
+#include <string.h>
+#include <linux/pci.h>
+
+#include "qdma.h"
+#include "qdma_version.h"
+#include "qdma_access_common.h"
+#include "qdma_mbox_protocol.h"
+#include "qdma_mbox.h"
+#include "qdma_devops.h"
+
+static int eth_qdma_vf_dev_init(struct rte_eth_dev *dev);
+static int eth_qdma_vf_dev_uninit(struct rte_eth_dev *dev);
+
+/*
+ * The set of PCI devices this driver supports
+ */
+static struct rte_pci_id qdma_vf_pci_id_tbl[] = {
+#define RTE_PCI_DEV_ID_DECL(vend, dev) {RTE_PCI_DEVICE(vend, dev)},
+#ifndef PCI_VENDOR_ID_VVDN
+#define PCI_VENDOR_ID_VVDN 0x1f44
+#endif
+
+       /** Gen 3 VF */
+       RTE_PCI_DEV_ID_DECL(PCI_VENDOR_ID_VVDN, 0x0281) /* VF on PF 0 */
+
+       { .vendor_id = 0, /* sentinel */ },
+};
+
+static int qdma_ethdev_online(struct rte_eth_dev *dev)
+{
+       int rv = 0;
+       int qbase = -1;
+       struct qdma_pci_dev *qdma_dev = dev->data->dev_private;
+       struct qdma_mbox_msg *m = qdma_mbox_msg_alloc();
+
+       if (!m)
+               return -ENOMEM;
+
+       qmda_mbox_compose_vf_online(qdma_dev->func_id, 0, &qbase, m->raw_data);
+
+       rv = qdma_mbox_msg_send(dev, m, MBOX_OP_RSP_TIMEOUT);
+       if (rv < 0)
+               PMD_DRV_LOG(ERR, "%x, send hello failed %d.\n",
+                           qdma_dev->func_id, rv);
+
+       rv = qdma_mbox_vf_dev_info_get(m->raw_data,
+                               &qdma_dev->dev_cap,
+                               &qdma_dev->dma_device_index);
+
+       if (rv < 0) {
+               PMD_DRV_LOG(ERR, "%x, failed to get dev info %d.\n",
+                               qdma_dev->func_id, rv);
+       } else {
+               qdma_mbox_msg_free(m);
+       }
+       return rv;
+}
+
+static int qdma_ethdev_offline(struct rte_eth_dev *dev)
+{
+       int rv;
+       struct qdma_pci_dev *qdma_dev = dev->data->dev_private;
+       struct qdma_mbox_msg *m = qdma_mbox_msg_alloc();
+
+       if (!m)
+               return -ENOMEM;
+
+       qdma_mbox_compose_vf_offline(qdma_dev->func_id, m->raw_data);
+
+       rv = qdma_mbox_msg_send(dev, m, 0);
+       if (rv < 0)
+               PMD_DRV_LOG(ERR, "%x, send bye failed %d.\n",
+                           qdma_dev->func_id, rv);
+
+       return rv;
+}
+
+/**
+ * DPDK callback to register a PCI device.
+ *
+ * This function creates an Ethernet device for each port of a given
+ * PCI device.
+ *
+ * @param[in] dev
+ *   Pointer to Ethernet device structure.
+ *
+ * @return
+ *   0 on success, negative errno value on failure.
+ */
+static int eth_qdma_vf_dev_init(struct rte_eth_dev *dev)
+{
+       struct qdma_pci_dev *dma_priv;
+       uint8_t *baseaddr;
+       int i, idx;
+       static bool once = true;
+       struct rte_pci_device *pci_dev;
+
+       /* sanity checks */
+       if (dev == NULL)
+               return -EINVAL;
+       if (dev->data == NULL)
+               return -EINVAL;
+       if (dev->data->dev_private == NULL)
+               return -EINVAL;
+
+       pci_dev = RTE_ETH_DEV_TO_PCI(dev);
+       if (pci_dev == NULL)
+               return -EINVAL;
+
+       /* for secondary processes, we don't initialise any further as primary
+        * has already done this work.
+        */
+       if (rte_eal_process_type() != RTE_PROC_PRIMARY)
+               return 0;
+
+       if (once) {
+               RTE_LOG(INFO, PMD, "QDMA PMD VERSION: %s\n", QDMA_PMD_VERSION);
+               once = false;
+       }
+
+       /* allocate space for a single Ethernet MAC address */
+       dev->data->mac_addrs = rte_zmalloc("qdma_vf",
+                       RTE_ETHER_ADDR_LEN * 1, 0);
+       if (dev->data->mac_addrs == NULL)
+               return -ENOMEM;
+
+       /* Copy some dummy Ethernet MAC address for XDMA device
+        * This will change in real NIC device...
+        */
+       for (i = 0; i < RTE_ETHER_ADDR_LEN; ++i)
+               dev->data->mac_addrs[0].addr_bytes[i] = 0x15 + i;
+
+       /* Init system & device */
+       dma_priv = (struct qdma_pci_dev *)dev->data->dev_private;
+       dma_priv->func_id = 0;
+       dma_priv->is_vf = 1;
+       dma_priv->timer_count = DEFAULT_TIMER_CNT_TRIG_MODE_TIMER;
+
+       dma_priv->en_desc_prefetch = 0;
+       dma_priv->cmpt_desc_len = DEFAULT_QDMA_CMPT_DESC_LEN;
+       dma_priv->c2h_bypass_mode = RTE_PMD_QDMA_RX_BYPASS_NONE;
+       dma_priv->h2c_bypass_mode = 0;
+
+       dma_priv->config_bar_idx = DEFAULT_VF_CONFIG_BAR;
+       dma_priv->bypass_bar_idx = BAR_ID_INVALID;
+       dma_priv->user_bar_idx = BAR_ID_INVALID;
+
+       if (qdma_check_kvargs(dev->device->devargs, dma_priv)) {
+               PMD_DRV_LOG(INFO, "devargs failed\n");
+               rte_free(dev->data->mac_addrs);
+               return -EINVAL;
+       }
+
+       /* Store BAR address and length of Config BAR */
+       baseaddr = (uint8_t *)
+                       pci_dev->mem_resource[dma_priv->config_bar_idx].addr;
+       dma_priv->bar_addr[dma_priv->config_bar_idx] = baseaddr;
+
+       /* Assigning QDMA access layer function pointers based on the HW design 
*/
+       dma_priv->hw_access = rte_zmalloc("vf_hwaccess",
+                       sizeof(struct qdma_hw_access), 0);
+       if (dma_priv->hw_access == NULL) {
+               rte_free(dev->data->mac_addrs);
+               return -ENOMEM;
+       }
+       idx = qdma_hw_access_init(dev, dma_priv->is_vf, dma_priv->hw_access);
+       if (idx < 0) {
+               rte_free(dma_priv->hw_access);
+               rte_free(dev->data->mac_addrs);
+               return -EINVAL;
+       }
+
+       idx = qdma_get_hw_version(dev);
+       if (idx < 0) {
+               rte_free(dma_priv->hw_access);
+               rte_free(dev->data->mac_addrs);
+               return -EINVAL;
+       }
+
+       idx = qdma_identify_bars(dev);
+       if (idx < 0) {
+               rte_free(dma_priv->hw_access);
+               rte_free(dev->data->mac_addrs);
+               return -EINVAL;
+       }
+
+       /* Store BAR address and length of AXI Master Lite BAR(user bar) */
+       if (dma_priv->user_bar_idx >= 0) {
+               baseaddr = (uint8_t *)
+                            pci_dev->mem_resource[dma_priv->user_bar_idx].addr;
+               dma_priv->bar_addr[dma_priv->user_bar_idx] = baseaddr;
+       }
+
+       if (dma_priv->ip_type == QDMA_VERSAL_HARD_IP)
+               dma_priv->dev_cap.mailbox_intr = 0;
+       else
+               dma_priv->dev_cap.mailbox_intr = 1;
+
+       qdma_mbox_init(dev);
+       idx = qdma_ethdev_online(dev);
+       if (idx < 0) {
+               rte_free(dma_priv->hw_access);
+               rte_free(dev->data->mac_addrs);
+               return -EINVAL;
+       }
+
+       if (dma_priv->dev_cap.cmpt_trig_count_timer) {
+               /* Setting default Mode to
+                * RTE_PMD_QDMA_TRIG_MODE_USER_TIMER_COUNT
+                */
+               dma_priv->trigger_mode =
+                               RTE_PMD_QDMA_TRIG_MODE_USER_TIMER_COUNT;
+       } else {
+               /* Setting default Mode to RTE_PMD_QDMA_TRIG_MODE_USER_TIMER */
+               dma_priv->trigger_mode = RTE_PMD_QDMA_TRIG_MODE_USER_TIMER;
+       }
+       if (dma_priv->trigger_mode == RTE_PMD_QDMA_TRIG_MODE_USER_TIMER_COUNT)
+               dma_priv->timer_count = DEFAULT_TIMER_CNT_TRIG_MODE_COUNT_TIMER;
+
+       dma_priv->reset_state = RESET_STATE_IDLE;
+
+       PMD_DRV_LOG(INFO, "VF-%d(DEVFN) QDMA device driver probe:",
+                               dma_priv->func_id);
+
+       return 0;
+}
+
+/**
+ * DPDK callback to deregister PCI device.
+ *
+ * @param[in] dev
+ *   Pointer to Ethernet device structure.
+ *
+ * @return
+ *   0 on success, negative errno value on failure.
+ */
+static int eth_qdma_vf_dev_uninit(struct rte_eth_dev *dev)
+{
+       struct qdma_pci_dev *qdma_dev = dev->data->dev_private;
+
+       /* only uninitialize in the primary process */
+       if (rte_eal_process_type() != RTE_PROC_PRIMARY)
+               return -EPERM;
+
+       qdma_ethdev_offline(dev);
+
+       if (qdma_dev->reset_state != RESET_STATE_RECV_PF_RESET_REQ)
+               qdma_mbox_uninit(dev);
+
+       dev->dev_ops = NULL;
+       dev->rx_pkt_burst = NULL;
+       dev->tx_pkt_burst = NULL;
+       dev->data->nb_rx_queues = 0;
+       dev->data->nb_tx_queues = 0;
+
+       if (dev->data->mac_addrs != NULL) {
+               rte_free(dev->data->mac_addrs);
+               dev->data->mac_addrs = NULL;
+       }
+
+       if (qdma_dev->q_info != NULL) {
+               rte_free(qdma_dev->q_info);
+               qdma_dev->q_info = NULL;
+       }
+
+       return 0;
+}
+
+static int eth_qdma_vf_pci_probe(struct rte_pci_driver *pci_drv __rte_unused,
+                                       struct rte_pci_device *pci_dev)
+{
+       return rte_eth_dev_pci_generic_probe(pci_dev,
+                                               sizeof(struct qdma_pci_dev),
+                                               eth_qdma_vf_dev_init);
+}
+
+/* Detach a ethdev interface */
+static int eth_qdma_vf_pci_remove(struct rte_pci_device *pci_dev)
+{
+       return rte_eth_dev_pci_generic_remove(pci_dev, eth_qdma_vf_dev_uninit);
+}
+
+static struct rte_pci_driver rte_qdma_vf_pmd = {
+       .id_table = qdma_vf_pci_id_tbl,
+       .drv_flags = RTE_PCI_DRV_NEED_MAPPING,
+       .probe = eth_qdma_vf_pci_probe,
+       .remove = eth_qdma_vf_pci_remove,
+};
+
+bool
+is_vf_device_supported(struct rte_eth_dev *dev)
+{
+       if (strcmp(dev->device->driver->name, rte_qdma_vf_pmd.driver.name))
+               return false;
+
+       return true;
+}
+
+RTE_PMD_REGISTER_PCI(net_qdma_vf, rte_qdma_vf_pmd);
+RTE_PMD_REGISTER_PCI_TABLE(net_qdma_vf, qdma_vf_pci_id_tbl);
-- 
2.36.1

Reply via email to