From: Danylo Vodopianov <dvo-...@napatech.com>

Implemented functionality for retrieving received packets from
virtual queues, supporting both SPLIT_RING and PACKED_RING types.

Updated sg_ops structure to include the new packet retrieval functions.

Signed-off-by: Danylo Vodopianov <dvo-...@napatech.com>
---
 doc/guides/nics/ntnic.rst                     |  44 ++++
 drivers/net/ntnic/dbsconfig/ntnic_dbsconfig.c | 216 ++++++++++++++++++
 2 files changed, 260 insertions(+)

diff --git a/doc/guides/nics/ntnic.rst b/doc/guides/nics/ntnic.rst
index 7ac92c891c..9ac8098735 100644
--- a/doc/guides/nics/ntnic.rst
+++ b/doc/guides/nics/ntnic.rst
@@ -51,6 +51,50 @@ and they are also supported.
 If vfio-pci is not required, kernel version 4.18 is supported.
 
 
+Configuration
+-------------
+
+Command line arguments
+~~~~~~~~~~~~~~~~~~~~~
+
+Following standard DPDK command line arguments are used by the PMD:
+
+    -a: Used to specifically define the NT adapter by PCI ID.
+    --iova-mode: Must be set to ‘pa’ for Physical Address mode.
+
+NTNIC specific arguments can be passed to the PMD in the PCI device parameter 
list::
+
+    <application> ... -a 0000:03:00.0[{,<NTNIC specific argument>}]
+
+The NTNIC specific argument format is::
+
+    <object>.<attribute>=[<object-ids>:]<value>
+
+Multiple arguments for the same device are separated by ‘,’ comma.
+<object-ids> can be a single value or a range.
+
+
+- ``rxqs`` parameter [int]
+
+    Specify number of RX queues to use.
+
+    To specify number of RX queues::
+
+        -a <domain>:<bus>:00.0,rxqs=4,txqs=4
+
+    By default, the value is set to 1.
+
+- ``txqs`` parameter [int]
+
+    Specify number of TX queues to use.
+
+    To specify number of TX queues::
+
+        -a <domain>:<bus>:00.0,rxqs=4,txqs=4
+
+    By default, the value is set to 1.
+
+
 Logging and Debugging
 ---------------------
 
diff --git a/drivers/net/ntnic/dbsconfig/ntnic_dbsconfig.c 
b/drivers/net/ntnic/dbsconfig/ntnic_dbsconfig.c
index 6d7862791d..a5051891e8 100644
--- a/drivers/net/ntnic/dbsconfig/ntnic_dbsconfig.c
+++ b/drivers/net/ntnic/dbsconfig/ntnic_dbsconfig.c
@@ -58,6 +58,15 @@
                }                                                               
                  \
        } while (0)
 
+#define inc_used(vq, num) do { \
+       struct nthw_virt_queue *temp_vq = (vq); \
+       temp_vq->next_used += (num); \
+       if (temp_vq->next_used >= temp_vq->queue_size) { \
+               temp_vq->next_used -= temp_vq->queue_size; \
+               temp_vq->used_wrap_count ^= 1; \
+       } \
+} while (0)
+
 struct __rte_aligned(8) virtq_avail {
        uint16_t flags;
        uint16_t idx;
@@ -107,6 +116,10 @@ struct nthw_virt_queue {
                        struct pvirtq_event_suppress *driver_event;
                        struct pvirtq_event_suppress *device_event;
                        struct pvirtq_desc *desc;
+                       struct {
+                               uint16_t next;
+                               uint16_t num;
+                       } outs;
                        /*
                         * when in-order release used Tx packets from FPGA it 
may collapse
                         * into a batch. When getting new Tx buffers we may 
only need
@@ -1097,6 +1110,107 @@ nthw_setup_mngd_tx_virt_queue(nthw_dbs_t *p_nthw_dbs,
        return NULL;
 }
 
+static uint16_t nthw_get_rx_packets(struct nthw_virt_queue *rxvq,
+       uint16_t n,
+       struct nthw_received_packets *rp,
+       uint16_t *nb_pkts)
+{
+       uint16_t segs = 0;
+       uint16_t pkts = 0;
+
+       if (rxvq->vq_type == SPLIT_RING) {
+               uint16_t i;
+               uint16_t entries_ready = (uint16_t)(rxvq->cached_idx - 
rxvq->used_idx);
+
+               if (entries_ready < n) {
+                       /* Look for more packets */
+                       rxvq->cached_idx = rxvq->p_used->idx;
+                       entries_ready = (uint16_t)(rxvq->cached_idx - 
rxvq->used_idx);
+
+                       if (entries_ready == 0) {
+                               *nb_pkts = 0;
+                               return 0;
+                       }
+
+                       if (n > entries_ready)
+                               n = entries_ready;
+               }
+
+               /*
+                * Give packets - make sure all packets are whole packets.
+                * Valid because queue_size is always 2^n
+                */
+               const uint16_t queue_mask = (uint16_t)(rxvq->queue_size - 1);
+               const uint32_t buf_len = rxvq->p_desc[0].len;
+
+               uint16_t used = rxvq->used_idx;
+
+               for (i = 0; i < n; ++i) {
+                       uint32_t id = rxvq->p_used->ring[used & queue_mask].id;
+                       rp[i].addr = rxvq->p_virtual_addr[id].virt_addr;
+                       rp[i].len = rxvq->p_used->ring[used & queue_mask].len;
+
+                       uint32_t pkt_len = ((struct _pkt_hdr_rx 
*)rp[i].addr)->cap_len;
+
+                       if (pkt_len > buf_len) {
+                               /* segmented */
+                               int nbsegs = (pkt_len + buf_len - 1) / buf_len;
+
+                               if (((int)i + nbsegs) > n) {
+                                       /* don't have enough segments - break 
out */
+                                       break;
+                               }
+
+                               int ii;
+
+                               for (ii = 1; ii < nbsegs; ii++) {
+                                       ++i;
+                                       id = rxvq->p_used->ring[(used + ii) & 
queue_mask].id;
+                                       rp[i].addr = 
rxvq->p_virtual_addr[id].virt_addr;
+                                       rp[i].len =
+                                               rxvq->p_used->ring[(used + ii) 
& queue_mask].len;
+                               }
+
+                               used += nbsegs;
+
+                       } else {
+                               ++used;
+                       }
+
+                       pkts++;
+                       segs = i + 1;
+               }
+
+               rxvq->used_idx = used;
+
+       } else if (rxvq->vq_type == PACKED_RING) {
+               /* This requires in-order behavior from FPGA */
+               int i;
+
+               for (i = 0; i < n; i++) {
+                       struct pvirtq_desc *desc = &rxvq->desc[rxvq->next_used];
+
+                       uint16_t flags = desc->flags;
+                       uint8_t avail = !!(flags & VIRTQ_DESC_F_AVAIL);
+                       uint8_t used = !!(flags & VIRTQ_DESC_F_USED);
+
+                       if (avail != rxvq->used_wrap_count || used != 
rxvq->used_wrap_count)
+                               break;
+
+                       rp[pkts].addr = 
rxvq->p_virtual_addr[desc->id].virt_addr;
+                       rp[pkts].len = desc->len;
+                       pkts++;
+
+                       inc_used(rxvq, 1);
+               }
+
+               segs = pkts;
+       }
+
+       *nb_pkts = pkts;
+       return segs;
+}
+
 /*
  * Put buffers back into Avail Ring
  */
@@ -1139,6 +1253,106 @@ static void nthw_release_rx_packets(struct 
nthw_virt_queue *rxvq, uint16_t n)
        }
 }
 
+static uint16_t nthw_get_tx_packets(struct nthw_virt_queue *txvq,
+       uint16_t n,
+       uint16_t *first_idx,
+       struct nthw_cvirtq_desc *cvq,
+       struct nthw_memory_descriptor **p_virt_addr)
+{
+       int m = 0;
+       uint16_t queue_mask =
+               (uint16_t)(txvq->queue_size - 1);       /* Valid because 
queue_size is always 2^n */
+       *p_virt_addr = txvq->p_virtual_addr;
+
+       if (txvq->vq_type == SPLIT_RING) {
+               cvq->s = txvq->p_desc;
+               cvq->vq_type = SPLIT_RING;
+
+               *first_idx = txvq->tx_descr_avail_idx;
+
+               uint16_t entries_used =
+                       (uint16_t)((txvq->tx_descr_avail_idx - 
txvq->cached_idx) & queue_mask);
+               uint16_t entries_ready = (uint16_t)(txvq->queue_size - 1 - 
entries_used);
+
+               vq_log_arg(txvq,
+                       "ask %i: descrAvail %i, cachedidx %i, used: %i, ready 
%i used->idx %i\n",
+                       n, txvq->tx_descr_avail_idx, txvq->cached_idx, 
entries_used, entries_ready,
+                       txvq->p_used->idx);
+
+               if (entries_ready < n) {
+                       /*
+                        * Look for more packets.
+                        * Using the used_idx in the avail ring since they are 
held synchronous
+                        * because of in-order
+                        */
+                       txvq->cached_idx =
+                               txvq->p_avail->ring[(txvq->p_used->idx - 1) & 
queue_mask];
+
+                       vq_log_arg(txvq, "Update: get cachedidx %i (used_idx-1 
%i)\n",
+                               txvq->cached_idx, (txvq->p_used->idx - 1) & 
queue_mask);
+                       entries_used =
+                               (uint16_t)((txvq->tx_descr_avail_idx - 
txvq->cached_idx)
+                               & queue_mask);
+                       entries_ready = (uint16_t)(txvq->queue_size - 1 - 
entries_used);
+                       vq_log_arg(txvq, "new used: %i, ready %i\n", 
entries_used, entries_ready);
+
+                       if (n > entries_ready)
+                               n = entries_ready;
+               }
+
+       } else if (txvq->vq_type == PACKED_RING) {
+               int i;
+
+               cvq->p = txvq->desc;
+               cvq->vq_type = PACKED_RING;
+
+               if (txvq->outs.num) {
+                       *first_idx = txvq->outs.next;
+                       uint16_t num = min(n, txvq->outs.num);
+                       txvq->outs.next = (txvq->outs.next + num) & queue_mask;
+                       txvq->outs.num -= num;
+
+                       if (n == num)
+                               return n;
+
+                       m = num;
+                       n -= num;
+
+               } else {
+                       *first_idx = txvq->next_used;
+               }
+
+               /* iterate the ring - this requires in-order behavior from FPGA 
*/
+               for (i = 0; i < n; i++) {
+                       struct pvirtq_desc *desc = &txvq->desc[txvq->next_used];
+
+                       uint16_t flags = desc->flags;
+                       uint8_t avail = !!(flags & VIRTQ_DESC_F_AVAIL);
+                       uint8_t used = !!(flags & VIRTQ_DESC_F_USED);
+
+                       if (avail != txvq->used_wrap_count || used != 
txvq->used_wrap_count) {
+                               n = i;
+                               break;
+                       }
+
+                       uint16_t incr = (desc->id - txvq->next_used) & 
queue_mask;
+                       i += incr;
+                       inc_used(txvq, incr + 1);
+               }
+
+               if (i > n) {
+                       int outs_num = i - n;
+                       txvq->outs.next = (txvq->next_used - outs_num) & 
queue_mask;
+                       txvq->outs.num = outs_num;
+               }
+
+       } else {
+               return 0;
+       }
+
+       return m + n;
+}
+
 static void nthw_release_tx_packets(struct nthw_virt_queue *txvq, uint16_t n, 
uint16_t n_segs[])
 {
        int i;
@@ -1200,7 +1414,9 @@ static struct sg_ops_s sg_ops = {
        .nthw_release_mngd_rx_virt_queue = nthw_release_mngd_rx_virt_queue,
        .nthw_setup_mngd_tx_virt_queue = nthw_setup_mngd_tx_virt_queue,
        .nthw_release_mngd_tx_virt_queue = nthw_release_mngd_tx_virt_queue,
+       .nthw_get_rx_packets = nthw_get_rx_packets,
        .nthw_release_rx_packets = nthw_release_rx_packets,
+       .nthw_get_tx_packets = nthw_get_tx_packets,
        .nthw_release_tx_packets = nthw_release_tx_packets,
        .nthw_virt_queue_init = nthw_virt_queue_init
 };
-- 
2.45.0

Reply via email to