Add support for eth_rx_descriptor_dump for cn9k and cn10k.
This patch dumps contents of receviced packet descriptor from CQ
for debug to file

Signed-off-by: Rakesh Kudurumalla <rkuduruma...@marvell.com>
---
 drivers/net/cnxk/cn10k_ethdev.c | 67 +++++++++++++++++++++++++++++++++
 drivers/net/cnxk/cn9k_ethdev.c  | 53 ++++++++++++++++++++++++++
 2 files changed, 120 insertions(+)

diff --git a/drivers/net/cnxk/cn10k_ethdev.c b/drivers/net/cnxk/cn10k_ethdev.c
index 4a4e97287c..7eee9b1da8 100644
--- a/drivers/net/cnxk/cn10k_ethdev.c
+++ b/drivers/net/cnxk/cn10k_ethdev.c
@@ -656,6 +656,72 @@ cn10k_nix_reassembly_conf_set(struct rte_eth_dev *eth_dev,
        return rc;
 }
 
+static int
+cn10k_nix_rx_avail_get(struct cn10k_eth_rxq *rxq)
+{
+       uint32_t qmask = rxq->qmask;
+       uint64_t reg, head, tail;
+       int available;
+
+       /* Use LDADDA version to avoid reorder */
+       reg = roc_atomic64_add_sync(rxq->wdata, rxq->cq_status);
+       /* CQ_OP_STATUS operation error */
+       if (reg & BIT_ULL(NIX_CQ_OP_STAT_OP_ERR) ||
+           reg & BIT_ULL(NIX_CQ_OP_STAT_CQ_ERR))
+               return 0;
+       tail = reg & 0xFFFFF;
+       head = (reg >> 20) & 0xFFFFF;
+       if (tail < head)
+               available = tail - head + qmask + 1;
+       else
+               available = tail - head;
+
+       return available;
+}
+
+static int
+cn10k_rx_descriptor_dump(const struct rte_eth_dev *eth_dev, uint16_t qid,
+                        uint16_t offset, uint16_t num, FILE *file)
+{
+       struct cn10k_eth_rxq *rxq = eth_dev->data->rx_queues[qid];
+       const uint64_t data_off = rxq->data_off;
+       const uint32_t qmask = rxq->qmask;
+       const uintptr_t desc = rxq->desc;
+       struct cpt_parse_hdr_s *cpth;
+       uint32_t head = rxq->head;
+       struct nix_cqe_hdr_s *cq;
+       uint16_t count = 0;
+       int availble_pkts;
+       uint64_t cq_w1;
+
+       availble_pkts = cn10k_nix_rx_avail_get(rxq);
+
+       if ((offset + num - 1) >= availble_pkts) {
+               plt_err("Invalid BD num=%u\n", num);
+               return -EINVAL;
+       }
+
+       while (count < num) {
+               cq = (struct nix_cqe_hdr_s *)(desc + CQE_SZ(head) +
+                                             count + offset);
+               cq_w1 = *((const uint64_t *)cq + 1);
+               if (cq_w1 & BIT(11)) {
+                       rte_iova_t buff = *((rte_iova_t *)((uint64_t *)cq + 9));
+                       struct rte_mbuf *mbuf =
+                               (struct rte_mbuf *)(buff - data_off);
+                       cpth = (struct cpt_parse_hdr_s *)
+                               ((uintptr_t)mbuf + (uint16_t)data_off);
+                       roc_cpt_parse_hdr_dump(file, cpth);
+               } else {
+                       roc_nix_cqe_dump(file, cq);
+               }
+
+               count++;
+               head &= qmask;
+       }
+       return 0;
+}
+
 static int
 cn10k_nix_tm_mark_vlan_dei(struct rte_eth_dev *eth_dev, int mark_green,
                           int mark_yellow, int mark_red,
@@ -794,6 +860,7 @@ nix_eth_dev_ops_override(void)
                        cn10k_nix_reassembly_capability_get;
        cnxk_eth_dev_ops.ip_reassembly_conf_get = cn10k_nix_reassembly_conf_get;
        cnxk_eth_dev_ops.ip_reassembly_conf_set = cn10k_nix_reassembly_conf_set;
+       cnxk_eth_dev_ops.eth_rx_descriptor_dump = cn10k_rx_descriptor_dump;
 }
 
 /* Update platform specific tm ops */
diff --git a/drivers/net/cnxk/cn9k_ethdev.c b/drivers/net/cnxk/cn9k_ethdev.c
index bae4dda5e2..e88631a02e 100644
--- a/drivers/net/cnxk/cn9k_ethdev.c
+++ b/drivers/net/cnxk/cn9k_ethdev.c
@@ -664,6 +664,58 @@ cn9k_nix_tm_mark_ip_dscp(struct rte_eth_dev *eth_dev, int 
mark_green,
        return rc;
 }
 
+static int
+cn9k_nix_rx_avail_get(struct cn9k_eth_rxq *rxq)
+{
+       uint32_t qmask = rxq->qmask;
+       uint64_t reg, head, tail;
+       int available;
+
+       /* Use LDADDA version to avoid reorder */
+       reg = roc_atomic64_add_sync(rxq->wdata, rxq->cq_status);
+       /* CQ_OP_STATUS operation error */
+       if (reg & BIT_ULL(NIX_CQ_OP_STAT_OP_ERR) ||
+           reg & BIT_ULL(NIX_CQ_OP_STAT_CQ_ERR))
+               return 0;
+       tail = reg & 0xFFFFF;
+       head = (reg >> 20) & 0xFFFFF;
+       if (tail < head)
+               available = tail - head + qmask + 1;
+       else
+               available = tail - head;
+
+       return available;
+}
+
+static int
+cn9k_rx_descriptor_dump(const struct rte_eth_dev *eth_dev, uint16_t qid,
+                        uint16_t offset, uint16_t num, FILE *file)
+{
+       struct cn9k_eth_rxq *rxq = eth_dev->data->rx_queues[qid];
+       const uint32_t qmask = rxq->qmask;
+       const uintptr_t desc = rxq->desc;
+       uint32_t head = rxq->head;
+       struct nix_cqe_hdr_s *cq;
+       uint16_t count = 0;
+       int availble_pkts;
+
+       availble_pkts = cn9k_nix_rx_avail_get(rxq);
+
+       if ((offset + num - 1) >= availble_pkts) {
+               plt_err("Invalid BD num=%u\n", num);
+               return -EINVAL;
+       }
+
+       while (count < num) {
+               cq = (struct nix_cqe_hdr_s *)(desc + CQE_SZ(head) +
+                                             count + offset);
+               roc_nix_cqe_dump(file, cq);
+               count++;
+               head &= qmask;
+       }
+       return 0;
+}
+
 /* Update platform specific eth dev ops */
 static void
 nix_eth_dev_ops_override(void)
@@ -687,6 +739,7 @@ nix_eth_dev_ops_override(void)
        cnxk_eth_dev_ops.mtr_ops_get = NULL;
        cnxk_eth_dev_ops.timesync_read_tx_timestamp =
                cn9k_nix_timesync_read_tx_timestamp;
+       cnxk_eth_dev_ops.eth_rx_descriptor_dump = cn9k_rx_descriptor_dump;
 }
 
 /* Update platform specific eth dev ops */
-- 
2.25.1

Reply via email to