From: Dawid Gorecki <d...@semihalf.com> ENA driver did not allow applications to call tx_cleanup. Freeing Tx mbufs was always done by the driver and it was not possible to manually request the driver to free mbufs.
Modify ena_tx_cleanup function to accept maximum number of packets to free and return number of packets that was freed. Signed-off-by: Dawid Gorecki <d...@semihalf.com> Reviewed-by: Michal Krawczyk <m...@semihalf.com> Reviewed-by: Shai Brandes <shaib...@amazon.com> --- doc/guides/nics/features/ena.ini | 1 + doc/guides/rel_notes/release_22_03.rst | 1 + drivers/net/ena/ena_ethdev.c | 27 ++++++++++++++++++-------- 3 files changed, 21 insertions(+), 8 deletions(-) diff --git a/doc/guides/nics/features/ena.ini b/doc/guides/nics/features/ena.ini index 55690aaf5a..59c1ae85fa 100644 --- a/doc/guides/nics/features/ena.ini +++ b/doc/guides/nics/features/ena.ini @@ -7,6 +7,7 @@ Link status = Y Link status event = Y Rx interrupt = Y +Free Tx mbuf on demand = Y MTU update = Y Scattered Rx = Y TSO = Y diff --git a/doc/guides/rel_notes/release_22_03.rst b/doc/guides/rel_notes/release_22_03.rst index 5c2521e870..95b31fc372 100644 --- a/doc/guides/rel_notes/release_22_03.rst +++ b/doc/guides/rel_notes/release_22_03.rst @@ -114,6 +114,7 @@ New Features * Added support for the link status configuration. * Added optimized memcpy support for the ARM platforms. * Added ENA admin queue support for the MP applications. + * Added free Tx mbuf on demand feature support. * **Updated Cisco enic driver.** diff --git a/drivers/net/ena/ena_ethdev.c b/drivers/net/ena/ena_ethdev.c index 73c1d63174..5e7b964995 100644 --- a/drivers/net/ena/ena_ethdev.c +++ b/drivers/net/ena/ena_ethdev.c @@ -170,7 +170,7 @@ static void ena_tx_map_mbuf(struct ena_ring *tx_ring, void **push_header, uint16_t *header_len); static int ena_xmit_mbuf(struct ena_ring *tx_ring, struct rte_mbuf *mbuf); -static void ena_tx_cleanup(struct ena_ring *tx_ring); +static int ena_tx_cleanup(void *txp, uint32_t free_pkt_cnt); static uint16_t eth_ena_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts); static uint16_t eth_ena_prep_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, @@ -276,6 +276,7 @@ static const struct eth_dev_ops ena_dev_ops = { .rx_queue_intr_disable = ena_rx_queue_intr_disable, .rss_hash_update = ena_rss_hash_update, .rss_hash_conf_get = ena_rss_hash_conf_get, + .tx_done_cleanup = ena_tx_cleanup, }; /********************************************************************* @@ -2990,16 +2991,22 @@ static int ena_xmit_mbuf(struct ena_ring *tx_ring, struct rte_mbuf *mbuf) return 0; } -static void ena_tx_cleanup(struct ena_ring *tx_ring) +static int ena_tx_cleanup(void *txp, uint32_t free_pkt_cnt) { + struct ena_ring *tx_ring = (struct ena_ring *)txp; unsigned int total_tx_descs = 0; + unsigned int total_tx_pkts = 0; uint16_t cleanup_budget; uint16_t next_to_clean = tx_ring->next_to_clean; - /* Attempt to release all Tx descriptors (ring_size - 1 -> size_mask) */ - cleanup_budget = tx_ring->size_mask; + /* + * If free_pkt_cnt is equal to 0, it means that the user requested + * full cleanup, so attempt to release all Tx descriptors + * (ring_size - 1 -> size_mask) + */ + cleanup_budget = (free_pkt_cnt == 0) ? tx_ring->size_mask : free_pkt_cnt; - while (likely(total_tx_descs < cleanup_budget)) { + while (likely(total_tx_pkts < cleanup_budget)) { struct rte_mbuf *mbuf; struct ena_tx_buffer *tx_info; uint16_t req_id; @@ -3021,6 +3028,7 @@ static void ena_tx_cleanup(struct ena_ring *tx_ring) tx_ring->empty_tx_reqs[next_to_clean] = req_id; total_tx_descs += tx_info->tx_descs; + total_tx_pkts++; /* Put back descriptor to the ring for reuse */ next_to_clean = ENA_IDX_NEXT_MASKED(next_to_clean, @@ -3034,8 +3042,11 @@ static void ena_tx_cleanup(struct ena_ring *tx_ring) ena_com_update_dev_comp_head(tx_ring->ena_com_io_cq); } - /* Notify completion handler that the cleanup was just called */ - tx_ring->last_cleanup_ticks = rte_get_timer_cycles(); + /* Notify completion handler that full cleanup was performed */ + if (free_pkt_cnt == 0 || total_tx_pkts < cleanup_budget) + tx_ring->last_cleanup_ticks = rte_get_timer_cycles(); + + return total_tx_pkts; } static uint16_t eth_ena_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, @@ -3056,7 +3067,7 @@ static uint16_t eth_ena_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, available_desc = ena_com_free_q_entries(tx_ring->ena_com_io_sq); if (available_desc < tx_ring->tx_free_thresh) - ena_tx_cleanup(tx_ring); + ena_tx_cleanup((void *)tx_ring, 0); for (sent_idx = 0; sent_idx < nb_pkts; sent_idx++) { if (ena_xmit_mbuf(tx_ring, tx_pkts[sent_idx])) -- 2.25.1