Remove CONFIG_RTE_LIBRTE_ICE_RX_ALLOW_BULK_ALLOC with below
consideration:

1. a default Rx path can always be selected by setting a proper
   rx_free_thresh value at runtime, see
   ice_check_rx_burst_bulk_alloc_preconditions.

2. its not a big deal to always reserve more space for desc ring.
   "ring_size = (uint16_t)(rxq->nb_rx_desc + ICE_RX_MAX_BURST);"

3. Fixes a potential invalid memory access in ice_reset_rx_queue.
   if CONFIG_RTE_LIBRTE_ICE_RX_ALLOW_BULK_ALLOC is turned on while
   ice_check_rx_burst_bulk_alloc_preconditions return fail.
   below code will have problem.

   for (i = 0; i < ICE_RX_MAX_BURST; ++i)
        rxq->sw_ring[rxq->nb_rx_desc + i].mbuf = &rxq->fake_mbuf;

Fixes: 50370662b727 ("net/ice: support device and queue ops")
Cc: sta...@dpdk.org

Signed-off-by: Qi Zhang <qi.z.zh...@intel.com>
---
 config/common_base         |  1 -
 drivers/net/ice/ice_rxtx.c | 64 ++++++++++------------------------------------
 2 files changed, 13 insertions(+), 52 deletions(-)

diff --git a/config/common_base b/config/common_base
index 7ca2f28b1..c31175f9d 100644
--- a/config/common_base
+++ b/config/common_base
@@ -337,7 +337,6 @@ CONFIG_RTE_LIBRTE_ICE_PMD=y
 CONFIG_RTE_LIBRTE_ICE_DEBUG_RX=n
 CONFIG_RTE_LIBRTE_ICE_DEBUG_TX=n
 CONFIG_RTE_LIBRTE_ICE_DEBUG_TX_FREE=n
-CONFIG_RTE_LIBRTE_ICE_RX_ALLOW_BULK_ALLOC=y
 CONFIG_RTE_LIBRTE_ICE_16BYTE_RX_DESC=n
 
 # Compile burst-oriented IAVF PMD driver
diff --git a/drivers/net/ice/ice_rxtx.c b/drivers/net/ice/ice_rxtx.c
index 60c411bfa..c7e5fc484 100644
--- a/drivers/net/ice/ice_rxtx.c
+++ b/drivers/net/ice/ice_rxtx.c
@@ -236,17 +236,15 @@ _ice_rx_queue_release_mbufs(struct ice_rx_queue *rxq)
                        rxq->sw_ring[i].mbuf = NULL;
                }
        }
-#ifdef RTE_LIBRTE_ICE_RX_ALLOW_BULK_ALLOC
-               if (rxq->rx_nb_avail == 0)
-                       return;
-               for (i = 0; i < rxq->rx_nb_avail; i++) {
-                       struct rte_mbuf *mbuf;
-
-                       mbuf = rxq->rx_stage[rxq->rx_next_avail + i];
-                       rte_pktmbuf_free_seg(mbuf);
-               }
-               rxq->rx_nb_avail = 0;
-#endif /* RTE_LIBRTE_ICE_RX_ALLOW_BULK_ALLOC */
+       if (rxq->rx_nb_avail == 0)
+               return;
+       for (i = 0; i < rxq->rx_nb_avail; i++) {
+               struct rte_mbuf *mbuf;
+
+               mbuf = rxq->rx_stage[rxq->rx_next_avail + i];
+               rte_pktmbuf_free_seg(mbuf);
+       }
+       rxq->rx_nb_avail = 0;
 }
 
 static void
@@ -309,16 +307,10 @@ ice_switch_rx_queue(struct ice_hw *hw, uint16_t q_idx, 
bool on)
 }
 
 static inline int
-#ifdef RTE_LIBRTE_ICE_RX_ALLOW_BULK_ALLOC
 ice_check_rx_burst_bulk_alloc_preconditions(struct ice_rx_queue *rxq)
-#else
-ice_check_rx_burst_bulk_alloc_preconditions
-       (__rte_unused struct ice_rx_queue *rxq)
-#endif
 {
        int ret = 0;
 
-#ifdef RTE_LIBRTE_ICE_RX_ALLOW_BULK_ALLOC
        if (!(rxq->rx_free_thresh >= ICE_RX_MAX_BURST)) {
                PMD_INIT_LOG(DEBUG, "Rx Burst Bulk Alloc Preconditions: "
                             "rxq->rx_free_thresh=%d, "
@@ -338,9 +330,6 @@ ice_check_rx_burst_bulk_alloc_preconditions
                             rxq->nb_rx_desc, rxq->rx_free_thresh);
                ret = -EINVAL;
        }
-#else
-       ret = -EINVAL;
-#endif
 
        return ret;
 }
@@ -357,17 +346,11 @@ ice_reset_rx_queue(struct ice_rx_queue *rxq)
                return;
        }
 
-#ifdef RTE_LIBRTE_ICE_RX_ALLOW_BULK_ALLOC
-       if (ice_check_rx_burst_bulk_alloc_preconditions(rxq) == 0)
-               len = (uint16_t)(rxq->nb_rx_desc + ICE_RX_MAX_BURST);
-       else
-#endif /* RTE_LIBRTE_ICE_RX_ALLOW_BULK_ALLOC */
-               len = rxq->nb_rx_desc;
+       len = (uint16_t)(rxq->nb_rx_desc + ICE_RX_MAX_BURST);
 
        for (i = 0; i < len * sizeof(union ice_rx_flex_desc); i++)
                ((volatile char *)rxq->rx_ring)[i] = 0;
 
-#ifdef RTE_LIBRTE_ICE_RX_ALLOW_BULK_ALLOC
        memset(&rxq->fake_mbuf, 0x0, sizeof(rxq->fake_mbuf));
        for (i = 0; i < ICE_RX_MAX_BURST; ++i)
                rxq->sw_ring[rxq->nb_rx_desc + i].mbuf = &rxq->fake_mbuf;
@@ -375,7 +358,6 @@ ice_reset_rx_queue(struct ice_rx_queue *rxq)
        rxq->rx_nb_avail = 0;
        rxq->rx_next_avail = 0;
        rxq->rx_free_trigger = (uint16_t)(rxq->rx_free_thresh - 1);
-#endif /* RTE_LIBRTE_ICE_RX_ALLOW_BULK_ALLOC */
 
        rxq->rx_tail = 0;
        rxq->nb_rx_hold = 0;
@@ -926,13 +908,11 @@ ice_rx_queue_setup(struct rte_eth_dev *dev,
        /* Allocate the maximun number of RX ring hardware descriptor. */
        len = ICE_MAX_RING_DESC;
 
-#ifdef RTE_LIBRTE_ICE_RX_ALLOW_BULK_ALLOC
        /**
         * Allocating a little more memory because vectorized/bulk_alloc Rx
         * functions doesn't check boundaries each time.
         */
        len += ICE_RX_MAX_BURST;
-#endif
 
        /* Allocate the maximum number of RX ring hardware descriptor. */
        ring_size = sizeof(union ice_rx_flex_desc) * len;
@@ -952,11 +932,8 @@ ice_rx_queue_setup(struct rte_eth_dev *dev,
        rxq->rx_ring_dma = rz->iova;
        rxq->rx_ring = rz->addr;
 
-#ifdef RTE_LIBRTE_ICE_RX_ALLOW_BULK_ALLOC
+       /* always reserve more for bulk alloc */
        len = (uint16_t)(nb_desc + ICE_RX_MAX_BURST);
-#else
-       len = nb_desc;
-#endif
 
        /* Allocate the software ring. */
        rxq->sw_ring = rte_zmalloc_socket(NULL,
@@ -977,17 +954,14 @@ ice_rx_queue_setup(struct rte_eth_dev *dev,
        use_def_burst_func = ice_check_rx_burst_bulk_alloc_preconditions(rxq);
 
        if (!use_def_burst_func) {
-#ifdef RTE_LIBRTE_ICE_RX_ALLOW_BULK_ALLOC
                PMD_INIT_LOG(DEBUG, "Rx Burst Bulk Alloc Preconditions are "
                             "satisfied. Rx Burst Bulk Alloc function will be "
                             "used on port=%d, queue=%d.",
                             rxq->port_id, rxq->queue_id);
-#endif /* RTE_LIBRTE_ICE_RX_ALLOW_BULK_ALLOC */
        } else {
                PMD_INIT_LOG(DEBUG, "Rx Burst Bulk Alloc Preconditions are "
-                            "not satisfied, Scattered Rx is requested, "
-                            "or RTE_LIBRTE_ICE_RX_ALLOW_BULK_ALLOC is "
-                            "not enabled on port=%d, queue=%d.",
+                            "not satisfied, Scattered Rx is requested. "
+                            "on port=%d, queue=%d.",
                             rxq->port_id, rxq->queue_id);
                ad->rx_bulk_alloc_allowed = false;
        }
@@ -1399,7 +1373,6 @@ ice_rxd_to_pkt_fields(struct rte_mbuf *mb,
 #endif
 }
 
-#ifdef RTE_LIBRTE_ICE_RX_ALLOW_BULK_ALLOC
 #define ICE_LOOK_AHEAD 8
 #if (ICE_LOOK_AHEAD != 8)
 #error "PMD ICE: ICE_LOOK_AHEAD must be 8\n"
@@ -1620,15 +1593,6 @@ ice_recv_pkts_bulk_alloc(void *rx_queue,
 
        return nb_rx;
 }
-#else
-static uint16_t
-ice_recv_pkts_bulk_alloc(void __rte_unused *rx_queue,
-                        struct rte_mbuf __rte_unused **rx_pkts,
-                        uint16_t __rte_unused nb_pkts)
-{
-       return 0;
-}
-#endif /* RTE_LIBRTE_ICE_RX_ALLOW_BULK_ALLOC */
 
 static uint16_t
 ice_recv_scattered_pkts(void *rx_queue,
@@ -1872,9 +1836,7 @@ ice_dev_supported_ptypes_get(struct rte_eth_dev *dev)
                ptypes = ptypes_os;
 
        if (dev->rx_pkt_burst == ice_recv_pkts ||
-#ifdef RTE_LIBRTE_ICE_RX_ALLOW_BULK_ALLOC
            dev->rx_pkt_burst == ice_recv_pkts_bulk_alloc ||
-#endif
            dev->rx_pkt_burst == ice_recv_scattered_pkts)
                return ptypes;
 
-- 
2.13.6

Reply via email to