From: Harish Patil
Add support for inner/outer L3/L4 TX csum offload for MPLS-in-UDP packets.
The driver checks for PKT_TX_TUNNEL_MPLSINUDP in Tx ol_flags and updates TX
BD elements with appropriate offsets/length of tunnel headers.
The pseudo csum calculation is removed from qede_xmit_prep_pkts() since its
not needed.
Note: Some lines are exceeding 80 columns and is intentionally
not fixed since it affects code readability.
Signed-off-by: Harish Patil
Signed-off-by: Rasesh Mody
---
drivers/net/qede/qede_rxtx.c | 197 +++---
drivers/net/qede/qede_rxtx.h |3 +-
2 files changed, 150 insertions(+), 50 deletions(-)
diff --git a/drivers/net/qede/qede_rxtx.c b/drivers/net/qede/qede_rxtx.c
index f65c833..9efb860 100644
--- a/drivers/net/qede/qede_rxtx.c
+++ b/drivers/net/qede/qede_rxtx.c
@@ -1445,7 +1445,6 @@ static inline uint32_t
qede_rx_cqe_to_tunn_pkt_type(uint16_t flags)
uint64_t ol_flags;
struct rte_mbuf *m;
uint16_t i;
- int ret;
for (i = 0; i < nb_pkts; i++) {
m = tx_pkts[i];
@@ -1478,14 +1477,6 @@ static inline uint32_t
qede_rx_cqe_to_tunn_pkt_type(uint16_t flags)
break;
}
#endif
- /* TBD: pseudo csum calcuation required if
-* ETH_TX_DATA_2ND_BD_L4_PSEUDO_CSUM_MODE not set?
-*/
- ret = rte_net_intel_cksum_prepare(m);
- if (ret != 0) {
- rte_errno = ret;
- break;
- }
}
#ifdef RTE_LIBRTE_QEDE_DEBUG_TX
@@ -1496,6 +1487,27 @@ static inline uint32_t
qede_rx_cqe_to_tunn_pkt_type(uint16_t flags)
return i;
}
+#define MPLSINUDP_HDR_SIZE (12)
+
+#ifdef RTE_LIBRTE_QEDE_DEBUG_TX
+static inline void
+qede_mpls_tunn_tx_sanity_check(struct rte_mbuf *mbuf,
+ struct qede_tx_queue *txq)
+{
+ if (((mbuf->outer_l2_len + mbuf->outer_l3_len) / 2) > 0xff)
+ PMD_TX_LOG(ERR, txq, "tunn_l4_hdr_start_offset overflow\n");
+ if (((mbuf->outer_l2_len + mbuf->outer_l3_len +
+ MPLSINUDP_HDR_SIZE) / 2) > 0xff)
+ PMD_TX_LOG(ERR, txq, "tunn_hdr_size overflow\n");
+ if (((mbuf->l2_len - MPLSINUDP_HDR_SIZE) / 2) >
+ ETH_TX_DATA_2ND_BD_TUNN_INNER_L2_HDR_SIZE_W_MASK)
+ PMD_TX_LOG(ERR, txq, "inner_l2_hdr_size overflow\n");
+ if (((mbuf->l2_len - MPLSINUDP_HDR_SIZE + mbuf->l3_len) / 2) >
+ ETH_TX_DATA_2ND_BD_L4_HDR_START_OFFSET_W_MASK)
+ PMD_TX_LOG(ERR, txq, "inner_l2_hdr_size overflow\n");
+}
+#endif
+
uint16_t
qede_xmit_pkts(void *p_txq, struct rte_mbuf **tx_pkts, uint16_t nb_pkts)
{
@@ -1510,9 +1522,10 @@ static inline uint32_t
qede_rx_cqe_to_tunn_pkt_type(uint16_t flags)
uint16_t nb_frags;
uint16_t nb_pkt_sent = 0;
uint8_t nbds;
- bool ipv6_ext_flg;
bool lso_flg;
+ bool mplsoudp_flg;
__rte_unused bool tunn_flg;
+ bool tunn_ipv6_ext_flg;
struct eth_tx_1st_bd *bd1;
struct eth_tx_2nd_bd *bd2;
struct eth_tx_3rd_bd *bd3;
@@ -1529,6 +1542,10 @@ static inline uint32_t
qede_rx_cqe_to_tunn_pkt_type(uint16_t flags)
uint16_t mss;
uint16_t bd3_bf;
+ uint8_t tunn_l4_hdr_start_offset;
+ uint8_t tunn_hdr_size;
+ uint8_t inner_l2_hdr_size;
+ uint16_t inner_l4_hdr_offset;
if (unlikely(txq->nb_tx_avail < txq->tx_free_thresh)) {
PMD_TX_LOG(DEBUG, txq, "send=%u avail=%u free_thresh=%u",
@@ -1540,7 +1557,6 @@ static inline uint32_t
qede_rx_cqe_to_tunn_pkt_type(uint16_t flags)
bd_prod = rte_cpu_to_le_16(ecore_chain_get_prod_idx(&txq->tx_pbl));
while (nb_tx_pkts--) {
/* Init flags/values */
- ipv6_ext_flg = false;
tunn_flg = false;
lso_flg = false;
nbds = 0;
@@ -1555,6 +1571,10 @@ static inline uint32_t
qede_rx_cqe_to_tunn_pkt_type(uint16_t flags)
bd2_bf2 = 0;
mss = 0;
bd3_bf = 0;
+ mplsoudp_flg = false;
+ tunn_ipv6_ext_flg = false;
+ tunn_hdr_size = 0;
+ tunn_l4_hdr_start_offset = 0;
mbuf = *tx_pkts++;
assert(mbuf);
@@ -1566,20 +1586,18 @@ static inline uint32_t
qede_rx_cqe_to_tunn_pkt_type(uint16_t flags)
tx_ol_flags = mbuf->ol_flags;
bd1_bd_flags_bf |= 1 << ETH_TX_1ST_BD_FLAGS_START_BD_SHIFT;
-#define RTE_ETH_IS_IPV6_HDR_EXT(ptype) ((ptype) & RTE_PTYPE_L3_IPV6_EXT)
- if (RTE_ETH_IS_IPV6_HDR_EXT(mbuf->packet_type)) {
- ipv6_ext_flg = true;
+ /* TX prepare would have already checked supported tunnel Tx
+* offloads. Don't rely on pkt_type marked by Rx, instead use
+* tx_ol_flags to decide.
+