Added TSO support for axgbe PMD. Initial Implementation for the TSO feature support Currently only headers transmitted to tester receiver side
Signed-off-by: Jesna K E <jesna....@amd.com> --- doc/guides/nics/features/axgbe.ini | 1 + drivers/net/axgbe/axgbe_common.h | 12 ++ drivers/net/axgbe/axgbe_dev.c | 13 ++ drivers/net/axgbe/axgbe_ethdev.c | 3 + drivers/net/axgbe/axgbe_ethdev.h | 1 + drivers/net/axgbe/axgbe_rxtx.c | 276 +++++++++++++++++------------ 6 files changed, 193 insertions(+), 113 deletions(-) diff --git a/doc/guides/nics/features/axgbe.ini b/doc/guides/nics/features/axgbe.ini index 5e2d6498e5..5c30c967bc 100644 --- a/doc/guides/nics/features/axgbe.ini +++ b/doc/guides/nics/features/axgbe.ini @@ -7,6 +7,7 @@ Speed capabilities = Y Link status = Y Scattered Rx = Y +TSO = Y Promiscuous mode = Y Allmulticast mode = Y RSS hash = Y diff --git a/drivers/net/axgbe/axgbe_common.h b/drivers/net/axgbe/axgbe_common.h index a5d11c5832..c30efe4c02 100644 --- a/drivers/net/axgbe/axgbe_common.h +++ b/drivers/net/axgbe/axgbe_common.h @@ -161,6 +161,10 @@ #define DMA_CH_CARBR_LO 0x5c #define DMA_CH_SR 0x60 +/* Setting MSS register entry bit positions and sizes for TSO */ +#define DMA_CH_CR_MSS_INDEX 0 +#define DMA_CH_CR_MSS_WIDTH 14 + /* DMA channel register entry bit positions and sizes */ #define DMA_CH_CR_PBLX8_INDEX 16 #define DMA_CH_CR_PBLX8_WIDTH 1 @@ -1232,6 +1236,14 @@ #define TX_CONTEXT_DESC3_VT_INDEX 0 #define TX_CONTEXT_DESC3_VT_WIDTH 16 +/* TSO related register entry bit positions and sizes*/ +#define TX_NORMAL_DESC3_TPL_INDEX 0 +#define TX_NORMAL_DESC3_TPL_WIDTH 18 +#define TX_NORMAL_DESC3_THL_INDEX 19 +#define TX_NORMAL_DESC3_THL_WIDTH 4 +#define TX_CONTEXT_DESC3_OSTC_INDEX 27 +#define TX_CONTEXT_DESC3_OSTC_WIDTH 1 + #define TX_NORMAL_DESC2_HL_B1L_INDEX 0 #define TX_NORMAL_DESC2_HL_B1L_WIDTH 14 #define TX_NORMAL_DESC2_IC_INDEX 31 diff --git a/drivers/net/axgbe/axgbe_dev.c b/drivers/net/axgbe/axgbe_dev.c index 6a7fddffca..eef453fab0 100644 --- a/drivers/net/axgbe/axgbe_dev.c +++ b/drivers/net/axgbe/axgbe_dev.c @@ -808,6 +808,18 @@ int axgbe_write_rss_lookup_table(struct axgbe_port *pdata) return 0; } +static void xgbe_config_tso_mode(struct axgbe_port *pdata) +{ + unsigned int i; + struct axgbe_tx_queue *txq; + + for (i = 0; i < pdata->eth_dev->data->nb_tx_queues; i++) { + txq = pdata->eth_dev->data->tx_queues[i]; + AXGMAC_DMA_IOWRITE_BITS(txq, DMA_CH_TCR, TSE, 1); + AXGMAC_DMA_IOWRITE_BITS(txq, DMA_CH_CR, MSS, 800); + } +} + static int axgbe_enable_rss(struct axgbe_port *pdata) { int ret; @@ -1314,6 +1326,7 @@ static int axgbe_init(struct axgbe_port *pdata) axgbe_config_rx_pbl_val(pdata); axgbe_config_rx_buffer_size(pdata); axgbe_config_rss(pdata); + xgbe_config_tso_mode(pdata); wrapper_tx_desc_init(pdata); ret = wrapper_rx_desc_init(pdata); if (ret) diff --git a/drivers/net/axgbe/axgbe_ethdev.c b/drivers/net/axgbe/axgbe_ethdev.c index 3717166384..0a4901aabc 100644 --- a/drivers/net/axgbe/axgbe_ethdev.c +++ b/drivers/net/axgbe/axgbe_ethdev.c @@ -12,6 +12,8 @@ #include "eal_filesystem.h" +#include <rte_vect.h> + #ifdef RTE_ARCH_X86 #include <cpuid.h> #else @@ -1237,6 +1239,7 @@ axgbe_dev_info_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info) RTE_ETH_TX_OFFLOAD_IPV4_CKSUM | RTE_ETH_TX_OFFLOAD_MULTI_SEGS | RTE_ETH_TX_OFFLOAD_UDP_CKSUM | + RTE_ETH_TX_OFFLOAD_TCP_TSO | RTE_ETH_TX_OFFLOAD_TCP_CKSUM; if (pdata->hw_feat.rss) { diff --git a/drivers/net/axgbe/axgbe_ethdev.h b/drivers/net/axgbe/axgbe_ethdev.h index 7f19321d88..31a583c2c6 100644 --- a/drivers/net/axgbe/axgbe_ethdev.h +++ b/drivers/net/axgbe/axgbe_ethdev.h @@ -583,6 +583,7 @@ struct axgbe_port { unsigned int tx_osp_mode; unsigned int tx_max_fifo_size; unsigned int multi_segs_tx; + unsigned int tso_tx; /* Rx settings */ unsigned int rx_sf_mode; diff --git a/drivers/net/axgbe/axgbe_rxtx.c b/drivers/net/axgbe/axgbe_rxtx.c index a9ff291cef..d7c97f3919 100644 --- a/drivers/net/axgbe/axgbe_rxtx.c +++ b/drivers/net/axgbe/axgbe_rxtx.c @@ -627,6 +627,9 @@ int axgbe_dev_tx_queue_setup(struct rte_eth_dev *dev, uint16_t queue_idx, RTE_ETH_TX_OFFLOAD_MULTI_SEGS)) pdata->multi_segs_tx = true; + if ((dev_data->dev_conf.txmode.offloads & + RTE_ETH_TX_OFFLOAD_TCP_TSO)) + pdata->tso_tx = true; return 0; } @@ -827,6 +830,7 @@ static int axgbe_xmit_hw(struct axgbe_tx_queue *txq, idx = AXGBE_GET_DESC_IDX(txq, txq->cur); desc = &txq->desc[idx]; + PMD_DRV_LOG(DEBUG, "tso:Inside axgbe_xmit_hw\n"); /* Update buffer address and length */ desc->baddr = rte_mbuf_data_iova(mbuf); @@ -873,7 +877,6 @@ static int axgbe_xmit_hw(struct axgbe_tx_queue *txq, AXGMAC_SET_BITS_LE(desc->desc3, TX_NORMAL_DESC3, OWN, 1); rte_wmb(); - /* Save mbuf */ txq->sw_ring[idx] = mbuf; /* Update current index*/ @@ -884,6 +887,7 @@ static int axgbe_xmit_hw(struct axgbe_tx_queue *txq, return 0; } + /* Tx Descriptor formation for segmented mbuf * Each mbuf will require multiple descriptors */ @@ -892,130 +896,175 @@ static int axgbe_xmit_hw_seg(struct axgbe_tx_queue *txq, struct rte_mbuf *mbuf) { - volatile struct axgbe_tx_desc *desc; - uint16_t idx; - uint64_t mask; - int start_index; - uint32_t pkt_len = 0; - int nb_desc_free; - struct rte_mbuf *tx_pkt; - - nb_desc_free = txq->nb_desc - (txq->cur - txq->dirty); - - if (mbuf->nb_segs > nb_desc_free) { - axgbe_xmit_cleanup_seg(txq); - nb_desc_free = txq->nb_desc - (txq->cur - txq->dirty); - if (unlikely(mbuf->nb_segs > nb_desc_free)) - return RTE_ETH_TX_DESC_UNAVAIL; - } - - idx = AXGBE_GET_DESC_IDX(txq, txq->cur); - desc = &txq->desc[idx]; - /* Saving the start index for setting the OWN bit finally */ - start_index = idx; - + volatile struct axgbe_tx_desc *desc; + uint16_t idx; + uint64_t mask; + int start_index; + uint32_t pkt_len = 0; + int nb_desc_free; + struct rte_mbuf *tx_pkt; + uint64_t l2_len = 0; + uint64_t l3_len = 0; + uint64_t l4_len = 0; + uint64_t tso_segsz = 0; + uint64_t total_hdr_len; + int tso = 0; + + /*Parameters required for tso*/ + l2_len = mbuf->l2_len; + l3_len = mbuf->l3_len; + l4_len = mbuf->l4_len; + tso_segsz = mbuf->tso_segsz; + total_hdr_len = l2_len + l3_len + l4_len; + + if (txq->pdata->tso_tx) + tso = 1; + else + tso = 0; + + PMD_DRV_LOG(DEBUG, "tso:l2_len = %ld,l3_len=%ld,l4_len=%ld,tso_segsz=%lu, + total_hdr_len=%lu\n", l2_len, l3_len, l4_len, tso_segsz, total_hdr_len); + + nb_desc_free = txq->nb_desc - (txq->cur - txq->dirty); + + PMD_DRV_LOG(DEBUG, "tso::Inside axgbe_xmit_hw_seg\n"); + if (mbuf->nb_segs > nb_desc_free) { + axgbe_xmit_cleanup_seg(txq); + nb_desc_free = txq->nb_desc - (txq->cur - txq->dirty); + if (unlikely(mbuf->nb_segs > nb_desc_free)) + return RTE_ETH_TX_DESC_UNAVAIL; + } + + idx = AXGBE_GET_DESC_IDX(txq, txq->cur); + desc = &txq->desc[idx]; + /* Saving the start index for setting the OWN bit finally */ + start_index = idx; tx_pkt = mbuf; - /* Max_pkt len = 9018 ; need to update it according to Jumbo pkt size */ - pkt_len = tx_pkt->pkt_len; + /* Max_pkt len = 9018 ; need to update it according to Jumbo pkt size */ + pkt_len = tx_pkt->pkt_len; - /* Update buffer address and length */ - desc->baddr = rte_mbuf_data_iova(tx_pkt); + /* Update buffer address and length */ + desc->baddr = rte_pktmbuf_iova_offset(mbuf,0); + /*For TSO first buffer contains the Header */ + if (tso) AXGMAC_SET_BITS_LE(desc->desc2, TX_NORMAL_DESC2, HL_B1L, - tx_pkt->data_len); - /* Total msg length to transmit */ - AXGMAC_SET_BITS_LE(desc->desc3, TX_NORMAL_DESC3, FL, - tx_pkt->pkt_len); - /* Timestamp enablement check */ - if (mbuf->ol_flags & RTE_MBUF_F_TX_IEEE1588_TMST) - AXGMAC_SET_BITS_LE(desc->desc2, TX_NORMAL_DESC2, TTSE, 1); + total_hdr_len); + else + AXGMAC_SET_BITS_LE(desc->desc2, TX_NORMAL_DESC2, HL_B1L, + tx_pkt->data_len); rte_wmb(); - /* Mark it as First Descriptor */ - AXGMAC_SET_BITS_LE(desc->desc3, TX_NORMAL_DESC3, FD, 1); - /* Mark it as a NORMAL descriptor */ - AXGMAC_SET_BITS_LE(desc->desc3, TX_NORMAL_DESC3, CTXT, 0); - /* configure h/w Offload */ - mask = mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK; - if (mask == RTE_MBUF_F_TX_TCP_CKSUM || mask == RTE_MBUF_F_TX_UDP_CKSUM) - AXGMAC_SET_BITS_LE(desc->desc3, TX_NORMAL_DESC3, CIC, 0x3); - else if (mbuf->ol_flags & RTE_MBUF_F_TX_IP_CKSUM) - AXGMAC_SET_BITS_LE(desc->desc3, TX_NORMAL_DESC3, CIC, 0x1); - rte_wmb(); - if (mbuf->ol_flags & (RTE_MBUF_F_TX_VLAN | RTE_MBUF_F_TX_QINQ)) { - /* Mark it as a CONTEXT descriptor */ - AXGMAC_SET_BITS_LE(desc->desc3, TX_CONTEXT_DESC3, - CTXT, 1); - /* Set the VLAN tag */ - AXGMAC_SET_BITS_LE(desc->desc3, TX_CONTEXT_DESC3, - VT, mbuf->vlan_tci); - /* Indicate this descriptor contains the VLAN tag */ - AXGMAC_SET_BITS_LE(desc->desc3, TX_CONTEXT_DESC3, - VLTV, 1); - AXGMAC_SET_BITS_LE(desc->desc2, TX_NORMAL_DESC2, VTIR, - TX_NORMAL_DESC2_VLAN_INSERT); - } else { - AXGMAC_SET_BITS_LE(desc->desc2, TX_NORMAL_DESC2, VTIR, 0x0); - } - rte_wmb(); + /* Timestamp enablement check */ + if (mbuf->ol_flags & RTE_MBUF_F_TX_IEEE1588_TMST) + AXGMAC_SET_BITS_LE(desc->desc2, TX_NORMAL_DESC2, TTSE, 1); + + rte_wmb(); + /* Mark it as First Descriptor */ + AXGMAC_SET_BITS_LE(desc->desc3, TX_NORMAL_DESC3, FD, 1); + /* Mark it as a NORMAL descriptor */ + AXGMAC_SET_BITS_LE(desc->desc3, TX_NORMAL_DESC3, CTXT, 0); + /* configure h/w Offload */ + mask = mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK; + if (mask == RTE_MBUF_F_TX_TCP_CKSUM || mask == RTE_MBUF_F_TX_UDP_CKSUM) + AXGMAC_SET_BITS_LE(desc->desc3, TX_NORMAL_DESC3, CIC, 0x3); + else if (mbuf->ol_flags & RTE_MBUF_F_TX_IP_CKSUM) + AXGMAC_SET_BITS_LE(desc->desc3, TX_NORMAL_DESC3, CIC, 0x1); + rte_wmb(); + + if (mbuf->ol_flags & (RTE_MBUF_F_TX_VLAN | RTE_MBUF_F_TX_QINQ)) { + /* Mark it as a CONTEXT descriptor */ + AXGMAC_SET_BITS_LE(desc->desc3, TX_CONTEXT_DESC3, + CTXT, 1); + /* Set the VLAN tag */ + AXGMAC_SET_BITS_LE(desc->desc3, TX_CONTEXT_DESC3, + VT, mbuf->vlan_tci); + /* Indicate this descriptor contains the VLAN tag */ + AXGMAC_SET_BITS_LE(desc->desc3, TX_CONTEXT_DESC3, + VLTV, 1); + AXGMAC_SET_BITS_LE(desc->desc2, TX_NORMAL_DESC2, VTIR, + TX_NORMAL_DESC2_VLAN_INSERT); + } else { + AXGMAC_SET_BITS_LE(desc->desc2, TX_NORMAL_DESC2, VTIR, 0x0); + } + rte_wmb(); + + /*Register settings for TSO*/ + if (tso) { + PMD_DRV_LOG(DEBUG, "tso : Inside TSO register settings\n"); + /* Enable TSO */ + AXGMAC_SET_BITS_LE(desc->desc3, TX_NORMAL_DESC3, TSE,1); + AXGMAC_SET_BITS_LE(desc->desc3, TX_NORMAL_DESC3, TPL, + ((mbuf->pkt_len)-total_hdr_len)); + AXGMAC_SET_BITS_LE(desc->desc3, TX_NORMAL_DESC3, THL, + l4_len); + } else { + /* Enable CRC and Pad Insertion */ + AXGMAC_SET_BITS_LE(desc->desc3, TX_NORMAL_DESC3, CPC, 0); + /* Total msg length to transmit */ + AXGMAC_SET_BITS_LE(desc->desc3, TX_NORMAL_DESC3, FL, + mbuf->pkt_len); + } /* Save mbuf */ - txq->sw_ring[idx] = tx_pkt; - /* Update current index*/ - txq->cur++; + txq->sw_ring[idx] = tx_pkt; + /* Update current index*/ + txq->cur++; tx_pkt = tx_pkt->next; while (tx_pkt != NULL) { - idx = AXGBE_GET_DESC_IDX(txq, txq->cur); - desc = &txq->desc[idx]; - - /* Update buffer address and length */ - desc->baddr = rte_mbuf_data_iova(tx_pkt); - - AXGMAC_SET_BITS_LE(desc->desc2, - TX_NORMAL_DESC2, HL_B1L, tx_pkt->data_len); - - rte_wmb(); - - /* Mark it as a NORMAL descriptor */ - AXGMAC_SET_BITS_LE(desc->desc3, TX_NORMAL_DESC3, CTXT, 0); - /* configure h/w Offload */ - mask = mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK; - if (mask == RTE_MBUF_F_TX_TCP_CKSUM || - mask == RTE_MBUF_F_TX_UDP_CKSUM) - AXGMAC_SET_BITS_LE(desc->desc3, - TX_NORMAL_DESC3, CIC, 0x3); - else if (mbuf->ol_flags & RTE_MBUF_F_TX_IP_CKSUM) - AXGMAC_SET_BITS_LE(desc->desc3, - TX_NORMAL_DESC3, CIC, 0x1); - - rte_wmb(); - - /* Set OWN bit */ - AXGMAC_SET_BITS_LE(desc->desc3, TX_NORMAL_DESC3, OWN, 1); - rte_wmb(); - - /* Save mbuf */ - txq->sw_ring[idx] = tx_pkt; - /* Update current index*/ - txq->cur++; - - tx_pkt = tx_pkt->next; - } - - /* Set LD bit for the last descriptor */ - AXGMAC_SET_BITS_LE(desc->desc3, TX_NORMAL_DESC3, LD, 1); - rte_wmb(); - - /* Update stats */ - txq->bytes += pkt_len; - - /* Set OWN bit for the first descriptor */ - desc = &txq->desc[start_index]; - AXGMAC_SET_BITS_LE(desc->desc3, TX_NORMAL_DESC3, OWN, 1); - rte_wmb(); - + idx = AXGBE_GET_DESC_IDX(txq, txq->cur); + desc = &txq->desc[idx]; + + if (tso) + desc->baddr = rte_pktmbuf_iova_offset(mbuf,total_hdr_len); + else + /* Update buffer address and length */ + desc->baddr = rte_mbuf_data_iova(tx_pkt); + + AXGMAC_SET_BITS_LE(desc->desc2, + TX_NORMAL_DESC2, HL_B1L, tx_pkt->data_len); + + rte_wmb(); + + /* Mark it as a NORMAL descriptor */ + AXGMAC_SET_BITS_LE(desc->desc3, TX_NORMAL_DESC3, CTXT, 0); + /* configure h/w Offload */ + mask = mbuf->ol_flags & RTE_MBUF_F_TX_L4_MASK; + if (mask == RTE_MBUF_F_TX_TCP_CKSUM || + mask == RTE_MBUF_F_TX_UDP_CKSUM) + AXGMAC_SET_BITS_LE(desc->desc3, + TX_NORMAL_DESC3, CIC, 0x3); + else if (mbuf->ol_flags & RTE_MBUF_F_TX_IP_CKSUM) + AXGMAC_SET_BITS_LE(desc->desc3, + TX_NORMAL_DESC3, CIC, 0x1); + + rte_wmb(); + + /* Set OWN bit */ + AXGMAC_SET_BITS_LE(desc->desc3, TX_NORMAL_DESC3, OWN, 1); + rte_wmb(); + + /* Save mbuf */ + txq->sw_ring[idx] = tx_pkt; + /* Update current index*/ + txq->cur++; + + tx_pkt = tx_pkt->next; + } + + /* Set LD bit for the last descriptor */ + AXGMAC_SET_BITS_LE(desc->desc3, TX_NORMAL_DESC3, LD, 1); + rte_wmb(); + + /* Update stats */ + txq->bytes += pkt_len; + + /* Set OWN bit for the first descriptor */ + desc = &txq->desc[start_index]; + AXGMAC_SET_BITS_LE(desc->desc3, TX_NORMAL_DESC3, OWN, 1); + rte_wmb(); return 0; } @@ -1061,6 +1110,7 @@ axgbe_xmit_pkts_seg(void *tx_queue, struct rte_mbuf **tx_pkts, idx * sizeof(struct axgbe_tx_desc)); /* Update tail reg with next immediate address to kick Tx DMA channel*/ AXGMAC_DMA_IOWRITE(txq, DMA_CH_TDTR_LO, tail_addr); + txq->pkts += nb_pkt_sent; return nb_pkt_sent; } -- 2.34.1