> From: Olivier Matz [mailto:olivier.matz at 6wind.com] > Sent: Wednesday, November 26, 2014 3:05 PM > To: dev at dpdk.org > Cc: olivier.matz at 6wind.com; Walukiewicz, Miroslaw; Liu, Jijiang; Liu, > Yong; jigsaw at gmail.com; Richardson, Bruce; Ananyev, Konstantin > Subject: [PATCH v4 12/13] testpmd: support TSO in csum forward engine > > Add two new commands in testpmd: > > - tso set <segsize> <portid> > - tso show <portid> > > These commands can be used enable TSO when transmitting TCP packets in > the csum forward engine. Ex: > > set fwd csum > tx_checksum set ip hw 0 > tso set 800 0 > start > > Signed-off-by: Olivier Matz <olivier.matz at 6wind.com>
Acked-by: Konstantin Ananyev <konstantin.ananyev at intel.com> > --- > app/test-pmd/cmdline.c | 92 > +++++++++++++++++++++++++++++++++++++++++++++++++ > app/test-pmd/csumonly.c | 64 ++++++++++++++++++++++++---------- > app/test-pmd/testpmd.h | 1 + > 3 files changed, 139 insertions(+), 18 deletions(-) > > diff --git a/app/test-pmd/cmdline.c b/app/test-pmd/cmdline.c > index 722cd76..2a8c260 100644 > --- a/app/test-pmd/cmdline.c > +++ b/app/test-pmd/cmdline.c > @@ -329,6 +329,14 @@ static void cmd_help_long_parsed(void *parsed_result, > "tx_checksum show (port_id)\n" > " Display tx checksum offload configuration\n\n" > > + "tso set (segsize) (portid)\n" > + " Enable TCP Segmentation Offload in csum forward" > + " engine.\n" > + " Please check the NIC datasheet for HW limits.\n\n" > + > + "tso show (portid)" > + " Display the status of TCP Segmentation > Offload.\n\n" > + > "set fwd (%s)\n" > " Set packet forwarding mode.\n\n" > > @@ -2984,6 +2992,88 @@ cmdline_parse_inst_t cmd_tx_cksum_show = { > }, > }; > > +/* *** ENABLE HARDWARE SEGMENTATION IN TX PACKETS *** */ > +struct cmd_tso_set_result { > + cmdline_fixed_string_t tso; > + cmdline_fixed_string_t mode; > + uint16_t tso_segsz; > + uint8_t port_id; > +}; > + > +static void > +cmd_tso_set_parsed(void *parsed_result, > + __attribute__((unused)) struct cmdline *cl, > + __attribute__((unused)) void *data) > +{ > + struct cmd_tso_set_result *res = parsed_result; > + struct rte_eth_dev_info dev_info; > + > + if (port_id_is_invalid(res->port_id)) > + return; > + > + if (!strcmp(res->mode, "set")) > + ports[res->port_id].tso_segsz = res->tso_segsz; > + > + if (ports[res->port_id].tso_segsz == 0) > + printf("TSO is disabled\n"); > + else > + printf("TSO segment size is %d\n", > + ports[res->port_id].tso_segsz); > + > + /* display warnings if configuration is not supported by the NIC */ > + rte_eth_dev_info_get(res->port_id, &dev_info); > + if ((ports[res->port_id].tso_segsz != 0) && > + (dev_info.tx_offload_capa & DEV_TX_OFFLOAD_TCP_TSO) == 0) { > + printf("Warning: TSO enabled but not " > + "supported by port %d\n", res->port_id); > + } > +} > + > +cmdline_parse_token_string_t cmd_tso_set_tso = > + TOKEN_STRING_INITIALIZER(struct cmd_tso_set_result, > + tso, "tso"); > +cmdline_parse_token_string_t cmd_tso_set_mode = > + TOKEN_STRING_INITIALIZER(struct cmd_tso_set_result, > + mode, "set"); > +cmdline_parse_token_num_t cmd_tso_set_tso_segsz = > + TOKEN_NUM_INITIALIZER(struct cmd_tso_set_result, > + tso_segsz, UINT16); > +cmdline_parse_token_num_t cmd_tso_set_portid = > + TOKEN_NUM_INITIALIZER(struct cmd_tso_set_result, > + port_id, UINT8); > + > +cmdline_parse_inst_t cmd_tso_set = { > + .f = cmd_tso_set_parsed, > + .data = NULL, > + .help_str = "Set TSO segment size for csum engine (0 to disable): " > + "tso set <tso_segsz> <port>", > + .tokens = { > + (void *)&cmd_tso_set_tso, > + (void *)&cmd_tso_set_mode, > + (void *)&cmd_tso_set_tso_segsz, > + (void *)&cmd_tso_set_portid, > + NULL, > + }, > +}; > + > +cmdline_parse_token_string_t cmd_tso_show_mode = > + TOKEN_STRING_INITIALIZER(struct cmd_tso_set_result, > + mode, "show"); > + > + > +cmdline_parse_inst_t cmd_tso_show = { > + .f = cmd_tso_set_parsed, > + .data = NULL, > + .help_str = "Show TSO segment size for csum engine: " > + "tso show <port>", > + .tokens = { > + (void *)&cmd_tso_set_tso, > + (void *)&cmd_tso_show_mode, > + (void *)&cmd_tso_set_portid, > + NULL, > + }, > +}; > + > /* *** ENABLE/DISABLE FLUSH ON RX STREAMS *** */ > struct cmd_set_flush_rx { > cmdline_fixed_string_t set; > @@ -8660,6 +8750,8 @@ cmdline_parse_ctx_t main_ctx[] = { > (cmdline_parse_inst_t *)&cmd_tx_vlan_set_pvid, > (cmdline_parse_inst_t *)&cmd_tx_cksum_set, > (cmdline_parse_inst_t *)&cmd_tx_cksum_show, > + (cmdline_parse_inst_t *)&cmd_tso_set, > + (cmdline_parse_inst_t *)&cmd_tso_show, > (cmdline_parse_inst_t *)&cmd_link_flow_control_set, > (cmdline_parse_inst_t *)&cmd_link_flow_control_set_rx, > (cmdline_parse_inst_t *)&cmd_link_flow_control_set_tx, > diff --git a/app/test-pmd/csumonly.c b/app/test-pmd/csumonly.c > index 9a5408e..490342f 100644 > --- a/app/test-pmd/csumonly.c > +++ b/app/test-pmd/csumonly.c > @@ -88,12 +88,12 @@ > #endif > > static uint16_t > -get_psd_sum(void *l3_hdr, uint16_t ethertype) > +get_psd_sum(void *l3_hdr, uint16_t ethertype, uint64_t ol_flags) > { > if (ethertype == _htons(ETHER_TYPE_IPv4)) > - return rte_ipv4_phdr_cksum(l3_hdr); > + return rte_ipv4_phdr_cksum(l3_hdr, ol_flags); > else /* assume ethertype == ETHER_TYPE_IPv6 */ > - return rte_ipv6_phdr_cksum(l3_hdr); > + return rte_ipv6_phdr_cksum(l3_hdr, ol_flags); > } > > static uint16_t > @@ -108,14 +108,15 @@ get_udptcp_checksum(void *l3_hdr, void *l4_hdr, > uint16_t ethertype) > /* > * Parse an ethernet header to fill the ethertype, l2_len, l3_len and > * ipproto. This function is able to recognize IPv4/IPv6 with one optional > vlan > - * header. > + * header. The l4_len argument is only set in case of TCP (useful for TSO). > */ > static void > parse_ethernet(struct ether_hdr *eth_hdr, uint16_t *ethertype, uint16_t > *l2_len, > - uint16_t *l3_len, uint8_t *l4_proto) > + uint16_t *l3_len, uint8_t *l4_proto, uint16_t *l4_len) > { > struct ipv4_hdr *ipv4_hdr; > struct ipv6_hdr *ipv6_hdr; > + struct tcp_hdr *tcp_hdr; > > *l2_len = sizeof(struct ether_hdr); > *ethertype = eth_hdr->ether_type; > @@ -143,6 +144,14 @@ parse_ethernet(struct ether_hdr *eth_hdr, uint16_t > *ethertype, uint16_t *l2_len, > *l4_proto = 0; > break; > } > + > + if (*l4_proto == IPPROTO_TCP) { > + tcp_hdr = (struct tcp_hdr *)((char *)eth_hdr + > + *l2_len + *l3_len); > + *l4_len = (tcp_hdr->data_off & 0xf0) >> 2; > + } > + else > + *l4_len = 0; > } > > /* modify the IPv4 or IPv4 source address of a packet */ > @@ -165,7 +174,7 @@ change_ip_addresses(void *l3_hdr, uint16_t ethertype) > * depending on the testpmd command line configuration */ > static uint64_t > process_inner_cksums(void *l3_hdr, uint16_t ethertype, uint16_t l3_len, > - uint8_t l4_proto, uint16_t testpmd_ol_flags) > + uint8_t l4_proto, uint16_t tso_segsz, uint16_t testpmd_ol_flags) > { > struct ipv4_hdr *ipv4_hdr = l3_hdr; > struct udp_hdr *udp_hdr; > @@ -177,11 +186,16 @@ process_inner_cksums(void *l3_hdr, uint16_t ethertype, > uint16_t l3_len, > ipv4_hdr = l3_hdr; > ipv4_hdr->hdr_checksum = 0; > > - if (testpmd_ol_flags & TESTPMD_TX_OFFLOAD_IP_CKSUM) > + if (tso_segsz != 0 && l4_proto == IPPROTO_TCP) { > ol_flags |= PKT_TX_IP_CKSUM; > - else > - ipv4_hdr->hdr_checksum = rte_ipv4_cksum(ipv4_hdr); > - > + } > + else { > + if (testpmd_ol_flags & TESTPMD_TX_OFFLOAD_IP_CKSUM) > + ol_flags |= PKT_TX_IP_CKSUM; > + else > + ipv4_hdr->hdr_checksum = > + rte_ipv4_cksum(ipv4_hdr); > + } > ol_flags |= PKT_TX_IPV4; > } > else if (ethertype == _htons(ETHER_TYPE_IPv6)) > @@ -197,7 +211,7 @@ process_inner_cksums(void *l3_hdr, uint16_t ethertype, > uint16_t l3_len, > if (testpmd_ol_flags & TESTPMD_TX_OFFLOAD_UDP_CKSUM) { > ol_flags |= PKT_TX_UDP_CKSUM; > udp_hdr->dgram_cksum = get_psd_sum(l3_hdr, > - ethertype); > + ethertype, ol_flags); > } > else { > udp_hdr->dgram_cksum = > @@ -209,9 +223,13 @@ process_inner_cksums(void *l3_hdr, uint16_t ethertype, > uint16_t l3_len, > else if (l4_proto == IPPROTO_TCP) { > tcp_hdr = (struct tcp_hdr *)((char *)l3_hdr + l3_len); > tcp_hdr->cksum = 0; > - if (testpmd_ol_flags & TESTPMD_TX_OFFLOAD_TCP_CKSUM) { > + if (tso_segsz != 0) { > + ol_flags |= PKT_TX_TCP_SEG; > + tcp_hdr->cksum = get_psd_sum(l3_hdr, ethertype, > ol_flags); > + } > + else if (testpmd_ol_flags & TESTPMD_TX_OFFLOAD_TCP_CKSUM) { > ol_flags |= PKT_TX_TCP_CKSUM; > - tcp_hdr->cksum = get_psd_sum(l3_hdr, ethertype); > + tcp_hdr->cksum = get_psd_sum(l3_hdr, ethertype, > ol_flags); > } > else { > tcp_hdr->cksum = > @@ -282,6 +300,8 @@ process_outer_cksums(void *outer_l3_hdr, uint16_t > outer_ethertype, > * - modify the IPs in inner headers and in outer headers if any > * - reprocess the checksum of all supported layers. This is done in SW > * or HW, depending on testpmd command line configuration > + * - if TSO is enabled in testpmd command line, also flag the mbuf for TCP > + * segmentation offload (this implies HW TCP checksum) > * Then transmit packets on the output port. > * > * (1) Supported packets are: > @@ -312,7 +332,9 @@ pkt_burst_checksum_forward(struct fwd_stream *fs) > uint16_t testpmd_ol_flags; > uint8_t l4_proto; > uint16_t ethertype = 0, outer_ethertype = 0; > - uint16_t l2_len = 0, l3_len = 0, outer_l2_len = 0, outer_l3_len = 0; > + uint16_t l2_len = 0, l3_len = 0, l4_len = 0; > + uint16_t outer_l2_len = 0, outer_l3_len = 0; > + uint16_t tso_segsz; > int tunnel = 0; > uint32_t rx_bad_ip_csum; > uint32_t rx_bad_l4_csum; > @@ -342,6 +364,7 @@ pkt_burst_checksum_forward(struct fwd_stream *fs) > > txp = &ports[fs->tx_port]; > testpmd_ol_flags = txp->tx_ol_flags; > + tso_segsz = txp->tso_segsz; > > for (i = 0; i < nb_rx; i++) { > > @@ -357,7 +380,8 @@ pkt_burst_checksum_forward(struct fwd_stream *fs) > * and inner headers */ > > eth_hdr = rte_pktmbuf_mtod(m, struct ether_hdr *); > - parse_ethernet(eth_hdr, ðertype, &l2_len, &l3_len, > &l4_proto); > + parse_ethernet(eth_hdr, ðertype, &l2_len, &l3_len, > + &l4_proto, &l4_len); > l3_hdr = (char *)eth_hdr + l2_len; > > /* check if it's a supported tunnel (only vxlan for now) */ > @@ -385,7 +409,7 @@ pkt_burst_checksum_forward(struct fwd_stream *fs) > sizeof(struct vxlan_hdr)); > > parse_ethernet(eth_hdr, ðertype, &l2_len, > - &l3_len, &l4_proto); > + &l3_len, &l4_proto, &l4_len); > l3_hdr = (char *)eth_hdr + l2_len; > } > } > @@ -399,11 +423,12 @@ pkt_burst_checksum_forward(struct fwd_stream *fs) > > /* step 3: depending on user command line configuration, > * recompute checksum either in software or flag the > - * mbuf to offload the calculation to the NIC */ > + * mbuf to offload the calculation to the NIC. If TSO > + * is configured, prepare the mbuf for TCP segmentation. */ > > /* process checksums of inner headers first */ > ol_flags |= process_inner_cksums(l3_hdr, ethertype, > - l3_len, l4_proto, testpmd_ol_flags); > + l3_len, l4_proto, tso_segsz, testpmd_ol_flags); > > /* Then process outer headers if any. Note that the software > * checksum will be wrong if one of the inner checksums is > @@ -432,6 +457,7 @@ pkt_burst_checksum_forward(struct fwd_stream *fs) > sizeof(struct udp_hdr) + > sizeof(struct vxlan_hdr) + l2_len; > m->l3_len = l3_len; > + m->l4_len = l4_len; > } > } else { > /* this is only useful if an offload flag is > @@ -439,7 +465,9 @@ pkt_burst_checksum_forward(struct fwd_stream *fs) > * case */ > m->l2_len = l2_len; > m->l3_len = l3_len; > + m->l4_len = l4_len; > } > + m->tso_segsz = tso_segsz; > m->ol_flags = ol_flags; > > } > diff --git a/app/test-pmd/testpmd.h b/app/test-pmd/testpmd.h > index 83311fa..16c6fbf 100644 > --- a/app/test-pmd/testpmd.h > +++ b/app/test-pmd/testpmd.h > @@ -149,6 +149,7 @@ struct rte_port { > struct fwd_stream *tx_stream; /**< Port TX stream, if unique */ > unsigned int socket_id; /**< For NUMA support */ > uint16_t tx_ol_flags;/**< TX Offload Flags > (TESTPMD_TX_OFFLOAD...). */ > + uint16_t tso_segsz; /**< MSS for segmentation offload. > */ > uint16_t tx_vlan_id; /**< Tag Id. in TX VLAN packets. */ > void *fwd_ctx; /**< Forwarding mode context */ > uint64_t rx_bad_ip_csum; /**< rx pkts with bad ip > checksum */ > -- > 2.1.0