> Use HW parsed packet type when ethdev supports necessary protocols. > If packet type is not supported, then register ethdev callbacks > for parse packet in SW. This is better for performance as it > effects fast path. > > Signed-off-by: Nithin Dabilpuram <ndabilpu...@marvell.com> > --- > examples/ipsec-secgw/ipsec-secgw.c | 259 > +++++++++++++++++++++++++++---------- > 1 file changed, 194 insertions(+), 65 deletions(-) > > diff --git a/examples/ipsec-secgw/ipsec-secgw.c > b/examples/ipsec-secgw/ipsec-secgw.c > index 76919e5..e8f9e90 100644 > --- a/examples/ipsec-secgw/ipsec-secgw.c > +++ b/examples/ipsec-secgw/ipsec-secgw.c > @@ -374,53 +374,30 @@ print_stats_cb(__rte_unused void *param) > static inline void > prepare_one_packet(struct rte_mbuf *pkt, struct ipsec_traffic *t) > { > + uint32_t ptype = pkt->packet_type; > const struct rte_ether_hdr *eth; > const struct rte_ipv4_hdr *iph4; > const struct rte_ipv6_hdr *iph6; > - const struct rte_udp_hdr *udp; > - uint16_t ip4_hdr_len; > - uint16_t nat_port; > + uint32_t tun_type, l3_type; > + > + tun_type = ptype & RTE_PTYPE_TUNNEL_MASK; > + l3_type = ptype & RTE_PTYPE_L3_MASK; > > eth = rte_pktmbuf_mtod(pkt, const struct rte_ether_hdr *); > - if (eth->ether_type == rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV4)) { > - > + if (l3_type == RTE_PTYPE_L3_IPV4) { > iph4 = (const struct rte_ipv4_hdr *)rte_pktmbuf_adj(pkt, > RTE_ETHER_HDR_LEN); > adjust_ipv4_pktlen(pkt, iph4, 0); > > - switch (iph4->next_proto_id) { > - case IPPROTO_ESP: > + if (tun_type == RTE_PTYPE_TUNNEL_ESP) { > t->ipsec.pkts[(t->ipsec.num)++] = pkt; > - break; > - case IPPROTO_UDP: > - if (app_sa_prm.udp_encap == 1) { > - ip4_hdr_len = ((iph4->version_ihl & > - RTE_IPV4_HDR_IHL_MASK) * > - RTE_IPV4_IHL_MULTIPLIER); > - udp = rte_pktmbuf_mtod_offset(pkt, > - struct rte_udp_hdr *, ip4_hdr_len); > - nat_port = rte_cpu_to_be_16(IPSEC_NAT_T_PORT); > - if (udp->src_port == nat_port || > - udp->dst_port == nat_port){ > - t->ipsec.pkts[(t->ipsec.num)++] = pkt; > - pkt->packet_type |= > - MBUF_PTYPE_TUNNEL_ESP_IN_UDP; > - break; > - } > - } > - /* Fall through */ > - default: > + } else { > t->ip4.data[t->ip4.num] = &iph4->next_proto_id; > t->ip4.pkts[(t->ip4.num)++] = pkt; > } > pkt->l2_len = 0; > pkt->l3_len = sizeof(*iph4); > - pkt->packet_type |= RTE_PTYPE_L3_IPV4; > - if (pkt->packet_type & RTE_PTYPE_L4_TCP) > - pkt->l4_len = sizeof(struct rte_tcp_hdr); > - else if (pkt->packet_type & RTE_PTYPE_L4_UDP) > - pkt->l4_len = sizeof(struct rte_udp_hdr); > - } else if (eth->ether_type == rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV6)) { > + } else if (l3_type & RTE_PTYPE_L3_IPV6) {
As a nit: RTE_ETH_IS_IPV6_HDR(l3_type) > int next_proto; > size_t l3len, ext_len; > uint8_t *p; > @@ -430,47 +407,37 @@ prepare_one_packet(struct rte_mbuf *pkt, struct > ipsec_traffic *t) > RTE_ETHER_HDR_LEN); > adjust_ipv6_pktlen(pkt, iph6, 0); > > - next_proto = iph6->proto; > - > - /* determine l3 header size up to ESP extension */ > l3len = sizeof(struct ip6_hdr); > - p = rte_pktmbuf_mtod(pkt, uint8_t *); > - while (next_proto != IPPROTO_ESP && l3len < pkt->data_len && > - (next_proto = rte_ipv6_get_next_ext(p + l3len, > - next_proto, &ext_len)) >= 0) > - l3len += ext_len; > > - /* drop packet when IPv6 header exceeds first segment length */ > - if (unlikely(l3len > pkt->data_len)) { > - free_pkts(&pkt, 1); > - return; > - } > - > - switch (next_proto) { > - case IPPROTO_ESP: > + if (tun_type == RTE_PTYPE_TUNNEL_ESP) { > t->ipsec.pkts[(t->ipsec.num)++] = pkt; > - break; > - case IPPROTO_UDP: > - if (app_sa_prm.udp_encap == 1) { > - udp = rte_pktmbuf_mtod_offset(pkt, > - struct rte_udp_hdr *, l3len); > - nat_port = rte_cpu_to_be_16(IPSEC_NAT_T_PORT); > - if (udp->src_port == nat_port || > - udp->dst_port == nat_port){ > - t->ipsec.pkts[(t->ipsec.num)++] = pkt; > - pkt->packet_type |= > - MBUF_PTYPE_TUNNEL_ESP_IN_UDP; > - break; > - } > - } > - /* Fall through */ > - default: > + } else { > t->ip6.data[t->ip6.num] = &iph6->proto; > t->ip6.pkts[(t->ip6.num)++] = pkt; > } > + > + /* Determine l3 header size up to ESP extension by walking > + * through extension headers. > + */ > + if (l3_type == RTE_PTYPE_L3_IPV6_EXT || > + l3_type == RTE_PTYPE_L3_IPV6_EXT_UNKNOWN) { > + p = rte_pktmbuf_mtod(pkt, uint8_t *); > + next_proto = iph6->proto; > + while (next_proto != IPPROTO_ESP && > + l3len < pkt->data_len && > + (next_proto = rte_ipv6_get_next_ext(p + l3len, > + next_proto, &ext_len)) >= 0) > + l3len += ext_len; > + > + /* Drop pkt when IPv6 header exceeds first seg size */ > + if (unlikely(l3len > pkt->data_len)) { > + free_pkts(&pkt, 1); > + return; > + } > + } > + > pkt->l2_len = 0; > pkt->l3_len = l3len; > - pkt->packet_type |= RTE_PTYPE_L3_IPV6; > } else { > /* Unknown/Unsupported type, drop the packet */ > RTE_LOG(ERR, IPSEC, "Unsupported packet type 0x%x\n", > @@ -479,6 +446,11 @@ prepare_one_packet(struct rte_mbuf *pkt, struct > ipsec_traffic *t) > return; > } > > + if (ptype & RTE_PTYPE_L4_TCP) I think it needs to be: If ((ptype & RTE_PTYPE_L4_MASK) == RTE_PTYPE_L4_TCP) Same for udp. Though it seems that it was there already since commit a7f32947a316c4757a315239752596ca1cf1b268. > + pkt->l4_len = sizeof(struct rte_tcp_hdr); > + else if (ptype & RTE_PTYPE_L4_UDP) > + pkt->l4_len = sizeof(struct rte_udp_hdr); > + > /* Check if the packet has been processed inline. For inline protocol > * processed packets, the metadata in the mbuf can be used to identify > * the security processing done on the packet. The metadata will be > @@ -2249,6 +2221,147 @@ cryptodevs_init(uint16_t req_queue_num) > return total_nb_qps; > } > > +static int > +check_ptype(int portid) > +{ > + int l3_ipv4 = 0, l3_ipv6 = 0, l4_udp = 0, tunnel_esp = 0; > + int i, nb_ptypes; > + uint32_t mask; > + > + mask = (RTE_PTYPE_L3_MASK | RTE_PTYPE_L4_MASK | > + RTE_PTYPE_TUNNEL_MASK); > + > + nb_ptypes = rte_eth_dev_get_supported_ptypes(portid, mask, NULL, 0); > + if (nb_ptypes <= 0) > + return 0; > + > + uint32_t ptypes[nb_ptypes]; > + > + nb_ptypes = rte_eth_dev_get_supported_ptypes(portid, mask, ptypes, > nb_ptypes); > + for (i = 0; i < nb_ptypes; ++i) { > + if (ptypes[i] & RTE_PTYPE_L3_IPV4) As nit: RTE_ETH_IS_IPV4_HDR(ptypes[i]) > + l3_ipv4 = 1; > + if (ptypes[i] & RTE_PTYPE_L3_IPV6) As nit: RTE_ETH_IS_IPV6_HDR(ptypes[i]) > + l3_ipv6 = 1; > + if (ptypes[i] & RTE_PTYPE_TUNNEL_ESP) if ((ptypes[i] & RTE_PTYPE_TUNNEL_MASK == RTE_PTYPE_TUNNEL_ESP) > + tunnel_esp = 1; > + if (ptypes[i] & RTE_PTYPE_L4_UDP) Same as above. > + l4_udp = 1; > + } > + > + if (l3_ipv4 == 0) > + printf("port %d cannot parse RTE_PTYPE_L3_IPV4\n", portid); > + > + if (l3_ipv6 == 0) > + printf("port %d cannot parse RTE_PTYPE_L3_IPV6\n", portid); > + > + if (l4_udp == 0) > + printf("port %d cannot parse RTE_PTYPE_L4_UDP\n", portid); > + > + if (tunnel_esp == 0) > + printf("port %d cannot parse RTE_PTYPE_TUNNEL_ESP\n", portid); > + > + if (l3_ipv4 && l3_ipv6 && l4_udp && tunnel_esp) > + return 1; > + > + return 0; > + > +} > + > +static inline void > +parse_ptype(struct rte_mbuf *m) > +{ > + uint32_t packet_type = RTE_PTYPE_UNKNOWN; > + const struct rte_ipv4_hdr *iph4; > + const struct rte_ipv6_hdr *iph6; > + const struct rte_ether_hdr *eth; > + const struct rte_udp_hdr *udp; > + uint16_t nat_port, ether_type; > + int next_proto = 0; > + size_t ext_len = 0; > + const uint8_t *p; > + uint32_t l3len; > + > + eth = rte_pktmbuf_mtod(m, struct rte_ether_hdr *); > + ether_type = eth->ether_type; > + > + if (ether_type == rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV4)) { > + iph4 = (const struct rte_ipv4_hdr *)(eth + 1); > + l3len = ((iph4->version_ihl & RTE_IPV4_HDR_IHL_MASK) * > + RTE_IPV4_IHL_MULTIPLIER); > + > + if (l3len == sizeof(struct rte_ipv4_hdr)) > + packet_type |= RTE_PTYPE_L3_IPV4; > + else > + packet_type |= RTE_PTYPE_L3_IPV4_EXT_UNKNOWN; > + > + next_proto = iph4->next_proto_id; > + p = (const uint8_t *)iph4; > + } else if (ether_type == rte_cpu_to_be_16(RTE_ETHER_TYPE_IPV6)) { > + iph6 = (const struct rte_ipv6_hdr *)(eth + 1); > + l3len = sizeof(struct ip6_hdr); > + > + /* determine l3 header size up to ESP extension */ > + next_proto = iph6->proto; > + p = (const uint8_t *)iph6; > + while (next_proto != IPPROTO_ESP && l3len < m->data_len && > + (next_proto = rte_ipv6_get_next_ext(p + l3len, > + next_proto, &ext_len)) >= 0) > + l3len += ext_len; > + > + /* Skip IPv6 header exceeds first segment length */ > + if (unlikely(l3len + RTE_ETHER_HDR_LEN > m->data_len)) > + goto exit; > + > + if (l3len == sizeof(struct ip6_hdr)) > + packet_type |= RTE_PTYPE_L3_IPV6; > + else > + packet_type |= RTE_PTYPE_L3_IPV6_EXT; > + } > + > + switch (next_proto) { > + case IPPROTO_ESP: > + packet_type |= RTE_PTYPE_TUNNEL_ESP; > + break; > + case IPPROTO_UDP: > + if (app_sa_prm.udp_encap == 1) { > + udp = (const struct rte_udp_hdr *)(p + l3len); > + nat_port = rte_cpu_to_be_16(IPSEC_NAT_T_PORT); > + if (udp->src_port == nat_port || > + udp->dst_port == nat_port) > + packet_type |= > + MBUF_PTYPE_TUNNEL_ESP_IN_UDP; > + } > + break; > + default: > + break; > + } > +exit: > + m->packet_type = packet_type; > +} > + > +static uint16_t > +parse_ptype_cb(uint16_t port __rte_unused, uint16_t queue __rte_unused, > + struct rte_mbuf *pkts[], uint16_t nb_pkts, > + uint16_t max_pkts __rte_unused, > + void *user_param __rte_unused) > +{ > + uint32_t i; > + > + if (unlikely(nb_pkts == 0)) > + return nb_pkts; > + > + rte_prefetch0(rte_pktmbuf_mtod(pkts[0], struct ether_hdr *)); > + for (i = 0; i < (unsigned int) (nb_pkts - 1); ++i) { > + rte_prefetch0(rte_pktmbuf_mtod(pkts[i+1], > + struct ether_hdr *)); > + parse_ptype(pkts[i]); > + } > + parse_ptype(pkts[i]); > + > + return nb_pkts; > +} > + > static void > port_init(uint16_t portid, uint64_t req_rx_offloads, uint64_t > req_tx_offloads) > { > @@ -2260,6 +2373,7 @@ port_init(uint16_t portid, uint64_t req_rx_offloads, > uint64_t req_tx_offloads) > struct lcore_conf *qconf; > struct rte_ether_addr ethaddr; > struct rte_eth_conf local_port_conf = port_conf; > + int ptype_supported; > > ret = rte_eth_dev_info_get(portid, &dev_info); > if (ret != 0) > @@ -2357,6 +2471,11 @@ port_init(uint16_t portid, uint64_t req_rx_offloads, > uint64_t req_tx_offloads) > rte_exit(EXIT_FAILURE, "Cannot adjust number of descriptors: " > "err=%d, port=%d\n", ret, portid); > > + /* Check if required ptypes are supported */ > + ptype_supported = check_ptype(portid); > + if (!ptype_supported) > + printf("Port %d: softly parse packet type info\n", portid); > + > /* init one TX queue per lcore */ > tx_queueid = 0; > for (lcore_id = 0; lcore_id < RTE_MAX_LCORE; lcore_id++) { > @@ -2418,6 +2537,16 @@ port_init(uint16_t portid, uint64_t req_rx_offloads, > uint64_t req_tx_offloads) > rte_exit(EXIT_FAILURE, > "rte_eth_rx_queue_setup: err=%d, " > "port=%d\n", ret, portid); > + > + /* Register Rx callback if ptypes are not supported */ > + if (!ptype_supported && > + !rte_eth_add_rx_callback(portid, queue, > + parse_ptype_cb, NULL)) { > + printf("Failed to add rx callback: port=%d, " > + "queue=%d\n", portid, queue); > + } > + > + > } > } > printf("\n"); > -- > 2.8.4