Adds the process callback function for ip6_lookup_fib node. Signed-off-by: Ankur Dwivedi <adwiv...@marvell.com> --- lib/node/ip6_lookup_fib.c | 158 ++++++++++++++++++++++++++++++++++++++ 1 file changed, 158 insertions(+)
diff --git a/lib/node/ip6_lookup_fib.c b/lib/node/ip6_lookup_fib.c index caf3e9d1fa..6ad975dddb 100644 --- a/lib/node/ip6_lookup_fib.c +++ b/lib/node/ip6_lookup_fib.c @@ -40,6 +40,163 @@ static struct ip6_lookup_fib_node_main ip6_lookup_fib_nm; #define IP6_LOOKUP_FIB_NODE_PRIV1_OFF(ctx) \ (((struct ip6_lookup_fib_node_ctx *)ctx)->mbuf_priv1_off) +static uint16_t +ip6_lookup_fib_node_process(struct rte_graph *graph, struct rte_node *node, void **objs, + uint16_t nb_objs) +{ + struct rte_mbuf *mbuf0, *mbuf1, *mbuf2, *mbuf3, **pkts; + struct rte_fib6 *fib = IP6_LOOKUP_FIB_NODE(node->ctx); + const int dyn = IP6_LOOKUP_FIB_NODE_PRIV1_OFF(node->ctx); + struct rte_ipv6_addr ip[nb_objs]; + struct rte_ipv6_hdr *ipv6_hdr; + uint64_t next_hop[nb_objs]; + uint16_t lookup_err = 0; + void **to_next, **from; + uint16_t last_spec = 0; + rte_edge_t next_index; + uint16_t n_left_from; + uint16_t held = 0; + uint32_t drop_nh; + uint16_t next; + int i, rc; + + /* Speculative next */ + next_index = RTE_NODE_IP6_LOOKUP_NEXT_REWRITE; + /* Drop node */ + drop_nh = ((uint32_t)RTE_NODE_IP6_LOOKUP_NEXT_PKT_DROP) << 16; + + pkts = (struct rte_mbuf **)objs; + from = objs; + n_left_from = nb_objs; + + /* Get stream for the speculated next node */ + to_next = rte_node_next_stream_get(graph, node, next_index, nb_objs); + + for (i = OBJS_PER_CLINE; i < RTE_GRAPH_BURST_SIZE; i += OBJS_PER_CLINE) + rte_prefetch0(&objs[i]); + +#if RTE_GRAPH_BURST_SIZE > 64 + for (i = 0; i < 4 && i < n_left_from; i++) { + rte_prefetch0(pkts[i]); + rte_prefetch0(rte_pktmbuf_mtod_offset(pkts[i], void *, + sizeof(struct rte_ether_hdr))); + } +#endif + + i = 0; + while (n_left_from >= 4) { +#if RTE_GRAPH_BURST_SIZE > 64 + if (likely(n_left_from > 7)) { + rte_prefetch0(pkts[4]); + rte_prefetch0(rte_pktmbuf_mtod_offset(pkts[4], void *, + sizeof(struct rte_ether_hdr))); + rte_prefetch0(pkts[5]); + rte_prefetch0(rte_pktmbuf_mtod_offset(pkts[5], void *, + sizeof(struct rte_ether_hdr))); + rte_prefetch0(pkts[6]); + rte_prefetch0(rte_pktmbuf_mtod_offset(pkts[6], void *, + sizeof(struct rte_ether_hdr))); + rte_prefetch0(pkts[7]); + rte_prefetch0(rte_pktmbuf_mtod_offset(pkts[7], void *, + sizeof(struct rte_ether_hdr))); + } +#endif + + mbuf0 = pkts[0]; + mbuf1 = pkts[1]; + mbuf2 = pkts[2]; + mbuf3 = pkts[3]; + pkts += 4; + n_left_from -= 4; + /* Extract DIP of mbuf0 */ + ipv6_hdr = rte_pktmbuf_mtod_offset(mbuf0, struct rte_ipv6_hdr *, + sizeof(struct rte_ether_hdr)); + /* Extract hop_limits as ipv6 hdr is in cache */ + node_mbuf_priv1(mbuf0, dyn)->ttl = ipv6_hdr->hop_limits; + + ip[i++] = ipv6_hdr->dst_addr; + + /* Extract DIP of mbuf1 */ + ipv6_hdr = rte_pktmbuf_mtod_offset(mbuf1, struct rte_ipv6_hdr *, + sizeof(struct rte_ether_hdr)); + /* Extract hop_limits as ipv6 hdr is in cache */ + node_mbuf_priv1(mbuf1, dyn)->ttl = ipv6_hdr->hop_limits; + + ip[i++] = ipv6_hdr->dst_addr; + + /* Extract DIP of mbuf2 */ + ipv6_hdr = rte_pktmbuf_mtod_offset(mbuf2, struct rte_ipv6_hdr *, + sizeof(struct rte_ether_hdr)); + /* Extract hop_limits as ipv6 hdr is in cache */ + node_mbuf_priv1(mbuf2, dyn)->ttl = ipv6_hdr->hop_limits; + + ip[i++] = ipv6_hdr->dst_addr; + + /* Extract DIP of mbuf3 */ + ipv6_hdr = rte_pktmbuf_mtod_offset(mbuf3, struct rte_ipv6_hdr *, + sizeof(struct rte_ether_hdr)); + /* Extract hop_limits as ipv6 hdr is in cache */ + node_mbuf_priv1(mbuf3, dyn)->ttl = ipv6_hdr->hop_limits; + + ip[i++] = ipv6_hdr->dst_addr; + } + while (n_left_from > 0) { + mbuf0 = pkts[0]; + pkts += 1; + n_left_from -= 1; + + /* Extract DIP of mbuf0 */ + ipv6_hdr = rte_pktmbuf_mtod_offset(mbuf0, struct rte_ipv6_hdr *, + sizeof(struct rte_ether_hdr)); + /* Extract hop_limits as ipv6 hdr is in cache */ + node_mbuf_priv1(mbuf0, dyn)->ttl = ipv6_hdr->hop_limits; + + ip[i++] = ipv6_hdr->dst_addr; + } + + rc = rte_fib6_lookup_bulk(fib, ip, next_hop, nb_objs); + if (unlikely(rc != 0)) + return 0; + + for (i = 0; i < nb_objs; i++) { + if (unlikely(next_hop[i] == FIB6_DEFAULT_NH)) { + next_hop[i] = drop_nh; + lookup_err += 1; + } + + mbuf0 = (struct rte_mbuf *)objs[i]; + node_mbuf_priv1(mbuf0, dyn)->nh = (uint16_t)next_hop[i]; + next = (uint16_t)(next_hop[i] >> 16); + + if (unlikely(next_index ^ next)) { + /* Copy things successfully speculated till now */ + rte_memcpy(to_next, from, last_spec * sizeof(from[0])); + from += last_spec; + to_next += last_spec; + held += last_spec; + last_spec = 0; + + rte_node_enqueue_x1(graph, node, next, from[0]); + from += 1; + } else { + last_spec += 1; + } + } + + /* !!! Home run !!! */ + if (likely(last_spec == nb_objs)) { + rte_node_next_stream_move(graph, node, next_index); + return nb_objs; + } + + NODE_INCREMENT_XSTAT_ID(node, 0, lookup_err != 0, lookup_err); + held += last_spec; + rte_memcpy(to_next, from, last_spec * sizeof(from[0])); + rte_node_next_stream_put(graph, node, next_index, held); + + return nb_objs; +} + RTE_EXPORT_EXPERIMENTAL_SYMBOL(rte_node_ip6_fib_route_add, 25.07) int rte_node_ip6_fib_route_add(const struct rte_ipv6_addr *ip, uint8_t depth, uint16_t next_hop, @@ -143,6 +300,7 @@ static struct rte_node_xstats ip6_lookup_fib_xstats = { }; static struct rte_node_register ip6_lookup_fib_node = { + .process = ip6_lookup_fib_node_process, .name = "ip6_lookup_fib", .init = ip6_lookup_fib_node_init, -- 2.25.1