Optimizing VxLAN tunneling performance in userspace datapath by offloading
the rx checksum validation on tunnel packets to the NIC when it is supported.
Signed-off-by: Sugesh Chandran <sugesh.chand...@intel.com>
---
 lib/dpif-netdev.c            | 73 ++++++++++++++++++++++++----------
 lib/dpif.c                   |  5 ++-
 lib/netdev-dpdk.c            | 31 +++++++++++++--
 lib/netdev-dpdk.h            | 17 ++++++++
 lib/netdev-provider.h        |  2 +-
 lib/netdev-vport.c           | 95 ++++++++++++++++++++++++++++++++++++++++++--
 lib/netdev.c                 |  5 ++-
 lib/netdev.h                 |  2 +-
 lib/odp-execute.c            |  6 +--
 lib/odp-execute.h            |  6 ++-
 ofproto/ofproto-dpif-xlate.c |  3 +-
 11 files changed, 205 insertions(+), 40 deletions(-)

diff --git a/lib/dpif-netdev.c b/lib/dpif-netdev.c
index 2870951..d8ab8b5 100644
--- a/lib/dpif-netdev.c
+++ b/lib/dpif-netdev.c
@@ -477,9 +477,11 @@ static void dp_netdev_execute_actions(struct 
dp_netdev_pmd_thread *pmd,
                                       struct dp_packet **, int c,
                                       bool may_steal,
                                       const struct nlattr *actions,
-                                      size_t actions_len);
+                                      size_t actions_len,
+                                      uint32_t rx_ol_flags);
 static void dp_netdev_input(struct dp_netdev_pmd_thread *,
-                            struct dp_packet **, int cnt, odp_port_t port_no);
+                                      struct dp_packet **, int cnt,
+                                      struct dp_netdev_port *port);
 static void dp_netdev_recirculate(struct dp_netdev_pmd_thread *,
                                   struct dp_packet **, int cnt);
 
@@ -2369,7 +2371,7 @@ dpif_netdev_execute(struct dpif *dpif, struct 
dpif_execute *execute)
 
     pp = execute->packet;
     dp_netdev_execute_actions(pmd, &pp, 1, false, execute->actions,
-                              execute->actions_len);
+                              execute->actions_len, 0);
     if (pmd->core_id == NON_PMD_CORE_ID) {
         dp_netdev_pmd_unref(pmd);
         ovs_mutex_unlock(&dp->port_mutex);
@@ -2572,7 +2574,7 @@ dp_netdev_process_rxq_port(struct dp_netdev_pmd_thread 
*pmd,
         *recirc_depth_get() = 0;
 
         cycles_count_start(pmd);
-        dp_netdev_input(pmd, packets, cnt, port->port_no);
+        dp_netdev_input(pmd, packets, cnt, port);
         cycles_count_end(pmd, PMD_CYCLES_PROCESSING);
     } else if (error != EAGAIN && error != EOPNOTSUPP) {
         static struct vlog_rate_limit rl = VLOG_RATE_LIMIT_INIT(1, 5);
@@ -3365,7 +3367,7 @@ packet_batch_init(struct packet_batch *batch, struct 
dp_netdev_flow *flow)
 static inline void
 packet_batch_execute(struct packet_batch *batch,
                      struct dp_netdev_pmd_thread *pmd,
-                     long long now)
+                     long long now, uint32_t rx_ol_flags)
 {
     struct dp_netdev_actions *actions;
     struct dp_netdev_flow *flow = batch->flow;
@@ -3376,7 +3378,7 @@ packet_batch_execute(struct packet_batch *batch,
     actions = dp_netdev_flow_get_actions(flow);
 
     dp_netdev_execute_actions(pmd, batch->packets, batch->packet_count, true,
-                              actions->actions, actions->size);
+                              actions->actions, actions->size, rx_ol_flags);
 }
 
 static inline void
@@ -3394,6 +3396,19 @@ dp_netdev_queue_batches(struct dp_packet *pkt,
     packet_batch_update(batch, pkt, mf);
 }
 
+static inline bool
+is_checksum_valid(struct dp_packet *packet) {
+#ifdef DPDK_NETDEV
+    if (packet->mbuf.ol_flags & (PKT_RX_IP_CKSUM_BAD |
+                                 PKT_RX_L4_CKSUM_BAD)) {
+        return 0;
+    }
+    return 1;
+#else
+    return 0;
+#endif
+}
+
 /* Try to process all ('cnt') the 'packets' using only the exact match cache
  * 'pmd->flow_cache'. If a flow is not found for a packet 'packets[i]', the
  * miniflow is copied into 'keys' and the packet pointer is moved at the
@@ -3409,7 +3424,8 @@ static inline size_t
 emc_processing(struct dp_netdev_pmd_thread *pmd, struct dp_packet **packets,
                size_t cnt, struct netdev_flow_key *keys,
                struct packet_batch batches[], size_t *n_batches,
-               bool md_is_valid, odp_port_t port_no)
+               bool md_is_valid, struct dp_netdev_port *port,
+               uint32_t rx_checksum_ofld_capa)
 {
     struct emc_cache *flow_cache = &pmd->flow_cache;
     struct netdev_flow_key *key = &keys[0];
@@ -3425,6 +3441,13 @@ emc_processing(struct dp_netdev_pmd_thread *pmd, struct 
dp_packet **packets,
             continue;
         }
 
+        if (OVS_UNLIKELY(rx_checksum_ofld_capa &&
+                         !is_checksum_valid(packet))) {
+            dp_packet_delete(packet);
+            n_dropped++;
+            continue;
+        }
+
         if (i != cnt - 1) {
             /* Prefetch next packet data and metadata. */
             OVS_PREFETCH(dp_packet_data(packets[i+1]));
@@ -3432,7 +3455,7 @@ emc_processing(struct dp_netdev_pmd_thread *pmd, struct 
dp_packet **packets,
         }
 
         if (!md_is_valid) {
-            pkt_metadata_init(&packet->md, port_no);
+            pkt_metadata_init(&packet->md, port->port_no);
         }
         miniflow_extract(packet, &key->mf);
         key->len = 0; /* Not computed yet. */
@@ -3462,7 +3485,8 @@ static inline void
 fast_path_processing(struct dp_netdev_pmd_thread *pmd,
                      struct dp_packet **packets, size_t cnt,
                      struct netdev_flow_key *keys,
-                     struct packet_batch batches[], size_t *n_batches)
+                     struct packet_batch batches[], size_t *n_batches,
+                     uint32_t rx_ol_flags)
 {
 #if !defined(__CHECKER__) && !defined(_WIN32)
     const size_t PKT_ARRAY_SIZE = cnt;
@@ -3541,7 +3565,7 @@ fast_path_processing(struct dp_netdev_pmd_thread *pmd,
              * the actions.  Otherwise, if there are any slow path actions,
              * we'll send the packet up twice. */
             dp_netdev_execute_actions(pmd, &packets[i], 1, true,
-                                      actions.data, actions.size);
+                                      actions.data, actions.size, rx_ol_flags);
 
             add_actions = put_actions.size ? &put_actions : &actions;
             if (OVS_LIKELY(error != ENOSPC)) {
@@ -3606,7 +3630,7 @@ fast_path_processing(struct dp_netdev_pmd_thread *pmd,
 static void
 dp_netdev_input__(struct dp_netdev_pmd_thread *pmd,
                   struct dp_packet **packets, int cnt,
-                  bool md_is_valid, odp_port_t port_no)
+                  bool md_is_valid, struct dp_netdev_port *port)
 {
 #if !defined(__CHECKER__) && !defined(_WIN32)
     const size_t PKT_ARRAY_SIZE = cnt;
@@ -3618,12 +3642,18 @@ dp_netdev_input__(struct dp_netdev_pmd_thread *pmd,
     struct packet_batch batches[PKT_ARRAY_SIZE];
     long long now = time_msec();
     size_t newcnt, n_batches, i;
+    uint32_t rx_ol_flags;
 
     n_batches = 0;
+    /* Collect the Rx offloading features only if its DPDK port */
+    rx_ol_flags = port && (packets[0]->source == DPBUF_DPDK) ?
+                            netdev_get_rx_offload_capa(port->netdev) :
+                            0;
     newcnt = emc_processing(pmd, packets, cnt, keys, batches, &n_batches,
-                            md_is_valid, port_no);
+                            md_is_valid, port, rx_ol_flags);
     if (OVS_UNLIKELY(newcnt)) {
-        fast_path_processing(pmd, packets, newcnt, keys, batches, &n_batches);
+        fast_path_processing(pmd, packets, newcnt, keys, batches,
+                                           &n_batches, rx_ol_flags);
     }
 
     for (i = 0; i < n_batches; i++) {
@@ -3631,16 +3661,16 @@ dp_netdev_input__(struct dp_netdev_pmd_thread *pmd,
     }
 
     for (i = 0; i < n_batches; i++) {
-        packet_batch_execute(&batches[i], pmd, now);
+        packet_batch_execute(&batches[i], pmd, now, rx_ol_flags);
     }
 }
 
 static void
 dp_netdev_input(struct dp_netdev_pmd_thread *pmd,
                 struct dp_packet **packets, int cnt,
-                odp_port_t port_no)
+                struct dp_netdev_port *port)
 {
-     dp_netdev_input__(pmd, packets, cnt, false, port_no);
+     dp_netdev_input__(pmd, packets, cnt, false, port);
 }
 
 static void
@@ -3716,7 +3746,7 @@ dp_netdev_clone_pkt_batch(struct dp_packet **dst_pkts,
 
 static void
 dp_execute_cb(void *aux_, struct dp_packet **packets, int cnt,
-              const struct nlattr *a, bool may_steal)
+              const struct nlattr *a, bool may_steal, uint32_t rx_ol_flags)
     OVS_NO_THREAD_SAFETY_ANALYSIS
 {
     struct dp_netdev_execute_aux *aux = aux_;
@@ -3776,7 +3806,7 @@ dp_execute_cb(void *aux_, struct dp_packet **packets, int 
cnt,
                    packets = tnl_pkt;
                 }
 
-                err = netdev_pop_header(p->netdev, packets, cnt);
+                err = netdev_pop_header(p->netdev, packets, cnt, rx_ol_flags);
                 if (!err) {
 
                     for (i = 0; i < cnt; i++) {
@@ -3816,7 +3846,7 @@ dp_execute_cb(void *aux_, struct dp_packet **packets, int 
cnt,
                                          NULL);
                 if (!error || error == ENOSPC) {
                     dp_netdev_execute_actions(pmd, &packets[i], 1, may_steal,
-                                              actions.data, actions.size);
+                                              actions.data, actions.size, 0);
                 } else if (may_steal) {
                     dp_packet_delete(packets[i]);
                 }
@@ -3878,12 +3908,13 @@ static void
 dp_netdev_execute_actions(struct dp_netdev_pmd_thread *pmd,
                           struct dp_packet **packets, int cnt,
                           bool may_steal,
-                          const struct nlattr *actions, size_t actions_len)
+                          const struct nlattr *actions, size_t actions_len,
+                          uint32_t rx_ol_flags)
 {
     struct dp_netdev_execute_aux aux = { pmd };
 
     odp_execute_actions(&aux, packets, cnt, may_steal, actions,
-                        actions_len, dp_execute_cb);
+                        actions_len, dp_execute_cb, rx_ol_flags);
 }
 
 const struct dpif_class dpif_netdev_class = {
diff --git a/lib/dpif.c b/lib/dpif.c
index a784de7..760719b 100644
--- a/lib/dpif.c
+++ b/lib/dpif.c
@@ -1088,7 +1088,8 @@ struct dpif_execute_helper_aux {
  * meaningful. */
 static void
 dpif_execute_helper_cb(void *aux_, struct dp_packet **packets, int cnt,
-                       const struct nlattr *action, bool may_steal OVS_UNUSED)
+                       const struct nlattr *action, bool may_steal OVS_UNUSED,
+                       uint32_t rx_ol_flags)
 {
     struct dpif_execute_helper_aux *aux = aux_;
     int type = nl_attr_type(action);
@@ -1167,7 +1168,7 @@ dpif_execute_with_help(struct dpif *dpif, struct 
dpif_execute *execute)
 
     pp = execute->packet;
     odp_execute_actions(&aux, &pp, 1, false, execute->actions,
-                        execute->actions_len, dpif_execute_helper_cb);
+                        execute->actions_len, dpif_execute_helper_cb, 0);
     return aux.error;
 }
 
diff --git a/lib/netdev-dpdk.c b/lib/netdev-dpdk.c
index c7217ea..5e951c7 100644
--- a/lib/netdev-dpdk.c
+++ b/lib/netdev-dpdk.c
@@ -321,6 +321,10 @@ struct netdev_dpdk {
     struct rte_eth_link link;
     int link_reset_cnt;
 
+    /* RX offload capability bitmask, RX checksum offloading is only
+     * in use now */
+    uint32_t rx_ol_flags;
+
     /* The user might request more txqs than the NIC has.  We remap those
      * ('up.n_txq') on these ('real_n_txq').
      * If the numbers match, 'txq_needs_locking' is false, otherwise it is
@@ -527,7 +531,8 @@ dpdk_watchdog(void *dummy OVS_UNUSED)
 }
 
 static int
-dpdk_eth_dev_queue_setup(struct netdev_dpdk *dev, int n_rxq, int n_txq)
+dpdk_eth_dev_queue_setup(struct netdev_dpdk *dev, int n_rxq, int n_txq,
+                         const struct rte_eth_conf *new_port_conf)
 {
     int diag = 0;
     int i;
@@ -542,7 +547,8 @@ dpdk_eth_dev_queue_setup(struct netdev_dpdk *dev, int 
n_rxq, int n_txq)
             VLOG_INFO("Retrying setup with (rxq:%d txq:%d)", n_rxq, n_txq);
         }
 
-        diag = rte_eth_dev_configure(dev->port_id, n_rxq, n_txq, &port_conf);
+        diag = rte_eth_dev_configure(dev->port_id, n_rxq, n_txq,
+                                     new_port_conf);
         if (diag) {
             break;
         }
@@ -596,6 +602,7 @@ dpdk_eth_dev_init(struct netdev_dpdk *dev) 
OVS_REQUIRES(dpdk_mutex)
     struct rte_pktmbuf_pool_private *mbp_priv;
     struct rte_eth_dev_info info;
     struct ether_addr eth_addr;
+    struct rte_eth_conf new_port_conf;
     int diag;
     int n_rxq, n_txq;
 
@@ -608,7 +615,17 @@ dpdk_eth_dev_init(struct netdev_dpdk *dev) 
OVS_REQUIRES(dpdk_mutex)
     n_rxq = MIN(info.max_rx_queues, dev->up.n_rxq);
     n_txq = MIN(info.max_tx_queues, dev->up.n_txq);
 
-    diag = dpdk_eth_dev_queue_setup(dev, n_rxq, n_txq);
+    new_port_conf = port_conf;
+    /* Enable rx checksum offload if it is supported by the NIC */
+    if (info.rx_offload_capa & DEV_RX_OFFLOAD_IPV4_CKSUM) {
+        dev->rx_ol_flags = info.rx_offload_capa &
+                           (DEV_RX_OFFLOAD_UDP_CKSUM |
+                            DEV_RX_OFFLOAD_TCP_CKSUM |
+                            DEV_RX_OFFLOAD_IPV4_CKSUM);
+        new_port_conf.rxmode.hw_ip_checksum = 1;
+    }
+
+    diag = dpdk_eth_dev_queue_setup(dev, n_rxq, n_txq, &new_port_conf);
     if (diag) {
         VLOG_ERR("Interface %s(rxq:%d txq:%d) configure error: %s",
                  dev->up.name, n_rxq, n_txq, rte_strerror(-diag));
@@ -725,6 +742,7 @@ netdev_dpdk_init(struct netdev *netdev, unsigned int 
port_no,
     netdev->n_rxq = NR_QUEUE;
     netdev->requested_n_rxq = NR_QUEUE;
     dev->real_n_txq = NR_QUEUE;
+    dev->rx_ol_flags = 0;
 
     if (type == DPDK_DEV_ETH) {
         netdev_dpdk_alloc_txq(dev, NR_QUEUE);
@@ -2927,3 +2945,10 @@ dpdk_thread_is_pmd(void)
 {
     return rte_lcore_id() != NON_PMD_CORE_ID;
 }
+
+uint32_t
+netdev_get_rx_offload_capa(struct netdev *netdev_) {
+    struct netdev_dpdk *netdev;
+    netdev = netdev_dpdk_cast(netdev_);
+    return netdev->rx_ol_flags;
+}
diff --git a/lib/netdev-dpdk.h b/lib/netdev-dpdk.h
index 646d3e2..341d560 100644
--- a/lib/netdev-dpdk.h
+++ b/lib/netdev-dpdk.h
@@ -19,6 +19,7 @@ struct dp_packet;
 #include <rte_spinlock.h>
 #include <rte_launch.h>
 #include <rte_malloc.h>
+#include "netdev-provider.h"
 
 #define NON_PMD_CORE_ID LCORE_ID_ANY
 
@@ -26,6 +27,12 @@ int dpdk_init(int argc, char **argv);
 void netdev_dpdk_register(void);
 void free_dpdk_buf(struct dp_packet *);
 int pmd_thread_setaffinity_cpu(unsigned cpu);
+uint32_t netdev_get_rx_offload_capa(struct netdev *netdev_);
+
+static inline uint32_t
+get_checksum_ofld_flags(uint32_t rx_ol_flags) {
+    return rx_ol_flags & DEV_RX_OFFLOAD_IPV4_CKSUM;
+}
 
 #else
 
@@ -60,5 +67,15 @@ pmd_thread_setaffinity_cpu(unsigned cpu OVS_UNUSED)
     return 0;
 }
 
+static uint32_t
+netdev_get_rx_offload_capa(struct netdev *netdev_) {
+    return 0;
+}
+
+static inline uint32_t
+get_checksum_ofld_flags(uint32_t rx_ol_flags) {
+    return 0;
+}
+
 #endif /* DPDK_NETDEV */
 #endif
diff --git a/lib/netdev-provider.h b/lib/netdev-provider.h
index cda25eb..8f829e6 100644
--- a/lib/netdev-provider.h
+++ b/lib/netdev-provider.h
@@ -276,7 +276,7 @@ struct netdev_class {
 
     /* Pop tunnel header from packet, build tunnel metadata and resize packet
      * for further processing. */
-    int (*pop_header)(struct dp_packet *packet);
+    int (*pop_header)(struct dp_packet *packet, uint32_t rx_ol_flags);
 
     /* Returns the id of the numa node the 'netdev' is on.  If there is no
      * such info, returns NETDEV_NUMA_UNSPEC. */
diff --git a/lib/netdev-vport.c b/lib/netdev-vport.c
index e398562..feae11b 100644
--- a/lib/netdev-vport.c
+++ b/lib/netdev-vport.c
@@ -873,6 +873,74 @@ ipv6_hdr(void *eth)
 }
 
 static void *
+ip_extract_tnl_md_with_valid_checksum(struct dp_packet *packet,
+                            struct flow_tnl *tnl, unsigned int *hlen)
+{
+    void *nh;
+    struct ip_header *ip;
+    struct ovs_16aligned_ip6_hdr *ip6;
+    void *l4;
+    int l3_size;
+
+    nh = dp_packet_l3(packet);
+    ip = nh;
+    ip6 = nh;
+    l4 = dp_packet_l4(packet);
+
+    if (!nh || !l4) {
+        return NULL;
+    }
+
+    *hlen = sizeof(struct eth_header);
+
+    l3_size = dp_packet_size(packet) -
+                ((char *)nh - (char *)dp_packet_data(packet));
+
+    if (IP_VER(ip->ip_ihl_ver) == 4) {
+
+        ovs_be32 ip_src, ip_dst;
+
+        if (ntohs(ip->ip_tot_len) > l3_size) {
+            VLOG_WARN_RL(&err_rl,
+                         "ip packet is truncated (IP length %d, actual %d)",
+                         ntohs(ip->ip_tot_len), l3_size);
+            return NULL;
+        }
+        if (IP_IHL(ip->ip_ihl_ver) * 4 > sizeof(struct ip_header)) {
+            VLOG_WARN_RL(&err_rl, "ip options not supported on tunnel packets "
+                                  "(%d bytes)", IP_IHL(ip->ip_ihl_ver) * 4);
+            return NULL;
+        }
+
+        ip_src = get_16aligned_be32(&ip->ip_src);
+        ip_dst = get_16aligned_be32(&ip->ip_dst);
+
+        tnl->ip_src = ip_src;
+        tnl->ip_dst = ip_dst;
+        tnl->ip_tos = ip->ip_tos;
+        tnl->ip_ttl = ip->ip_ttl;
+
+        *hlen += IP_HEADER_LEN;
+
+    } else if (IP_VER(ip->ip_ihl_ver) == 6) {
+
+        memcpy(tnl->ipv6_src.s6_addr, ip6->ip6_src.be16, sizeof ip6->ip6_src);
+        memcpy(tnl->ipv6_dst.s6_addr, ip6->ip6_dst.be16, sizeof ip6->ip6_dst);
+        tnl->ip_tos = 0;
+        tnl->ip_ttl = ip6->ip6_hlim;
+
+        *hlen += IPV6_HEADER_LEN;
+
+    } else {
+        VLOG_WARN_RL(&err_rl, "ipv4 packet has invalid version (%d)",
+                                              IP_VER(ip->ip_ihl_ver));
+        return NULL;
+    }
+
+    return l4;
+}
+
+static void *
 ip_extract_tnl_md(struct dp_packet *packet, struct flow_tnl *tnl,
                   unsigned int *hlen)
 {
@@ -989,6 +1057,23 @@ push_ip_header(struct dp_packet *packet,
 }
 
 static void *
+udp_extract_tnl_md_with_valid_checksum(struct dp_packet *packet,
+        struct flow_tnl *tnl, unsigned int *hlen)
+{
+    struct udp_header *udp;
+
+    udp = ip_extract_tnl_md_with_valid_checksum(packet, tnl, hlen);
+    if (!udp) {
+        return NULL;
+    }
+    tnl->flags |= FLOW_TNL_F_CSUM;
+    tnl->tp_src = udp->udp_src;
+    tnl->tp_dst = udp->udp_dst;
+
+    return udp + 1;
+}
+
+static void *
 udp_extract_tnl_md(struct dp_packet *packet, struct flow_tnl *tnl,
                    unsigned int *hlen)
 {
@@ -1183,7 +1268,7 @@ pkt_metadata_init_tnl(struct pkt_metadata *md)
 }
 
 static int
-netdev_gre_pop_header(struct dp_packet *packet)
+netdev_gre_pop_header(struct dp_packet *packet, uint32_t rx_ol_flags)
 {
     struct pkt_metadata *md = &packet->md;
     struct flow_tnl *tnl = &md->tunnel;
@@ -1280,19 +1365,21 @@ netdev_gre_build_header(const struct netdev *netdev,
 }
 
 static int
-netdev_vxlan_pop_header(struct dp_packet *packet)
+netdev_vxlan_pop_header(struct dp_packet *packet, uint32_t rx_ol_flags)
 {
     struct pkt_metadata *md = &packet->md;
     struct flow_tnl *tnl = &md->tunnel;
     struct vxlanhdr *vxh;
     unsigned int hlen;
 
+    rx_ol_flags = get_checksum_ofld_flags(rx_ol_flags);
     pkt_metadata_init_tnl(md);
     if (VXLAN_HLEN > dp_packet_l4_size(packet)) {
         return EINVAL;
     }
 
-    vxh = udp_extract_tnl_md(packet, tnl, &hlen);
+    vxh = rx_ol_flags ? udp_extract_tnl_md_with_valid_checksum(packet, tnl,
+                        &hlen) : udp_extract_tnl_md(packet, tnl, &hlen);
     if (!vxh) {
         return EINVAL;
     }
@@ -1338,7 +1425,7 @@ netdev_vxlan_build_header(const struct netdev *netdev,
 }
 
 static int
-netdev_geneve_pop_header(struct dp_packet *packet)
+netdev_geneve_pop_header(struct dp_packet *packet, uint32_t rx_ol_flags)
 {
     struct pkt_metadata *md = &packet->md;
     struct flow_tnl *tnl = &md->tunnel;
diff --git a/lib/netdev.c b/lib/netdev.c
index 3e50694..d1494fd 100644
--- a/lib/netdev.c
+++ b/lib/netdev.c
@@ -767,7 +767,8 @@ netdev_send(struct netdev *netdev, int qid, struct 
dp_packet **buffers,
 }
 
 int
-netdev_pop_header(struct netdev *netdev, struct dp_packet **buffers, int cnt)
+netdev_pop_header(struct netdev *netdev, struct dp_packet **buffers, int cnt,
+                  uint32_t rx_ol_flags)
 {
     int i;
 
@@ -778,7 +779,7 @@ netdev_pop_header(struct netdev *netdev, struct dp_packet 
**buffers, int cnt)
     for (i = 0; i < cnt; i++) {
         int err;
 
-        err = netdev->netdev_class->pop_header(buffers[i]);
+        err = netdev->netdev_class->pop_header(buffers[i], rx_ol_flags);
         if (err) {
             dp_packet_clear(buffers[i]);
         }
diff --git a/lib/netdev.h b/lib/netdev.h
index 05968b2..6de0381 100644
--- a/lib/netdev.h
+++ b/lib/netdev.h
@@ -193,7 +193,7 @@ int netdev_push_header(const struct netdev *netdev,
                        struct dp_packet **buffers, int cnt,
                        const struct ovs_action_push_tnl *data);
 int netdev_pop_header(struct netdev *netdev, struct dp_packet **buffers,
-                      int cnt);
+                      int cnt, uint32_t rx_ol_flags);
 
 /* Hardware address. */
 int netdev_set_etheraddr(struct netdev *, const struct eth_addr mac);
diff --git a/lib/odp-execute.c b/lib/odp-execute.c
index b5204b2..06a0c30 100644
--- a/lib/odp-execute.c
+++ b/lib/odp-execute.c
@@ -477,7 +477,7 @@ odp_execute_sample(void *dp, struct dp_packet *packet, bool 
steal,
     }
 
     odp_execute_actions(dp, &packet, 1, steal, nl_attr_get(subactions),
-                        nl_attr_get_size(subactions), dp_execute_action);
+                        nl_attr_get_size(subactions), dp_execute_action, 0);
 }
 
 static bool
@@ -516,7 +516,7 @@ requires_datapath_assistance(const struct nlattr *a)
 void
 odp_execute_actions(void *dp, struct dp_packet **packets, int cnt, bool steal,
                     const struct nlattr *actions, size_t actions_len,
-                    odp_execute_cb dp_execute_action)
+                    odp_execute_cb dp_execute_action, uint32_t rx_ol_flags)
 {
     const struct nlattr *a;
     unsigned int left;
@@ -532,7 +532,7 @@ odp_execute_actions(void *dp, struct dp_packet **packets, 
int cnt, bool steal,
                  * not need it any more. */
                 bool may_steal = steal && last_action;
 
-                dp_execute_action(dp, packets, cnt, a, may_steal);
+                dp_execute_action(dp, packets, cnt, a, may_steal, rx_ol_flags);
 
                 if (last_action) {
                     /* We do not need to free the packets. dp_execute_actions()
diff --git a/lib/odp-execute.h b/lib/odp-execute.h
index c602bb4..49e587b 100644
--- a/lib/odp-execute.h
+++ b/lib/odp-execute.h
@@ -28,7 +28,8 @@ struct dp_packet;
 struct pkt_metadata;
 
 typedef void (*odp_execute_cb)(void *dp, struct dp_packet **packets, int cnt,
-                               const struct nlattr *action, bool may_steal);
+                               const struct nlattr *action, bool may_steal,
+                               uint32_t rx_ol_flags);
 
 /* Actions that need to be executed in the context of a datapath are handed
  * to 'dp_execute_action', if non-NULL.  Currently this is called only for
@@ -37,5 +38,6 @@ typedef void (*odp_execute_cb)(void *dp, struct dp_packet 
**packets, int cnt,
 void odp_execute_actions(void *dp, struct dp_packet **packets, int cnt,
                          bool steal,
                          const struct nlattr *actions, size_t actions_len,
-                         odp_execute_cb dp_execute_action);
+                         odp_execute_cb dp_execute_action,
+                         uint32_t rx_ol_flags);
 #endif
diff --git a/ofproto/ofproto-dpif-xlate.c b/ofproto/ofproto-dpif-xlate.c
index a02dc24..848a94a 100644
--- a/ofproto/ofproto-dpif-xlate.c
+++ b/ofproto/ofproto-dpif-xlate.c
@@ -3593,7 +3593,8 @@ execute_controller_action(struct xlate_ctx *ctx, int len,
     packet = dp_packet_clone(ctx->xin->packet);
 
     odp_execute_actions(NULL, &packet, 1, false,
-                        ctx->odp_actions->data, ctx->odp_actions->size, NULL);
+                        ctx->odp_actions->data, ctx->odp_actions->size, NULL,
+                        0);
 
     /* A packet sent by an action in a table-miss rule is considered an
      * explicit table miss.  OpenFlow before 1.3 doesn't have that concept so
-- 
2.5.0

_______________________________________________
dev mailing list
dev@openvswitch.org
http://openvswitch.org/mailman/listinfo/dev

Reply via email to