This patch skips calling BPF program in the receive path if
the queue is associated with UMEM that is not shared and
bound to an AF_XDP socket that has enabled skip bpf during
bind() call.

Signed-off-by: Sridhar Samudrala <sridhar.samudr...@intel.com>
---
 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c | 20 +++++++++++++++++--
 drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c  | 16 +++++++++++++--
 2 files changed, 32 insertions(+), 4 deletions(-)

diff --git a/drivers/net/ethernet/intel/ixgbe/ixgbe_main.c 
b/drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
index dc7b128c780e..594792860cdd 100644
--- a/drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
+++ b/drivers/net/ethernet/intel/ixgbe/ixgbe_main.c
@@ -2197,6 +2197,7 @@ static struct sk_buff *ixgbe_run_xdp(struct ixgbe_adapter 
*adapter,
        int err, result = IXGBE_XDP_PASS;
        struct bpf_prog *xdp_prog;
        struct xdp_frame *xdpf;
+       struct xdp_umem *umem;
        u32 act;
 
        rcu_read_lock();
@@ -2207,6 +2208,13 @@ static struct sk_buff *ixgbe_run_xdp(struct 
ixgbe_adapter *adapter,
 
        prefetchw(xdp->data_hard_start); /* xdp_frame write */
 
+       umem = xdp_get_umem_from_qid(rx_ring->netdev, rx_ring->queue_index);
+       if (xsk_umem_skip_bpf(umem)) {
+               err = xsk_umem_rcv(umem, xdp);
+               result = !err ? IXGBE_XDP_REDIR : IXGBE_XDP_CONSUMED;
+               goto xdp_out;
+       }
+
        act = bpf_prog_run_xdp(xdp_prog, xdp);
        switch (act) {
        case XDP_PASS:
@@ -2400,8 +2408,16 @@ static int ixgbe_clean_rx_irq(struct ixgbe_q_vector 
*q_vector,
                total_rx_packets++;
        }
 
-       if (xdp_xmit & IXGBE_XDP_REDIR)
-               xdp_do_flush_map();
+       if (xdp_xmit & IXGBE_XDP_REDIR) {
+               struct xdp_umem *umem;
+
+               umem = xdp_get_umem_from_qid(rx_ring->netdev,
+                                            rx_ring->queue_index);
+               if (xsk_umem_skip_bpf(umem))
+                       xsk_umem_flush(umem);
+               else
+                       xdp_do_flush_map();
+       }
 
        if (xdp_xmit & IXGBE_XDP_TX) {
                struct ixgbe_ring *ring = adapter->xdp_ring[smp_processor_id()];
diff --git a/drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c 
b/drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c
index 6b609553329f..9ea8a769d7a8 100644
--- a/drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c
+++ b/drivers/net/ethernet/intel/ixgbe/ixgbe_xsk.c
@@ -148,6 +148,12 @@ static int ixgbe_run_xdp_zc(struct ixgbe_adapter *adapter,
        struct xdp_frame *xdpf;
        u32 act;
 
+       if (xsk_umem_skip_bpf(rx_ring->xsk_umem)) {
+               err = xsk_umem_rcv(rx_ring->xsk_umem, xdp);
+               result = !err ? IXGBE_XDP_REDIR : IXGBE_XDP_CONSUMED;
+               return result;
+       }
+
        rcu_read_lock();
        xdp_prog = READ_ONCE(rx_ring->xdp_prog);
        act = bpf_prog_run_xdp(xdp_prog, xdp);
@@ -527,8 +533,14 @@ int ixgbe_clean_rx_irq_zc(struct ixgbe_q_vector *q_vector,
                ixgbe_rx_skb(q_vector, skb);
        }
 
-       if (xdp_xmit & IXGBE_XDP_REDIR)
-               xdp_do_flush_map();
+       if (xdp_xmit & IXGBE_XDP_REDIR) {
+               struct xdp_umem *umem = rx_ring->xsk_umem;
+
+               if (xsk_umem_skip_bpf(umem))
+                       xsk_umem_flush(umem);
+               else
+                       xdp_do_flush_map();
+       }
 
        if (xdp_xmit & IXGBE_XDP_TX) {
                struct ixgbe_ring *ring = adapter->xdp_ring[smp_processor_id()];
-- 
2.20.1

Reply via email to