On 4/11/2019 12:00 PM, Cian Ferriter wrote:
> It can be useful to use pcap files for some rudimental performance
> testing. This patch enables this functionality in the pcap driver.
> 
> At a high level, this works by creaing a ring of sufficient size to
> store the packets in the pcap file passed to the application. When the
> rx function for this mode is called, packets are dequeued from the ring
> for use by the application and also enqueued back on to the ring to be
> "received" again.
> 
> A tx_drop mode is also added since transmitting to a tx_pcap file isn't
> desirable at a high traffic rate.
> 
> Jumbo frames are not supported in this mode. When filling the ring at rx
> queue setup time, the presence of multi segment mbufs is checked for.
> The PMD will exit on detection of these multi segment mbufs.
> 
> Signed-off-by: Cian Ferriter <cian.ferri...@intel.com>
> ---
>  doc/guides/nics/pcap_ring.rst   |  15 +++
>  drivers/net/pcap/rte_eth_pcap.c | 209 ++++++++++++++++++++++++++++++--
>  2 files changed, 215 insertions(+), 9 deletions(-)
> 
> diff --git a/doc/guides/nics/pcap_ring.rst b/doc/guides/nics/pcap_ring.rst
> index c1ef9196b..45f55a345 100644
> --- a/doc/guides/nics/pcap_ring.rst
> +++ b/doc/guides/nics/pcap_ring.rst
> @@ -106,6 +106,21 @@ Runtime Config Options
>  
>     --vdev 'net_pcap0,iface=eth0,phy_mac=1'
>  
> +- Use the RX PCAP file to infinitely receive packets
> +
> + In case ``rx_pcap=`` configuration is set, user may want to use the 
> selected PCAP file for rudimental
> + performance testing. This can be done with a ``devarg`` ``infinite_rx``, 
> for example::
> +
> +   --vdev 'net_pcap0,rx_pcap=file_rx.pcap,infinite_rx=1,tx_drop=1'
> +
> + When this mode is used, it is recommended to use the ``tx_drop`` ``devarg``.
> +
> +- Drop all packets on transmit
> +
> + The user may want to drop all packets on tx for a device. This can be done 
> with the ``tx_drop`` ``devarg``, for example::
> +
> +   --vdev 'net_pcap0,rx_pcap=file_rx.pcap,tx_drop=1'

Is there a performance drop when tx files are directed to 'loopback' interface,
like: --vdev 'net_pcap0,rx_pcap=file_rx.pcap,tx_iface=lo' ?
If no performance drop I am for using 'lo' interface instead of adding new 
devargs.

> +
>  Examples of Usage
>  ^^^^^^^^^^^^^^^^^
>  
> diff --git a/drivers/net/pcap/rte_eth_pcap.c b/drivers/net/pcap/rte_eth_pcap.c
> index 353538f16..b72db973a 100644
> --- a/drivers/net/pcap/rte_eth_pcap.c
> +++ b/drivers/net/pcap/rte_eth_pcap.c
<...>

> @@ -178,6 +186,40 @@ eth_pcap_gather_data(unsigned char *data, struct 
> rte_mbuf *mbuf)
>       }
>  }
>  
> +static uint16_t
> +eth_pcap_rx_infinite(void *queue, struct rte_mbuf **bufs, uint16_t nb_pkts)
> +{
> +     int i;
> +     struct pcap_rx_queue *pcap_q = queue;
> +
> +     if (unlikely(nb_pkts == 0))
> +             return 0;
> +
> +     if (rte_pktmbuf_alloc_bulk(pcap_q->mb_pool, bufs, nb_pkts) != 0)
> +             return 0;
> +
> +     for (i = 0; i < nb_pkts; i++) {
> +             struct rte_mbuf *pcap_buf;
> +             int err = rte_ring_dequeue(pcap_q->pkts, (void **)&pcap_buf);
> +             if (err)
> +                     rte_panic("failed to dequeue from pcap pkts ring\n");

Please don't have any rte_panic() in the driver code, there are a few more in 
patch.

> +
> +             rte_memcpy(rte_pktmbuf_mtod(bufs[i], void *),
> +                             rte_pktmbuf_mtod(pcap_buf, void *),
> +                             pcap_buf->data_len);
> +             bufs[i]->data_len = pcap_buf->data_len;
> +             bufs[i]->pkt_len = pcap_buf->pkt_len;
> +             bufs[i]->port = pcap_q->port_id;
> +
> +             /* enqueue packet back on ring to allow infinite rx */
> +             rte_ring_enqueue(pcap_q->pkts, pcap_buf);
> +     }
> +
> +     pcap_q->rx_stat.pkts += i;

For consistency, can you please update "pcap_q->rx_stat.bytes" too?

> +
> +     return i;
> +}
> +
>  static uint16_t
>  eth_pcap_rx(void *queue, struct rte_mbuf **bufs, uint16_t nb_pkts)
>  {

<...>

> @@ -447,6 +509,24 @@ open_single_rx_pcap(const char *pcap_filename, pcap_t 
> **pcap)
>       return 0;
>  }
>  
> +static uint64_t
> +count_packets_in_pcaps(pcap_t **pcap, struct pcap_rx_queue *pcap_q)
> +{
> +     const u_char *packet;
> +     struct pcap_pkthdr header;
> +     uint64_t pcap_pkt_count;

Compiler is complaining about uninitialized 'pcap_pkt_count'.

> +
> +     while ((packet = pcap_next(*pcap, &header)))
> +             pcap_pkt_count++;

It seems there is no quicker way to get number of packets from a pcap file, if
anyone know a way, comment is welcome.

> +
> +     /* the pcap is reopened so it can be used as normal later */
> +     pcap_close(*pcap);
> +     *pcap = NULL;
> +     open_single_rx_pcap(pcap_q->name, pcap);
> +
> +     return pcap_pkt_count;
> +}
> +
>  static int
>  eth_dev_start(struct rte_eth_dev *dev)
>  {

<...>

> @@ -671,6 +752,49 @@ eth_rx_queue_setup(struct rte_eth_dev *dev,
>       pcap_q->queue_id = rx_queue_id;
>       dev->data->rx_queues[rx_queue_id] = pcap_q;
>  
> +     if (internals->infinite_rx) {
> +             struct pmd_process_private *pp;
> +             char ring_name[NAME_MAX];
> +             uint16_t num_rx = 0;
> +             static uint32_t ring_number;
> +             uint64_t pcap_pkt_count = 0;
> +             pcap_t **pcap;
> +
> +             pp = rte_eth_devices[pcap_q->port_id].process_private;
> +             pcap = &pp->rx_pcap[pcap_q->queue_id];
> +
> +             if (unlikely(*pcap == NULL))
> +                     rte_panic("no packets in pcap to fill ring with\n");
> +
> +             pcap_pkt_count = count_packets_in_pcaps(pcap, pcap_q);
> +
> +             snprintf(ring_name, sizeof(ring_name), "PCAP_RING%" PRIu16,
> +                             ring_number);
> +
> +             pcap_q->pkts = rte_ring_create(ring_name,
> +                             rte_align64pow2(pcap_pkt_count + 1), 0,
> +                             RING_F_SP_ENQ | RING_F_SC_DEQ);
> +             ring_number++;
> +             if (!pcap_q->pkts)
> +                     rte_panic("failed to alloc ring\n");
> +
> +
> +             struct rte_mbuf *bufs[pcap_pkt_count];

Can you please move deceleration to the beginning of the function.

> +             if (rte_pktmbuf_alloc_bulk(mb_pool, bufs, pcap_pkt_count) != 0)
> +                     return 0;

Should this return error?

> +
> +             num_rx = eth_pcap_rx(pcap_q, bufs, pcap_pkt_count);
> +
> +             /* check for multiseg mbufs */
> +             for (i = 0; i < num_rx; i++) {
> +                     if (bufs[i]->nb_segs != 1)
> +                             rte_panic("jumbo frames are not supported in 
> infinite rx mode\n");

This need to be replaced with returning error, and please remember to cleanup in
that case.

> +             }
> +
> +             rte_ring_enqueue_bulk(pcap_q->pkts, (void * const *)bufs,
> +                             num_rx, NULL);
> +     }
> +
>       return 0;
>  }
>  
<...>

> @@ -1132,9 +1285,18 @@ eth_from_pcaps(struct rte_vdev_device *vdev,
>               }
>       }
>  
> -     eth_dev->rx_pkt_burst = eth_pcap_rx;
> +     /* assign rx ops */
> +     if (infinite_rx) {
> +             eth_dev->rx_pkt_burst = eth_pcap_rx_infinite;
> +             internals->infinite_rx = infinite_rx;

Since "infinite_rx" can be 0 or 1, isn't it better to set
"internals->infinite_rx" out of this check, although functionally it will be
same (since internals is all zero by default), logically better I think

> +     } else {
> +             eth_dev->rx_pkt_burst = eth_pcap_rx;
> +     }
>  
> -     if (using_dumpers)
> +     /* assign tx ops */
> +     if (infinite_rx)
> +             eth_dev->tx_pkt_burst = eth_tx_drop;

Shouldn't set 'tx_pkt_burst' with 'infinite_rx' check, I guess intention was
'tx_drop' check.

> +     else if (using_dumpers)
>               eth_dev->tx_pkt_burst = eth_pcap_tx_dumper;
>       else
>               eth_dev->tx_pkt_burst = eth_pcap_tx;

<...>

> @@ -1217,6 +1380,17 @@ pmd_pcap_probe(struct rte_vdev_device *dev)
>       pcaps.num_of_queue = 0;
>  
>       if (is_rx_pcap) {
> +             /*
> +              * We check whether we want to infinitely rx the pcap file
> +              */
> +             if (rte_kvargs_count(kvlist, ETH_PCAP_INFINITE_RX_ARG) == 1) {

What do you think printing a warning if user provided the value more than once?

> +                     ret = rte_kvargs_process(kvlist,
> +                                     ETH_PCAP_INFINITE_RX_ARG,
> +                                     &get_infinite_rx_arg, &infinite_rx);
> +                     if (ret < 0)
> +                             goto free_kvlist;
> +             }
> +
>               ret = rte_kvargs_process(kvlist, ETH_PCAP_RX_PCAP_ARG,
>                               &open_rx_pcap, &pcaps);
>       } else {
> @@ -1228,18 +1402,30 @@ pmd_pcap_probe(struct rte_vdev_device *dev)
>               goto free_kvlist;
>  
>       /*
> -      * We check whether we want to open a TX stream to a real NIC or a
> -      * pcap file
> +      * We check whether we want to open a TX stream to a real NIC,
> +      * a pcap file, or drop packets on tx
>        */
>       is_tx_pcap = rte_kvargs_count(kvlist, ETH_PCAP_TX_PCAP_ARG) ? 1 : 0;
>       dumpers.num_of_queue = 0;
>  
> -     if (is_tx_pcap)
> +     tx_drop = rte_kvargs_count(kvlist, ETH_PCAP_TX_DROP_ARG) ? 1 : 0;

Can drop this line, below already getting the 'tx_drop' value, overwriting this 
one.

> +     if (rte_kvargs_count(kvlist, ETH_PCAP_TX_DROP_ARG) == 1) {
> +             ret = rte_kvargs_process(kvlist, ETH_PCAP_TX_DROP_ARG,
> +                             &get_tx_drop_arg, &tx_drop);
> +             if (ret < 0)
> +                     goto free_kvlist;
> +     }
> +
> +     if (tx_drop) {
> +             /* add dummy queue which counts dropped packets */
> +             ret = add_queue(&dumpers, "dummy", "tx_drop", NULL, NULL);

This will break the multiple queue use case. Instead of adding a single queue,
need to find a way to add multiple queues as user want and all will just ignore
the packtets.

> +     } else if (is_tx_pcap) {
>               ret = rte_kvargs_process(kvlist, ETH_PCAP_TX_PCAP_ARG,
>                               &open_tx_pcap, &dumpers);
> -     else
> +     } else {
>               ret = rte_kvargs_process(kvlist, ETH_PCAP_TX_IFACE_ARG,
>                               &open_tx_iface, &dumpers);
> +     }
>  
>       if (ret < 0)
>               goto free_kvlist;

<...>

Reply via email to