On 4/26/20 4:19 AM, Marvin Liu wrote:
> Previously, virtio split ring vectorized path was enabled by default.
> This is not suitable for everyone because that path dose not follow
> virtio spec. Add new devarg for virtio vectorized path selection. By
> default vectorized path is disabled.
>
> Signed-off-by: Marvin Liu <yong....@intel.com>
> Reviewed-by: Maxime Coquelin <maxime.coque...@redhat.com>
>
> diff --git a/doc/guides/nics/virtio.rst b/doc/guides/nics/virtio.rst
> index 6286286db..902a1f0cf 100644
> --- a/doc/guides/nics/virtio.rst
> +++ b/doc/guides/nics/virtio.rst
> @@ -363,6 +363,13 @@ Below devargs are supported by the PCI virtio driver:
> rte_eth_link_get_nowait function.
> (Default: 10000 (10G))
>
> +#. ``vectorized``:
> +
> + It is used to specify whether virtio device perfer to use vectorized
> path.
s/perfer/prefers/
> + Afterwards, dependencies of vectorized path will be checked in path
> + election.
> + (Default: 0 (disabled))
> +
> Below devargs are supported by the virtio-user vdev:
>
> #. ``path``:
> diff --git a/drivers/net/virtio/virtio_ethdev.c
> b/drivers/net/virtio/virtio_ethdev.c
> index 37766cbb6..0a69a4db1 100644
> --- a/drivers/net/virtio/virtio_ethdev.c
> +++ b/drivers/net/virtio/virtio_ethdev.c
> @@ -48,7 +48,8 @@ static int virtio_dev_allmulticast_disable(struct
> rte_eth_dev *dev);
> static uint32_t virtio_dev_speed_capa_get(uint32_t speed);
> static int virtio_dev_devargs_parse(struct rte_devargs *devargs,
> int *vdpa,
> - uint32_t *speed);
> + uint32_t *speed,
> + int *vectorized);
> static int virtio_dev_info_get(struct rte_eth_dev *dev,
> struct rte_eth_dev_info *dev_info);
> static int virtio_dev_link_update(struct rte_eth_dev *dev,
> @@ -1551,8 +1552,8 @@ set_rxtx_funcs(struct rte_eth_dev *eth_dev)
> eth_dev->rx_pkt_burst = &virtio_recv_pkts_packed;
> }
> } else {
> - if (hw->use_simple_rx) {
> - PMD_INIT_LOG(INFO, "virtio: using simple Rx path on
> port %u",
> + if (hw->use_vec_rx) {
> + PMD_INIT_LOG(INFO, "virtio: using vectorized Rx path on
> port %u",
> eth_dev->data->port_id);
> eth_dev->rx_pkt_burst = virtio_recv_pkts_vec;
> } else if (hw->use_inorder_rx) {
> @@ -1886,6 +1887,7 @@ eth_virtio_dev_init(struct rte_eth_dev *eth_dev)
> {
> struct virtio_hw *hw = eth_dev->data->dev_private;
> uint32_t speed = SPEED_UNKNOWN;
> + int vectorized = 0;
> int ret;
>
> if (sizeof(struct virtio_net_hdr_mrg_rxbuf) > RTE_PKTMBUF_HEADROOM) {
> @@ -1912,7 +1914,7 @@ eth_virtio_dev_init(struct rte_eth_dev *eth_dev)
> return 0;
> }
> ret = virtio_dev_devargs_parse(eth_dev->device->devargs,
> - NULL, &speed);
> + NULL, &speed, &vectorized);
> if (ret < 0)
> return ret;
> hw->speed = speed;
> @@ -1949,6 +1951,11 @@ eth_virtio_dev_init(struct rte_eth_dev *eth_dev)
> if (ret < 0)
> goto err_virtio_init;
>
> + if (vectorized) {
> + if (!vtpci_packed_queue(hw))
> + hw->use_vec_rx = 1;
> + }
> +
> hw->opened = true;
>
> return 0;
> @@ -2021,9 +2028,20 @@ virtio_dev_speed_capa_get(uint32_t speed)
> }
> }
>
> +static int vectorized_check_handler(__rte_unused const char *key,
> + const char *value, void *ret_val)
> +{
> + if (strcmp(value, "1") == 0)
> + *(int *)ret_val = 1;
> + else
> + *(int *)ret_val = 0;
> +
> + return 0;
> +}
>
> #define VIRTIO_ARG_SPEED "speed"
> #define VIRTIO_ARG_VDPA "vdpa"
> +#define VIRTIO_ARG_VECTORIZED "vectorized"
>
>
> static int
> @@ -2045,7 +2063,7 @@ link_speed_handler(const char *key __rte_unused,
>
> static int
> virtio_dev_devargs_parse(struct rte_devargs *devargs, int *vdpa,
> - uint32_t *speed)
> + uint32_t *speed, int *vectorized)
> {
> struct rte_kvargs *kvlist;
> int ret = 0;
> @@ -2081,6 +2099,18 @@ virtio_dev_devargs_parse(struct rte_devargs *devargs,
> int *vdpa,
> }
> }
>
> + if (vectorized &&
> + rte_kvargs_count(kvlist, VIRTIO_ARG_VECTORIZED) == 1) {
> + ret = rte_kvargs_process(kvlist,
> + VIRTIO_ARG_VECTORIZED,
> + vectorized_check_handler, vectorized);
> + if (ret < 0) {
> + PMD_INIT_LOG(ERR, "Failed to parse %s",
> + VIRTIO_ARG_VECTORIZED);
> + goto exit;
> + }
> + }
> +
> exit:
> rte_kvargs_free(kvlist);
> return ret;
> @@ -2092,7 +2122,8 @@ static int eth_virtio_pci_probe(struct rte_pci_driver
> *pci_drv __rte_unused,
> int vdpa = 0;
> int ret = 0;
>
> - ret = virtio_dev_devargs_parse(pci_dev->device.devargs, &vdpa, NULL);
> + ret = virtio_dev_devargs_parse(pci_dev->device.devargs, &vdpa, NULL,
> + NULL);
> if (ret < 0) {
> PMD_INIT_LOG(ERR, "devargs parsing is failed");
> return ret;
> @@ -2257,33 +2288,31 @@ virtio_dev_configure(struct rte_eth_dev *dev)
> return -EBUSY;
> }
>
> - hw->use_simple_rx = 1;
> -
> if (vtpci_with_feature(hw, VIRTIO_F_IN_ORDER)) {
> hw->use_inorder_tx = 1;
> hw->use_inorder_rx = 1;
> - hw->use_simple_rx = 0;
> + hw->use_vec_rx = 0;
> }
>
> if (vtpci_packed_queue(hw)) {
> - hw->use_simple_rx = 0;
> + hw->use_vec_rx = 0;
> hw->use_inorder_rx = 0;
> }
>
> #if defined RTE_ARCH_ARM64 || defined RTE_ARCH_ARM
> if (!rte_cpu_get_flag_enabled(RTE_CPUFLAG_NEON)) {
> - hw->use_simple_rx = 0;
> + hw->use_vec_rx = 0;
> }
> #endif
> if (vtpci_with_feature(hw, VIRTIO_NET_F_MRG_RXBUF)) {
> - hw->use_simple_rx = 0;
> + hw->use_vec_rx = 0;
> }
>
> if (rx_offloads & (DEV_RX_OFFLOAD_UDP_CKSUM |
> DEV_RX_OFFLOAD_TCP_CKSUM |
> DEV_RX_OFFLOAD_TCP_LRO |
> DEV_RX_OFFLOAD_VLAN_STRIP))
> - hw->use_simple_rx = 0;
> + hw->use_vec_rx = 0;
>
> return 0;
> }
> diff --git a/drivers/net/virtio/virtio_pci.h b/drivers/net/virtio/virtio_pci.h
> index bd89357e4..668e688e1 100644
> --- a/drivers/net/virtio/virtio_pci.h
> +++ b/drivers/net/virtio/virtio_pci.h
> @@ -253,7 +253,8 @@ struct virtio_hw {
> uint8_t vlan_strip;
> uint8_t use_msix;
> uint8_t modern;
> - uint8_t use_simple_rx;
> + uint8_t use_vec_rx;
> + uint8_t use_vec_tx;
> uint8_t use_inorder_rx;
> uint8_t use_inorder_tx;
> uint8_t weak_barriers;
> diff --git a/drivers/net/virtio/virtio_rxtx.c
> b/drivers/net/virtio/virtio_rxtx.c
> index e450477e8..84f4cf946 100644
> --- a/drivers/net/virtio/virtio_rxtx.c
> +++ b/drivers/net/virtio/virtio_rxtx.c
> @@ -996,7 +996,7 @@ virtio_dev_rx_queue_setup_finish(struct rte_eth_dev *dev,
> uint16_t queue_idx)
> /* Allocate blank mbufs for the each rx descriptor */
> nbufs = 0;
>
> - if (hw->use_simple_rx) {
> + if (hw->use_vec_rx && !vtpci_packed_queue(hw)) {
> for (desc_idx = 0; desc_idx < vq->vq_nentries;
> desc_idx++) {
> vq->vq_split.ring.avail->ring[desc_idx] = desc_idx;
> @@ -1014,7 +1014,7 @@ virtio_dev_rx_queue_setup_finish(struct rte_eth_dev
> *dev, uint16_t queue_idx)
> &rxvq->fake_mbuf;
> }
>
> - if (hw->use_simple_rx) {
> + if (hw->use_vec_rx && !vtpci_packed_queue(hw)) {
> while (vq->vq_free_cnt >= RTE_VIRTIO_VPMD_RX_REARM_THRESH) {
> virtio_rxq_rearm_vec(rxvq);
> nbufs += RTE_VIRTIO_VPMD_RX_REARM_THRESH;
> diff --git a/drivers/net/virtio/virtio_user_ethdev.c
> b/drivers/net/virtio/virtio_user_ethdev.c
> index 953f00d72..150a8d987 100644
> --- a/drivers/net/virtio/virtio_user_ethdev.c
> +++ b/drivers/net/virtio/virtio_user_ethdev.c
> @@ -525,7 +525,7 @@ virtio_user_eth_dev_alloc(struct rte_vdev_device *vdev)
> */
> hw->use_msix = 1;
> hw->modern = 0;
> - hw->use_simple_rx = 0;
> + hw->use_vec_rx = 0;
> hw->use_inorder_rx = 0;
> hw->use_inorder_tx = 0;
> hw->virtio_user_dev = dev;
> diff --git a/drivers/net/virtio/virtqueue.c b/drivers/net/virtio/virtqueue.c
> index 0b4e3bf3e..ca23180de 100644
> --- a/drivers/net/virtio/virtqueue.c
> +++ b/drivers/net/virtio/virtqueue.c
> @@ -32,7 +32,8 @@ virtqueue_detach_unused(struct virtqueue *vq)
> end = (vq->vq_avail_idx + vq->vq_free_cnt) & (vq->vq_nentries - 1);
>
> for (idx = 0; idx < vq->vq_nentries; idx++) {
> - if (hw->use_simple_rx && type == VTNET_RQ) {
> + if (hw->use_vec_rx && !vtpci_packed_queue(hw) &&
> + type == VTNET_RQ) {
> if (start <= end && idx >= start && idx < end)
> continue;
> if (start > end && (idx >= start || idx < end))
> @@ -97,7 +98,7 @@ virtqueue_rxvq_flush_split(struct virtqueue *vq)
> for (i = 0; i < nb_used; i++) {
> used_idx = vq->vq_used_cons_idx & (vq->vq_nentries - 1);
> uep = &vq->vq_split.ring.used->ring[used_idx];
> - if (hw->use_simple_rx) {
> + if (hw->use_vec_rx) {
> desc_idx = used_idx;
> rte_pktmbuf_free(vq->sw_ring[desc_idx]);
> vq->vq_free_cnt++;
> @@ -121,7 +122,7 @@ virtqueue_rxvq_flush_split(struct virtqueue *vq)
> vq->vq_used_cons_idx++;
> }
>
> - if (hw->use_simple_rx) {
> + if (hw->use_vec_rx) {
> while (vq->vq_free_cnt >= RTE_VIRTIO_VPMD_RX_REARM_THRESH) {
> virtio_rxq_rearm_vec(rxq);
> if (virtqueue_kick_prepare(vq))
>