[dpdk-dev] [PATCH v3 1/4] net/ixgbe: support VLAN strip per queue offloading in PF
VLAN strip is a per queue offloading in PF. With this patch it can be enabled or disabled on any Rx queue in PF. Signed-off-by: Wei Dai --- drivers/net/ixgbe/ixgbe_ethdev.c | 109 +-- drivers/net/ixgbe/ixgbe_ethdev.h | 4 +- drivers/net/ixgbe/ixgbe_pf.c | 5 +- drivers/net/ixgbe/ixgbe_rxtx.c | 1 + drivers/net/ixgbe/ixgbe_rxtx.h | 1 + 5 files changed, 51 insertions(+), 69 deletions(-) diff --git a/drivers/net/ixgbe/ixgbe_ethdev.c b/drivers/net/ixgbe/ixgbe_ethdev.c index 4483258..73755d2 100644 --- a/drivers/net/ixgbe/ixgbe_ethdev.c +++ b/drivers/net/ixgbe/ixgbe_ethdev.c @@ -2001,64 +2001,6 @@ ixgbe_vlan_hw_strip_enable(struct rte_eth_dev *dev, uint16_t queue) ixgbe_vlan_hw_strip_bitmap_set(dev, queue, 1); } -void -ixgbe_vlan_hw_strip_disable_all(struct rte_eth_dev *dev) -{ - struct ixgbe_hw *hw = - IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private); - uint32_t ctrl; - uint16_t i; - struct ixgbe_rx_queue *rxq; - - PMD_INIT_FUNC_TRACE(); - - if (hw->mac.type == ixgbe_mac_82598EB) { - ctrl = IXGBE_READ_REG(hw, IXGBE_VLNCTRL); - ctrl &= ~IXGBE_VLNCTRL_VME; - IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, ctrl); - } else { - /* Other 10G NIC, the VLAN strip can be setup per queue in RXDCTL */ - for (i = 0; i < dev->data->nb_rx_queues; i++) { - rxq = dev->data->rx_queues[i]; - ctrl = IXGBE_READ_REG(hw, IXGBE_RXDCTL(rxq->reg_idx)); - ctrl &= ~IXGBE_RXDCTL_VME; - IXGBE_WRITE_REG(hw, IXGBE_RXDCTL(rxq->reg_idx), ctrl); - - /* record those setting for HW strip per queue */ - ixgbe_vlan_hw_strip_bitmap_set(dev, i, 0); - } - } -} - -void -ixgbe_vlan_hw_strip_enable_all(struct rte_eth_dev *dev) -{ - struct ixgbe_hw *hw = - IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private); - uint32_t ctrl; - uint16_t i; - struct ixgbe_rx_queue *rxq; - - PMD_INIT_FUNC_TRACE(); - - if (hw->mac.type == ixgbe_mac_82598EB) { - ctrl = IXGBE_READ_REG(hw, IXGBE_VLNCTRL); - ctrl |= IXGBE_VLNCTRL_VME; - IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, ctrl); - } else { - /* Other 10G NIC, the VLAN strip can be setup per queue in RXDCTL */ - for (i = 0; i < dev->data->nb_rx_queues; i++) { - rxq = dev->data->rx_queues[i]; - ctrl = IXGBE_READ_REG(hw, IXGBE_RXDCTL(rxq->reg_idx)); - ctrl |= IXGBE_RXDCTL_VME; - IXGBE_WRITE_REG(hw, IXGBE_RXDCTL(rxq->reg_idx), ctrl); - - /* record those setting for HW strip per queue */ - ixgbe_vlan_hw_strip_bitmap_set(dev, i, 1); - } - } -} - static void ixgbe_vlan_hw_extend_disable(struct rte_eth_dev *dev) { @@ -2114,14 +2056,57 @@ ixgbe_vlan_hw_extend_enable(struct rte_eth_dev *dev) */ } +void +ixgbe_vlan_hw_strip_config(struct rte_eth_dev *dev) +{ + struct ixgbe_hw *hw = + IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private); + struct rte_eth_rxmode *rxmode = &dev->data->dev_conf.rxmode; + uint32_t ctrl; + uint16_t i; + struct ixgbe_rx_queue *rxq; + bool on; + + PMD_INIT_FUNC_TRACE(); + + if (hw->mac.type == ixgbe_mac_82598EB) { + if (rxmode->offloads & DEV_RX_OFFLOAD_VLAN_STRIP) { + ctrl = IXGBE_READ_REG(hw, IXGBE_VLNCTRL); + ctrl |= IXGBE_VLNCTRL_VME; + IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, ctrl); + } else { + ctrl = IXGBE_READ_REG(hw, IXGBE_VLNCTRL); + ctrl &= ~IXGBE_VLNCTRL_VME; + IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, ctrl); + } + } else { + /* +* Other 10G NIC, the VLAN strip can be setup +* per queue in RXDCTL +*/ + for (i = 0; i < dev->data->nb_rx_queues; i++) { + rxq = dev->data->rx_queues[i]; + ctrl = IXGBE_READ_REG(hw, IXGBE_RXDCTL(rxq->reg_idx)); + if (rxq->offloads & DEV_RX_OFFLOAD_VLAN_STRIP) { + ctrl |= IXGBE_RXDCTL_VME; + on = TRUE; + } else { + ctrl &= ~IXGBE_RXDCTL_VME; + on = FALSE; + } + IXGBE_WRITE_REG(hw, IXGBE_RXDCTL(rxq->reg_idx), ctrl); + + /* record those setting for HW strip per queue */ + ixgbe_vlan_hw_strip_bitmap_set(dev, i, on); + } + } +} + static int ixgbe_vlan_offload_set(
[dpdk-dev] [PATCH v3 2/4] net/ixgbe: support VLAN strip per queue offloading in VF
VLAN strip is a per queue offloading in VF. With this patch it can be enabled or disabled on any Rx queue in VF. Signed-off-by: Wei Dai --- drivers/net/ixgbe/ixgbe_ethdev.c | 8 +--- 1 file changed, 5 insertions(+), 3 deletions(-) diff --git a/drivers/net/ixgbe/ixgbe_ethdev.c b/drivers/net/ixgbe/ixgbe_ethdev.c index 73755d2..8bb67ba 100644 --- a/drivers/net/ixgbe/ixgbe_ethdev.c +++ b/drivers/net/ixgbe/ixgbe_ethdev.c @@ -5215,15 +5215,17 @@ ixgbevf_vlan_offload_set(struct rte_eth_dev *dev, int mask) { struct ixgbe_hw *hw = IXGBE_DEV_PRIVATE_TO_HW(dev->data->dev_private); + struct ixgbe_rx_queue *rxq; uint16_t i; int on = 0; /* VF function only support hw strip feature, others are not support */ if (mask & ETH_VLAN_STRIP_MASK) { - on = !!(dev->data->dev_conf.rxmode.hw_vlan_strip); - - for (i = 0; i < hw->mac.max_rx_queues; i++) + for (i = 0; i < hw->mac.max_rx_queues; i++) { + rxq = dev->data->rx_queues[i]; + on = !!(rxq->offloads & DEV_RX_OFFLOAD_VLAN_STRIP); ixgbevf_vlan_strip_queue_set(dev, i, on); + } } return 0; -- 2.7.5
[dpdk-dev] [PATCH v3 0/4] net/ixgbe: convert to new offloads API
This patch set adds support of per queue VLAN strip offloading in ixgbe PF and VF. This patch support new offloads API in ixgbe PF and VF. --- v3: Rx header spliting capability is only enabled in #ifdef RTE_HEADER_SPLIT_ENABLE. Tx vector tranmit function only work without any Tx offloads. v2: improve error checking Wei Dai (4): net/ixgbe: support VLAN strip per queue offloading in PF net/ixgbe: support VLAN strip per queue offloading in VF net/ixgbe: convert to new Rx offloads API net/ixgbe: convert to new Tx offloads API drivers/net/ixgbe/ixgbe_ethdev.c | 264 ++ drivers/net/ixgbe/ixgbe_ethdev.h | 4 +- drivers/net/ixgbe/ixgbe_ipsec.c | 13 +- drivers/net/ixgbe/ixgbe_pf.c | 5 +- drivers/net/ixgbe/ixgbe_rxtx.c| 245 --- drivers/net/ixgbe/ixgbe_rxtx.h| 7 + drivers/net/ixgbe/ixgbe_rxtx_vec_common.h | 2 +- drivers/net/ixgbe/ixgbe_rxtx_vec_neon.c | 2 +- 8 files changed, 370 insertions(+), 172 deletions(-) -- 2.7.5
[dpdk-dev] [PATCH v3 3/4] net/ixgbe: convert to new Rx offloads API
Ethdev Rx offloads API has changed since: commit ce17eddefc20 ("ethdev: introduce Rx queue offloads API") This commit support the new Rx offloads API. Signed-off-by: Wei Dai --- drivers/net/ixgbe/ixgbe_ethdev.c | 93 + drivers/net/ixgbe/ixgbe_ipsec.c | 8 +- drivers/net/ixgbe/ixgbe_rxtx.c| 165 +++--- drivers/net/ixgbe/ixgbe_rxtx.h| 3 + drivers/net/ixgbe/ixgbe_rxtx_vec_common.h | 2 +- drivers/net/ixgbe/ixgbe_rxtx_vec_neon.c | 2 +- 6 files changed, 207 insertions(+), 66 deletions(-) diff --git a/drivers/net/ixgbe/ixgbe_ethdev.c b/drivers/net/ixgbe/ixgbe_ethdev.c index 8bb67ba..9437f05 100644 --- a/drivers/net/ixgbe/ixgbe_ethdev.c +++ b/drivers/net/ixgbe/ixgbe_ethdev.c @@ -2105,19 +2105,22 @@ ixgbe_vlan_hw_strip_config(struct rte_eth_dev *dev) static int ixgbe_vlan_offload_set(struct rte_eth_dev *dev, int mask) { + struct rte_eth_rxmode *rxmode; + rxmode = &dev->data->dev_conf.rxmode; + if (mask & ETH_VLAN_STRIP_MASK) { ixgbe_vlan_hw_strip_config(dev); } if (mask & ETH_VLAN_FILTER_MASK) { - if (dev->data->dev_conf.rxmode.hw_vlan_filter) + if (rxmode->offloads & DEV_RX_OFFLOAD_VLAN_FILTER) ixgbe_vlan_hw_filter_enable(dev); else ixgbe_vlan_hw_filter_disable(dev); } if (mask & ETH_VLAN_EXTEND_MASK) { - if (dev->data->dev_conf.rxmode.hw_vlan_extend) + if (rxmode->offloads & DEV_RX_OFFLOAD_VLAN_EXTEND) ixgbe_vlan_hw_extend_enable(dev); else ixgbe_vlan_hw_extend_disable(dev); @@ -2332,6 +2335,8 @@ ixgbe_dev_configure(struct rte_eth_dev *dev) IXGBE_DEV_PRIVATE_TO_INTR(dev->data->dev_private); struct ixgbe_adapter *adapter = (struct ixgbe_adapter *)dev->data->dev_private; + struct rte_eth_dev_info dev_info; + uint64_t rx_offloads; int ret; PMD_INIT_FUNC_TRACE(); @@ -2343,6 +2348,15 @@ ixgbe_dev_configure(struct rte_eth_dev *dev) return ret; } + ixgbe_dev_info_get(dev, &dev_info); + rx_offloads = dev->data->dev_conf.rxmode.offloads; + if ((rx_offloads & dev_info.rx_offload_capa) != rx_offloads) { + PMD_DRV_LOG(ERR, "Some Rx offloads are not supported " + "requested 0x%" PRIx64 " supported 0x%" PRIx64, + rx_offloads, dev_info.rx_offload_capa); + return -ENOTSUP; + } + /* set flag to update link status after init */ intr->flags |= IXGBE_FLAG_NEED_LINK_UPDATE; @@ -3632,30 +3646,9 @@ ixgbe_dev_info_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info) else dev_info->max_vmdq_pools = ETH_64_POOLS; dev_info->vmdq_queue_num = dev_info->max_rx_queues; - dev_info->rx_offload_capa = - DEV_RX_OFFLOAD_VLAN_STRIP | - DEV_RX_OFFLOAD_IPV4_CKSUM | - DEV_RX_OFFLOAD_UDP_CKSUM | - DEV_RX_OFFLOAD_TCP_CKSUM | - DEV_RX_OFFLOAD_CRC_STRIP; - - /* -* RSC is only supported by 82599 and x540 PF devices in a non-SR-IOV -* mode. -*/ - if ((hw->mac.type == ixgbe_mac_82599EB || -hw->mac.type == ixgbe_mac_X540) && - !RTE_ETH_DEV_SRIOV(dev).active) - dev_info->rx_offload_capa |= DEV_RX_OFFLOAD_TCP_LRO; - - if (hw->mac.type == ixgbe_mac_82599EB || - hw->mac.type == ixgbe_mac_X540) - dev_info->rx_offload_capa |= DEV_RX_OFFLOAD_MACSEC_STRIP; - - if (hw->mac.type == ixgbe_mac_X550 || - hw->mac.type == ixgbe_mac_X550EM_x || - hw->mac.type == ixgbe_mac_X550EM_a) - dev_info->rx_offload_capa |= DEV_RX_OFFLOAD_OUTER_IPV4_CKSUM; + dev_info->rx_queue_offload_capa = ixgbe_get_rx_queue_offloads(dev); + dev_info->rx_offload_capa = (ixgbe_get_rx_port_offloads(dev) | +dev_info->rx_queue_offload_capa); dev_info->tx_offload_capa = DEV_TX_OFFLOAD_VLAN_INSERT | @@ -3675,10 +3668,8 @@ ixgbe_dev_info_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info) dev_info->tx_offload_capa |= DEV_TX_OFFLOAD_OUTER_IPV4_CKSUM; #ifdef RTE_LIBRTE_SECURITY - if (dev->security_ctx) { - dev_info->rx_offload_capa |= DEV_RX_OFFLOAD_SECURITY; + if (dev->security_ctx) dev_info->tx_offload_capa |= DEV_TX_OFFLOAD_SECURITY; - } #endif dev_info->default_rxconf = (struct rte_eth_rxconf) { @@ -3689,6 +3680,7 @@ ixgbe_dev_info_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info) }, .rx_free_thresh = IXGBE_DEFAULT_RX_FREE_THRESH, .rx_drop_en = 0, +
[dpdk-dev] [PATCH v3 4/4] net/ixgbe: convert to new Tx offloads API
Ethdev Tx offloads API has changed since: commit cba7f53b717d ("ethdev: introduce Tx queue offloads API") This commit support the new Tx offloads API. Signed-off-by: Wei Dai --- drivers/net/ixgbe/ixgbe_ethdev.c | 56 +--- drivers/net/ixgbe/ixgbe_ipsec.c | 5 ++- drivers/net/ixgbe/ixgbe_rxtx.c | 79 ++-- drivers/net/ixgbe/ixgbe_rxtx.h | 3 ++ 4 files changed, 108 insertions(+), 35 deletions(-) diff --git a/drivers/net/ixgbe/ixgbe_ethdev.c b/drivers/net/ixgbe/ixgbe_ethdev.c index 9437f05..6288690 100644 --- a/drivers/net/ixgbe/ixgbe_ethdev.c +++ b/drivers/net/ixgbe/ixgbe_ethdev.c @@ -2337,6 +2337,7 @@ ixgbe_dev_configure(struct rte_eth_dev *dev) (struct ixgbe_adapter *)dev->data->dev_private; struct rte_eth_dev_info dev_info; uint64_t rx_offloads; + uint64_t tx_offloads; int ret; PMD_INIT_FUNC_TRACE(); @@ -2356,6 +2357,13 @@ ixgbe_dev_configure(struct rte_eth_dev *dev) rx_offloads, dev_info.rx_offload_capa); return -ENOTSUP; } + tx_offloads = dev->data->dev_conf.txmode.offloads; + if ((tx_offloads & dev_info.tx_offload_capa) != tx_offloads) { + PMD_DRV_LOG(ERR, "Some Tx offloads are not supported " + "requested 0x%" PRIx64 " supported 0x%" PRIx64, + tx_offloads, dev_info.tx_offload_capa); + return -ENOTSUP; + } /* set flag to update link status after init */ intr->flags |= IXGBE_FLAG_NEED_LINK_UPDATE; @@ -3649,28 +3657,8 @@ ixgbe_dev_info_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info) dev_info->rx_queue_offload_capa = ixgbe_get_rx_queue_offloads(dev); dev_info->rx_offload_capa = (ixgbe_get_rx_port_offloads(dev) | dev_info->rx_queue_offload_capa); - - dev_info->tx_offload_capa = - DEV_TX_OFFLOAD_VLAN_INSERT | - DEV_TX_OFFLOAD_IPV4_CKSUM | - DEV_TX_OFFLOAD_UDP_CKSUM | - DEV_TX_OFFLOAD_TCP_CKSUM | - DEV_TX_OFFLOAD_SCTP_CKSUM | - DEV_TX_OFFLOAD_TCP_TSO; - - if (hw->mac.type == ixgbe_mac_82599EB || - hw->mac.type == ixgbe_mac_X540) - dev_info->tx_offload_capa |= DEV_TX_OFFLOAD_MACSEC_INSERT; - - if (hw->mac.type == ixgbe_mac_X550 || - hw->mac.type == ixgbe_mac_X550EM_x || - hw->mac.type == ixgbe_mac_X550EM_a) - dev_info->tx_offload_capa |= DEV_TX_OFFLOAD_OUTER_IPV4_CKSUM; - -#ifdef RTE_LIBRTE_SECURITY - if (dev->security_ctx) - dev_info->tx_offload_capa |= DEV_TX_OFFLOAD_SECURITY; -#endif + dev_info->tx_queue_offload_capa = ixgbe_get_tx_queue_offloads(dev); + dev_info->tx_offload_capa = ixgbe_get_tx_port_offloads(dev); dev_info->default_rxconf = (struct rte_eth_rxconf) { .rx_thresh = { @@ -3692,7 +3680,9 @@ ixgbe_dev_info_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info) .tx_free_thresh = IXGBE_DEFAULT_TX_FREE_THRESH, .tx_rs_thresh = IXGBE_DEFAULT_TX_RSBIT_THRESH, .txq_flags = ETH_TXQ_FLAGS_NOMULTSEGS | - ETH_TXQ_FLAGS_NOOFFLOADS, +ETH_TXQ_FLAGS_NOOFFLOADS | +ETH_TXQ_FLAGS_IGNORE, + .offloads = 0, }; dev_info->rx_desc_lim = rx_desc_lim; @@ -3776,12 +3766,8 @@ ixgbevf_dev_info_get(struct rte_eth_dev *dev, dev_info->rx_queue_offload_capa = ixgbe_get_rx_queue_offloads(dev); dev_info->rx_offload_capa = (ixgbe_get_rx_port_offloads(dev) | dev_info->rx_queue_offload_capa); - dev_info->tx_offload_capa = DEV_TX_OFFLOAD_VLAN_INSERT | - DEV_TX_OFFLOAD_IPV4_CKSUM | - DEV_TX_OFFLOAD_UDP_CKSUM | - DEV_TX_OFFLOAD_TCP_CKSUM | - DEV_TX_OFFLOAD_SCTP_CKSUM | - DEV_TX_OFFLOAD_TCP_TSO; + dev_info->tx_queue_offload_capa = ixgbe_get_tx_queue_offloads(dev); + dev_info->tx_offload_capa = ixgbe_get_tx_port_offloads(dev); dev_info->default_rxconf = (struct rte_eth_rxconf) { .rx_thresh = { @@ -3803,7 +3789,9 @@ ixgbevf_dev_info_get(struct rte_eth_dev *dev, .tx_free_thresh = IXGBE_DEFAULT_TX_FREE_THRESH, .tx_rs_thresh = IXGBE_DEFAULT_TX_RSBIT_THRESH, .txq_flags = ETH_TXQ_FLAGS_NOMULTSEGS | - ETH_TXQ_FLAGS_NOOFFLOADS, +ETH_TXQ_FLAGS_NOOFFLOADS | +ETH_TXQ_FLAGS_IGNORE, + .offloads = 0, }; dev_info->rx_desc_lim = rx_desc_lim; @@ -4941,6 +4929,7 @@ ixgbevf_dev_configure(struct r
[dpdk-dev] [PATCH] net/sfc: fix mbuf data alignment calculation
Unlike ffs() rte_bsf32() counts bit position from 0. Fixes: 0c7a0c35f24c ("net/sfc: calculate Rx buffer size which may be used") Cc: sta...@dpdk.org Signed-off-by: Andrew Rybchenko --- drivers/net/sfc/sfc_rx.c | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/drivers/net/sfc/sfc_rx.c b/drivers/net/sfc/sfc_rx.c index abc53fb..64bc9d4 100644 --- a/drivers/net/sfc/sfc_rx.c +++ b/drivers/net/sfc/sfc_rx.c @@ -887,7 +887,7 @@ sfc_rx_mbuf_data_alignment(struct rte_mempool *mb_pool) order = MIN(order, rte_bsf32(data_off)); - return 1u << (order - 1); + return 1u << order; } static uint16_t -- 2.7.4
Re: [dpdk-dev] [PATCH 01/17] net/virtio: vring init for packed queues
On Fri, Mar 16, 2018 at 04:21:04PM +0100, Jens Freimann wrote: > Add and initialize descriptor data structures. > > Signed-off-by: Jens Freimann > --- [...] > --- a/drivers/net/virtio/virtio_ring.h > +++ b/drivers/net/virtio/virtio_ring.h > @@ -9,6 +9,7 @@ > > #include > > + There is no need to add this blank line. > /* This marks a buffer as continuing via the next field. */ > #define VRING_DESC_F_NEXT 1 > /* This marks a buffer as write-only (otherwise read-only). */ > @@ -54,11 +55,23 @@ struct vring_used { > struct vring_used_elem ring[0]; > }; [...] > struct vring { > unsigned int num; > struct vring_desc *desc; > struct vring_avail *avail; > struct vring_used *used; > + struct vring_desc_packed *desc_packed; Maybe it's better to use anonymous union here. > }; > > /* The standard layout for the ring is a continuous chunk of memory which > @@ -95,10 +108,13 @@ struct vring { > #define vring_avail_event(vr) (*(uint16_t *)&(vr)->used->ring[(vr)->num]) > > static inline size_t > -vring_size(unsigned int num, unsigned long align) > +vring_size(struct virtio_hw *hw, unsigned int num, unsigned long align) > { > size_t size; > > + if (vtpci_packed_queue(hw)) > + return num * sizeof(struct vring_desc_packed); > + Besides the descriptors, the ring also contains event suppression structures. Thanks
Re: [dpdk-dev] [PATCH 02/17] net/virtio: don't call virtio_disable_intr() for packed queues
On Fri, Mar 16, 2018 at 04:21:05PM +0100, Jens Freimann wrote: > When VIRTIO_F_PACKED_RING is set, don't call virtio_disable_intr(). > This function accesses data structures which are not > available when packed virtqueues are enabled. Packed ring has event suppression structures. Thanks > > Signed-off-by: Jens Freimann > --- > drivers/net/virtio/virtio_ethdev.c | 12 ++-- > 1 file changed, 6 insertions(+), 6 deletions(-) > > diff --git a/drivers/net/virtio/virtio_ethdev.c > b/drivers/net/virtio/virtio_ethdev.c > index ce4f53d..19536eb 100644 > --- a/drivers/net/virtio/virtio_ethdev.c > +++ b/drivers/net/virtio/virtio_ethdev.c > @@ -313,12 +313,11 @@ struct rte_virtio_xstats_name_off { > vq->vq_desc_tail_idx = (uint16_t)(vq->vq_nentries - 1); > > vring_desc_init(vr->desc, size); > + /* > + * Disable device(host) interrupting guest > + */ > + virtqueue_disable_intr(vq); > } > - > - /* > - * Disable device(host) interrupting guest > - */ > - virtqueue_disable_intr(vq); > } > > static int > @@ -736,7 +735,8 @@ struct rte_virtio_xstats_name_off { > struct virtnet_rx *rxvq = dev->data->rx_queues[queue_id]; > struct virtqueue *vq = rxvq->vq; > > - virtqueue_disable_intr(vq); > + if (!vtpci_packed_queue(vq->hw)) > + virtqueue_disable_intr(vq); > return 0; > } > > -- > 1.8.3.1 >
Re: [dpdk-dev] [PATCH 03/17] net/virtio: add virtio 1.1 defines
On Fri, Mar 16, 2018 at 04:21:06PM +0100, Jens Freimann wrote: > Signed-off-by: Jens Freimann > --- > drivers/net/virtio/virtio_ring.h | 6 ++ > 1 file changed, 6 insertions(+) > > diff --git a/drivers/net/virtio/virtio_ring.h > b/drivers/net/virtio/virtio_ring.h > index fc45e34..6eb0077 100644 > --- a/drivers/net/virtio/virtio_ring.h > +++ b/drivers/net/virtio/virtio_ring.h > @@ -17,6 +17,12 @@ > /* This means the buffer contains a list of buffer descriptors. */ > #define VRING_DESC_F_INDIRECT 4 > > +#define VIRTQ_DESC_F_AVAIL 7 > +#define VIRTQ_DESC_F_USED 15 It's better to use consistent prefix (VRING_DESC_F_ instead of VIRTQ_DESC_F_). Besides, maybe it's better to define them as (1ULL << 7) and (1ULL << 15) to make them consistent with the definitions of other VRING_DESC_F_* flags. > +#define DESC_USED (1ULL << VIRTQ_DESC_F_USED) > +#define DESC_AVAIL (1ULL << VIRTQ_DESC_F_AVAIL) > + > + There is no need to add two blank lines here. Thanks > /* The Host uses this in used->flags to advise the Guest: don't kick me > * when you add a buffer. It's unreliable, so it's simply an > * optimization. Guest will still kick if it's out of buffers. */ > -- > 1.8.3.1 >
Re: [dpdk-dev] [PATCH 04/17] net/virtio: add packed virtqueue helpers
On Fri, Mar 16, 2018 at 04:21:07PM +0100, Jens Freimann wrote: > Add helper functions to set/clear and check descriptor flags. > > Signed-off-by: Jens Freimann > --- [...] > --- a/drivers/net/virtio/virtio_ring.h > +++ b/drivers/net/virtio/virtio_ring.h > @@ -74,12 +74,45 @@ struct vring_desc_packed { > > struct vring { > unsigned int num; > + unsigned int avail_wrap_counter; > struct vring_desc *desc; > struct vring_avail *avail; > struct vring_used *used; > struct vring_desc_packed *desc_packed; Maybe it's better to use anonymous union. > }; [...] > diff --git a/drivers/net/virtio/virtqueue.c b/drivers/net/virtio/virtqueue.c > index a7d0a9c..6806056 100644 > --- a/drivers/net/virtio/virtqueue.c > +++ b/drivers/net/virtio/virtqueue.c > @@ -65,6 +65,9 @@ struct rte_mbuf * Please make sure the diff contains function name. > uint16_t used_idx, desc_idx; > uint16_t nb_used, i; > > + if (vtpci_packed_queue(vq->hw)) > + return; I guess packed-ring also needs to support virtqueue_rxvq_flush(). Thanks > + > nb_used = VIRTQUEUE_NUSED(vq); > > for (i = 0; i < nb_used; i++) { > -- > 1.8.3.1 >
Re: [dpdk-dev] [PATCH 05/17] net/virtio: don't dump split virtqueue data
On Fri, Mar 16, 2018 at 04:21:08PM +0100, Jens Freimann wrote: > VIRTQUEUE_DUMP access split virtqueue data which is not > correct when packed virtqueues are used. > > Signed-off-by: Jens Freimann > --- > drivers/net/virtio/virtqueue.h | 4 +++- > 1 file changed, 3 insertions(+), 1 deletion(-) > > diff --git a/drivers/net/virtio/virtqueue.h b/drivers/net/virtio/virtqueue.h > index cc2e7c0..82160ca 100644 > --- a/drivers/net/virtio/virtqueue.h > +++ b/drivers/net/virtio/virtqueue.h > @@ -359,7 +359,9 @@ struct virtio_tx_region { > } > > #ifdef RTE_LIBRTE_VIRTIO_DEBUG_DUMP > -#define VIRTQUEUE_DUMP(vq) do { \ > +#define VIRTQUEUE_DUMP(vq) \ > + do { \ > + if (vtpci_packed_queue((vq)->hw)) break; \ Maybe it's better to make VIRTQUEUE_DUMP() support packed ring. Thanks > uint16_t used_idx, nused; \ > used_idx = (vq)->vq_ring.used->idx; \ > nused = (uint16_t)(used_idx - (vq)->vq_used_cons_idx); \ > -- > 1.8.3.1 >
Re: [dpdk-dev] [PATCH] net/mlx5: add supported hash function check
On Sun, Mar 18, 2018 at 03:37:20PM +0800, Xueming Li wrote: > Add supported RSS hash function check in device configuration to > have better error verbosity for application developers. > > Signed-off-by: Xueming Li > --- > drivers/net/mlx5/mlx5_ethdev.c | 8 > 1 file changed, 8 insertions(+) > > diff --git a/drivers/net/mlx5/mlx5_ethdev.c b/drivers/net/mlx5/mlx5_ethdev.c > index b73cb53..175a1ff 100644 > --- a/drivers/net/mlx5/mlx5_ethdev.c > +++ b/drivers/net/mlx5/mlx5_ethdev.c > @@ -346,6 +346,14 @@ struct ethtool_link_settings { > rx_offloads, supp_rx_offloads); > return ENOTSUP; > } > + if (dev->data->dev_conf.rx_adv_conf.rss_conf.rss_hf & > + MLX5_RSS_HF_MASK) { > + ERROR("Some RSS hash function not supported " > + "requested 0x%" PRIx64 " supported 0x%" PRIx64, > + dev->data->dev_conf.rx_adv_conf.rss_conf.rss_hf, > + (uint64_t)(~MLX5_RSS_HF_MASK)); > + return ENOTSUP; > + } > if (use_app_rss_key && > (dev->data->dev_conf.rx_adv_conf.rss_conf.rss_key_len != >rss_hash_default_key_len)) { > -- > 1.8.3.1 > I would answer than an application should not try to configure something not advertise by the device. This information is present in struct rte_eth_dev_info returned by mlx5_dev_infos_get() and thus the devops of the device. Seems rte_eth_dev_configure() should be fixed to avoid configuring wrong values. Regards, -- Nélio Laranjeiro 6WIND
Re: [dpdk-dev] [PATCH 06/17] net/virtio-user: add option to use packed queues
On Fri, Mar 16, 2018 at 04:21:09PM +0100, Jens Freimann wrote: [...] > diff --git a/drivers/net/virtio/virtio_user_ethdev.c > b/drivers/net/virtio/virtio_user_ethdev.c > index 2636490..ee291b3 100644 > --- a/drivers/net/virtio/virtio_user_ethdev.c > +++ b/drivers/net/virtio/virtio_user_ethdev.c > @@ -278,6 +278,8 @@ > VIRTIO_USER_ARG_QUEUE_SIZE, > #define VIRTIO_USER_ARG_INTERFACE_NAME "iface" > VIRTIO_USER_ARG_INTERFACE_NAME, > +#define VIRTIO_USER_ARG_VERSION_1_1 "version_1_1" > + VIRTIO_USER_ARG_VERSION_1_1, Maybe we can enable packed-ring by default for virtio-user. If we really need a flag to enable it, the devarg name should be packed_ring instead of version_1_1. Thanks
Re: [dpdk-dev] [PATCH v2 00/41] Memory Hotplug for DPDK
Hi Anatoly, On Wed, Mar 7, 2018 at 10:26 PM, Anatoly Burakov wrote: > This patchset introduces dynamic memory allocation for DPDK (aka memory > hotplug). Based upon RFC submitted in December [1]. > > Dependencies (to be applied in specified order): > - IPC bugfixes patchset [2] > - IPC improvements patchset [3] > - IPC asynchronous request API patch [4] > - Function to return number of sockets [5] > > Deprecation notices relevant to this patchset: > - General outline of memory hotplug changes [6] > - EAL NUMA node count changes [7] > > The vast majority of changes are in the EAL and malloc, the external API > disruption is minimal: a new set of API's are added for contiguous memory > allocation for rte_memzone, and a few API additions in rte_memory due to > switch to memseg_lists as opposed to memsegs. Every other API change is > internal to EAL, and all of the memory allocation/freeing is handled > through rte_malloc, with no externally visible API changes. > > Quick outline of all changes done as part of this patchset: > > * Malloc heap adjusted to handle holes in address space > * Single memseg list replaced by multiple memseg lists > * VA space for hugepages is preallocated in advance > * Added alloc/free for pages happening as needed on rte_malloc/rte_free > * Added contiguous memory allocation API's for rte_memzone > * Integrated Pawel Wodkowski's patch for registering/unregistering memory >with VFIO [8] > * Callbacks for registering memory allocations > * Multiprocess support done via DPDK IPC introduced in 18.02 > > The biggest difference is a "memseg" now represents a single page (as opposed > to > being a big contiguous block of pages). As a consequence, both memzones and > malloc elements are no longer guaranteed to be physically contiguous, unless > the user asks for it at reserve time. To preserve whatever functionality that > was dependent on previous behavior, a legacy memory option is also provided, > however it is expected (or perhaps vainly hoped) to be temporary solution. > > Why multiple memseg lists instead of one? Since memseg is a single page now, > the list of memsegs will get quite big, and we need to locate pages somehow > when we allocate and free them. We could of course just walk the list and > allocate one contiguous chunk of VA space for memsegs, but this > implementation uses separate lists instead in order to speed up many > operations with memseg lists. > > For v1 and v2, the following limitations are present: > - FreeBSD does not even compile, let alone run > - No 32-bit support I just read on annou...@dpdk.org [1] that an early merge to this series is expected. So, would this limitation be fixed before merge? Or, has it already been fixed in github repo? [1] http://dpdk.org/ml/archives/announce/2018-March/000182.html [...] - Shreyansh
[dpdk-dev] [PATCH v1] net/mlx4: disable CRC stripping
Previous to this commit mlx4 CRC stripping was executed by default and there was no verbs API to disable it. Since OFED version 4.3-1.5.0.0 the API query_device_ex() indicates if CRC stripping capability is supported in HW and if so CRC stripping can be disabled during WQ initialization. This commit uses these new APIs to allow disabling CRC stripping through rte configuration. A user can specify --disable-crc-stipping in testpmd command line or 'port config all crc-strip off' in testpmd interactive shell. Signed-off-by: Ophir Munk --- drivers/net/mlx4/mlx4.c | 4 drivers/net/mlx4/mlx4.h | 1 + drivers/net/mlx4/mlx4_rxq.c | 33 +++-- drivers/net/mlx4/mlx4_rxtx.h | 1 + 4 files changed, 37 insertions(+), 2 deletions(-) diff --git a/drivers/net/mlx4/mlx4.c b/drivers/net/mlx4/mlx4.c index ee93daf..95e01a6 100644 --- a/drivers/net/mlx4/mlx4.c +++ b/drivers/net/mlx4/mlx4.c @@ -578,6 +578,10 @@ mlx4_pci_probe(struct rte_pci_driver *pci_drv, struct rte_pci_device *pci_dev) } DEBUG("supported RSS hash fields mask: %016" PRIx64, priv->hw_rss_sup); + priv->hw_fcs_strip = !!(device_attr_ex.raw_packet_caps & +IBV_RAW_PACKET_CAP_SCATTER_FCS); + DEBUG("FCS stripping configuration is %ssupported", + (priv->hw_fcs_strip ? "" : "not ")); /* Configure the first MAC address by default. */ if (mlx4_get_mac(priv, &mac.addr_bytes)) { ERROR("cannot get MAC address, is mlx4_en loaded?" diff --git a/drivers/net/mlx4/mlx4.h b/drivers/net/mlx4/mlx4.h index 19c8a22..7de896a 100644 --- a/drivers/net/mlx4/mlx4.h +++ b/drivers/net/mlx4/mlx4.h @@ -105,6 +105,7 @@ struct priv { uint32_t isolated:1; /**< Toggle isolated mode. */ uint32_t hw_csum:1; /**< Checksum offload is supported. */ uint32_t hw_csum_l2tun:1; /**< Checksum support for L2 tunnels. */ + uint32_t hw_fcs_strip:1; /**< FCS stripping is supported. */ uint64_t hw_rss_sup; /**< Supported RSS hash fields (Verbs format). */ struct rte_intr_handle intr_handle; /**< Port interrupt handle. */ struct mlx4_drop *drop; /**< Shared resources for drop flow rules. */ diff --git a/drivers/net/mlx4/mlx4_rxq.c b/drivers/net/mlx4/mlx4_rxq.c index 7a036ed..6748355 100644 --- a/drivers/net/mlx4/mlx4_rxq.c +++ b/drivers/net/mlx4/mlx4_rxq.c @@ -491,6 +491,8 @@ mlx4_rxq_attach(struct rxq *rxq) const char *msg; struct ibv_cq *cq = NULL; struct ibv_wq *wq = NULL; + unsigned int create_flags = 0; + unsigned int comp_mask = 0; volatile struct mlx4_wqe_data_seg (*wqes)[]; unsigned int i; int ret; @@ -503,6 +505,11 @@ mlx4_rxq_attach(struct rxq *rxq) msg = "CQ creation failure"; goto error; } + /* By default, FCS (CRC) is stripped by hardware. */ + if (rxq->crc_present) { + create_flags |= IBV_WQ_FLAGS_SCATTER_FCS; + comp_mask |= IBV_WQ_INIT_ATTR_FLAGS; + } wq = mlx4_glue->create_wq (priv->ctx, &(struct ibv_wq_init_attr){ @@ -511,6 +518,8 @@ mlx4_rxq_attach(struct rxq *rxq) .max_sge = sges_n, .pd = priv->pd, .cq = cq, + .comp_mask = comp_mask, + .create_flags = create_flags, }); if (!wq) { ret = errno ? errno : EINVAL; @@ -649,9 +658,10 @@ mlx4_rxq_detach(struct rxq *rxq) uint64_t mlx4_get_rx_queue_offloads(struct priv *priv) { - uint64_t offloads = DEV_RX_OFFLOAD_SCATTER | - DEV_RX_OFFLOAD_CRC_STRIP; + uint64_t offloads = DEV_RX_OFFLOAD_SCATTER; + if (priv->hw_fcs_strip) + offloads |= DEV_RX_OFFLOAD_CRC_STRIP; if (priv->hw_csum) offloads |= DEV_RX_OFFLOAD_CHECKSUM; return offloads; @@ -781,6 +791,24 @@ mlx4_rx_queue_setup(struct rte_eth_dev *dev, uint16_t idx, uint16_t desc, (void *)dev, idx); return -rte_errno; } + /* By default, FCS (CRC) is stripped by hardware. */ + unsigned char crc_present; + if (conf->offloads & DEV_RX_OFFLOAD_CRC_STRIP) { + crc_present = 0; + } else if (priv->hw_fcs_strip) { + crc_present = 1; + } else { + WARN("%p: CRC stripping has been disabled but will still" +" be performed by hardware, make sure MLNX_OFED and" +" firmware are up to date", +(void *)dev); + crc_present = 0; + } + DEBUG("%p: CRC stripping is %s, %u bytes will be subtracted from" + " incoming frames to hide it", + (void *)dev, + crc_present ? "
Re: [dpdk-dev] [PATCH 07/17] net/virtio: implement transmit path for packed queues
On Fri, Mar 16, 2018 at 04:21:10PM +0100, Jens Freimann wrote: [...] > diff --git a/drivers/net/virtio/Makefile b/drivers/net/virtio/Makefile > index 6c2c996..aa1e600 100644 > --- a/drivers/net/virtio/Makefile > +++ b/drivers/net/virtio/Makefile > @@ -28,6 +28,7 @@ LIBABIVER := 1 > SRCS-$(CONFIG_RTE_LIBRTE_VIRTIO_PMD) += virtqueue.c > SRCS-$(CONFIG_RTE_LIBRTE_VIRTIO_PMD) += virtio_pci.c > SRCS-$(CONFIG_RTE_LIBRTE_VIRTIO_PMD) += virtio_rxtx.c > +SRCS-$(CONFIG_RTE_LIBRTE_VIRTIO_PMD) += virtio_rxtx_1.1.c There is no need to introduce this file just for Tx. > SRCS-$(CONFIG_RTE_LIBRTE_VIRTIO_PMD) += virtio_ethdev.c > SRCS-$(CONFIG_RTE_LIBRTE_VIRTIO_PMD) += virtio_rxtx_simple.c [...] > @@ -603,7 +605,8 @@ struct rte_virtio_xstats_name_off { > } > > vtpci_reset(hw); > - virtio_dev_free_mbufs(dev); > + if (!vtpci_packed_queue(hw)) > + virtio_dev_free_mbufs(dev); I think we also need to free mbufs for packed ring. > virtio_free_queues(hw); > } [...] > +/* Cleanup from completed transmits. */ > +static void > +virtio_xmit_cleanup(struct virtqueue *vq) > +{ > + uint16_t idx; > + uint16_t size = vq->vq_nentries; > + struct vring_desc_packed *desc = vq->vq_ring.desc_packed; > + > + idx = vq->vq_used_cons_idx & (size - 1); > + while (desc_is_used(&desc[idx]) && > +vq->vq_free_cnt < size) { > + while (desc[idx].flags & VRING_DESC_F_NEXT) { We can't use VRING_DESC_F_NEXT when handling used descriptors. Thanks
Re: [dpdk-dev] [RFC 0/5] Introduce Intel FPGA BUS
Hi Rosen, On Thu, Mar 15, 2018 at 03:11:06PM +0800, Rosen Xu wrote: > Intel FPGA BUS in DPDK > - > > RFC [1]: > http://www.dpdk.org/ml/archives/dev/2018-March/092297.html > http://www.dpdk.org/ml/archives/dev/2018-March/092298.html > http://www.dpdk.org/ml/archives/dev/2018-March/092299.html > http://www.dpdk.org/ml/archives/dev/2018-March/092300.html > http://www.dpdk.org/ml/archives/dev/2018-March/092301.html > Have you had the time to read the section titled "Contribute by sending patches" of the Development page from dpdk.org [1]? You need to generate the new thread with an increased version number (-v option for git format-patch), generate the patchset using --thread, and send the new version as a reply to the origin thread cover letter. These points are important for working over emails and allowing us to have a clear view of your work. [1]: http://dpdk.org/dev Regards, -- Gaëtan Rivet 6WIND
Re: [dpdk-dev] [RFC 0/5] Introduce Intel FPGA BUS
Got, many thanks for your reminder. -Original Message- From: Gaëtan Rivet [mailto:gaetan.ri...@6wind.com] Sent: Monday, March 19, 2018 17:07 To: Xu, Rosen Cc: dev@dpdk.org; Doherty, Declan ; Richardson, Bruce ; shreyansh.j...@nxp.com; Zhang, Tianfei ; Wu, Hao Subject: Re: [RFC 0/5] Introduce Intel FPGA BUS Hi Rosen, On Thu, Mar 15, 2018 at 03:11:06PM +0800, Rosen Xu wrote: > Intel FPGA BUS in DPDK > - > > RFC [1]: > http://www.dpdk.org/ml/archives/dev/2018-March/092297.html > http://www.dpdk.org/ml/archives/dev/2018-March/092298.html > http://www.dpdk.org/ml/archives/dev/2018-March/092299.html > http://www.dpdk.org/ml/archives/dev/2018-March/092300.html > http://www.dpdk.org/ml/archives/dev/2018-March/092301.html > Have you had the time to read the section titled "Contribute by sending patches" of the Development page from dpdk.org [1]? You need to generate the new thread with an increased version number (-v option for git format-patch), generate the patchset using --thread, and send the new version as a reply to the origin thread cover letter. These points are important for working over emails and allowing us to have a clear view of your work. [1]: http://dpdk.org/dev Regards, -- Gaëtan Rivet 6WIND
Re: [dpdk-dev] [PATCH v1] net/mlx5: fix existing file removal
Friday, March 16, 2018 5:38 PM, Nélio Laranjeiro: > On Fri, Mar 16, 2018 at 11:22:27PM +0800, Xueming Li wrote: > > There is no guarantee that the file won't be removed by external > > user/application between the stat() and remove() syscalls, remove() > > will fail if the file no longer exists. > > > > Fixes: f8b9a3bad467 ("net/mlx5: install a socket to exchange a file > > descriptor") > > > > Signed-off-by: Xueming Li > > Acked-by: Nelio Laranjeiro Applied to next-net-mlx, thanks.
Re: [dpdk-dev] [PATCH 07/17] net/virtio: implement transmit path for packed queues
On Mon, Mar 19, 2018 at 05:04:43PM +0800, Tiwei Bie wrote: On Fri, Mar 16, 2018 at 04:21:10PM +0100, Jens Freimann wrote: [...] diff --git a/drivers/net/virtio/Makefile b/drivers/net/virtio/Makefile index 6c2c996..aa1e600 100644 --- a/drivers/net/virtio/Makefile +++ b/drivers/net/virtio/Makefile @@ -28,6 +28,7 @@ LIBABIVER := 1 SRCS-$(CONFIG_RTE_LIBRTE_VIRTIO_PMD) += virtqueue.c SRCS-$(CONFIG_RTE_LIBRTE_VIRTIO_PMD) += virtio_pci.c SRCS-$(CONFIG_RTE_LIBRTE_VIRTIO_PMD) += virtio_rxtx.c +SRCS-$(CONFIG_RTE_LIBRTE_VIRTIO_PMD) += virtio_rxtx_1.1.c There is no need to introduce this file just for Tx. I agree, this is a leftover from the prototype. I'll merge it into virtio_rxtx.c SRCS-$(CONFIG_RTE_LIBRTE_VIRTIO_PMD) += virtio_ethdev.c SRCS-$(CONFIG_RTE_LIBRTE_VIRTIO_PMD) += virtio_rxtx_simple.c [...] @@ -603,7 +605,8 @@ struct rte_virtio_xstats_name_off { } vtpci_reset(hw); - virtio_dev_free_mbufs(dev); + if (!vtpci_packed_queue(hw)) + virtio_dev_free_mbufs(dev); I think we also need to free mbufs for packed ring. yes, will fix virtio_free_queues(hw); } [...] +/* Cleanup from completed transmits. */ +static void +virtio_xmit_cleanup(struct virtqueue *vq) +{ + uint16_t idx; + uint16_t size = vq->vq_nentries; + struct vring_desc_packed *desc = vq->vq_ring.desc_packed; + + idx = vq->vq_used_cons_idx & (size - 1); + while (desc_is_used(&desc[idx]) && + vq->vq_free_cnt < size) { + while (desc[idx].flags & VRING_DESC_F_NEXT) { We can't use VRING_DESC_F_NEXT when handling used descriptors. Already fixes this but must have been lost in a rebase. Will send it in v3. thanks for the review! regards, Jens Thanks
[dpdk-dev] [PATCH 1/2] drivers: add common folder
Add driver/common folder and skeleton makefile for adding commonly used functions across mempool, event and net devices. Signed-off-by: Pavan Nikhilesh --- drivers/Makefile | 14 -- drivers/common/Makefile| 7 +++ drivers/common/meson.build | 6 ++ drivers/meson.build| 9 + 4 files changed, 26 insertions(+), 10 deletions(-) create mode 100644 drivers/common/Makefile create mode 100644 drivers/common/meson.build diff --git a/drivers/Makefile b/drivers/Makefile index ee65c87b0..bd83ad9f3 100644 --- a/drivers/Makefile +++ b/drivers/Makefile @@ -4,17 +4,19 @@ include $(RTE_SDK)/mk/rte.vars.mk DIRS-y += bus +DIRS-y += common +DEPDIRS-common := bus DIRS-y += mempool -DEPDIRS-mempool := bus +DEPDIRS-mempool := bus common DIRS-y += net -DEPDIRS-net := bus mempool +DEPDIRS-net := bus common mempool DIRS-$(CONFIG_RTE_LIBRTE_BBDEV) += bbdev -DEPDIRS-bbdev := bus mempool +DEPDIRS-bbdev := bus common mempool DIRS-$(CONFIG_RTE_LIBRTE_CRYPTODEV) += crypto -DEPDIRS-crypto := bus mempool +DEPDIRS-crypto := bus common mempool DIRS-$(CONFIG_RTE_LIBRTE_EVENTDEV) += event -DEPDIRS-event := bus mempool net +DEPDIRS-event := bus common mempool net DIRS-$(CONFIG_RTE_LIBRTE_RAWDEV) += raw -DEPDIRS-raw := bus mempool net event +DEPDIRS-raw := bus common mempool net event include $(RTE_SDK)/mk/rte.subdir.mk diff --git a/drivers/common/Makefile b/drivers/common/Makefile new file mode 100644 index 0..192066307 --- /dev/null +++ b/drivers/common/Makefile @@ -0,0 +1,7 @@ +# SPDX-License-Identifier: BSD-3-Clause +# Copyright(c) 2018 Cavium, Inc +# + +include $(RTE_SDK)/mk/rte.vars.mk + +include $(RTE_SDK)/mk/rte.subdir.mk diff --git a/drivers/common/meson.build b/drivers/common/meson.build new file mode 100644 index 0..20d4a30ba --- /dev/null +++ b/drivers/common/meson.build @@ -0,0 +1,6 @@ +# SPDX-License-Identifier: BSD-3-Clause +# Copyright(c) 2018 Cavium, Inc + +std_deps = ['bus_pci'] +config_flag_fmt = 'RTE_LIBRTE_@0@_COMMON' +driver_name_fmt = 'rte_@0@_common' diff --git a/drivers/meson.build b/drivers/meson.build index b41a0f18e..9494e5d76 100644 --- a/drivers/meson.build +++ b/drivers/meson.build @@ -3,10 +3,11 @@ # Defines the order in which the drivers are buit. driver_classes = ['bus', - 'mempool', # depends on bus. - 'net', # depends on bus and mempool. - 'crypto', # depenss on bus, mempool (net in future). - 'event'] # depends on bus, mempool and net. + 'common', # depends on bus. + 'mempool', # depends on bus and common. + 'net', # depends on bus, common and mempool. + 'crypto', # depenss on bus, common and mempool (net in future). + 'event'] # depends on bus, common, mempool and net. foreach class:driver_classes drivers = [] -- 2.16.2
[dpdk-dev] [PATCH 2/2] octeontx: move mbox to common folder
Move commonly used functions across mempool, event and net devices to a common folder in drivers. Signed-off-by: Pavan Nikhilesh --- drivers/common/Makefile| 4 drivers/common/meson.build | 1 + drivers/common/octeontx/Makefile | 25 ++ drivers/common/octeontx/meson.build| 8 +++ .../{mempool => common}/octeontx/octeontx_mbox.c | 12 ++- .../{mempool => common}/octeontx/octeontx_mbox.h | 11 ++ .../{mempool => common}/octeontx/octeontx_ssovf.c | 1 - .../octeontx/rte_octeontx_common_version.map | 7 ++ drivers/event/octeontx/Makefile| 3 ++- drivers/event/octeontx/meson.build | 2 +- drivers/mempool/octeontx/Makefile | 5 ++--- drivers/mempool/octeontx/meson.build | 6 ++ drivers/mempool/octeontx/octeontx_fpavf.c | 4 drivers/mempool/octeontx/octeontx_pool_logs.h | 9 .../octeontx/rte_mempool_octeontx_version.map | 6 -- drivers/net/octeontx/Makefile | 3 ++- mk/rte.app.mk | 5 + 17 files changed, 81 insertions(+), 31 deletions(-) create mode 100644 drivers/common/octeontx/Makefile create mode 100644 drivers/common/octeontx/meson.build rename drivers/{mempool => common}/octeontx/octeontx_mbox.c (95%) rename drivers/{mempool => common}/octeontx/octeontx_mbox.h (72%) rename drivers/{mempool => common}/octeontx/octeontx_ssovf.c (99%) create mode 100644 drivers/common/octeontx/rte_octeontx_common_version.map diff --git a/drivers/common/Makefile b/drivers/common/Makefile index 192066307..0fd223761 100644 --- a/drivers/common/Makefile +++ b/drivers/common/Makefile @@ -4,4 +4,8 @@ include $(RTE_SDK)/mk/rte.vars.mk +ifeq ($(CONFIG_RTE_LIBRTE_PMD_OCTEONTX_SSOVF)$(CONFIG_RTE_LIBRTE_OCTEONTX_MEMPOOL),yy) +DIRS-y += octeontx +endif + include $(RTE_SDK)/mk/rte.subdir.mk diff --git a/drivers/common/meson.build b/drivers/common/meson.build index 20d4a30ba..558fe7186 100644 --- a/drivers/common/meson.build +++ b/drivers/common/meson.build @@ -1,6 +1,7 @@ # SPDX-License-Identifier: BSD-3-Clause # Copyright(c) 2018 Cavium, Inc +drivers = ['octeontx'] std_deps = ['bus_pci'] config_flag_fmt = 'RTE_LIBRTE_@0@_COMMON' driver_name_fmt = 'rte_@0@_common' diff --git a/drivers/common/octeontx/Makefile b/drivers/common/octeontx/Makefile new file mode 100644 index 0..1fe0ca35c --- /dev/null +++ b/drivers/common/octeontx/Makefile @@ -0,0 +1,25 @@ +# SPDX-License-Identifier: BSD-3-Clause +# Copyright(c) 2018 Cavium, Inc +# + +include $(RTE_SDK)/mk/rte.vars.mk + +# +# library name +# +LIB = librte_octeontx_common.a + +CFLAGS += $(WERROR_FLAGS) +EXPORT_MAP := rte_octeontx_common_version.map + +LIBABIVER := 1 + +# +# all source are stored in SRCS-y +# +SRCS-y += octeontx_ssovf.c +SRCS-y += octeontx_mbox.c + +LDLIBS += -lrte_eal -lrte_bus_pci + +include $(RTE_SDK)/mk/rte.lib.mk diff --git a/drivers/common/octeontx/meson.build b/drivers/common/octeontx/meson.build new file mode 100644 index 0..654eb482e --- /dev/null +++ b/drivers/common/octeontx/meson.build @@ -0,0 +1,8 @@ +# SPDX-License-Identifier: BSD-3-Clause +# Copyright(c) 2018 Cavium, Inc +# + +std_deps += ['bus_pci'] +sources = files('octeontx_mbox.c', + 'octeontx_ssovf.c' +) diff --git a/drivers/mempool/octeontx/octeontx_mbox.c b/drivers/common/octeontx/octeontx_mbox.c similarity index 95% rename from drivers/mempool/octeontx/octeontx_mbox.c rename to drivers/common/octeontx/octeontx_mbox.c index f8cb6a453..23e90711c 100644 --- a/drivers/mempool/octeontx/octeontx_mbox.c +++ b/drivers/common/octeontx/octeontx_mbox.c @@ -11,7 +11,6 @@ #include #include "octeontx_mbox.h" -#include "octeontx_pool_logs.h" /* Mbox operation timeout in seconds */ #define MBOX_WAIT_TIME_SEC 3 @@ -60,6 +59,17 @@ struct mbox_ram_hdr { }; }; +int octeontx_logtype_mbox; + +RTE_INIT(otx_init_log); +static void +otx_init_log(void) +{ + octeontx_logtype_mbox = rte_log_register("pmd.octeontx.mbox"); + if (octeontx_logtype_mbox >= 0) + rte_log_set_level(octeontx_logtype_mbox, RTE_LOG_NOTICE); +} + static inline void mbox_msgcpy(volatile uint8_t *d, volatile const uint8_t *s, uint16_t size) { diff --git a/drivers/mempool/octeontx/octeontx_mbox.h b/drivers/common/octeontx/octeontx_mbox.h similarity index 72% rename from drivers/mempool/octeontx/octeontx_mbox.h rename to drivers/common/octeontx/octeontx_mbox.h index 1b056071a..f40b8fb5f 100644 --- a/drivers/mempool/octeontx/octeontx_mbox.h +++ b/drivers/common/octeontx/octeontx_mbox.h @@ -10,6 +10,17 @@ #define SSOW_BAR4_LEN (64 * 1024) #define SSO_VHGRP_PF_MBOX(x) (0x200ULL | ((x) << 3)) +#define MBOX_LOG(level, fmt, args...) \ + rte_log(RTE_LOG_ ## level, octeontx_logtype_mbox,\ +
[dpdk-dev] [PATCH v3 0/5] vhost: support selective datapath
This patch set introduces support for selective datapath in DPDK vhost-user lib. vDPA stands for vhost Data Path Acceleration. The idea is to enable various types of virtio-compatible devices to do data transfer with virtio driver directly to enable acceleration. The default datapath is the existing software implementation, more options will be available when new engines are added. Design details An engine is a group of virtio-compatible devices. The definition of engine is as follows: struct rte_vdpa_eng_addr { union { uint8_t __dummy[64]; struct rte_pci_addr pci_addr; }; }; struct rte_vdpa_eng_info { char name[MAX_VDPA_NAME_LEN]; struct rte_vdpa_eng_addr *addr; }; struct rte_vdpa_dev_ops { vdpa_dev_conf_tdev_conf; vdpa_dev_close_t dev_close; vdpa_vring_state_set_t vring_state_set; vdpa_feature_set_t feature_set; vdpa_migration_done_t migration_done; vdpa_get_vfio_group_fd_t get_vfio_group_fd; vdpa_get_vfio_device_fd_t get_vfio_device_fd; vdpa_get_notify_area_tget_notify_area; }; struct rte_vdpa_eng_ops { vdpa_eng_init_t eng_init; vdpa_eng_uninit_t eng_uninit; vdpa_info_query_t info_query; }; struct rte_vdpa_eng_driver { const char *name; struct rte_vdpa_eng_ops eng_ops; struct rte_vdpa_dev_ops dev_ops; } __rte_cache_aligned; struct rte_vdpa_engine { struct rte_vdpa_eng_infoeng_info; struct rte_vdpa_eng_driver *eng_drv; } __rte_cache_aligned; A set of engine ops is defined in rte_vdpa_eng_ops for engine init, uninit, and attributes reporting. The attributes are defined as follows: struct rte_vdpa_eng_attr { uint64_t features; uint64_t protocol_features; uint32_t queue_num; uint32_t dev_num; }; A set of device ops is defined in rte_vdpa_dev_ops for each virtio device in the engine to do device specific operations. Changes to the current vhost-user lib are: 1. Make vhost device capabilities configurable to adopt various engines. Such capabilities include supported features, protocol features, queue number. APIs are introduced to let app configure these capabilities. 2. In addition to the existing vhost framework, a set of callbacks is added for vhost to call the driver for device operations at the right time: a. dev_conf: Called to configure the actual device when the virtio device becomes ready. b. dev_close: Called to close the actual device when the virtio device is stopped. c. vring_state_set: Called to change the state of the vring in the actual device when vring state changes. d. feature_set: Called to set the negotiated features to device. e. migration_done: Called to allow the device to response to RARP sending. f. get_vfio_group_fd: Called to get the VFIO group fd of the device. g. get_vfio_device_fd: Called to get the VFIO device fd of the device. h. get_notify_area: Called to get the notify area info of the queue. 3. To make vhost aware of its own type, an engine id (eid) and a device id (did) are added into the vhost data structure to identify the actual device. APIs are introduced to let app configure them. When the default software datapath is used, eid and did are set to -1. When alternative datapath is used, eid and did are set by app to specify which device to use. Each vhost-user socket can have only 1 connection in this case. Working process: 1. Register driver during DPDK initialization. 2. Register engine with driver name and address. 3. Get engine attributes. 4. For vhost device creation: a. Register vhost-user socket. b. Set eid and did of the vhost-user socket. c. Register vhost-user callbacks. d. Start to wait for connection. 4. When connection comes and virtio device data structure is negotiated, the device will be configured with all needed info. --- Changes in v3: 1. Keep macro names the same as in the spec. 2. Export new APIs where they're introduced. --- Changes in v2: 1. Ensure negotiated capabilities are supported in vhost-user lib. 2. Add APIs for live migration. 3. Configure the data path at the right time. 4. Add VFIO related vDPA device ops. 5. Rebase on dpdk-next-virtio. Zhihong Wang (5): vhost: export vhost feature definitions vhost: support selective datapath vhost: add apis for datapath configuration vhost: adapt vhost lib for selective datapath vhost: add apis for live migration lib/librte_vhost/Makefile | 4 +- lib/librte_vhost/rte_vdpa.h| 126 +++ lib/librte_vhost/rte_vhost.h | 178 + lib/librte_vhost/rte_vhost_version.map | 19 lib/librte_vhost/socket.c | 141 +- lib/
[dpdk-dev] [PATCH v3 3/5] vhost: add apis for datapath configuration
This patch adds APIs for datapath configuration. The eid and did of the vhost-user socket can be configured to identify the actual device. When the default software datapath is used, eid and did are set to -1. When alternative datapath is used, eid and did are set by app to specify which device to use. Each vhost-user socket can have only 1 connection in this case. Signed-off-by: Zhihong Wang --- lib/librte_vhost/rte_vhost.h | 70 ++ lib/librte_vhost/rte_vhost_version.map | 6 +++ lib/librte_vhost/socket.c | 65 +++ lib/librte_vhost/vhost.c | 50 lib/librte_vhost/vhost.h | 10 + 5 files changed, 201 insertions(+) diff --git a/lib/librte_vhost/rte_vhost.h b/lib/librte_vhost/rte_vhost.h index 671ea5053..7aa57ca87 100644 --- a/lib/librte_vhost/rte_vhost.h +++ b/lib/librte_vhost/rte_vhost.h @@ -200,6 +200,54 @@ int rte_vhost_driver_register(const char *path, uint64_t flags); int rte_vhost_driver_unregister(const char *path); /** + * Set the engine id, enforce single connection per socket + * + * @param path + * The vhost-user socket file path + * @param eid + * Engine id + * @return + * 0 on success, -1 on failure + */ +int __rte_experimental +rte_vhost_driver_set_vdpa_eid(const char *path, int eid); + +/** + * Set the device id, enforce single connection per socket + * + * @param path + * The vhost-user socket file path + * @param did + * Device id + * @return + * 0 on success, -1 on failure + */ +int __rte_experimental +rte_vhost_driver_set_vdpa_did(const char *path, int did); + +/** + * Get the engine id + * + * @param path + * The vhost-user socket file path + * @return + * Engine id, -1 on failure + */ +int __rte_experimental +rte_vhost_driver_get_vdpa_eid(const char *path); + +/** + * Get the device id + * + * @param path + * The vhost-user socket file path + * @return + * Device id, -1 on failure + */ +int __rte_experimental +rte_vhost_driver_get_vdpa_did(const char *path); + +/** * Set the feature bits the vhost-user driver supports. * * @param path @@ -464,6 +512,28 @@ int rte_vhost_vring_call(int vid, uint16_t vring_idx); */ uint32_t rte_vhost_rx_queue_count(int vid, uint16_t qid); +/** + * Get vdpa engine id for vhost device. + * + * @param vid + * vhost device ID + * @return + * engine id + */ +int __rte_experimental +rte_vhost_get_vdpa_eid(int vid); + +/** + * Get vdpa device id for vhost device. + * + * @param vid + * vhost device ID + * @return + * device id + */ +int __rte_experimental +rte_vhost_get_vdpa_did(int vid); + #ifdef __cplusplus } #endif diff --git a/lib/librte_vhost/rte_vhost_version.map b/lib/librte_vhost/rte_vhost_version.map index 57a3edd01..c505596c5 100644 --- a/lib/librte_vhost/rte_vhost_version.map +++ b/lib/librte_vhost/rte_vhost_version.map @@ -66,4 +66,10 @@ EXPERIMENTAL { rte_vdpa_find_engine_id; rte_vdpa_info_query; rte_vdpa_register_driver; + rte_vhost_driver_set_vdpa_eid; + rte_vhost_driver_set_vdpa_did; + rte_vhost_driver_get_vdpa_eid; + rte_vhost_driver_get_vdpa_did; + rte_vhost_get_vdpa_eid; + rte_vhost_get_vdpa_did; } DPDK_18.02; diff --git a/lib/librte_vhost/socket.c b/lib/librte_vhost/socket.c index cfc31e179..8551eb58c 100644 --- a/lib/librte_vhost/socket.c +++ b/lib/librte_vhost/socket.c @@ -52,6 +52,13 @@ struct vhost_user_socket { uint64_t supported_features; uint64_t features; + /* engine and device id to identify a certain port on a specific +* backend, both are set to -1 for sw. when used, one socket can +* have 1 connection only. +*/ + int eid; + int did; + struct vhost_device_ops const *notify_ops; }; @@ -545,6 +552,64 @@ find_vhost_user_socket(const char *path) } int +rte_vhost_driver_set_vdpa_eid(const char *path, int eid) +{ + struct vhost_user_socket *vsocket; + + pthread_mutex_lock(&vhost_user.mutex); + vsocket = find_vhost_user_socket(path); + if (vsocket) + vsocket->eid = eid; + pthread_mutex_unlock(&vhost_user.mutex); + + return vsocket ? 0 : -1; +} + +int +rte_vhost_driver_set_vdpa_did(const char *path, int did) +{ + struct vhost_user_socket *vsocket; + + pthread_mutex_lock(&vhost_user.mutex); + vsocket = find_vhost_user_socket(path); + if (vsocket) + vsocket->did = did; + pthread_mutex_unlock(&vhost_user.mutex); + + return vsocket ? 0 : -1; +} + +int +rte_vhost_driver_get_vdpa_eid(const char *path) +{ + struct vhost_user_socket *vsocket; + int eid = -1; + + pthread_mutex_lock(&vhost_user.mutex); + vsocket = find_vhost_user_socket(path); + if (vsocket) + eid = vsocket->eid; + pthread_mutex_unlock(&vhost_user.mutex); + + return eid; +} + +int +rte_vhost_driver_get_vdpa_did(const char
[dpdk-dev] [PATCH v3 4/5] vhost: adapt vhost lib for selective datapath
This patch adapts vhost lib for selective datapath by calling device ops at the corresponding stage. Signed-off-by: Zhihong Wang --- Changes in v2: 1. Ensure negotiated capabilities are supported in vhost-user lib. 2. Configure the data path at the right time. lib/librte_vhost/rte_vhost.h | 27 lib/librte_vhost/rte_vhost_version.map | 2 + lib/librte_vhost/socket.c | 76 +- lib/librte_vhost/vhost.c | 3 ++ lib/librte_vhost/vhost.h | 2 + lib/librte_vhost/vhost_user.c | 56 ++--- 6 files changed, 158 insertions(+), 8 deletions(-) diff --git a/lib/librte_vhost/rte_vhost.h b/lib/librte_vhost/rte_vhost.h index 7aa57ca87..77c2a1a8b 100644 --- a/lib/librte_vhost/rte_vhost.h +++ b/lib/librte_vhost/rte_vhost.h @@ -303,6 +303,33 @@ int rte_vhost_driver_disable_features(const char *path, uint64_t features); int rte_vhost_driver_get_features(const char *path, uint64_t *features); /** + * Get the protocol feature bits before feature negotiation. + * + * @param path + * The vhost-user socket file path + * @param protocol_features + * A pointer to store the queried protocol feature bits + * @return + * 0 on success, -1 on failure + */ +int __rte_experimental +rte_vhost_driver_get_protocol_features(const char *path, + uint64_t *protocol_features); + +/** + * Get the queue number bits before feature negotiation. + * + * @param path + * The vhost-user socket file path + * @param queue_num + * A pointer to store the queried queue number bits + * @return + * 0 on success, -1 on failure + */ +int __rte_experimental +rte_vhost_driver_get_queue_num(const char *path, uint32_t *queue_num); + +/** * Get the feature bits after negotiation * * @param vid diff --git a/lib/librte_vhost/rte_vhost_version.map b/lib/librte_vhost/rte_vhost_version.map index c505596c5..8ef2a396c 100644 --- a/lib/librte_vhost/rte_vhost_version.map +++ b/lib/librte_vhost/rte_vhost_version.map @@ -72,4 +72,6 @@ EXPERIMENTAL { rte_vhost_driver_get_vdpa_did; rte_vhost_get_vdpa_eid; rte_vhost_get_vdpa_did; + rte_vhost_driver_get_protocol_features; + rte_vhost_driver_get_queue_num; } DPDK_18.02; diff --git a/lib/librte_vhost/socket.c b/lib/librte_vhost/socket.c index 8551eb58c..14dce2a73 100644 --- a/lib/librte_vhost/socket.c +++ b/lib/librte_vhost/socket.c @@ -216,6 +216,9 @@ vhost_user_add_connection(int fd, struct vhost_user_socket *vsocket) vhost_set_builtin_virtio_net(vid, vsocket->use_builtin_virtio_net); + vhost_set_vdpa_eid(vid, vsocket->eid); + vhost_set_vdpa_did(vid, vsocket->did); + if (vsocket->dequeue_zero_copy) vhost_enable_dequeue_zero_copy(vid); @@ -677,11 +680,80 @@ int rte_vhost_driver_get_features(const char *path, uint64_t *features) { struct vhost_user_socket *vsocket; + struct rte_vdpa_eng_attr attr; + int eid = -1; pthread_mutex_lock(&vhost_user.mutex); vsocket = find_vhost_user_socket(path); - if (vsocket) - *features = vsocket->features; + if (vsocket) { + eid = vsocket->eid; + if (rte_vdpa_info_query(eid, &attr) < 0) + *features = vsocket->features; + else + *features = vsocket->features & attr.features; + + } + pthread_mutex_unlock(&vhost_user.mutex); + + if (!vsocket) { + RTE_LOG(ERR, VHOST_CONFIG, + "socket file %s is not registered yet.\n", path); + return -1; + } else { + return 0; + } +} + +int +rte_vhost_driver_get_protocol_features(const char *path, + uint64_t *protocol_features) +{ + struct vhost_user_socket *vsocket; + struct rte_vdpa_eng_attr attr; + int eid = -1; + + pthread_mutex_lock(&vhost_user.mutex); + vsocket = find_vhost_user_socket(path); + if (vsocket) { + eid = vsocket->eid; + if (rte_vdpa_info_query(eid, &attr) < 0) + *protocol_features = VHOST_USER_PROTOCOL_FEATURES; + else + *protocol_features = VHOST_USER_PROTOCOL_FEATURES + & attr.protocol_features; + + } + pthread_mutex_unlock(&vhost_user.mutex); + + if (!vsocket) { + RTE_LOG(ERR, VHOST_CONFIG, + "socket file %s is not registered yet.\n", path); + return -1; + } else { + return 0; + } +} + +int +rte_vhost_driver_get_queue_num(const char *path, + uint32_t *queue_num) +{ + struct vhost_user_socket *vsocket; + struct rte_vdpa_eng_attr attr; + int eid = -1; + + pthread_mutex_lock(&vhost_user.mutex); + vsocket = find_vhost_user_socket(path); + if (vsocket) { +
[dpdk-dev] [PATCH v3 1/5] vhost: export vhost feature definitions
This patch exports vhost-user protocol features to support device driver development. Signed-off-by: Zhihong Wang --- Changes in v3: 1. Keep macro names the same as in the spec. lib/librte_vhost/rte_vhost.h | 30 ++ lib/librte_vhost/vhost.h | 2 -- lib/librte_vhost/vhost_user.h | 7 --- 3 files changed, 30 insertions(+), 9 deletions(-) diff --git a/lib/librte_vhost/rte_vhost.h b/lib/librte_vhost/rte_vhost.h index d33206997..671ea5053 100644 --- a/lib/librte_vhost/rte_vhost.h +++ b/lib/librte_vhost/rte_vhost.h @@ -29,6 +29,36 @@ extern "C" { #define RTE_VHOST_USER_DEQUEUE_ZERO_COPY (1ULL << 2) #define RTE_VHOST_USER_IOMMU_SUPPORT (1ULL << 3) +/** Protocol features. */ +#ifndef VHOST_USER_PROTOCOL_F_MQ +#define VHOST_USER_PROTOCOL_F_MQ 0 +#endif + +#ifndef VHOST_USER_PROTOCOL_F_LOG_SHMFD +#define VHOST_USER_PROTOCOL_F_LOG_SHMFD1 +#endif + +#ifndef VHOST_USER_PROTOCOL_F_RARP +#define VHOST_USER_PROTOCOL_F_RARP 2 +#endif + +#ifndef VHOST_USER_PROTOCOL_F_REPLY_ACK +#define VHOST_USER_PROTOCOL_F_REPLY_ACK3 +#endif + +#ifndef VHOST_USER_PROTOCOL_F_NET_MTU +#define VHOST_USER_PROTOCOL_F_NET_MTU 4 +#endif + +#ifndef VHOST_USER_PROTOCOL_F_SLAVE_REQ +#define VHOST_USER_PROTOCOL_F_SLAVE_REQ5 +#endif + +/** Indicate whether protocol features negotiation is supported. */ +#ifndef VHOST_USER_F_PROTOCOL_FEATURES +#define VHOST_USER_F_PROTOCOL_FEATURES 30 +#endif + /** * Information relating to memory regions including offsets to * addresses in QEMUs memory file. diff --git a/lib/librte_vhost/vhost.h b/lib/librte_vhost/vhost.h index 58aec2e0d..2e28e4026 100644 --- a/lib/librte_vhost/vhost.h +++ b/lib/librte_vhost/vhost.h @@ -174,8 +174,6 @@ struct vhost_msg { #define VIRTIO_F_VERSION_1 32 #endif -#define VHOST_USER_F_PROTOCOL_FEATURES 30 - /* Features supported by this builtin vhost-user net driver. */ #define VIRTIO_NET_SUPPORTED_FEATURES ((1ULL << VIRTIO_NET_F_MRG_RXBUF) | \ (1ULL << VIRTIO_F_ANY_LAYOUT) | \ diff --git a/lib/librte_vhost/vhost_user.h b/lib/librte_vhost/vhost_user.h index 0fafbe6e0..97afabc47 100644 --- a/lib/librte_vhost/vhost_user.h +++ b/lib/librte_vhost/vhost_user.h @@ -14,13 +14,6 @@ #define VHOST_MEMORY_MAX_NREGIONS 8 -#define VHOST_USER_PROTOCOL_F_MQ 0 -#define VHOST_USER_PROTOCOL_F_LOG_SHMFD1 -#define VHOST_USER_PROTOCOL_F_RARP 2 -#define VHOST_USER_PROTOCOL_F_REPLY_ACK3 -#define VHOST_USER_PROTOCOL_F_NET_MTU 4 -#define VHOST_USER_PROTOCOL_F_SLAVE_REQ 5 - #define VHOST_USER_PROTOCOL_FEATURES ((1ULL << VHOST_USER_PROTOCOL_F_MQ) | \ (1ULL << VHOST_USER_PROTOCOL_F_LOG_SHMFD) |\ (1ULL << VHOST_USER_PROTOCOL_F_RARP) | \ -- 2.13.6
[dpdk-dev] [PATCH v3 2/5] vhost: support selective datapath
This patch introduces support for selective datapath in DPDK vhost-user lib to enable various types of virtio-compatible devices to do data transfer with virtio driver directly to enable acceleration. The default datapath is the existing software implementation, more options will be available when new engines are registered. An engine is a group of virtio-compatible devices under a single address. The engine driver includes: 1. A set of engine ops is defined in rte_vdpa_eng_ops to perform engine init, uninit, and attributes reporting. 2. A set of device ops is defined in rte_vdpa_dev_ops for virtio devices in the engine to do device specific operations: a. dev_conf: Called to configure the actual device when the virtio device becomes ready. b. dev_close: Called to close the actual device when the virtio device is stopped. c. vring_state_set: Called to change the state of the vring in the actual device when vring state changes. d. feature_set: Called to set the negotiated features to device. e. migration_done: Called to allow the device to response to RARP sending. f. get_vfio_group_fd: Called to get the VFIO group fd of the device. g. get_vfio_device_fd: Called to get the VFIO device fd of the device. h. get_notify_area: Called to get the notify area info of the queue. Signed-off-by: Zhihong Wang --- Changes in v2: 1. Add VFIO related vDPA device ops. lib/librte_vhost/Makefile | 4 +- lib/librte_vhost/rte_vdpa.h| 126 + lib/librte_vhost/rte_vhost_version.map | 8 +++ lib/librte_vhost/vdpa.c| 124 4 files changed, 260 insertions(+), 2 deletions(-) create mode 100644 lib/librte_vhost/rte_vdpa.h create mode 100644 lib/librte_vhost/vdpa.c diff --git a/lib/librte_vhost/Makefile b/lib/librte_vhost/Makefile index 5d6c6abae..37044ac03 100644 --- a/lib/librte_vhost/Makefile +++ b/lib/librte_vhost/Makefile @@ -22,9 +22,9 @@ LDLIBS += -lrte_eal -lrte_mempool -lrte_mbuf -lrte_ethdev -lrte_net # all source are stored in SRCS-y SRCS-$(CONFIG_RTE_LIBRTE_VHOST) := fd_man.c iotlb.c socket.c vhost.c \ - vhost_user.c virtio_net.c + vhost_user.c virtio_net.c vdpa.c # install includes -SYMLINK-$(CONFIG_RTE_LIBRTE_VHOST)-include += rte_vhost.h +SYMLINK-$(CONFIG_RTE_LIBRTE_VHOST)-include += rte_vhost.h rte_vdpa.h include $(RTE_SDK)/mk/rte.lib.mk diff --git a/lib/librte_vhost/rte_vdpa.h b/lib/librte_vhost/rte_vdpa.h new file mode 100644 index 0..23fb471be --- /dev/null +++ b/lib/librte_vhost/rte_vdpa.h @@ -0,0 +1,126 @@ +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright(c) 2018 Intel Corporation + */ + +#ifndef _RTE_VDPA_H_ +#define _RTE_VDPA_H_ + +/** + * @file + * + * Device specific vhost lib + */ + +#include +#include "rte_vhost.h" + +#define MAX_VDPA_ENGINE_NUM 128 +#define MAX_VDPA_NAME_LEN 128 + +struct rte_vdpa_eng_addr { + union { + uint8_t __dummy[64]; + struct rte_pci_addr pci_addr; + }; +}; + +struct rte_vdpa_eng_info { + struct rte_vdpa_eng_addr *addr; + char name[MAX_VDPA_NAME_LEN]; +}; + +struct rte_vdpa_eng_attr { + uint64_t features; + uint64_t protocol_features; + uint32_t queue_num; + uint32_t dev_num; +}; + +/* register/remove engine */ +typedef int (*vdpa_eng_init_t)(int eid, struct rte_vdpa_eng_addr *addr); +typedef int (*vdpa_eng_uninit_t)(int eid); + +/* query info of this engine */ +typedef int (*vdpa_info_query_t)(int eid, + struct rte_vdpa_eng_attr *attr); + +/* driver configure/close the port based on connection */ +typedef int (*vdpa_dev_conf_t)(int vid); +typedef int (*vdpa_dev_close_t)(int vid); + +/* enable/disable this vring */ +typedef int (*vdpa_vring_state_set_t)(int vid, int vring, int state); + +/* set features when changed */ +typedef int (*vdpa_feature_set_t)(int vid); + +/* destination operations when migration done, e.g. send rarp */ +typedef int (*vdpa_migration_done_t)(int vid); + +/* get the vfio group fd */ +typedef int (*vdpa_get_vfio_group_fd_t)(int vid); + +/* get the vfio device fd */ +typedef int (*vdpa_get_vfio_device_fd_t)(int vid); + +/* get the notify area info of the queue */ +typedef int (*vdpa_get_notify_area_t)(int vid, int qid, uint64_t *offset, + uint64_t *size); +/* device ops */ +struct rte_vdpa_dev_ops { + vdpa_dev_conf_t dev_conf; + vdpa_dev_close_t dev_close; + vdpa_vring_state_set_tvring_state_set; + vdpa_feature_set_tfeature_set; + vdpa_migration_done_t migration_done; + vdpa_get_vfio_group_fd_t get_vfio_group_fd; + vdpa_get_vfio_device_fd_t get_vfio_device_fd; + vdpa_get_notify_area_tget_notify_area; +}; + +/* engine ops */ +struct rte_vdpa_eng_ops { + vdpa
[dpdk-dev] [PATCH v3 5/5] vhost: add apis for live migration
This patch adds APIs to enable live migration for non-builtin data paths. At src side, last_avail/used_idx from the device need to be set into the virtio_net structure, and the log_base and log_size from the virtio_net structure need to be set into the device. At dst side, last_avail/used_idx need to be read from the virtio_net structure and set into the device. Signed-off-by: Zhihong Wang --- lib/librte_vhost/rte_vhost.h | 51 +++ lib/librte_vhost/rte_vhost_version.map | 3 ++ lib/librte_vhost/vhost.c | 63 ++ 3 files changed, 117 insertions(+) diff --git a/lib/librte_vhost/rte_vhost.h b/lib/librte_vhost/rte_vhost.h index 77c2a1a8b..0f9303949 100644 --- a/lib/librte_vhost/rte_vhost.h +++ b/lib/librte_vhost/rte_vhost.h @@ -540,6 +540,57 @@ int rte_vhost_vring_call(int vid, uint16_t vring_idx); uint32_t rte_vhost_rx_queue_count(int vid, uint16_t qid); /** + * Get log base and log size of the vhost device + * + * @param vid + * vhost device ID + * @param log_base + * vhost log base + * @param log_size + * vhost log size + * @return + * 0 on success, -1 on failure + */ +int __rte_experimental +rte_vhost_get_log_base(int vid, uint64_t *log_base, uint64_t *log_size); + +/** + * Get last_avail/used_idx of the vhost virtqueue + * + * @param vid + * vhost device ID + * @param queue_id + * vhost queue index + * @param last_avail_idx + * vhost last_avail_idx to get + * @param last_used_idx + * vhost last_used_idx to get + * @return + * 0 on success, -1 on failure + */ +int __rte_experimental +rte_vhost_get_vring_base(int vid, uint16_t queue_id, + uint16_t *last_avail_idx, uint16_t *last_used_idx); + +/** + * Set last_avail/used_idx of the vhost virtqueue + * + * @param vid + * vhost device ID + * @param queue_id + * vhost queue index + * @param last_avail_idx + * last_avail_idx to set + * @param last_used_idx + * last_used_idx to set + * @return + * 0 on success, -1 on failure + */ +int __rte_experimental +rte_vhost_set_vring_base(int vid, uint16_t queue_id, + uint16_t last_avail_idx, uint16_t last_used_idx); + +/** * Get vdpa engine id for vhost device. * * @param vid diff --git a/lib/librte_vhost/rte_vhost_version.map b/lib/librte_vhost/rte_vhost_version.map index 8ef2a396c..88bf6cb54 100644 --- a/lib/librte_vhost/rte_vhost_version.map +++ b/lib/librte_vhost/rte_vhost_version.map @@ -74,4 +74,7 @@ EXPERIMENTAL { rte_vhost_get_vdpa_did; rte_vhost_driver_get_protocol_features; rte_vhost_driver_get_queue_num; + rte_vhost_get_log_base; + rte_vhost_get_vring_base; + rte_vhost_set_vring_base; } DPDK_18.02; diff --git a/lib/librte_vhost/vhost.c b/lib/librte_vhost/vhost.c index f8a5a1c42..c7332c557 100644 --- a/lib/librte_vhost/vhost.c +++ b/lib/librte_vhost/vhost.c @@ -667,3 +667,66 @@ int rte_vhost_get_vdpa_did(int vid) return dev->did; } + +int rte_vhost_get_log_base(int vid, uint64_t *log_base, + uint64_t *log_size) +{ + struct virtio_net *dev = get_device(vid); + + if (!dev) + return -1; + + if (unlikely(!(dev->flags & VIRTIO_DEV_BUILTIN_VIRTIO_NET))) { + RTE_LOG(ERR, VHOST_DATA, + "(%d) %s: built-in vhost net backend is disabled.\n", + dev->vid, __func__); + return -1; + } + + *log_base = dev->log_base; + *log_size = dev->log_size; + + return 0; +} + +int rte_vhost_get_vring_base(int vid, uint16_t queue_id, + uint16_t *last_avail_idx, uint16_t *last_used_idx) +{ + struct virtio_net *dev = get_device(vid); + + if (!dev) + return -1; + + if (unlikely(!(dev->flags & VIRTIO_DEV_BUILTIN_VIRTIO_NET))) { + RTE_LOG(ERR, VHOST_DATA, + "(%d) %s: built-in vhost net backend is disabled.\n", + dev->vid, __func__); + return -1; + } + + *last_avail_idx = dev->virtqueue[queue_id]->last_avail_idx; + *last_used_idx = dev->virtqueue[queue_id]->last_used_idx; + + return 0; +} + +int rte_vhost_set_vring_base(int vid, uint16_t queue_id, + uint16_t last_avail_idx, uint16_t last_used_idx) +{ + struct virtio_net *dev = get_device(vid); + + if (!dev) + return -1; + + if (unlikely(!(dev->flags & VIRTIO_DEV_BUILTIN_VIRTIO_NET))) { + RTE_LOG(ERR, VHOST_DATA, + "(%d) %s: built-in vhost net backend is disabled.\n", + dev->vid, __func__); + return -1; + } + + dev->virtqueue[queue_id]->last_avail_idx = last_avail_idx; + dev->virtqueue[queue_id]->last_used_idx = last_used_idx; + + return 0; +} -- 2.13.6
Re: [dpdk-dev] [PATCH 08/17] net/virtio: implement receive path for packed queues
On Fri, Mar 16, 2018 at 04:21:11PM +0100, Jens Freimann wrote: [...] > diff --git a/drivers/net/virtio/virtio_ethdev.c > b/drivers/net/virtio/virtio_ethdev.c > index 722a2cd..888cc49 100644 > --- a/drivers/net/virtio/virtio_ethdev.c > +++ b/drivers/net/virtio/virtio_ethdev.c > @@ -1352,6 +1352,8 @@ static int virtio_dev_xstats_get_names(struct > rte_eth_dev *dev, > PMD_INIT_LOG(INFO, "virtio: using simple Rx path on port %u", > eth_dev->data->port_id); > eth_dev->rx_pkt_burst = virtio_recv_pkts_vec; > + } else if (vtpci_packed_queue(hw)) { > + eth_dev->rx_pkt_burst = &virtio_recv_pkts_packed; If MRG_RXBUF isn't supported at this point, then we will need to make sure that RING_PACKED and MRG_RXBUF won't be negotiated at the same time. Otherwise this commit will break the driver. > } else if (vtpci_with_feature(hw, VIRTIO_NET_F_MRG_RXBUF)) { > PMD_INIT_LOG(INFO, > "virtio: using mergeable buffer Rx path on port %u", > @@ -1507,7 +1509,8 @@ static int virtio_dev_xstats_get_names(struct > rte_eth_dev *dev, > > /* Setting up rx_header size for the device */ > if (vtpci_with_feature(hw, VIRTIO_NET_F_MRG_RXBUF) || > - vtpci_with_feature(hw, VIRTIO_F_VERSION_1)) > + vtpci_with_feature(hw, VIRTIO_F_VERSION_1) || > + vtpci_with_feature(hw, VIRTIO_F_RING_PACKED)) > hw->vtnet_hdr_size = sizeof(struct virtio_net_hdr_mrg_rxbuf); > else > hw->vtnet_hdr_size = sizeof(struct virtio_net_hdr); [...] > #ifdef RTE_LIBRTE_VIRTIO_DEBUG_DUMP > #define VIRTIO_DUMP_PACKET(m, len) rte_pktmbuf_dump(stdout, m, len) > @@ -428,6 +429,34 @@ > > PMD_INIT_FUNC_TRACE(); > > + if (vtpci_packed_queue(hw)) { > + struct vring_desc_packed *desc; > + struct vq_desc_extra *dxp; > + > + for (desc_idx = 0; desc_idx < vq->vq_nentries; > + desc_idx++) { > + m = rte_mbuf_raw_alloc(rxvq->mpool); > + if (unlikely(m == NULL)) > + return -ENOMEM; > + > + dxp = &vq->vq_descx[desc_idx]; > + dxp->cookie = m; > + dxp->ndescs = 1; > + > + desc = &vq->vq_ring.desc_packed[desc_idx]; > + desc->addr = VIRTIO_MBUF_ADDR(m, vq) + > + RTE_PKTMBUF_HEADROOM - hw->vtnet_hdr_size; > + desc->len = m->buf_len - RTE_PKTMBUF_HEADROOM + > + hw->vtnet_hdr_size; > + desc->flags |= VRING_DESC_F_WRITE; > + rte_smp_wmb(); > + set_desc_avail(&vq->vq_ring, desc); > + } > + toggle_wrap_counter(&vq->vq_ring); > + > + return 0; Maybe it's better to give the debug code (which is at the end of this function) a chance to run. Thanks
Re: [dpdk-dev] [PATCH 10/17] vhost: vring address setup for packed queues
On Fri, Mar 16, 2018 at 04:21:13PM +0100, Jens Freimann wrote: > From: Yuanhan Liu > > Add code to set up packed queues when enabled. > > Signed-off-by: Yuanhan Liu > Signed-off-by: Jens Freimann > --- > lib/librte_vhost/vhost.c | 1 + > lib/librte_vhost/vhost.h | 1 + > lib/librte_vhost/vhost_user.c | 17 - > 3 files changed, 18 insertions(+), 1 deletion(-) > > diff --git a/lib/librte_vhost/vhost.c b/lib/librte_vhost/vhost.c > index a407067..a300812 100644 > --- a/lib/librte_vhost/vhost.c > +++ b/lib/librte_vhost/vhost.c > @@ -567,6 +567,7 @@ struct virtio_net * > return -1; > } > > + There is no need to add this blank line. > dev->virtqueue[queue_id]->used->flags = VRING_USED_F_NO_NOTIFY; > return 0; > } > diff --git a/lib/librte_vhost/vhost.h b/lib/librte_vhost/vhost.h > index 06e973e..d35c4b1 100644 > --- a/lib/librte_vhost/vhost.h > +++ b/lib/librte_vhost/vhost.h > @@ -70,6 +70,7 @@ struct batch_copy_elem { > */ > struct vhost_virtqueue { > struct vring_desc *desc; > + struct vring_desc_packed *desc_packed; > struct vring_avail *avail; > struct vring_used *used; It's better to use union here. > diff --git a/lib/librte_vhost/vhost_user.c b/lib/librte_vhost/vhost_user.c > index 90ed211..bd1e393 100644 > --- a/lib/librte_vhost/vhost_user.c > +++ b/lib/librte_vhost/vhost_user.c > @@ -415,6 +415,19 @@ > struct vhost_virtqueue *vq = dev->virtqueue[vq_index]; > struct vhost_vring_addr *addr = &vq->ring_addrs; > > + if (dev->features & (1ULL << VIRTIO_F_RING_PACKED)) { > + vq->desc_packed = (struct vring_desc_packed *) ring_addr_to_vva > + (dev, vq, addr->desc_user_addr, > sizeof(vq->desc_packed)); > + vq->desc = NULL; > + vq->avail = NULL; > + vq->used = NULL; > + vq->log_guest_addr = 0; > + > + assert(vq->last_used_idx == 0); As a library, we shouldn't crash the process. Thanks
Re: [dpdk-dev] [PATCH 1/2] drivers: add common folder
Hi Pavan, On Mon, Mar 19, 2018 at 2:57 PM, Pavan Nikhilesh wrote: > Add driver/common folder and skeleton makefile for adding commonly used > functions across mempool, event and net devices. > > Signed-off-by: Pavan Nikhilesh > --- > drivers/Makefile | 14 -- > drivers/common/Makefile| 7 +++ > drivers/common/meson.build | 6 ++ > drivers/meson.build| 9 + > 4 files changed, 26 insertions(+), 10 deletions(-) > create mode 100644 drivers/common/Makefile > create mode 100644 drivers/common/meson.build > > diff --git a/drivers/Makefile b/drivers/Makefile > index ee65c87b0..bd83ad9f3 100644 > --- a/drivers/Makefile > +++ b/drivers/Makefile > @@ -4,17 +4,19 @@ > include $(RTE_SDK)/mk/rte.vars.mk > > DIRS-y += bus > +DIRS-y += common > +DEPDIRS-common := bus Why should common be dependent on bus? Shouldn't common be independent in itself? > DIRS-y += mempool > -DEPDIRS-mempool := bus > +DEPDIRS-mempool := bus common > DIRS-y += net > -DEPDIRS-net := bus mempool > +DEPDIRS-net := bus common mempool > DIRS-$(CONFIG_RTE_LIBRTE_BBDEV) += bbdev > -DEPDIRS-bbdev := bus mempool > +DEPDIRS-bbdev := bus common mempool > DIRS-$(CONFIG_RTE_LIBRTE_CRYPTODEV) += crypto > -DEPDIRS-crypto := bus mempool > +DEPDIRS-crypto := bus common mempool > DIRS-$(CONFIG_RTE_LIBRTE_EVENTDEV) += event > -DEPDIRS-event := bus mempool net > +DEPDIRS-event := bus common mempool net > DIRS-$(CONFIG_RTE_LIBRTE_RAWDEV) += raw > -DEPDIRS-raw := bus mempool net event > +DEPDIRS-raw := bus common mempool net event [...] - Shreyansh
Re: [dpdk-dev] [PATCH 11/17] vhost: add helpers for packed virtqueues
On Fri, Mar 16, 2018 at 04:21:14PM +0100, Jens Freimann wrote: > Add some helper functions to set/check descriptor flags > and toggle the used wrap counter. > > Signed-off-by: Jens Freimann [...] > > +static inline void > +toggle_wrap_counter(struct vhost_virtqueue *vq) > +{ > + vq->used_wrap_counter ^= 1; > +} > + > +static inline int > +desc_is_avail(struct vhost_virtqueue *vq, struct vring_desc_packed *desc) > +{ > + if (unlikely(!vq)) > + return -1; Maybe it's better to let the caller make sure the vq won't be NULL. > + > + if (vq->used_wrap_counter == 1) > + if ((desc->flags & DESC_AVAIL) && !(desc->flags & DESC_USED)) > + return 1; > + if (vq->used_wrap_counter == 0) Maybe it's better to use '} else {' here. Thanks > + if (!(desc->flags & DESC_AVAIL) && (desc->flags & DESC_USED)) > + return 1; > + return 0; > +}
[dpdk-dev] [PATCH] net/mlx5: fix ICC compilation
Remove the second declaration of device_attr [1] inside the loop as well as the query_device_ex() which has already been done outside of the loop. Fixes: 9a761de8ea14 ("net/mlx5: flow counter support") Cc: or...@mellanox.com Reported-by: Ferruh Yigit Signed-off-by: Nelio Laranjeiro [1] https://dpdk.org/ml/archives/dev/2018-March/091744.html --- drivers/net/mlx5/mlx5.c | 2 -- 1 file changed, 2 deletions(-) diff --git a/drivers/net/mlx5/mlx5.c b/drivers/net/mlx5/mlx5.c index a2acac703..b8b362a20 100644 --- a/drivers/net/mlx5/mlx5.c +++ b/drivers/net/mlx5/mlx5.c @@ -723,7 +723,6 @@ mlx5_pci_probe(struct rte_pci_driver *pci_drv __rte_unused, struct rte_eth_dev *eth_dev = NULL; struct ibv_device_attr_ex device_attr_ex; struct ether_addr mac; - struct ibv_device_attr_ex device_attr; struct mlx5_dev_config config = { .cqe_comp = cqe_comp, .mps = mps, @@ -780,7 +779,6 @@ mlx5_pci_probe(struct rte_pci_driver *pci_drv __rte_unused, err = ENODEV; goto port_error; } - mlx5_glue->query_device_ex(ctx, NULL, &device_attr); /* Check port status. */ err = mlx5_glue->query_port(ctx, port, &port_attr); if (err) { -- 2.11.0
Re: [dpdk-dev] [PATCH 12/17] vhost: dequeue for packed queues
On Fri, Mar 16, 2018 at 04:21:15PM +0100, Jens Freimann wrote: > Implement code to dequeue and process descriptors from > the vring if VIRTIO_F_PACKED is enabled. VIRTIO_F_RING_PACKED. > > Check if descriptor was made available by driver by looking at > VIRTIO_F_DESC_AVAIL flag in descriptor. If so dequeue and set > the used flag VIRTIO_F_DESC_USED to the current value of the > used wrap counter. > > Used ring wrap counter needs to be toggled when last descriptor is > written out. This allows the host/guest to detect new descriptors even > after the ring has wrapped. > > Signed-off-by: Jens Freimann > --- > lib/librte_vhost/vhost.c | 1 + > lib/librte_vhost/vhost.h | 1 + > lib/librte_vhost/virtio_net.c | 228 > ++ > 3 files changed, 230 insertions(+) > > diff --git a/lib/librte_vhost/vhost.c b/lib/librte_vhost/vhost.c > index a300812..8cba10d 100644 > --- a/lib/librte_vhost/vhost.c > +++ b/lib/librte_vhost/vhost.c > @@ -198,6 +198,7 @@ struct virtio_net * > > vq->kickfd = VIRTIO_UNINITIALIZED_EVENTFD; > vq->callfd = VIRTIO_UNINITIALIZED_EVENTFD; > + vq->used_wrap_counter = 1; > > vhost_user_iotlb_init(dev, vring_idx); > /* Backends are set to -1 indicating an inactive device. */ > diff --git a/lib/librte_vhost/vhost.h b/lib/librte_vhost/vhost.h > index d35c4b1..f77fefe 100644 > --- a/lib/librte_vhost/vhost.h > +++ b/lib/librte_vhost/vhost.h > @@ -108,6 +108,7 @@ struct vhost_virtqueue { > > struct batch_copy_elem *batch_copy_elems; > uint16_tbatch_copy_nb_elems; > + uint32_tused_wrap_counter; I didn't look into this, is uint32_t the best choice for defining used_wrap_counter compared with uint16_t or uint8_t? > > rte_rwlock_tiotlb_lock; > rte_rwlock_tiotlb_pending_lock; > diff --git a/lib/librte_vhost/virtio_net.c b/lib/librte_vhost/virtio_net.c > index 700aca7..8f59e4f 100644 > --- a/lib/librte_vhost/virtio_net.c > +++ b/lib/librte_vhost/virtio_net.c > @@ -19,6 +19,7 @@ > > #include "iotlb.h" > #include "vhost.h" > +#include "virtio-1.1.h" > > #define MAX_PKT_BURST 32 > > @@ -1118,6 +1119,233 @@ > } > } > > +static inline uint16_t > +dequeue_desc(struct virtio_net *dev, struct vhost_virtqueue *vq, It's better to have the word "packed" in the function name. > + struct rte_mempool *mbuf_pool, struct rte_mbuf *m, > + struct vring_desc_packed *descs) > +{ > + struct vring_desc_packed *desc; > + uint64_t desc_addr; > + uint32_t desc_avail, desc_offset; > + uint32_t mbuf_avail, mbuf_offset; > + uint32_t cpy_len; > + struct rte_mbuf *cur = m, *prev = m; > + struct virtio_net_hdr *hdr = NULL; > + uint16_t head_idx = vq->last_used_idx & (vq->size - 1); The ring size may not be a power of 2. > + int wrap_counter = vq->used_wrap_counter; > + int rc = 0; > + > + rte_spinlock_lock(&vq->access_lock); > + > + if (unlikely(vq->enabled == 0)) > + goto out; > + > + if (dev->features & (1ULL << VIRTIO_F_IOMMU_PLATFORM)) > + vhost_user_iotlb_rd_lock(vq); > + > + desc = &descs[vq->last_used_idx & (vq->size - 1)]; > + if (unlikely((desc->len < dev->vhost_hlen)) || > + (desc->flags & VRING_DESC_F_INDIRECT)) { > + RTE_LOG(ERR, VHOST_DATA, > + "INDIRECT not supported yet\n"); > + rc = -1; > + goto out; If INDIRECT isn't supported, we will need to make sure that INDIRECT and RING_PACKED won't be negotiated at the same time. Thanks
Re: [dpdk-dev] [PATCH 13/17] vhost: packed queue enqueue path
On Fri, Mar 16, 2018 at 04:21:16PM +0100, Jens Freimann wrote: [...] > +static inline uint32_t __attribute__((always_inline)) > +vhost_enqueue_burst_packed(struct virtio_net *dev, uint16_t queue_id, > + struct rte_mbuf **pkts, uint32_t count) > +{ > + struct vhost_virtqueue *vq; > + struct vring_desc_packed *descs; > + uint16_t idx; > + uint16_t mask; > + uint16_t i; > + > + vq = dev->virtqueue[queue_id]; > + > + rte_spinlock_lock(&vq->access_lock); > + > + if (unlikely(vq->enabled == 0)) { > + i = 0; > + goto out_access_unlock; > + } > + > + if (dev->features & (1ULL << VIRTIO_F_IOMMU_PLATFORM)) > + vhost_user_iotlb_rd_lock(vq); > + > + descs = vq->desc_packed; > + mask = vq->size - 1; Queue size may not be a power of 2 in packed ring. > + > + for (i = 0; i < count; i++) { [...] > + mbuf_avail = rte_pktmbuf_data_len(m); > + mbuf_offset = 0; > + while (mbuf_avail != 0 || m->next != NULL) { > + /* done with current mbuf, fetch next */ > + if (mbuf_avail == 0) { > + m = m->next; > + > + mbuf_offset = 0; > + mbuf_avail = rte_pktmbuf_data_len(m); > + } > + > + /* done with current desc buf, fetch next */ > + if (desc_avail == 0) { > + if ((desc->flags & VRING_DESC_F_NEXT) == 0) { > + /* Room in vring buffer is not enough */ > + goto out; > + } > + > + idx = (idx + 1); > + desc = &descs[idx]; Need to check whether idx >= queue size. Thanks
Re: [dpdk-dev] [PATCH v1 18/18] app/testpmd: add show device command
On Thu, Mar 15, 2018 at 06:49:48PM +0100, Gaetan Rivet wrote: > A new interactive command is offered: > >show device > > This commands lists all rte_device element matching the device > description. e.g.: > >show device bus=pci >show device bus=vdev >show device bus=vdev,class=eth >show device bus=vdev,class=eth,name=net_ring0 Silly mistake, it should read instead: show device bus=pci show device bus=vdev show device bus=vdev/class=eth show device bus=vdev/class=eth,name=net_ring0 > > These devices may not be otherwise useful, some buses will spawn devices > to keep track of their assets without having a driver to use them. > > Signed-off-by: Gaetan Rivet > --- > app/test-pmd/cmdline.c | 52 > ++ > 1 file changed, 52 insertions(+) > > diff --git a/app/test-pmd/cmdline.c b/app/test-pmd/cmdline.c > index 40b31ad7e..8ac5fb3ca 100644 > --- a/app/test-pmd/cmdline.c > +++ b/app/test-pmd/cmdline.c > @@ -6701,6 +6701,57 @@ cmdline_parse_inst_t cmd_showportall = { > }, > }; > > +/* *** SHOW DEVICE INFO *** */ > +struct cmd_showdevice_result { > + cmdline_fixed_string_t show; > + cmdline_fixed_string_t device; > + cmdline_fixed_string_t filter; > +}; > + > +static void > +cmd_showdevice_dump_device(const struct rte_device *dev) > +{ > + const struct rte_driver *drv = dev->driver; > + > + printf("0x%p: %s:%s\n", (const void *)dev, dev->name, > + drv ? drv->name : ""); > +} > + > +static void cmd_showdevice_parsed(void *parsed_result, > + __attribute__((unused)) struct cmdline *cl, > + __attribute__((unused)) void *data) > +{ > + struct cmd_showdevice_result *res = parsed_result; > + struct rte_dev_iterator it; > + const struct rte_device *dev; > + > + RTE_DEV_FOREACH(dev, res->filter, &it) > + cmd_showdevice_dump_device(dev); > +} > + > +cmdline_parse_token_string_t cmd_showdevice_show = > + TOKEN_STRING_INITIALIZER(struct cmd_showdevice_result, > + show, "show"); > +cmdline_parse_token_string_t cmd_showdevice_device = > + TOKEN_STRING_INITIALIZER(struct cmd_showdevice_result, > + device, "device"); > +cmdline_parse_token_string_t cmd_showdevice_filter = > + TOKEN_STRING_INITIALIZER(struct cmd_showdevice_result, > + filter, NULL); > + > +cmdline_parse_inst_t cmd_showdevice = { > + .f = cmd_showdevice_parsed, > + .data = NULL, > + .help_str = "show device " > + "", > + .tokens = { > + (void *)&cmd_showdevice_show, > + (void *)&cmd_showdevice_device, > + (void *)&cmd_showdevice_filter, > + NULL, > + }, > +}; > + > /* *** SHOW PORT INFO *** */ > struct cmd_showport_result { > cmdline_fixed_string_t show; > @@ -16038,6 +16089,7 @@ cmdline_parse_ctx_t main_ctx[] = { > (cmdline_parse_inst_t *)&cmd_help_long, > (cmdline_parse_inst_t *)&cmd_quit, > (cmdline_parse_inst_t *)&cmd_load_from_file, > + (cmdline_parse_inst_t *)&cmd_showdevice, > (cmdline_parse_inst_t *)&cmd_showport, > (cmdline_parse_inst_t *)&cmd_showqueue, > (cmdline_parse_inst_t *)&cmd_showportall, > -- > 2.11.0 > -- Gaëtan Rivet 6WIND
[dpdk-dev] [PATCH v5 01/19] crypto/ccp: add AMD ccp skeleton PMD
Signed-off-by: Ravi Kumar --- MAINTAINERS| 6 + config/common_base | 5 + doc/guides/rel_notes/release_18_05.rst | 5 + drivers/crypto/Makefile| 1 + drivers/crypto/ccp/Makefile| 29 drivers/crypto/ccp/rte_ccp_pmd.c | 36 ++ drivers/crypto/ccp/rte_pmd_ccp_version.map | 4 mk/rte.app.mk | 2 ++ 8 files changed, 88 insertions(+) create mode 100644 drivers/crypto/ccp/Makefile create mode 100644 drivers/crypto/ccp/rte_ccp_pmd.c create mode 100644 drivers/crypto/ccp/rte_pmd_ccp_version.map diff --git a/MAINTAINERS b/MAINTAINERS index a646ca3..8481731 100644 --- a/MAINTAINERS +++ b/MAINTAINERS @@ -640,6 +640,12 @@ M: Pablo de Lara T: git://dpdk.org/next/dpdk-next-crypto F: doc/guides/cryptodevs/features/default.ini +AMD CCP Crypto PMD +M: Ravi Kumar +F: drivers/crypto/ccp/ +F: doc/guides/cryptodevs/ccp.rst +F: doc/guides/cryptodevs/features/ccp.ini + ARMv8 Crypto M: Jerin Jacob F: drivers/crypto/armv8/ diff --git a/config/common_base b/config/common_base index ad03cf4..28237f0 100644 --- a/config/common_base +++ b/config/common_base @@ -529,6 +529,11 @@ CONFIG_RTE_LIBRTE_PMD_CRYPTO_SCHEDULER_DEBUG=n CONFIG_RTE_LIBRTE_PMD_NULL_CRYPTO=y # +# Compile PMD for AMD CCP crypto device +# +CONFIG_RTE_LIBRTE_PMD_CCP=n + +# # Compile PMD for Marvell Crypto device # CONFIG_RTE_LIBRTE_PMD_MRVL_CRYPTO=n diff --git a/doc/guides/rel_notes/release_18_05.rst b/doc/guides/rel_notes/release_18_05.rst index 3923dc2..c5b2854 100644 --- a/doc/guides/rel_notes/release_18_05.rst +++ b/doc/guides/rel_notes/release_18_05.rst @@ -41,6 +41,11 @@ New Features Also, make sure to start the actual text at the margin. = +* **Added a new crypto poll mode driver for AMD CCP devices.** + + Added the new ``ccp`` crypto driver for AMD CCP devices. See the + :doc:`../cryptodevs/ccp` crypto driver guide for more details on + this new driver. API Changes --- diff --git a/drivers/crypto/Makefile b/drivers/crypto/Makefile index 26e503e..9fbd986 100644 --- a/drivers/crypto/Makefile +++ b/drivers/crypto/Makefile @@ -20,5 +20,6 @@ endif ifeq ($(CONFIG_RTE_LIBRTE_DPAA_BUS),y) DIRS-$(CONFIG_RTE_LIBRTE_PMD_DPAA_SEC) += dpaa_sec endif +DIRS-$(CONFIG_RTE_LIBRTE_PMD_CCP) += ccp include $(RTE_SDK)/mk/rte.subdir.mk diff --git a/drivers/crypto/ccp/Makefile b/drivers/crypto/ccp/Makefile new file mode 100644 index 000..3b8442d --- /dev/null +++ b/drivers/crypto/ccp/Makefile @@ -0,0 +1,29 @@ +# SPDX-License-Identifier: BSD-3-Clause +# Copyright(c) 2018 Advanced Micro Devices, Inc. All rights reserved. + +include $(RTE_SDK)/mk/rte.vars.mk + +# library name +LIB = librte_pmd_ccp.a + +# build flags +CFLAGS += -O3 +CFLAGS += -I$(SRCDIR) +CFLAGS += $(WERROR_FLAGS) + +# library version +LIBABIVER := 1 + +# external library include paths +LDLIBS += -lcrypto +LDLIBS += -lrte_eal -lrte_mbuf -lrte_mempool -lrte_ring +LDLIBS += -lrte_cryptodev +LDLIBS += -lrte_pci -lrte_bus_pci + +# versioning export map +EXPORT_MAP := rte_pmd_ccp_version.map + +# library source files +SRCS-$(CONFIG_RTE_LIBRTE_PMD_CCP) += rte_ccp_pmd.c + +include $(RTE_SDK)/mk/rte.lib.mk diff --git a/drivers/crypto/ccp/rte_ccp_pmd.c b/drivers/crypto/ccp/rte_ccp_pmd.c new file mode 100644 index 000..71e7023 --- /dev/null +++ b/drivers/crypto/ccp/rte_ccp_pmd.c @@ -0,0 +1,36 @@ +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright(c) 2018 Advanced Micro Devices, Inc. All rights reserved. + */ + +#include +#include +#include + +uint8_t ccp_cryptodev_driver_id; + +/** Remove ccp pmd */ +static int +cryptodev_ccp_remove(struct rte_vdev_device *dev __rte_unused) +{ + return 0; +} + +/** Probe ccp pmd */ +static int +cryptodev_ccp_probe(struct rte_vdev_device *vdev __rte_unused) +{ + return 0; +} + +static struct rte_vdev_driver cryptodev_ccp_pmd_drv = { + .probe = cryptodev_ccp_probe, + .remove = cryptodev_ccp_remove +}; + +static struct cryptodev_driver ccp_crypto_drv; + +RTE_PMD_REGISTER_VDEV(CRYPTODEV_NAME_CCP_PMD, cryptodev_ccp_pmd_drv); +RTE_PMD_REGISTER_PARAM_STRING(CRYPTODEV_NAME_CCP_PMD, + "max_nb_queue_pairs= max_nb_sessions= socket_id="); +RTE_PMD_REGISTER_CRYPTO_DRIVER(ccp_crypto_drv, cryptodev_ccp_pmd_drv, + ccp_cryptodev_driver_id); diff --git a/drivers/crypto/ccp/rte_pmd_ccp_version.map b/drivers/crypto/ccp/rte_pmd_ccp_version.map new file mode 100644 index 000..9b9ab1a --- /dev/null +++ b/drivers/crypto/ccp/rte_pmd_ccp_version.map @@ -0,0 +1,4 @@ +DPDK_18.05 { + + local: *; +}; diff --git a/mk/rte.app.mk b/mk/rte.app.mk index 94525dc..6f50a15 100644 --- a/mk/rte.app.mk +++ b/mk/rte.app.mk @@ -221,6 +221,8 @@ ifeq ($(CONFIG_RTE_LIBRTE_DPAA_BUS),y) _LDLIBS-$(CONFIG_RTE_LIBRTE_PMD_DPAA_SEC) += -lr
[dpdk-dev] [PATCH v5 04/19] crypto/ccp: support session related crypto pmd ops
Signed-off-by: Ravi Kumar --- drivers/crypto/ccp/Makefile | 3 +- drivers/crypto/ccp/ccp_crypto.c | 203 + drivers/crypto/ccp/ccp_crypto.h | 241 +++ drivers/crypto/ccp/ccp_dev.h | 129 + drivers/crypto/ccp/ccp_pmd_ops.c | 61 +- 5 files changed, 633 insertions(+), 4 deletions(-) create mode 100644 drivers/crypto/ccp/ccp_crypto.c create mode 100644 drivers/crypto/ccp/ccp_crypto.h diff --git a/drivers/crypto/ccp/Makefile b/drivers/crypto/ccp/Makefile index f121712..3f8330d 100644 --- a/drivers/crypto/ccp/Makefile +++ b/drivers/crypto/ccp/Makefile @@ -25,8 +25,9 @@ EXPORT_MAP := rte_pmd_ccp_version.map # library source files SRCS-$(CONFIG_RTE_LIBRTE_PMD_CCP) += rte_ccp_pmd.c -SRCS-$(CONFIG_RTE_LIBRTE_PMD_CCP) += ccp_pmd_ops.c +SRCS-$(CONFIG_RTE_LIBRTE_PMD_CCP) += ccp_crypto.c SRCS-$(CONFIG_RTE_LIBRTE_PMD_CCP) += ccp_dev.c SRCS-$(CONFIG_RTE_LIBRTE_PMD_CCP) += ccp_pci.c +SRCS-$(CONFIG_RTE_LIBRTE_PMD_CCP) += ccp_pmd_ops.c include $(RTE_SDK)/mk/rte.lib.mk diff --git a/drivers/crypto/ccp/ccp_crypto.c b/drivers/crypto/ccp/ccp_crypto.c new file mode 100644 index 000..8bf4ce1 --- /dev/null +++ b/drivers/crypto/ccp/ccp_crypto.c @@ -0,0 +1,203 @@ +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright(c) 2018 Advanced Micro Devices, Inc. All rights reserved. + */ + +#include +#include +#include +#include +#include +#include +#include +#include + +#include +#include +#include +#include +#include +#include +#include + +#include "ccp_dev.h" +#include "ccp_crypto.h" +#include "ccp_pci.h" +#include "ccp_pmd_private.h" + +static enum ccp_cmd_order +ccp_get_cmd_id(const struct rte_crypto_sym_xform *xform) +{ + enum ccp_cmd_order res = CCP_CMD_NOT_SUPPORTED; + + if (xform == NULL) + return res; + if (xform->type == RTE_CRYPTO_SYM_XFORM_AUTH) { + if (xform->next == NULL) + return CCP_CMD_AUTH; + else if (xform->next->type == RTE_CRYPTO_SYM_XFORM_CIPHER) + return CCP_CMD_HASH_CIPHER; + } + if (xform->type == RTE_CRYPTO_SYM_XFORM_CIPHER) { + if (xform->next == NULL) + return CCP_CMD_CIPHER; + else if (xform->next->type == RTE_CRYPTO_SYM_XFORM_AUTH) + return CCP_CMD_CIPHER_HASH; + } + if (xform->type == RTE_CRYPTO_SYM_XFORM_AEAD) + return CCP_CMD_COMBINED; + return res; +} + +/* configure session */ +static int +ccp_configure_session_cipher(struct ccp_session *sess, +const struct rte_crypto_sym_xform *xform) +{ + const struct rte_crypto_cipher_xform *cipher_xform = NULL; + + cipher_xform = &xform->cipher; + + /* set cipher direction */ + if (cipher_xform->op == RTE_CRYPTO_CIPHER_OP_ENCRYPT) + sess->cipher.dir = CCP_CIPHER_DIR_ENCRYPT; + else + sess->cipher.dir = CCP_CIPHER_DIR_DECRYPT; + + /* set cipher key */ + sess->cipher.key_length = cipher_xform->key.length; + rte_memcpy(sess->cipher.key, cipher_xform->key.data, + cipher_xform->key.length); + + /* set iv parameters */ + sess->iv.offset = cipher_xform->iv.offset; + sess->iv.length = cipher_xform->iv.length; + + switch (cipher_xform->algo) { + default: + CCP_LOG_ERR("Unsupported cipher algo"); + return -1; + } + + + switch (sess->cipher.engine) { + default: + CCP_LOG_ERR("Invalid CCP Engine"); + return -ENOTSUP; + } + return 0; +} + +static int +ccp_configure_session_auth(struct ccp_session *sess, + const struct rte_crypto_sym_xform *xform) +{ + const struct rte_crypto_auth_xform *auth_xform = NULL; + + auth_xform = &xform->auth; + + sess->auth.digest_length = auth_xform->digest_length; + if (auth_xform->op == RTE_CRYPTO_AUTH_OP_GENERATE) + sess->auth.op = CCP_AUTH_OP_GENERATE; + else + sess->auth.op = CCP_AUTH_OP_VERIFY; + switch (auth_xform->algo) { + default: + CCP_LOG_ERR("Unsupported hash algo"); + return -ENOTSUP; + } + return 0; +} + +static int +ccp_configure_session_aead(struct ccp_session *sess, + const struct rte_crypto_sym_xform *xform) +{ + const struct rte_crypto_aead_xform *aead_xform = NULL; + + aead_xform = &xform->aead; + + sess->cipher.key_length = aead_xform->key.length; + rte_memcpy(sess->cipher.key, aead_xform->key.data, + aead_xform->key.length); + + if (aead_xform->op == RTE_CRYPTO_AEAD_OP_ENCRYPT) { + sess->cipher.dir = CCP_CIPHER_DIR_ENCRYPT; + sess->auth.op = CCP_AUTH_OP_GENERATE; + } else { + sess->cipher
[dpdk-dev] [PATCH v5 03/19] crypto/ccp: support basic pmd ops
Signed-off-by: Ravi Kumar --- drivers/crypto/ccp/ccp_dev.c | 9 ++ drivers/crypto/ccp/ccp_dev.h | 9 ++ drivers/crypto/ccp/ccp_pmd_ops.c | 61 +--- drivers/crypto/ccp/ccp_pmd_private.h | 43 + 4 files changed, 117 insertions(+), 5 deletions(-) diff --git a/drivers/crypto/ccp/ccp_dev.c b/drivers/crypto/ccp/ccp_dev.c index 42cff75..546f0a7 100644 --- a/drivers/crypto/ccp/ccp_dev.c +++ b/drivers/crypto/ccp/ccp_dev.c @@ -26,6 +26,15 @@ struct ccp_list ccp_list = TAILQ_HEAD_INITIALIZER(ccp_list); static int ccp_dev_id; +int +ccp_dev_start(struct rte_cryptodev *dev) +{ + struct ccp_private *priv = dev->data->dev_private; + + priv->last_dev = TAILQ_FIRST(&ccp_list); + return 0; +} + static const struct rte_memzone * ccp_queue_dma_zone_reserve(const char *queue_name, uint32_t queue_size, diff --git a/drivers/crypto/ccp/ccp_dev.h b/drivers/crypto/ccp/ccp_dev.h index f594e91..c360fe0 100644 --- a/drivers/crypto/ccp/ccp_dev.h +++ b/drivers/crypto/ccp/ccp_dev.h @@ -78,6 +78,10 @@ #define LSB_ITEM_SIZE 32 #define SLSB_MAP_SIZE (MAX_LSB_CNT * LSB_SIZE) +/* General CCP Defines */ + +#define CCP_SB_BYTES32 + /* bitmap */ enum { BITS_PER_WORD = sizeof(unsigned long) * CHAR_BIT @@ -273,6 +277,11 @@ high32_value(unsigned long addr) return ((uint64_t)addr >> 32) & 0x0; } +/* + * Start CCP device + */ +int ccp_dev_start(struct rte_cryptodev *dev); + /** * Detect ccp platform and initialize all ccp devices * diff --git a/drivers/crypto/ccp/ccp_pmd_ops.c b/drivers/crypto/ccp/ccp_pmd_ops.c index 099eb21..99b8ca5 100644 --- a/drivers/crypto/ccp/ccp_pmd_ops.c +++ b/drivers/crypto/ccp/ccp_pmd_ops.c @@ -2,18 +2,69 @@ * Copyright(c) 2018 Advanced Micro Devices, Inc. All rights reserved. */ +#include + +#include #include +#include + +#include "ccp_pmd_private.h" +#include "ccp_dev.h" + +static const struct rte_cryptodev_capabilities ccp_pmd_capabilities[] = { + RTE_CRYPTODEV_END_OF_CAPABILITIES_LIST() +}; + +static int +ccp_pmd_config(struct rte_cryptodev *dev __rte_unused, + struct rte_cryptodev_config *config __rte_unused) +{ + return 0; +} + +static int +ccp_pmd_start(struct rte_cryptodev *dev) +{ + return ccp_dev_start(dev); +} + +static void +ccp_pmd_stop(struct rte_cryptodev *dev __rte_unused) +{ + +} + +static int +ccp_pmd_close(struct rte_cryptodev *dev __rte_unused) +{ + return 0; +} + +static void +ccp_pmd_info_get(struct rte_cryptodev *dev, +struct rte_cryptodev_info *dev_info) +{ + struct ccp_private *internals = dev->data->dev_private; + + if (dev_info != NULL) { + dev_info->driver_id = dev->driver_id; + dev_info->feature_flags = dev->feature_flags; + dev_info->capabilities = ccp_pmd_capabilities; + dev_info->max_nb_queue_pairs = internals->max_nb_qpairs; + dev_info->sym.max_nb_sessions = internals->max_nb_sessions; + } +} struct rte_cryptodev_ops ccp_ops = { - .dev_configure = NULL, - .dev_start = NULL, - .dev_stop = NULL, - .dev_close = NULL, + .dev_configure = ccp_pmd_config, + .dev_start = ccp_pmd_start, + .dev_stop = ccp_pmd_stop, + .dev_close = ccp_pmd_close, .stats_get = NULL, .stats_reset= NULL, - .dev_infos_get = NULL, + .dev_infos_get = ccp_pmd_info_get, .queue_pair_setup = NULL, .queue_pair_release = NULL, diff --git a/drivers/crypto/ccp/ccp_pmd_private.h b/drivers/crypto/ccp/ccp_pmd_private.h index 0da9b92..47c4fb2 100644 --- a/drivers/crypto/ccp/ccp_pmd_private.h +++ b/drivers/crypto/ccp/ccp_pmd_private.h @@ -34,13 +34,56 @@ #define CCP_NB_MAX_DESCRIPTORS 1024 #define CCP_MAX_BURST 64 +#include "ccp_dev.h" + /* private data structure for each CCP crypto device */ struct ccp_private { unsigned int max_nb_qpairs; /**< Max number of queue pairs */ unsigned int max_nb_sessions; /**< Max number of sessions */ uint8_t crypto_num_dev; /**< Number of working crypto devices */ + struct ccp_device *last_dev;/**< Last working crypto device */ }; +/* CCP batch info */ +struct ccp_batch_info { + struct rte_crypto_op *op[CCP_MAX_BURST]; + /**< optable populated at enque time from app*/ + int op_idx; + struct ccp_queue *cmd_q; + uint16_t opcnt; + /**< no. of crypto ops in batch*/ + int desccnt; + /**< no. of ccp queue descriptors*/ + uint32_t head_offset; + /**< ccp queue
[dpdk-dev] [PATCH v5 02/19] crypto/ccp: support ccp device initialization and deintialization
Signed-off-by: Ravi Kumar --- drivers/crypto/ccp/Makefile | 3 + drivers/crypto/ccp/ccp_dev.c | 761 +++ drivers/crypto/ccp/ccp_dev.h | 284 + drivers/crypto/ccp/ccp_pci.c | 236 +++ drivers/crypto/ccp/ccp_pci.h | 27 ++ drivers/crypto/ccp/ccp_pmd_ops.c | 29 ++ drivers/crypto/ccp/ccp_pmd_private.h | 56 +++ drivers/crypto/ccp/rte_ccp_pmd.c | 151 ++- 8 files changed, 1545 insertions(+), 2 deletions(-) create mode 100644 drivers/crypto/ccp/ccp_dev.c create mode 100644 drivers/crypto/ccp/ccp_dev.h create mode 100644 drivers/crypto/ccp/ccp_pci.c create mode 100644 drivers/crypto/ccp/ccp_pci.h create mode 100644 drivers/crypto/ccp/ccp_pmd_ops.c create mode 100644 drivers/crypto/ccp/ccp_pmd_private.h diff --git a/drivers/crypto/ccp/Makefile b/drivers/crypto/ccp/Makefile index 3b8442d..f121712 100644 --- a/drivers/crypto/ccp/Makefile +++ b/drivers/crypto/ccp/Makefile @@ -25,5 +25,8 @@ EXPORT_MAP := rte_pmd_ccp_version.map # library source files SRCS-$(CONFIG_RTE_LIBRTE_PMD_CCP) += rte_ccp_pmd.c +SRCS-$(CONFIG_RTE_LIBRTE_PMD_CCP) += ccp_pmd_ops.c +SRCS-$(CONFIG_RTE_LIBRTE_PMD_CCP) += ccp_dev.c +SRCS-$(CONFIG_RTE_LIBRTE_PMD_CCP) += ccp_pci.c include $(RTE_SDK)/mk/rte.lib.mk diff --git a/drivers/crypto/ccp/ccp_dev.c b/drivers/crypto/ccp/ccp_dev.c new file mode 100644 index 000..42cff75 --- /dev/null +++ b/drivers/crypto/ccp/ccp_dev.c @@ -0,0 +1,761 @@ +/* SPDX-License-Identifier: BSD-3-Clause + * Copyright(c) 2018 Advanced Micro Devices, Inc. All rights reserved. + */ + +#include +#include +#include +#include +#include +#include +#include +#include +#include + +#include +#include +#include +#include +#include +#include + +#include "ccp_dev.h" +#include "ccp_pci.h" +#include "ccp_pmd_private.h" + +struct ccp_list ccp_list = TAILQ_HEAD_INITIALIZER(ccp_list); +static int ccp_dev_id; + +static const struct rte_memzone * +ccp_queue_dma_zone_reserve(const char *queue_name, + uint32_t queue_size, + int socket_id) +{ + const struct rte_memzone *mz; + unsigned int memzone_flags = 0; + const struct rte_memseg *ms; + + mz = rte_memzone_lookup(queue_name); + if (mz != 0) + return mz; + + ms = rte_eal_get_physmem_layout(); + switch (ms[0].hugepage_sz) { + case(RTE_PGSIZE_2M): + memzone_flags = RTE_MEMZONE_2MB; + break; + case(RTE_PGSIZE_1G): + memzone_flags = RTE_MEMZONE_1GB; + break; + case(RTE_PGSIZE_16M): + memzone_flags = RTE_MEMZONE_16MB; + break; + case(RTE_PGSIZE_16G): + memzone_flags = RTE_MEMZONE_16GB; + break; + default: + memzone_flags = RTE_MEMZONE_SIZE_HINT_ONLY; + } + + return rte_memzone_reserve_aligned(queue_name, + queue_size, + socket_id, + memzone_flags, + queue_size); +} + +/* bitmap support apis */ +static inline void +ccp_set_bit(unsigned long *bitmap, int n) +{ + __sync_fetch_and_or(&bitmap[WORD_OFFSET(n)], (1UL << BIT_OFFSET(n))); +} + +static inline void +ccp_clear_bit(unsigned long *bitmap, int n) +{ + __sync_fetch_and_and(&bitmap[WORD_OFFSET(n)], ~(1UL << BIT_OFFSET(n))); +} + +static inline uint32_t +ccp_get_bit(unsigned long *bitmap, int n) +{ + return ((bitmap[WORD_OFFSET(n)] & (1 << BIT_OFFSET(n))) != 0); +} + + +static inline uint32_t +ccp_ffz(unsigned long word) +{ + unsigned long first_zero; + + first_zero = __builtin_ffsl(~word); + return first_zero ? (first_zero - 1) : + BITS_PER_WORD; +} + +static inline uint32_t +ccp_find_first_zero_bit(unsigned long *addr, uint32_t limit) +{ + uint32_t i; + uint32_t nwords = 0; + + nwords = (limit - 1) / BITS_PER_WORD + 1; + for (i = 0; i < nwords; i++) { + if (addr[i] == 0UL) + return i * BITS_PER_WORD; + if (addr[i] < ~(0UL)) + break; + } + return (i == nwords) ? limit : i * BITS_PER_WORD + ccp_ffz(addr[i]); +} + +static void +ccp_bitmap_set(unsigned long *map, unsigned int start, int len) +{ + unsigned long *p = map + WORD_OFFSET(start); + const unsigned int size = start + len; + int bits_to_set = BITS_PER_WORD - (start % BITS_PER_WORD); + unsigned long mask_to_set = CCP_BITMAP_FIRST_WORD_MASK(start); + + while (len - bits_to_set >= 0) { + *p |= mask_to_set; + len -= bits_to_set; + bits_to_set = BITS_PER_WORD; + mask_to_set = ~0UL; + p++; + } + if (len) { + mask_to_set &= CCP_BITMAP_LAST_WORD_MASK(
[dpdk-dev] [PATCH v5 05/19] crypto/ccp: support queue pair related pmd ops
Signed-off-by: Ravi Kumar --- drivers/crypto/ccp/ccp_pmd_ops.c | 149 +-- 1 file changed, 144 insertions(+), 5 deletions(-) diff --git a/drivers/crypto/ccp/ccp_pmd_ops.c b/drivers/crypto/ccp/ccp_pmd_ops.c index 0560f68..bd0aea4 100644 --- a/drivers/crypto/ccp/ccp_pmd_ops.c +++ b/drivers/crypto/ccp/ccp_pmd_ops.c @@ -56,6 +56,145 @@ ccp_pmd_info_get(struct rte_cryptodev *dev, } } +static int +ccp_pmd_qp_release(struct rte_cryptodev *dev, uint16_t qp_id) +{ + struct ccp_qp *qp; + + if (dev->data->queue_pairs[qp_id] != NULL) { + qp = (struct ccp_qp *)dev->data->queue_pairs[qp_id]; + rte_ring_free(qp->processed_pkts); + rte_mempool_free(qp->batch_mp); + rte_free(qp); + dev->data->queue_pairs[qp_id] = NULL; + } + return 0; +} + +static int +ccp_pmd_qp_set_unique_name(struct rte_cryptodev *dev, + struct ccp_qp *qp) +{ + unsigned int n = snprintf(qp->name, sizeof(qp->name), + "ccp_pmd_%u_qp_%u", + dev->data->dev_id, qp->id); + + if (n > sizeof(qp->name)) + return -1; + + return 0; +} + +static struct rte_ring * +ccp_pmd_qp_create_batch_info_ring(struct ccp_qp *qp, + unsigned int ring_size, int socket_id) +{ + struct rte_ring *r; + + r = rte_ring_lookup(qp->name); + if (r) { + if (r->size >= ring_size) { + CCP_LOG_INFO( + "Reusing ring %s for processed packets", + qp->name); + return r; + } + CCP_LOG_INFO( + "Unable to reuse ring %s for processed packets", +qp->name); + return NULL; + } + + return rte_ring_create(qp->name, ring_size, socket_id, + RING_F_SP_ENQ | RING_F_SC_DEQ); +} + +static int +ccp_pmd_qp_setup(struct rte_cryptodev *dev, uint16_t qp_id, +const struct rte_cryptodev_qp_conf *qp_conf, +int socket_id, struct rte_mempool *session_pool) +{ + struct ccp_private *internals = dev->data->dev_private; + struct ccp_qp *qp; + int retval = 0; + + if (qp_id >= internals->max_nb_qpairs) { + CCP_LOG_ERR("Invalid qp_id %u, should be less than %u", + qp_id, internals->max_nb_qpairs); + return (-EINVAL); + } + + /* Free memory prior to re-allocation if needed. */ + if (dev->data->queue_pairs[qp_id] != NULL) + ccp_pmd_qp_release(dev, qp_id); + + /* Allocate the queue pair data structure. */ + qp = rte_zmalloc_socket("CCP Crypto PMD Queue Pair", sizeof(*qp), + RTE_CACHE_LINE_SIZE, socket_id); + if (qp == NULL) { + CCP_LOG_ERR("Failed to allocate queue pair memory"); + return (-ENOMEM); + } + + qp->dev = dev; + qp->id = qp_id; + dev->data->queue_pairs[qp_id] = qp; + + retval = ccp_pmd_qp_set_unique_name(dev, qp); + if (retval) { + CCP_LOG_ERR("Failed to create unique name for ccp qp"); + goto qp_setup_cleanup; + } + + qp->processed_pkts = ccp_pmd_qp_create_batch_info_ring(qp, + qp_conf->nb_descriptors, socket_id); + if (qp->processed_pkts == NULL) { + CCP_LOG_ERR("Failed to create batch info ring"); + goto qp_setup_cleanup; + } + + qp->sess_mp = session_pool; + + /* mempool for batch info */ + qp->batch_mp = rte_mempool_create( + qp->name, + qp_conf->nb_descriptors, + sizeof(struct ccp_batch_info), + RTE_CACHE_LINE_SIZE, + 0, NULL, NULL, NULL, NULL, + SOCKET_ID_ANY, 0); + if (qp->batch_mp == NULL) + goto qp_setup_cleanup; + memset(&qp->qp_stats, 0, sizeof(qp->qp_stats)); + return 0; + +qp_setup_cleanup: + dev->data->queue_pairs[qp_id] = NULL; + if (qp) + rte_free(qp); + return -1; +} + +static int +ccp_pmd_qp_start(struct rte_cryptodev *dev __rte_unused, +uint16_t queue_pair_id __rte_unused) +{ + return -ENOTSUP; +} + +static int +ccp_pmd_qp_stop(struct rte_cryptodev *dev __rte_unused, + uint16_t queue_pair_id __rte_unused) +{ + return -ENOTSUP; +} + +static uint32_t +ccp_pmd_qp_count(struct rte_cryptodev *dev) +{ + return dev->data->nb_queue_pairs; +} + static unsigned ccp_pmd_session_get_size(struct rte_cryptodev *dev __rte_unused) { @@ -121,11 +260,11 @@ struct rte_cryptodev_ops ccp_ops = { .dev_infos_get = ccp_p
[dpdk-dev] [PATCH v5 06/19] crypto/ccp: support crypto enqueue and dequeue burst api
Signed-off-by: Ravi Kumar --- drivers/crypto/ccp/ccp_crypto.c | 360 +++ drivers/crypto/ccp/ccp_crypto.h | 35 drivers/crypto/ccp/ccp_dev.c | 27 +++ drivers/crypto/ccp/ccp_dev.h | 9 + drivers/crypto/ccp/rte_ccp_pmd.c | 64 ++- 5 files changed, 488 insertions(+), 7 deletions(-) diff --git a/drivers/crypto/ccp/ccp_crypto.c b/drivers/crypto/ccp/ccp_crypto.c index 8bf4ce1..7ced435 100644 --- a/drivers/crypto/ccp/ccp_crypto.c +++ b/drivers/crypto/ccp/ccp_crypto.c @@ -201,3 +201,363 @@ ccp_set_session_parameters(struct ccp_session *sess, } return ret; } + +/* calculate CCP descriptors requirement */ +static inline int +ccp_cipher_slot(struct ccp_session *session) +{ + int count = 0; + + switch (session->cipher.algo) { + default: + CCP_LOG_ERR("Unsupported cipher algo %d", + session->cipher.algo); + } + return count; +} + +static inline int +ccp_auth_slot(struct ccp_session *session) +{ + int count = 0; + + switch (session->auth.algo) { + default: + CCP_LOG_ERR("Unsupported auth algo %d", + session->auth.algo); + } + + return count; +} + +static int +ccp_aead_slot(struct ccp_session *session) +{ + int count = 0; + + switch (session->aead_algo) { + default: + CCP_LOG_ERR("Unsupported aead algo %d", + session->aead_algo); + } + return count; +} + +int +ccp_compute_slot_count(struct ccp_session *session) +{ + int count = 0; + + switch (session->cmd_id) { + case CCP_CMD_CIPHER: + count = ccp_cipher_slot(session); + break; + case CCP_CMD_AUTH: + count = ccp_auth_slot(session); + break; + case CCP_CMD_CIPHER_HASH: + case CCP_CMD_HASH_CIPHER: + count = ccp_cipher_slot(session); + count += ccp_auth_slot(session); + break; + case CCP_CMD_COMBINED: + count = ccp_aead_slot(session); + break; + default: + CCP_LOG_ERR("Unsupported cmd_id"); + + } + + return count; +} + +static inline int +ccp_crypto_cipher(struct rte_crypto_op *op, + struct ccp_queue *cmd_q __rte_unused, + struct ccp_batch_info *b_info __rte_unused) +{ + int result = 0; + struct ccp_session *session; + + session = (struct ccp_session *)get_session_private_data( +op->sym->session, +ccp_cryptodev_driver_id); + + switch (session->cipher.algo) { + default: + CCP_LOG_ERR("Unsupported cipher algo %d", + session->cipher.algo); + return -ENOTSUP; + } + return result; +} + +static inline int +ccp_crypto_auth(struct rte_crypto_op *op, + struct ccp_queue *cmd_q __rte_unused, + struct ccp_batch_info *b_info __rte_unused) +{ + + int result = 0; + struct ccp_session *session; + + session = (struct ccp_session *)get_session_private_data( +op->sym->session, + ccp_cryptodev_driver_id); + + switch (session->auth.algo) { + default: + CCP_LOG_ERR("Unsupported auth algo %d", + session->auth.algo); + return -ENOTSUP; + } + + return result; +} + +static inline int +ccp_crypto_aead(struct rte_crypto_op *op, + struct ccp_queue *cmd_q __rte_unused, + struct ccp_batch_info *b_info __rte_unused) +{ + int result = 0; + struct ccp_session *session; + + session = (struct ccp_session *)get_session_private_data( +op->sym->session, + ccp_cryptodev_driver_id); + + switch (session->aead_algo) { + default: + CCP_LOG_ERR("Unsupported aead algo %d", + session->aead_algo); + return -ENOTSUP; + } + return result; +} + +int +process_ops_to_enqueue(const struct ccp_qp *qp, + struct rte_crypto_op **op, + struct ccp_queue *cmd_q, + uint16_t nb_ops, + int slots_req) +{ + int i, result = 0; + struct ccp_batch_info *b_info; + struct ccp_session *session; + + if (rte_mempool_get(qp->batch_mp, (void **)&b_info)) { + CCP_LOG_ERR("batch info allocation failed"); + return 0; + } + /* populate batch info necessary for dequeue */ + b_info->op_idx = 0; + b_info->lsb_buf_idx = 0; + b_info->desccnt = 0; + b_info->cmd_q = cmd_q; + b_info->lsb_buf_phys = +
[dpdk-dev] [PATCH v5 07/19] crypto/ccp: support sessionless operations
Signed-off-by: Ravi Kumar --- drivers/crypto/ccp/rte_ccp_pmd.c | 33 +++-- 1 file changed, 31 insertions(+), 2 deletions(-) diff --git a/drivers/crypto/ccp/rte_ccp_pmd.c b/drivers/crypto/ccp/rte_ccp_pmd.c index fd7d2d3..b3b2651 100644 --- a/drivers/crypto/ccp/rte_ccp_pmd.c +++ b/drivers/crypto/ccp/rte_ccp_pmd.c @@ -23,7 +23,7 @@ static unsigned int ccp_pmd_init_done; uint8_t ccp_cryptodev_driver_id; static struct ccp_session * -get_ccp_session(struct ccp_qp *qp __rte_unused, struct rte_crypto_op *op) +get_ccp_session(struct ccp_qp *qp, struct rte_crypto_op *op) { struct ccp_session *sess = NULL; @@ -35,6 +35,27 @@ get_ccp_session(struct ccp_qp *qp __rte_unused, struct rte_crypto_op *op) get_session_private_data( op->sym->session, ccp_cryptodev_driver_id); + } else if (op->sess_type == RTE_CRYPTO_OP_SESSIONLESS) { + void *_sess; + void *_sess_private_data = NULL; + + if (rte_mempool_get(qp->sess_mp, &_sess)) + return NULL; + if (rte_mempool_get(qp->sess_mp, (void **)&_sess_private_data)) + return NULL; + + sess = (struct ccp_session *)_sess_private_data; + + if (unlikely(ccp_set_session_parameters(sess, + op->sym->xform) != 0)) { + rte_mempool_put(qp->sess_mp, _sess); + rte_mempool_put(qp->sess_mp, _sess_private_data); + sess = NULL; + } + op->sym->session = (struct rte_cryptodev_sym_session *)_sess; + set_session_private_data(op->sym->session, +ccp_cryptodev_driver_id, +_sess_private_data); } return sess; @@ -82,10 +103,18 @@ ccp_pmd_dequeue_burst(void *queue_pair, struct rte_crypto_op **ops, uint16_t nb_ops) { struct ccp_qp *qp = queue_pair; - uint16_t nb_dequeued = 0; + uint16_t nb_dequeued = 0, i; nb_dequeued = process_ops_to_dequeue(qp, ops, nb_ops); + /* Free session if a session-less crypto op */ + for (i = 0; i < nb_dequeued; i++) + if (unlikely(ops[i]->sess_type == +RTE_CRYPTO_OP_SESSIONLESS)) { + rte_mempool_put(qp->sess_mp, + ops[i]->sym->session); + ops[i]->sym->session = NULL; + } qp->qp_stats.dequeued_count += nb_dequeued; return nb_dequeued; -- 2.7.4
[dpdk-dev] [PATCH v5 09/19] crypto/ccp: support ccp hwrng feature
Signed-off-by: Ravi Kumar --- drivers/crypto/ccp/ccp_dev.c | 20 drivers/crypto/ccp/ccp_dev.h | 11 +++ 2 files changed, 31 insertions(+) diff --git a/drivers/crypto/ccp/ccp_dev.c b/drivers/crypto/ccp/ccp_dev.c index 2e702de..55cfcdd 100644 --- a/drivers/crypto/ccp/ccp_dev.c +++ b/drivers/crypto/ccp/ccp_dev.c @@ -62,6 +62,26 @@ ccp_allot_queue(struct rte_cryptodev *cdev, int slot_req) return NULL; } +int +ccp_read_hwrng(uint32_t *value) +{ + struct ccp_device *dev; + + TAILQ_FOREACH(dev, &ccp_list, next) { + void *vaddr = (void *)(dev->pci.mem_resource[2].addr); + + while (dev->hwrng_retries++ < CCP_MAX_TRNG_RETRIES) { + *value = CCP_READ_REG(vaddr, TRNG_OUT_REG); + if (*value) { + dev->hwrng_retries = 0; + return 0; + } + } + dev->hwrng_retries = 0; + } + return -1; +} + static const struct rte_memzone * ccp_queue_dma_zone_reserve(const char *queue_name, uint32_t queue_size, diff --git a/drivers/crypto/ccp/ccp_dev.h b/drivers/crypto/ccp/ccp_dev.h index abc788c..de375ee 100644 --- a/drivers/crypto/ccp/ccp_dev.h +++ b/drivers/crypto/ccp/ccp_dev.h @@ -21,6 +21,7 @@ /**< CCP sspecific */ #define MAX_HW_QUEUES 5 +#define CCP_MAX_TRNG_RETRIES 10 /**< CCP Register Mappings */ #define Q_MASK_REG 0x000 @@ -197,6 +198,8 @@ struct ccp_device { /**< protection for shared lsb region allocation */ int qidx; /**< current queue index */ + int hwrng_retries; + /**< retry counter for CCP TRNG */ } __rte_cache_aligned; /**< CCP H/W engine related */ @@ -428,4 +431,12 @@ int ccp_probe_devices(const struct rte_pci_id *ccp_id); */ struct ccp_queue *ccp_allot_queue(struct rte_cryptodev *dev, int slot_req); +/** + * read hwrng value + * + * @param trng_value data pointer to write RNG value + * @return 0 on success otherwise -1 + */ +int ccp_read_hwrng(uint32_t *trng_value); + #endif /* _CCP_DEV_H_ */ -- 2.7.4
[dpdk-dev] [PATCH v5 08/19] crypto/ccp: support stats related crypto pmd ops
Signed-off-by: Ravi Kumar --- drivers/crypto/ccp/ccp_pmd_ops.c | 34 -- 1 file changed, 32 insertions(+), 2 deletions(-) diff --git a/drivers/crypto/ccp/ccp_pmd_ops.c b/drivers/crypto/ccp/ccp_pmd_ops.c index bd0aea4..d3708f4 100644 --- a/drivers/crypto/ccp/ccp_pmd_ops.c +++ b/drivers/crypto/ccp/ccp_pmd_ops.c @@ -42,6 +42,36 @@ ccp_pmd_close(struct rte_cryptodev *dev __rte_unused) } static void +ccp_pmd_stats_get(struct rte_cryptodev *dev, + struct rte_cryptodev_stats *stats) +{ + int qp_id; + + for (qp_id = 0; qp_id < dev->data->nb_queue_pairs; qp_id++) { + struct ccp_qp *qp = dev->data->queue_pairs[qp_id]; + + stats->enqueued_count += qp->qp_stats.enqueued_count; + stats->dequeued_count += qp->qp_stats.dequeued_count; + + stats->enqueue_err_count += qp->qp_stats.enqueue_err_count; + stats->dequeue_err_count += qp->qp_stats.dequeue_err_count; + } + +} + +static void +ccp_pmd_stats_reset(struct rte_cryptodev *dev) +{ + int qp_id; + + for (qp_id = 0; qp_id < dev->data->nb_queue_pairs; qp_id++) { + struct ccp_qp *qp = dev->data->queue_pairs[qp_id]; + + memset(&qp->qp_stats, 0, sizeof(qp->qp_stats)); + } +} + +static void ccp_pmd_info_get(struct rte_cryptodev *dev, struct rte_cryptodev_info *dev_info) { @@ -255,8 +285,8 @@ struct rte_cryptodev_ops ccp_ops = { .dev_stop = ccp_pmd_stop, .dev_close = ccp_pmd_close, - .stats_get = NULL, - .stats_reset= NULL, + .stats_get = ccp_pmd_stats_get, + .stats_reset= ccp_pmd_stats_reset, .dev_infos_get = ccp_pmd_info_get, -- 2.7.4
[dpdk-dev] [PATCH v5 10/19] crypto/ccp: support aes cipher algo
Signed-off-by: Ravi Kumar --- drivers/crypto/ccp/ccp_crypto.c | 197 ++- drivers/crypto/ccp/ccp_crypto.h | 13 +++ drivers/crypto/ccp/ccp_dev.h | 53 +++ drivers/crypto/ccp/ccp_pmd_ops.c | 60 4 files changed, 321 insertions(+), 2 deletions(-) diff --git a/drivers/crypto/ccp/ccp_crypto.c b/drivers/crypto/ccp/ccp_crypto.c index 7ced435..3c46ac3 100644 --- a/drivers/crypto/ccp/ccp_crypto.c +++ b/drivers/crypto/ccp/ccp_crypto.c @@ -54,6 +54,7 @@ ccp_configure_session_cipher(struct ccp_session *sess, const struct rte_crypto_sym_xform *xform) { const struct rte_crypto_cipher_xform *cipher_xform = NULL; + size_t i; cipher_xform = &xform->cipher; @@ -73,6 +74,21 @@ ccp_configure_session_cipher(struct ccp_session *sess, sess->iv.length = cipher_xform->iv.length; switch (cipher_xform->algo) { + case RTE_CRYPTO_CIPHER_AES_CTR: + sess->cipher.algo = CCP_CIPHER_ALGO_AES_CTR; + sess->cipher.um.aes_mode = CCP_AES_MODE_CTR; + sess->cipher.engine = CCP_ENGINE_AES; + break; + case RTE_CRYPTO_CIPHER_AES_ECB: + sess->cipher.algo = CCP_CIPHER_ALGO_AES_CBC; + sess->cipher.um.aes_mode = CCP_AES_MODE_ECB; + sess->cipher.engine = CCP_ENGINE_AES; + break; + case RTE_CRYPTO_CIPHER_AES_CBC: + sess->cipher.algo = CCP_CIPHER_ALGO_AES_CBC; + sess->cipher.um.aes_mode = CCP_AES_MODE_CBC; + sess->cipher.engine = CCP_ENGINE_AES; + break; default: CCP_LOG_ERR("Unsupported cipher algo"); return -1; @@ -80,10 +96,27 @@ ccp_configure_session_cipher(struct ccp_session *sess, switch (sess->cipher.engine) { + case CCP_ENGINE_AES: + if (sess->cipher.key_length == 16) + sess->cipher.ut.aes_type = CCP_AES_TYPE_128; + else if (sess->cipher.key_length == 24) + sess->cipher.ut.aes_type = CCP_AES_TYPE_192; + else if (sess->cipher.key_length == 32) + sess->cipher.ut.aes_type = CCP_AES_TYPE_256; + else { + CCP_LOG_ERR("Invalid cipher key length"); + return -1; + } + for (i = 0; i < sess->cipher.key_length ; i++) + sess->cipher.key_ccp[sess->cipher.key_length - i - 1] = + sess->cipher.key[i]; + break; default: CCP_LOG_ERR("Invalid CCP Engine"); return -ENOTSUP; } + sess->cipher.nonce_phys = rte_mem_virt2phy(sess->cipher.nonce); + sess->cipher.key_phys = rte_mem_virt2phy(sess->cipher.key_ccp); return 0; } @@ -209,6 +242,18 @@ ccp_cipher_slot(struct ccp_session *session) int count = 0; switch (session->cipher.algo) { + case CCP_CIPHER_ALGO_AES_CBC: + count = 2; + /**< op + passthrough for iv */ + break; + case CCP_CIPHER_ALGO_AES_ECB: + count = 1; + /**cipher.algo); @@ -271,10 +316,146 @@ ccp_compute_slot_count(struct ccp_session *session) return count; } +static void +ccp_perform_passthru(struct ccp_passthru *pst, +struct ccp_queue *cmd_q) +{ + struct ccp_desc *desc; + union ccp_function function; + + desc = &cmd_q->qbase_desc[cmd_q->qidx]; + + CCP_CMD_ENGINE(desc) = CCP_ENGINE_PASSTHRU; + + CCP_CMD_SOC(desc) = 0; + CCP_CMD_IOC(desc) = 0; + CCP_CMD_INIT(desc) = 0; + CCP_CMD_EOM(desc) = 0; + CCP_CMD_PROT(desc) = 0; + + function.raw = 0; + CCP_PT_BYTESWAP(&function) = pst->byte_swap; + CCP_PT_BITWISE(&function) = pst->bit_mod; + CCP_CMD_FUNCTION(desc) = function.raw; + + CCP_CMD_LEN(desc) = pst->len; + + if (pst->dir) { + CCP_CMD_SRC_LO(desc) = (uint32_t)(pst->src_addr); + CCP_CMD_SRC_HI(desc) = high32_value(pst->src_addr); + CCP_CMD_SRC_MEM(desc) = CCP_MEMTYPE_SYSTEM; + + CCP_CMD_DST_LO(desc) = (uint32_t)(pst->dest_addr); + CCP_CMD_DST_HI(desc) = 0; + CCP_CMD_DST_MEM(desc) = CCP_MEMTYPE_SB; + + if (pst->bit_mod != CCP_PASSTHRU_BITWISE_NOOP) + CCP_CMD_LSB_ID(desc) = cmd_q->sb_key; + } else { + + CCP_CMD_SRC_LO(desc) = (uint32_t)(pst->src_addr); + CCP_CMD_SRC_HI(desc) = 0; + CCP_CMD_SRC_MEM(desc) = CCP_MEMTYPE_SB; + + CCP_CMD_DST_LO(desc) = (uint32_t)(pst->dest_addr); + CCP_CMD_DST_HI(desc) = high32_value(pst->dest_addr); + CCP_CMD_DST_MEM(desc) = CCP_MEMTYPE_SYSTEM; + } + + cmd_q->qidx = (cmd_q->qidx + 1) % COMMANDS_PER_
[dpdk-dev] [PATCH v5 11/19] crypto/ccp: support 3des cipher algo
Signed-off-by: Ravi Kumar --- drivers/crypto/ccp/ccp_crypto.c | 132 ++- drivers/crypto/ccp/ccp_crypto.h | 3 + drivers/crypto/ccp/ccp_pmd_ops.c | 20 ++ 3 files changed, 154 insertions(+), 1 deletion(-) diff --git a/drivers/crypto/ccp/ccp_crypto.c b/drivers/crypto/ccp/ccp_crypto.c index 3c46ac3..a946f42 100644 --- a/drivers/crypto/ccp/ccp_crypto.c +++ b/drivers/crypto/ccp/ccp_crypto.c @@ -54,7 +54,7 @@ ccp_configure_session_cipher(struct ccp_session *sess, const struct rte_crypto_sym_xform *xform) { const struct rte_crypto_cipher_xform *cipher_xform = NULL; - size_t i; + size_t i, j, x; cipher_xform = &xform->cipher; @@ -89,6 +89,11 @@ ccp_configure_session_cipher(struct ccp_session *sess, sess->cipher.um.aes_mode = CCP_AES_MODE_CBC; sess->cipher.engine = CCP_ENGINE_AES; break; + case RTE_CRYPTO_CIPHER_3DES_CBC: + sess->cipher.algo = CCP_CIPHER_ALGO_3DES_CBC; + sess->cipher.um.des_mode = CCP_DES_MODE_CBC; + sess->cipher.engine = CCP_ENGINE_3DES; + break; default: CCP_LOG_ERR("Unsupported cipher algo"); return -1; @@ -111,6 +116,20 @@ ccp_configure_session_cipher(struct ccp_session *sess, sess->cipher.key_ccp[sess->cipher.key_length - i - 1] = sess->cipher.key[i]; break; + case CCP_ENGINE_3DES: + if (sess->cipher.key_length == 16) + sess->cipher.ut.des_type = CCP_DES_TYPE_128; + else if (sess->cipher.key_length == 24) + sess->cipher.ut.des_type = CCP_DES_TYPE_192; + else { + CCP_LOG_ERR("Invalid cipher key length"); + return -1; + } + for (j = 0, x = 0; j < sess->cipher.key_length/8; j++, x += 8) + for (i = 0; i < 8; i++) + sess->cipher.key_ccp[(8 + x) - i - 1] = + sess->cipher.key[i + x]; + break; default: CCP_LOG_ERR("Invalid CCP Engine"); return -ENOTSUP; @@ -254,6 +273,10 @@ ccp_cipher_slot(struct ccp_session *session) count = 2; /**< op + passthrough for iv */ break; + case CCP_CIPHER_ALGO_3DES_CBC: + count = 2; + /**< op + passthrough for iv */ + break; default: CCP_LOG_ERR("Unsupported cipher algo %d", session->cipher.algo); @@ -452,6 +475,109 @@ ccp_perform_aes(struct rte_crypto_op *op, return 0; } +static int +ccp_perform_3des(struct rte_crypto_op *op, + struct ccp_queue *cmd_q, + struct ccp_batch_info *b_info) +{ + struct ccp_session *session; + union ccp_function function; + unsigned char *lsb_buf; + struct ccp_passthru pst; + struct ccp_desc *desc; + uint32_t tail; + uint8_t *iv; + phys_addr_t src_addr, dest_addr, key_addr; + + session = (struct ccp_session *)get_session_private_data( +op->sym->session, + ccp_cryptodev_driver_id); + + iv = rte_crypto_op_ctod_offset(op, uint8_t *, session->iv.offset); + switch (session->cipher.um.des_mode) { + case CCP_DES_MODE_CBC: + lsb_buf = &(b_info->lsb_buf[b_info->lsb_buf_idx*CCP_SB_BYTES]); + b_info->lsb_buf_idx++; + + rte_memcpy(lsb_buf + (CCP_SB_BYTES - session->iv.length), + iv, session->iv.length); + + pst.src_addr = (phys_addr_t)rte_mem_virt2phy((void *) lsb_buf); + pst.dest_addr = (phys_addr_t)(cmd_q->sb_iv * CCP_SB_BYTES); + pst.len = CCP_SB_BYTES; + pst.dir = 1; + pst.bit_mod = CCP_PASSTHRU_BITWISE_NOOP; + pst.byte_swap = CCP_PASSTHRU_BYTESWAP_256BIT; + ccp_perform_passthru(&pst, cmd_q); + break; + case CCP_DES_MODE_CFB: + case CCP_DES_MODE_ECB: + CCP_LOG_ERR("Unsupported DES cipher mode"); + return -ENOTSUP; + } + + src_addr = rte_pktmbuf_mtophys_offset(op->sym->m_src, + op->sym->cipher.data.offset); + if (unlikely(op->sym->m_dst != NULL)) + dest_addr = + rte_pktmbuf_mtophys_offset(op->sym->m_dst, + op->sym->cipher.data.offset); + else + dest_addr = src_addr; + + key_addr = rte_mem_virt2phy(session->cipher.key_ccp); + + desc = &cmd_q->qbase_desc[cmd_q->qidx]; + + memset(desc, 0, Q_DESC_SIZE);
[dpdk-dev] [PATCH v5 12/19] crypto/ccp: support aes-cmac auth algo
Signed-off-by: Ravi Kumar --- drivers/crypto/ccp/ccp_crypto.c | 277 ++- drivers/crypto/ccp/ccp_pmd_ops.c | 20 +++ 2 files changed, 295 insertions(+), 2 deletions(-) diff --git a/drivers/crypto/ccp/ccp_crypto.c b/drivers/crypto/ccp/ccp_crypto.c index a946f42..6d8a0d6 100644 --- a/drivers/crypto/ccp/ccp_crypto.c +++ b/drivers/crypto/ccp/ccp_crypto.c @@ -10,6 +10,8 @@ #include #include #include +#include /*sub key apis*/ +#include /*sub key apis*/ #include #include @@ -48,6 +50,84 @@ ccp_get_cmd_id(const struct rte_crypto_sym_xform *xform) return res; } +/* prepare temporary keys K1 and K2 */ +static void prepare_key(unsigned char *k, unsigned char *l, int bl) +{ + int i; + /* Shift block to left, including carry */ + for (i = 0; i < bl; i++) { + k[i] = l[i] << 1; + if (i < bl - 1 && l[i + 1] & 0x80) + k[i] |= 1; + } + /* If MSB set fixup with R */ + if (l[0] & 0x80) + k[bl - 1] ^= bl == 16 ? 0x87 : 0x1b; +} + +/* subkeys K1 and K2 generation for CMAC */ +static int +generate_cmac_subkeys(struct ccp_session *sess) +{ + const EVP_CIPHER *algo; + EVP_CIPHER_CTX *ctx; + unsigned char *ccp_ctx; + size_t i; + int dstlen, totlen; + unsigned char zero_iv[AES_BLOCK_SIZE] = {0}; + unsigned char dst[2 * AES_BLOCK_SIZE] = {0}; + unsigned char k1[AES_BLOCK_SIZE] = {0}; + unsigned char k2[AES_BLOCK_SIZE] = {0}; + + if (sess->auth.ut.aes_type == CCP_AES_TYPE_128) + algo = EVP_aes_128_cbc(); + else if (sess->auth.ut.aes_type == CCP_AES_TYPE_192) + algo = EVP_aes_192_cbc(); + else if (sess->auth.ut.aes_type == CCP_AES_TYPE_256) + algo = EVP_aes_256_cbc(); + else { + CCP_LOG_ERR("Invalid CMAC type length"); + return -1; + } + + ctx = EVP_CIPHER_CTX_new(); + if (!ctx) { + CCP_LOG_ERR("ctx creation failed"); + return -1; + } + if (EVP_EncryptInit(ctx, algo, (unsigned char *)sess->auth.key, + (unsigned char *)zero_iv) <= 0) + goto key_generate_err; + if (EVP_CIPHER_CTX_set_padding(ctx, 0) <= 0) + goto key_generate_err; + if (EVP_EncryptUpdate(ctx, dst, &dstlen, zero_iv, + AES_BLOCK_SIZE) <= 0) + goto key_generate_err; + if (EVP_EncryptFinal_ex(ctx, dst + dstlen, &totlen) <= 0) + goto key_generate_err; + + memset(sess->auth.pre_compute, 0, CCP_SB_BYTES * 2); + + ccp_ctx = (unsigned char *)(sess->auth.pre_compute + CCP_SB_BYTES - 1); + prepare_key(k1, dst, AES_BLOCK_SIZE); + for (i = 0; i < AES_BLOCK_SIZE; i++, ccp_ctx--) + *ccp_ctx = k1[i]; + + ccp_ctx = (unsigned char *)(sess->auth.pre_compute + + (2 * CCP_SB_BYTES) - 1); + prepare_key(k2, k1, AES_BLOCK_SIZE); + for (i = 0; i < AES_BLOCK_SIZE; i++, ccp_ctx--) + *ccp_ctx = k2[i]; + + EVP_CIPHER_CTX_free(ctx); + + return 0; + +key_generate_err: + CCP_LOG_ERR("CMAC Init failed"); + return -1; +} + /* configure session */ static int ccp_configure_session_cipher(struct ccp_session *sess, @@ -144,6 +224,7 @@ ccp_configure_session_auth(struct ccp_session *sess, const struct rte_crypto_sym_xform *xform) { const struct rte_crypto_auth_xform *auth_xform = NULL; + size_t i; auth_xform = &xform->auth; @@ -153,6 +234,33 @@ ccp_configure_session_auth(struct ccp_session *sess, else sess->auth.op = CCP_AUTH_OP_VERIFY; switch (auth_xform->algo) { + case RTE_CRYPTO_AUTH_AES_CMAC: + sess->auth.algo = CCP_AUTH_ALGO_AES_CMAC; + sess->auth.engine = CCP_ENGINE_AES; + sess->auth.um.aes_mode = CCP_AES_MODE_CMAC; + sess->auth.key_length = auth_xform->key.length; + /**auth.ctx_len = CCP_SB_BYTES << 1; + sess->auth.offset = AES_BLOCK_SIZE; + sess->auth.block_size = AES_BLOCK_SIZE; + if (sess->auth.key_length == 16) + sess->auth.ut.aes_type = CCP_AES_TYPE_128; + else if (sess->auth.key_length == 24) + sess->auth.ut.aes_type = CCP_AES_TYPE_192; + else if (sess->auth.key_length == 32) + sess->auth.ut.aes_type = CCP_AES_TYPE_256; + else { + CCP_LOG_ERR("Invalid CMAC key length"); + return -1; + } + rte_memcpy(sess->auth.key, auth_xform->key.data, + sess->auth.key_length); + for (i = 0; i < sess->auth.key_length; i++) + sess->auth.key_ccp[ses
[dpdk-dev] [PATCH v5 13/19] crypto/ccp: support aes-gcm aead algo
Signed-off-by: Ravi Kumar --- drivers/crypto/ccp/ccp_crypto.c | 235 ++- drivers/crypto/ccp/ccp_pmd_ops.c | 30 + 2 files changed, 261 insertions(+), 4 deletions(-) diff --git a/drivers/crypto/ccp/ccp_crypto.c b/drivers/crypto/ccp/ccp_crypto.c index 6d8a0d6..3a14b77 100644 --- a/drivers/crypto/ccp/ccp_crypto.c +++ b/drivers/crypto/ccp/ccp_crypto.c @@ -273,6 +273,7 @@ ccp_configure_session_aead(struct ccp_session *sess, const struct rte_crypto_sym_xform *xform) { const struct rte_crypto_aead_xform *aead_xform = NULL; + size_t i; aead_xform = &xform->aead; @@ -287,6 +288,7 @@ ccp_configure_session_aead(struct ccp_session *sess, sess->cipher.dir = CCP_CIPHER_DIR_DECRYPT; sess->auth.op = CCP_AUTH_OP_VERIFY; } + sess->aead_algo = aead_xform->algo; sess->auth.aad_length = aead_xform->aad_length; sess->auth.digest_length = aead_xform->digest_length; @@ -295,10 +297,37 @@ ccp_configure_session_aead(struct ccp_session *sess, sess->iv.length = aead_xform->iv.length; switch (aead_xform->algo) { + case RTE_CRYPTO_AEAD_AES_GCM: + sess->cipher.algo = CCP_CIPHER_ALGO_AES_GCM; + sess->cipher.um.aes_mode = CCP_AES_MODE_GCTR; + sess->cipher.engine = CCP_ENGINE_AES; + if (sess->cipher.key_length == 16) + sess->cipher.ut.aes_type = CCP_AES_TYPE_128; + else if (sess->cipher.key_length == 24) + sess->cipher.ut.aes_type = CCP_AES_TYPE_192; + else if (sess->cipher.key_length == 32) + sess->cipher.ut.aes_type = CCP_AES_TYPE_256; + else { + CCP_LOG_ERR("Invalid aead key length"); + return -1; + } + for (i = 0; i < sess->cipher.key_length; i++) + sess->cipher.key_ccp[sess->cipher.key_length - i - 1] = + sess->cipher.key[i]; + sess->auth.algo = CCP_AUTH_ALGO_AES_GCM; + sess->auth.engine = CCP_ENGINE_AES; + sess->auth.um.aes_mode = CCP_AES_MODE_GHASH; + sess->auth.ctx_len = CCP_SB_BYTES; + sess->auth.offset = 0; + sess->auth.block_size = AES_BLOCK_SIZE; + sess->cmd_id = CCP_CMD_COMBINED; + break; default: CCP_LOG_ERR("Unsupported aead algo"); return -ENOTSUP; } + sess->cipher.nonce_phys = rte_mem_virt2phy(sess->cipher.nonce); + sess->cipher.key_phys = rte_mem_virt2phy(sess->cipher.key_ccp); return 0; } @@ -421,10 +450,27 @@ ccp_aead_slot(struct ccp_session *session) int count = 0; switch (session->aead_algo) { + case RTE_CRYPTO_AEAD_AES_GCM: + break; default: CCP_LOG_ERR("Unsupported aead algo %d", session->aead_algo); } + switch (session->auth.algo) { + case CCP_AUTH_ALGO_AES_GCM: + count = 5; + /** +* 1. Passthru iv +* 2. Hash AAD +* 3. GCTR +* 4. Reload passthru +* 5. Hash Final tag +*/ + break; + default: + CCP_LOG_ERR("Unsupported combined auth ALGO %d", + session->auth.algo); + } return count; } @@ -847,6 +893,179 @@ ccp_perform_3des(struct rte_crypto_op *op, return 0; } +static int +ccp_perform_aes_gcm(struct rte_crypto_op *op, struct ccp_queue *cmd_q) +{ + struct ccp_session *session; + union ccp_function function; + uint8_t *iv; + struct ccp_passthru pst; + struct ccp_desc *desc; + uint32_t tail; + uint64_t *temp; + phys_addr_t src_addr, dest_addr, key_addr, aad_addr; + phys_addr_t digest_dest_addr; + int length, non_align_len; + + session = (struct ccp_session *)get_session_private_data( +op->sym->session, +ccp_cryptodev_driver_id); + iv = rte_crypto_op_ctod_offset(op, uint8_t *, session->iv.offset); + key_addr = session->cipher.key_phys; + + src_addr = rte_pktmbuf_mtophys_offset(op->sym->m_src, + op->sym->aead.data.offset); + if (unlikely(op->sym->m_dst != NULL)) + dest_addr = rte_pktmbuf_mtophys_offset(op->sym->m_dst, + op->sym->aead.data.offset); + else + dest_addr = src_addr; + rte_pktmbuf_append(op->sym->m_src, session->auth.ctx_len); + digest_dest_addr = op->sym->aead.digest.phys_addr; + temp = (uint64_t *)(op->sym->aead.digest.data + AES_BLOCK_SIZE);
[dpdk-dev] [PATCH v5 14/19] crypto/ccp: support sha1 authentication algo
Signed-off-by: Ravi Kumar --- drivers/crypto/ccp/ccp_crypto.c | 367 +++ drivers/crypto/ccp/ccp_crypto.h | 23 +++ drivers/crypto/ccp/ccp_pmd_ops.c | 42 + 3 files changed, 432 insertions(+) diff --git a/drivers/crypto/ccp/ccp_crypto.c b/drivers/crypto/ccp/ccp_crypto.c index 3a14b77..517c284 100644 --- a/drivers/crypto/ccp/ccp_crypto.c +++ b/drivers/crypto/ccp/ccp_crypto.c @@ -10,6 +10,7 @@ #include #include #include +#include #include /*sub key apis*/ #include /*sub key apis*/ @@ -26,6 +27,14 @@ #include "ccp_pci.h" #include "ccp_pmd_private.h" +/* SHA initial context values */ +static uint32_t ccp_sha1_init[SHA_COMMON_DIGEST_SIZE / sizeof(uint32_t)] = { + SHA1_H4, SHA1_H3, + SHA1_H2, SHA1_H1, + SHA1_H0, 0x0U, + 0x0U, 0x0U, +}; + static enum ccp_cmd_order ccp_get_cmd_id(const struct rte_crypto_sym_xform *xform) { @@ -50,6 +59,59 @@ ccp_get_cmd_id(const struct rte_crypto_sym_xform *xform) return res; } +/* partial hash using openssl */ +static int partial_hash_sha1(uint8_t *data_in, uint8_t *data_out) +{ + SHA_CTX ctx; + + if (!SHA1_Init(&ctx)) + return -EFAULT; + SHA1_Transform(&ctx, data_in); + rte_memcpy(data_out, &ctx, SHA_DIGEST_LENGTH); + return 0; +} + +static int generate_partial_hash(struct ccp_session *sess) +{ + + uint8_t ipad[sess->auth.block_size]; + uint8_t opad[sess->auth.block_size]; + uint8_t *ipad_t, *opad_t; + uint32_t *hash_value_be32, hash_temp32[8]; + int i, count; + + opad_t = ipad_t = (uint8_t *)sess->auth.key; + + hash_value_be32 = (uint32_t *)((uint8_t *)sess->auth.pre_compute); + + /* considering key size is always equal to block size of algorithm */ + for (i = 0; i < sess->auth.block_size; i++) { + ipad[i] = (ipad_t[i] ^ HMAC_IPAD_VALUE); + opad[i] = (opad_t[i] ^ HMAC_OPAD_VALUE); + } + + switch (sess->auth.algo) { + case CCP_AUTH_ALGO_SHA1_HMAC: + count = SHA1_DIGEST_SIZE >> 2; + + if (partial_hash_sha1(ipad, (uint8_t *)hash_temp32)) + return -1; + for (i = 0; i < count; i++, hash_value_be32++) + *hash_value_be32 = hash_temp32[count - 1 - i]; + + hash_value_be32 = (uint32_t *)((uint8_t *)sess->auth.pre_compute + + sess->auth.ctx_len); + if (partial_hash_sha1(opad, (uint8_t *)hash_temp32)) + return -1; + for (i = 0; i < count; i++, hash_value_be32++) + *hash_value_be32 = hash_temp32[count - 1 - i]; + return 0; + default: + CCP_LOG_ERR("Invalid auth algo"); + return -1; + } +} + /* prepare temporary keys K1 and K2 */ static void prepare_key(unsigned char *k, unsigned char *l, int bl) { @@ -234,6 +296,31 @@ ccp_configure_session_auth(struct ccp_session *sess, else sess->auth.op = CCP_AUTH_OP_VERIFY; switch (auth_xform->algo) { + case RTE_CRYPTO_AUTH_SHA1: + sess->auth.engine = CCP_ENGINE_SHA; + sess->auth.algo = CCP_AUTH_ALGO_SHA1; + sess->auth.ut.sha_type = CCP_SHA_TYPE_1; + sess->auth.ctx = (void *)ccp_sha1_init; + sess->auth.ctx_len = CCP_SB_BYTES; + sess->auth.offset = CCP_SB_BYTES - SHA1_DIGEST_SIZE; + break; + case RTE_CRYPTO_AUTH_SHA1_HMAC: + if (auth_xform->key.length > SHA1_BLOCK_SIZE) + return -1; + sess->auth.engine = CCP_ENGINE_SHA; + sess->auth.algo = CCP_AUTH_ALGO_SHA1_HMAC; + sess->auth.ut.sha_type = CCP_SHA_TYPE_1; + sess->auth.ctx_len = CCP_SB_BYTES; + sess->auth.offset = CCP_SB_BYTES - SHA1_DIGEST_SIZE; + sess->auth.block_size = SHA1_BLOCK_SIZE; + sess->auth.key_length = auth_xform->key.length; + memset(sess->auth.key, 0, sess->auth.block_size); + memset(sess->auth.pre_compute, 0, sess->auth.ctx_len << 1); + rte_memcpy(sess->auth.key, auth_xform->key.data, + auth_xform->key.length); + if (generate_partial_hash(sess)) + return -1; + break; case RTE_CRYPTO_AUTH_AES_CMAC: sess->auth.algo = CCP_AUTH_ALGO_AES_CMAC; sess->auth.engine = CCP_ENGINE_AES; @@ -427,6 +514,13 @@ ccp_auth_slot(struct ccp_session *session) int count = 0; switch (session->auth.algo) { + case CCP_AUTH_ALGO_SHA1: + count = 3; + /**< op + lsb passthrough cpy to/from*/ + break; + case CCP_AUTH_ALGO_SHA1_HMAC: + count = 6; + break; case CCP_AUTH_ALGO_AES_CMAC:
[dpdk-dev] [PATCH v5 16/19] crypto/ccp: support sha3 family authentication algo
Signed-off-by: Ravi Kumar --- drivers/crypto/ccp/ccp_crypto.c | 667 +- drivers/crypto/ccp/ccp_crypto.h | 22 ++ drivers/crypto/ccp/ccp_pmd_ops.c | 168 + lib/librte_cryptodev/rte_crypto_sym.h | 17 + 4 files changed, 873 insertions(+), 1 deletion(-) diff --git a/drivers/crypto/ccp/ccp_crypto.c b/drivers/crypto/ccp/ccp_crypto.c index 8b26ad3..53e731b 100644 --- a/drivers/crypto/ccp/ccp_crypto.c +++ b/drivers/crypto/ccp/ccp_crypto.c @@ -63,6 +63,74 @@ uint64_t ccp_sha512_init[SHA512_DIGEST_SIZE / sizeof(uint64_t)] = { SHA512_H1, SHA512_H0, }; +#if defined(_MSC_VER) +#define SHA3_CONST(x) x +#else +#define SHA3_CONST(x) x##L +#endif + +/** 'Words' here refers to uint64_t */ +#define SHA3_KECCAK_SPONGE_WORDS \ + (((1600) / 8) / sizeof(uint64_t)) +typedef struct sha3_context_ { + uint64_t saved; + /** +* The portion of the input message that we +* didn't consume yet +*/ + union { + uint64_t s[SHA3_KECCAK_SPONGE_WORDS]; + /* Keccak's state */ + uint8_t sb[SHA3_KECCAK_SPONGE_WORDS * 8]; + /**total 200 ctx size**/ + }; + unsigned int byteIndex; + /** +* 0..7--the next byte after the set one +* (starts from 0; 0--none are buffered) +*/ + unsigned int wordIndex; + /** +* 0..24--the next word to integrate input +* (starts from 0) +*/ + unsigned int capacityWords; + /** +* the double size of the hash output in +* words (e.g. 16 for Keccak 512) +*/ +} sha3_context; + +#ifndef SHA3_ROTL64 +#define SHA3_ROTL64(x, y) \ + (((x) << (y)) | ((x) >> ((sizeof(uint64_t)*8) - (y +#endif + +static const uint64_t keccakf_rndc[24] = { + SHA3_CONST(0x0001UL), SHA3_CONST(0x8082UL), + SHA3_CONST(0x8000808aUL), SHA3_CONST(0x800080008000UL), + SHA3_CONST(0x808bUL), SHA3_CONST(0x8001UL), + SHA3_CONST(0x800080008081UL), SHA3_CONST(0x80008009UL), + SHA3_CONST(0x008aUL), SHA3_CONST(0x0088UL), + SHA3_CONST(0x80008009UL), SHA3_CONST(0x800aUL), + SHA3_CONST(0x8000808bUL), SHA3_CONST(0x808bUL), + SHA3_CONST(0x80008089UL), SHA3_CONST(0x80008003UL), + SHA3_CONST(0x80008002UL), SHA3_CONST(0x8080UL), + SHA3_CONST(0x800aUL), SHA3_CONST(0x8000800aUL), + SHA3_CONST(0x800080008081UL), SHA3_CONST(0x80008080UL), + SHA3_CONST(0x8001UL), SHA3_CONST(0x800080008008UL) +}; + +static const unsigned int keccakf_rotc[24] = { + 1, 3, 6, 10, 15, 21, 28, 36, 45, 55, 2, 14, 27, 41, 56, 8, 25, 43, 62, + 18, 39, 61, 20, 44 +}; + +static const unsigned int keccakf_piln[24] = { + 10, 7, 11, 17, 18, 3, 5, 16, 8, 21, 24, 4, 15, 23, 19, 13, 12, 2, 20, + 14, 22, 9, 6, 1 +}; + static enum ccp_cmd_order ccp_get_cmd_id(const struct rte_crypto_sym_xform *xform) { @@ -147,6 +215,223 @@ static int partial_hash_sha512(uint8_t *data_in, uint8_t *data_out) return 0; } +static void +keccakf(uint64_t s[25]) +{ + int i, j, round; + uint64_t t, bc[5]; +#define KECCAK_ROUNDS 24 + + for (round = 0; round < KECCAK_ROUNDS; round++) { + + /* Theta */ + for (i = 0; i < 5; i++) + bc[i] = s[i] ^ s[i + 5] ^ s[i + 10] ^ s[i + 15] ^ + s[i + 20]; + + for (i = 0; i < 5; i++) { + t = bc[(i + 4) % 5] ^ SHA3_ROTL64(bc[(i + 1) % 5], 1); + for (j = 0; j < 25; j += 5) + s[j + i] ^= t; + } + + /* Rho Pi */ + t = s[1]; + for (i = 0; i < 24; i++) { + j = keccakf_piln[i]; + bc[0] = s[j]; + s[j] = SHA3_ROTL64(t, keccakf_rotc[i]); + t = bc[0]; + } + + /* Chi */ + for (j = 0; j < 25; j += 5) { + for (i = 0; i < 5; i++) + bc[i] = s[j + i]; + for (i = 0; i < 5; i++) + s[j + i] ^= (~bc[(i + 1) % 5]) & + bc[(i + 2) % 5]; + } + + /* Iota */ + s[0] ^= keccakf_rndc[round]; + } +} + +static void +sha3_Init224(void *priv) +{ + sha3_context *ctx = (sha3_context *) priv; + + memset(ctx, 0, sizeof(*ctx)); + ctx->capacityWords = 2 * 224 / (8 * sizeof(uint64_t)); +} + +static void +sha3_Init256(void *priv) +{ + sha3_context *ctx = (sha3_context *) priv; + + memset(ctx, 0, sizeof(*ctx)); + ctx->capacityWords = 2 * 2
[dpdk-dev] [PATCH v5 15/19] crypto/ccp: support sha2 family authentication algo
Signed-off-by: Ravi Kumar --- drivers/crypto/ccp/ccp_crypto.c | 270 +++ drivers/crypto/ccp/ccp_crypto.h | 48 +++ drivers/crypto/ccp/ccp_pmd_ops.c | 168 3 files changed, 486 insertions(+) diff --git a/drivers/crypto/ccp/ccp_crypto.c b/drivers/crypto/ccp/ccp_crypto.c index 517c284..8b26ad3 100644 --- a/drivers/crypto/ccp/ccp_crypto.c +++ b/drivers/crypto/ccp/ccp_crypto.c @@ -35,6 +35,34 @@ static uint32_t ccp_sha1_init[SHA_COMMON_DIGEST_SIZE / sizeof(uint32_t)] = { 0x0U, 0x0U, }; +uint32_t ccp_sha224_init[SHA256_DIGEST_SIZE / sizeof(uint32_t)] = { + SHA224_H7, SHA224_H6, + SHA224_H5, SHA224_H4, + SHA224_H3, SHA224_H2, + SHA224_H1, SHA224_H0, +}; + +uint32_t ccp_sha256_init[SHA256_DIGEST_SIZE / sizeof(uint32_t)] = { + SHA256_H7, SHA256_H6, + SHA256_H5, SHA256_H4, + SHA256_H3, SHA256_H2, + SHA256_H1, SHA256_H0, +}; + +uint64_t ccp_sha384_init[SHA512_DIGEST_SIZE / sizeof(uint64_t)] = { + SHA384_H7, SHA384_H6, + SHA384_H5, SHA384_H4, + SHA384_H3, SHA384_H2, + SHA384_H1, SHA384_H0, +}; + +uint64_t ccp_sha512_init[SHA512_DIGEST_SIZE / sizeof(uint64_t)] = { + SHA512_H7, SHA512_H6, + SHA512_H5, SHA512_H4, + SHA512_H3, SHA512_H2, + SHA512_H1, SHA512_H0, +}; + static enum ccp_cmd_order ccp_get_cmd_id(const struct rte_crypto_sym_xform *xform) { @@ -71,6 +99,54 @@ static int partial_hash_sha1(uint8_t *data_in, uint8_t *data_out) return 0; } +static int partial_hash_sha224(uint8_t *data_in, uint8_t *data_out) +{ + SHA256_CTX ctx; + + if (!SHA224_Init(&ctx)) + return -EFAULT; + SHA256_Transform(&ctx, data_in); + rte_memcpy(data_out, &ctx, + SHA256_DIGEST_LENGTH); + return 0; +} + +static int partial_hash_sha256(uint8_t *data_in, uint8_t *data_out) +{ + SHA256_CTX ctx; + + if (!SHA256_Init(&ctx)) + return -EFAULT; + SHA256_Transform(&ctx, data_in); + rte_memcpy(data_out, &ctx, + SHA256_DIGEST_LENGTH); + return 0; +} + +static int partial_hash_sha384(uint8_t *data_in, uint8_t *data_out) +{ + SHA512_CTX ctx; + + if (!SHA384_Init(&ctx)) + return -EFAULT; + SHA512_Transform(&ctx, data_in); + rte_memcpy(data_out, &ctx, + SHA512_DIGEST_LENGTH); + return 0; +} + +static int partial_hash_sha512(uint8_t *data_in, uint8_t *data_out) +{ + SHA512_CTX ctx; + + if (!SHA512_Init(&ctx)) + return -EFAULT; + SHA512_Transform(&ctx, data_in); + rte_memcpy(data_out, &ctx, + SHA512_DIGEST_LENGTH); + return 0; +} + static int generate_partial_hash(struct ccp_session *sess) { @@ -78,11 +154,13 @@ static int generate_partial_hash(struct ccp_session *sess) uint8_t opad[sess->auth.block_size]; uint8_t *ipad_t, *opad_t; uint32_t *hash_value_be32, hash_temp32[8]; + uint64_t *hash_value_be64, hash_temp64[8]; int i, count; opad_t = ipad_t = (uint8_t *)sess->auth.key; hash_value_be32 = (uint32_t *)((uint8_t *)sess->auth.pre_compute); + hash_value_be64 = (uint64_t *)((uint8_t *)sess->auth.pre_compute); /* considering key size is always equal to block size of algorithm */ for (i = 0; i < sess->auth.block_size; i++) { @@ -106,6 +184,66 @@ static int generate_partial_hash(struct ccp_session *sess) for (i = 0; i < count; i++, hash_value_be32++) *hash_value_be32 = hash_temp32[count - 1 - i]; return 0; + case CCP_AUTH_ALGO_SHA224_HMAC: + count = SHA256_DIGEST_SIZE >> 2; + + if (partial_hash_sha224(ipad, (uint8_t *)hash_temp32)) + return -1; + for (i = 0; i < count; i++, hash_value_be32++) + *hash_value_be32 = hash_temp32[count - 1 - i]; + + hash_value_be32 = (uint32_t *)((uint8_t *)sess->auth.pre_compute + + sess->auth.ctx_len); + if (partial_hash_sha224(opad, (uint8_t *)hash_temp32)) + return -1; + for (i = 0; i < count; i++, hash_value_be32++) + *hash_value_be32 = hash_temp32[count - 1 - i]; + return 0; + case CCP_AUTH_ALGO_SHA256_HMAC: + count = SHA256_DIGEST_SIZE >> 2; + + if (partial_hash_sha256(ipad, (uint8_t *)hash_temp32)) + return -1; + for (i = 0; i < count; i++, hash_value_be32++) + *hash_value_be32 = hash_temp32[count - 1 - i]; + + hash_value_be32 = (uint32_t *)((uint8_t *)sess->auth.pre_compute + + sess->auth.ctx_len); + if (partial_hash_sha256(opad, (uint8_t *)hash_temp32)) +
[dpdk-dev] [PATCH v5 17/19] crypto/ccp: support cpu based md5 and sha2 family authentication algo
Signed-off-by: Ravi Kumar --- config/common_base | 1 + drivers/crypto/ccp/ccp_crypto.c | 282 ++- drivers/crypto/ccp/ccp_crypto.h | 5 +- drivers/crypto/ccp/ccp_pmd_ops.c | 23 +++ drivers/crypto/ccp/ccp_pmd_private.h | 10 ++ 5 files changed, 316 insertions(+), 5 deletions(-) diff --git a/config/common_base b/config/common_base index 28237f0..65e34ae 100644 --- a/config/common_base +++ b/config/common_base @@ -532,6 +532,7 @@ CONFIG_RTE_LIBRTE_PMD_NULL_CRYPTO=y # Compile PMD for AMD CCP crypto device # CONFIG_RTE_LIBRTE_PMD_CCP=n +CONFIG_RTE_LIBRTE_PMD_CCP_CPU_AUTH=n # # Compile PMD for Marvell Crypto device diff --git a/drivers/crypto/ccp/ccp_crypto.c b/drivers/crypto/ccp/ccp_crypto.c index 53e731b..a0809e4 100644 --- a/drivers/crypto/ccp/ccp_crypto.c +++ b/drivers/crypto/ccp/ccp_crypto.c @@ -27,6 +27,12 @@ #include "ccp_pci.h" #include "ccp_pmd_private.h" +#ifdef RTE_LIBRTE_PMD_CCP_CPU_AUTH +#include +#include +#include +#endif + /* SHA initial context values */ static uint32_t ccp_sha1_init[SHA_COMMON_DIGEST_SIZE / sizeof(uint32_t)] = { SHA1_H4, SHA1_H3, @@ -760,6 +766,17 @@ ccp_configure_session_auth(struct ccp_session *sess, else sess->auth.op = CCP_AUTH_OP_VERIFY; switch (auth_xform->algo) { +#ifdef RTE_LIBRTE_PMD_CCP_CPU_AUTH + case RTE_CRYPTO_AUTH_MD5_HMAC: + sess->auth.algo = CCP_AUTH_ALGO_MD5_HMAC; + sess->auth.offset = (CCP_SB_BYTES << 1) - MD5_DIGEST_SIZE; + sess->auth.key_length = auth_xform->key.length; + sess->auth.block_size = MD5_BLOCK_SIZE; + memset(sess->auth.key, 0, sess->auth.block_size); + rte_memcpy(sess->auth.key, auth_xform->key.data, + auth_xform->key.length); + break; +#endif case RTE_CRYPTO_AUTH_SHA1: sess->auth.engine = CCP_ENGINE_SHA; sess->auth.algo = CCP_AUTH_ALGO_SHA1; @@ -769,6 +786,17 @@ ccp_configure_session_auth(struct ccp_session *sess, sess->auth.offset = CCP_SB_BYTES - SHA1_DIGEST_SIZE; break; case RTE_CRYPTO_AUTH_SHA1_HMAC: +#ifdef RTE_LIBRTE_PMD_CCP_CPU_AUTH + if (auth_xform->key.length > SHA1_BLOCK_SIZE) + return -1; + sess->auth.algo = CCP_AUTH_ALGO_SHA1_HMAC; + sess->auth.offset = CCP_SB_BYTES - SHA1_DIGEST_SIZE; + sess->auth.block_size = SHA1_BLOCK_SIZE; + sess->auth.key_length = auth_xform->key.length; + memset(sess->auth.key, 0, sess->auth.block_size); + rte_memcpy(sess->auth.key, auth_xform->key.data, + auth_xform->key.length); +#else if (auth_xform->key.length > SHA1_BLOCK_SIZE) return -1; sess->auth.engine = CCP_ENGINE_SHA; @@ -784,6 +812,7 @@ ccp_configure_session_auth(struct ccp_session *sess, auth_xform->key.length); if (generate_partial_hash(sess)) return -1; +#endif break; case RTE_CRYPTO_AUTH_SHA224: sess->auth.algo = CCP_AUTH_ALGO_SHA224; @@ -794,6 +823,17 @@ ccp_configure_session_auth(struct ccp_session *sess, sess->auth.offset = CCP_SB_BYTES - SHA224_DIGEST_SIZE; break; case RTE_CRYPTO_AUTH_SHA224_HMAC: +#ifdef RTE_LIBRTE_PMD_CCP_CPU_AUTH + if (auth_xform->key.length > SHA224_BLOCK_SIZE) + return -1; + sess->auth.algo = CCP_AUTH_ALGO_SHA224_HMAC; + sess->auth.offset = CCP_SB_BYTES - SHA224_DIGEST_SIZE; + sess->auth.block_size = SHA224_BLOCK_SIZE; + sess->auth.key_length = auth_xform->key.length; + memset(sess->auth.key, 0, sess->auth.block_size); + rte_memcpy(sess->auth.key, auth_xform->key.data, + auth_xform->key.length); +#else if (auth_xform->key.length > SHA224_BLOCK_SIZE) return -1; sess->auth.algo = CCP_AUTH_ALGO_SHA224_HMAC; @@ -809,6 +849,7 @@ ccp_configure_session_auth(struct ccp_session *sess, auth_xform->key.length); if (generate_partial_hash(sess)) return -1; +#endif break; case RTE_CRYPTO_AUTH_SHA3_224: sess->auth.algo = CCP_AUTH_ALGO_SHA3_224; @@ -843,6 +884,17 @@ ccp_configure_session_auth(struct ccp_session *sess, sess->auth.offset = CCP_SB_BYTES - SHA256_DIGEST_SIZE; break; case RTE_CRYPTO_AUTH_SHA256_HMAC: +#ifdef RTE_LIBRTE_PMD_CCP_CPU_AUTH + if (auth_xform->key.length > SHA256_BLOCK_SIZE) + return -1; + sess->auth.algo = CCP_AUTH_ALGO_SHA256_HMAC; +
[dpdk-dev] [PATCH v5 18/19] test/crypto: add test for AMD CCP crypto poll mode
Signed-off-by: Ravi Kumar --- test/test/test_cryptodev.c | 161 +++ test/test/test_cryptodev.h | 1 + test/test/test_cryptodev_aes_test_vectors.h | 93 ++-- test/test/test_cryptodev_blockcipher.c | 9 +- test/test/test_cryptodev_blockcipher.h | 1 + test/test/test_cryptodev_des_test_vectors.h | 42 --- test/test/test_cryptodev_hash_test_vectors.h | 60 ++ 7 files changed, 301 insertions(+), 66 deletions(-) diff --git a/test/test/test_cryptodev.c b/test/test/test_cryptodev.c index 1417482..d1d7925 100644 --- a/test/test/test_cryptodev.c +++ b/test/test/test_cryptodev.c @@ -338,6 +338,23 @@ testsuite_setup(void) } } + /* Create an CCP device if required */ + if (gbl_driver_id == rte_cryptodev_driver_id_get( + RTE_STR(CRYPTODEV_NAME_CCP_PMD))) { + nb_devs = rte_cryptodev_device_count_by_driver( + rte_cryptodev_driver_id_get( + RTE_STR(CRYPTODEV_NAME_CCP_PMD))); + if (nb_devs < 1) { + ret = rte_vdev_init( + RTE_STR(CRYPTODEV_NAME_CCP_PMD), + NULL); + + TEST_ASSERT(ret == 0, "Failed to create " + "instance of pmd : %s", + RTE_STR(CRYPTODEV_NAME_CCP_PMD)); + } + } + #ifdef RTE_LIBRTE_PMD_CRYPTO_SCHEDULER if (gbl_driver_id == rte_cryptodev_driver_id_get( RTE_STR(CRYPTODEV_NAME_SCHEDULER_PMD))) { @@ -1727,6 +1744,44 @@ test_AES_cipheronly_openssl_all(void) } static int +test_AES_chain_ccp_all(void) +{ + struct crypto_testsuite_params *ts_params = &testsuite_params; + int status; + + status = test_blockcipher_all_tests(ts_params->mbuf_pool, + ts_params->op_mpool, + ts_params->session_mpool, + ts_params->valid_devs[0], + rte_cryptodev_driver_id_get( + RTE_STR(CRYPTODEV_NAME_CCP_PMD)), + BLKCIPHER_AES_CHAIN_TYPE); + + TEST_ASSERT_EQUAL(status, 0, "Test failed"); + + return TEST_SUCCESS; +} + +static int +test_AES_cipheronly_ccp_all(void) +{ + struct crypto_testsuite_params *ts_params = &testsuite_params; + int status; + + status = test_blockcipher_all_tests(ts_params->mbuf_pool, + ts_params->op_mpool, + ts_params->session_mpool, + ts_params->valid_devs[0], + rte_cryptodev_driver_id_get( + RTE_STR(CRYPTODEV_NAME_CCP_PMD)), + BLKCIPHER_AES_CIPHERONLY_TYPE); + + TEST_ASSERT_EQUAL(status, 0, "Test failed"); + + return TEST_SUCCESS; +} + +static int test_AES_chain_qat_all(void) { struct crypto_testsuite_params *ts_params = &testsuite_params; @@ -1898,6 +1953,25 @@ test_authonly_openssl_all(void) } static int +test_authonly_ccp_all(void) +{ + struct crypto_testsuite_params *ts_params = &testsuite_params; + int status; + + status = test_blockcipher_all_tests(ts_params->mbuf_pool, + ts_params->op_mpool, + ts_params->session_mpool, + ts_params->valid_devs[0], + rte_cryptodev_driver_id_get( + RTE_STR(CRYPTODEV_NAME_CCP_PMD)), + BLKCIPHER_AUTHONLY_TYPE); + + TEST_ASSERT_EQUAL(status, 0, "Test failed"); + + return TEST_SUCCESS; +} + +static int test_AES_chain_armv8_all(void) { struct crypto_testsuite_params *ts_params = &testsuite_params; @@ -4973,6 +5047,44 @@ test_3DES_cipheronly_dpaa2_sec_all(void) } static int +test_3DES_chain_ccp_all(void) +{ + struct crypto_testsuite_params *ts_params = &testsuite_params; + int status; + + status = test_blockcipher_all_tests(ts_params->mbuf_pool, + ts_params->op_mpool, + ts_params->session_mpool, + ts_params->valid_devs[0], + rte_cryptodev_driver_id_get( + RTE_STR(CRYPTODEV_NAME_CCP_PMD)), + BLKCIPHER_3DES_CHAIN_TYPE); + + TEST_ASSERT_EQUAL(status, 0, "Test failed"); + + return TEST_SUCCESS; +} + +static int +test_3DES_cipheronly_ccp_all(void) +{ + struct crypto_testsuite_params *ts_params = &testsuite_params; + int status; + + status = test_blockcipher_all_tests(ts_params->mbuf_pool, + ts_params->op_mpool, + ts_params->session_mpool, + ts_params->valid_devs[0], + rte_cryptodev_driver_id_get( + RTE_STR(CRYPTODEV_NAME_CCP_PMD)), + BLKCIPHER_3DES_CIPHERONLY_TYPE); + + TEST_ASSERT_EQUAL(status, 0, "Test failed"); + + return TEST_SUCCESS; +} + +static int test_3DES_cipheronly_qat_all(void) { struct crypto_testsuite_params *ts_p
[dpdk-dev] [PATCH v5 19/19] doc: add document for AMD CCP crypto poll mode driver
Signed-off-by: Ravi Kumar --- doc/guides/cryptodevs/ccp.rst | 102 + doc/guides/cryptodevs/features/ccp.ini | 57 doc/guides/cryptodevs/features/default.ini | 12 doc/guides/cryptodevs/index.rst| 1 + 4 files changed, 172 insertions(+) create mode 100644 doc/guides/cryptodevs/ccp.rst create mode 100644 doc/guides/cryptodevs/features/ccp.ini diff --git a/doc/guides/cryptodevs/ccp.rst b/doc/guides/cryptodevs/ccp.rst new file mode 100644 index 000..1fcd462 --- /dev/null +++ b/doc/guides/cryptodevs/ccp.rst @@ -0,0 +1,102 @@ +.. SPDX-License-Identifier: BSD-3-Clause + Copyright(c) 2018 Advanced Micro Devices, Inc. All rights reserved. + +AMD CCP Poll Mode Driver + + +This code provides the initial implementation of the ccp poll mode driver. +The CCP poll mode driver library (librte_pmd_ccp) implements support for +AMD’s cryptographic co-processor (CCP). The CCP PMD is a virtual crypto +poll mode driver which schedules crypto operations to one or more available +CCP hardware engines on the platform. The CCP PMD provides poll mode crypto +driver support for the following hardware accelerator devices:: + + AMD Cryptographic Co-processor (0x1456) + AMD Cryptographic Co-processor (0x1468) + +Features + + +CCP crypto PMD has support for: + +Cipher algorithms: + +* ``RTE_CRYPTO_CIPHER_AES_CBC`` +* ``RTE_CRYPTO_CIPHER_AES_ECB`` +* ``RTE_CRYPTO_CIPHER_AES_CTR`` +* ``RTE_CRYPTO_CIPHER_3DES_CBC`` + +Hash algorithms: + +* ``RTE_CRYPTO_AUTH_SHA1`` +* ``RTE_CRYPTO_AUTH_SHA1_HMAC`` +* ``RTE_CRYPTO_AUTH_SHA224`` +* ``RTE_CRYPTO_AUTH_SHA224_HMAC`` +* ``RTE_CRYPTO_AUTH_SHA256`` +* ``RTE_CRYPTO_AUTH_SHA256_HMAC`` +* ``RTE_CRYPTO_AUTH_SHA384`` +* ``RTE_CRYPTO_AUTH_SHA384_HMAC`` +* ``RTE_CRYPTO_AUTH_SHA512`` +* ``RTE_CRYPTO_AUTH_SHA512_HMAC`` +* ``RTE_CRYPTO_AUTH_MD5_HMAC`` +* ``RTE_CRYPTO_AUTH_AES_CMAC`` +* ``RTE_CRYPTO_AUTH_SHA3_224`` +* ``RTE_CRYPTO_AUTH_SHA3_224_HMAC`` +* ``RTE_CRYPTO_AUTH_SHA3_256`` +* ``RTE_CRYPTO_AUTH_SHA3_256_HMAC`` +* ``RTE_CRYPTO_AUTH_SHA3_384`` +* ``RTE_CRYPTO_AUTH_SHA3_384_HMAC`` +* ``RTE_CRYPTO_AUTH_SHA3_512`` +* ``RTE_CRYPTO_AUTH_SHA3_512_HMAC`` + +AEAD algorithms: + +* ``RTE_CRYPTO_AEAD_AES_GCM`` + +Installation + + +To compile CCP PMD, it has to be enabled in the config/common_base file. +* ``CONFIG_RTE_LIBRTE_PMD_CCP=y`` + +The CCP PMD also supports computing authentication over CPU with cipher offloaded +to CCP. To enable this feature, enable following in the configuration. +* ``CONFIG_RTE_LIBRTE_PMD_CCP_CPU_AUTH=y`` + +This code was verified on Ubuntu 16.04. + +Initialization +-- + +Bind the CCP devices to DPDK UIO driver module before running the CCP PMD stack. +e.g. for the 0x1456 device:: + + cd to the top-level DPDK directory + modprobe uio + insmod ./build/kmod/igb_uio.ko + echo "1022 1456" > /sys/bus/pci/drivers/igb_uio/new_id + +Another way to bind the CCP devices to DPDK UIO driver is by using the ``dpdk-devbind.py`` script. +The following command assumes ``BFD`` of ``:09:00.2``:: + + cd to the top-level DPDK directory + ./usertools/dpdk-devbind.py -b igb_uio :09:00.2 + +To verify real traffic l2fwd-crypto example can be used with following command: + +.. code-block:: console + + sudo ./build/l2fwd-crypto -l 1 -n 4 --vdev "crypto_ccp" -- -p 0x1 + --chain CIPHER_HASH --cipher_op ENCRYPT --cipher_algo AES_CBC + --cipher_key 00:01:02:03:04:05:06:07:08:09:0a:0b:0c:0d:0e:0f + --iv 00:01:02:03:04:05:06:07:08:09:0a:0b:0c:0d:0e:ff + --auth_op GENERATE --auth_algo SHA1_HMAC + --auth_key 11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11 + :11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11 + :11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11:11 + +Limitations +--- + +* Chained mbufs are not supported +* MD5_HMAC is supported only if ``CONFIG_RTE_LIBRTE_PMD_CCP_CPU_AUTH=y`` is enabled in configuration diff --git a/doc/guides/cryptodevs/features/ccp.ini b/doc/guides/cryptodevs/features/ccp.ini new file mode 100644 index 000..add4bd8 --- /dev/null +++ b/doc/guides/cryptodevs/features/ccp.ini @@ -0,0 +1,57 @@ +; +; Supported features of the 'ccp' crypto poll mode driver. +; +; Refer to default.ini for the full list of available PMD features. +; +[Features] +Symmetric crypto = Y +Sym operation chaining = Y +HW Accelerated = Y + +; +; Supported crypto algorithms of the 'ccp' crypto driver. +; +[Cipher] +AES CBC (128) = Y +AES CBC (192) = Y +AES CBC (256) = Y +AES ECB (128) = Y +AES ECB (192) = Y +AES ECB (256) = Y +AES CTR (128) = Y +AES CTR (192) = Y +AES CTR (256) = Y +3DES CBC = Y + +; +; Supported authentication algorithms of the 'ccp' crypto driver. +; +[Auth] +MD5 HMAC = Y +SHA1 = Y +SHA1 HMAC = Y +SHA224= Y +SHA224 HMAC= Y +SHA256
Re: [dpdk-dev] [PATCH v3 01/18] net/axgbe: add minimal dev init and uninit support
> > >-Original Message- >From: Ferruh Yigit [mailto:ferruh.yi...@intel.com] >Sent: Friday, March 16, 2018 11:12 PM >To: Kumar, Ravi1 ; dev@dpdk.org >Subject: Re: [PATCH v3 01/18] net/axgbe: add minimal dev init and uninit >support > >On 3/9/2018 8:42 AM, Ravi Kumar wrote: >> Signed-off-by: Ravi Kumar > ><...> > >> @@ -412,6 +412,12 @@ CONFIG_RTE_PMD_RING_MAX_TX_RINGS=16 >> CONFIG_RTE_LIBRTE_PMD_SOFTNIC=y >> >> # >> +# Compile AMD PMD >> +# >> +CONFIG_RTE_LIBRTE_AXGBE_PMD=y >> +CONFIG_RTE_LIBRTE_AXGBE_DEBUG_INIT=n > >Config file recently updated to have PMD options sorted alphabetically, can >you please move config option according. > ><...> > >> +/* The set of PCI devices this driver supports */ >> +#define AMD_PCI_VENDOR_ID 0x1022 >> +#define AMD_PCI_AXGBE_DEVICE_ID1 0x1458 #define >> +AMD_PCI_AXGBE_DEVICE_ID2 0x1459 > >More descriptive name can help others to find which devices are supported, >instead of id1 and id2. > ><...> > >> +static int >> +eth_axgbe_dev_init(struct rte_eth_dev *eth_dev) { >> +PMD_INIT_FUNC_TRACE(); >> +struct axgbe_port *pdata; >> +struct rte_pci_device *pci_dev; >> + >> +pdata = (struct axgbe_port *)eth_dev->data->dev_private; >> +pdata->eth_dev = eth_dev; > >Since there is already a RTE_PROC_PRIMARY check below, these should be below >that check because secondary doesn't need to update pdata->eth_dev > ><...> > >> +RTE_PMD_REGISTER_PCI(net_axgbe, rte_axgbe_pmd); >> +RTE_PMD_REGISTER_PCI_TABLE(net_axgbe, pci_id_axgbe_map); >> +RTE_PMD_REGISTER_KMOD_DEP(net_axgbe, "* igb_uio | uio_pci_generic"); > >Is vfio-pci intentionally not included as dependency? > >> + >> +RTE_INIT(axgbe_init_log); >> +static void >> +axgbe_init_log(void) >> +{ >> +axgbe_logtype_init = rte_log_register("pmd.axgbe.init"); > >The log string syntax updated, now it should be "pmd.net.axgbe.*" > ><...> > >> +#ifdef RTE_LIBRTE_AXGBE_DEBUG_INIT >> +#define PMD_INIT_FUNC_TRACE() PMD_INIT_LOG(DEBUG, " >>") #else >> +#define PMD_INIT_FUNC_TRACE() do { } while (0) #endif > >Do we need this config option? The idea of dynamic logging is get rid of them. >If you want to control trace logs seperately, you can create a new logtype for >it can control dynamically. As long as it is not in datapath we can remove log >config options. > >> +extern int axgbe_logtype_driver; >> +#define PMD_DRV_LOG_RAW(level, fmt, args...) \ >> +rte_log(RTE_LOG_ ## level, axgbe_logtype_driver, "%s(): " fmt, \ >> +__func__, ## args) >> + >> +#define PMD_DRV_LOG(level, fmt, args...) \ >> +PMD_DRV_LOG_RAW(level, fmt "\n", ## args) > >Do you need interim PMD_DRV_LOG_RAW? Why not directly define PMD_DRV_LOG? > ><...> Hi Ferruh, Thanks a lot for the review comments. We are working on these and will upload a patch-set that will address these and the previous SPDX licensing comments. Regards, Ravi
[dpdk-dev] [PATCH] net/mrvl: fix typo in log message
Show appropriate log message in case Tx offloads are either not supported or missing. Fixes: 7d8f6c20cc7c ("net/mrvl: switch to the new Tx offload API") Cc: sta...@dpdk.org Signed-off-by: Tomasz Duszynski --- drivers/net/mrvl/mrvl_ethdev.c | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/drivers/net/mrvl/mrvl_ethdev.c b/drivers/net/mrvl/mrvl_ethdev.c index fac924f..61c1223 100644 --- a/drivers/net/mrvl/mrvl_ethdev.c +++ b/drivers/net/mrvl/mrvl_ethdev.c @@ -1700,14 +1700,14 @@ mrvl_tx_queue_offloads_okay(struct rte_eth_dev *dev, uint64_t requested) uint64_t missing = mandatory & ~requested; if (unsupported) { - RTE_LOG(ERR, PMD, "Some Rx offloads are not supported. " + RTE_LOG(ERR, PMD, "Some Tx offloads are not supported. " "Requested 0x%" PRIx64 " supported 0x%" PRIx64 ".\n", requested, supported); return 0; } if (missing) { - RTE_LOG(ERR, PMD, "Some Rx offloads are missing. " + RTE_LOG(ERR, PMD, "Some Tx offloads are missing. " "Requested 0x%" PRIx64 " missing 0x%" PRIx64 ".\n", requested, missing); return 0; -- 2.7.4
[dpdk-dev] [PATCH v4 0/3] app/testpmd: add new commands to test new Tx/Rx offload
Existed testpmd commands can't support per queue offload configuration. And there are different commands to enable or disable different offloading. This patch set add following commands to support new Tx/Rx offloading API test. To get Rx offload capability of a port, please run: testpmd > rx_offload get capability To get current Rx offload per queue and per port configuration of a port, run: tesstpmd > rx_offload get configuration To enable or disable a Rx per port offloading, please run: testpmd > rx_offload enable|disable per_port vlan_strip|ipv4_cksum|... This command will set|clear the associated bit in dev->dev_conf.rxmode.offloads for rte_eth_dev_configure and tx_conf->offloads of all Rx queues for rte_eth_rx_queue_setup( ). To enable or disable a Tx per port offloading, please run: testpmd > rx_offload enable|disable per_queue vlan_strip|ipv4_cksum|... Same commands like "tx_offload ..." are also added to support new Tx offload API test. --- v4: improve testpmd command per port offload to set or clear the port configuration and the queue configuration of all queues. v3: add enum rte_eth_rx_offload_type and enum rte_eth_tx_offload_type free memory of port->rx_offloads and port->tx_offloads when testpmd is existed v2: use rte_eth_dev_rx_offload_name() and rte_eth_dev_tx_offload_name(). remove static const strings of Rx/Tx offload names. Wei Dai (3): ethdev: add enum type for loop on Rx/Tx offloads app/testpmd: add commands to test new Rx offload API pp/testpmd: add commands to test new Tx offload API app/test-pmd/cmdline.c| 753 ++ app/test-pmd/testpmd.c| 34 +- app/test-pmd/testpmd.h| 2 + lib/librte_ether/rte_ethdev.h | 44 +++ 4 files changed, 829 insertions(+), 4 deletions(-) -- 2.7.5
[dpdk-dev] [PATCH v4 1/3] ethdev: add enum type for loop on Rx/Tx offloads
This patch adds enum rte_eth_rx_offload_type and enum rte_eth_tx_offload_type. For a loop on all Rx offloads, it is convenient to begin with the first enum member ETH_RX_OFFLOAD_FIRST_FEATURE and to end at ETH_RX_OFFLOAD_TOTAL_NUM. A loop on all Tx offloads can begin with ETH_TX_OFFLOAD_FIRST_FEATURE and end at ETH_TX_OFFLOAD_TOTAL_NUM. Signed-off-by: Wei Dai --- lib/librte_ether/rte_ethdev.h | 44 +++ 1 file changed, 44 insertions(+) diff --git a/lib/librte_ether/rte_ethdev.h b/lib/librte_ether/rte_ethdev.h index 0361533..0089ea3 100644 --- a/lib/librte_ether/rte_ethdev.h +++ b/lib/librte_ether/rte_ethdev.h @@ -946,6 +946,27 @@ struct rte_eth_conf { DEV_RX_OFFLOAD_VLAN_FILTER | \ DEV_RX_OFFLOAD_VLAN_EXTEND) +enum rte_eth_rx_offload_type { + ETH_RX_OFFLOAD_FIRST_FEATURE = 0, + ETH_RX_OFFLOAD_VLAN_STRIP = ETH_RX_OFFLOAD_FIRST_FEATURE, + ETH_RX_OFFLOAD_IPV4_CKSUM, + ETH_RX_OFFLOAD_UDP_CKSUM, + ETH_RX_OFFLOAD_TCP_CKSUM, + ETH_RX_OFFLOAD_TCP_LRO, + ETH_RX_OFFLOAD_QINQ_STRIP, + ETH_RX_OFFLOAD_OUTER_IPV4_CKSUM, + ETH_RX_OFFLOAD_MACSEC_STRIP, + ETH_RX_OFFLOAD_HEADER_SPLIT, + ETH_RX_OFFLOAD_VLAN_FILTER, + ETH_RX_OFFLOAD_VLAN_EXTEND, + ETH_RX_OFFLOAD_JUMBO_FRAME, + ETH_RX_OFFLOAD_CRC_STRIP, + ETH_RX_OFFLOAD_SCATTER, + ETH_RX_OFFLOAD_TIMESTAMP, + ETH_RX_OFFLOAD_SECURITY, + ETH_RX_OFFLOAD_TOTAL_NUM +}; + /* * If new Rx offload capabilities are defined, they also must be * mentioned in rte_rx_offload_names in rte_ethdev.c file. @@ -981,6 +1002,29 @@ struct rte_eth_conf { */ #define DEV_TX_OFFLOAD_SECURITY 0x0002 +enum rte_eth_tx_offload_type { + ETH_TX_OFFLOAD_FIRST_FEATURE = 0, + ETH_TX_OFFLOAD_VLAN_INSERT = ETH_TX_OFFLOAD_FIRST_FEATURE, + ETH_TX_OFFLOAD_IPV4_CKSUM, + ETH_TX_OFFLOAD_UDP_CKSUM, + ETH_TX_OFFLOAD_TCP_CKSUM, + ETH_TX_OFFLOAD_SCTP_CKSUM, + ETH_TX_OFFLOAD_TCP_TSO, + ETH_TX_OFFLOAD_UDP_TSO, + ETH_TX_OFFLOAD_OUTER_IPV4_CKSUM, + ETH_TX_OFFLOAD_QINQ_INSERT, + ETH_TX_OFFLOAD_VXLAN_TNL_TSO, + ETH_TX_OFFLOAD_GRE_TNL_TSO, + ETH_TX_OFFLOAD_IPIP_TNL_TSO, + ETH_TX_OFFLOAD_GENEVE_TNL_TSO, + ETH_TX_OFFLOAD_MACSEC_INSERT, + ETH_TX_OFFLOAD_MT_LOCKFREE, + ETH_TX_OFFLOAD_MULTI_SEGS, + ETH_TX_OFFLOAD_MBUF_FAST_FREE, + ETH_TX_OFFLOAD_SECURITY, + ETH_TX_OFFLOAD_TOTAL_NUM +}; + /* * If new Tx offload capabilities are defined, they also must be * mentioned in rte_tx_offload_names in rte_ethdev.c file. -- 2.7.5
[dpdk-dev] [PATCH v4 2/3] app/testpmd: add commands to test new Rx offload API
Add following testpmd run-time commands to support test of new Rx offload API: rx_offload get capability rx_offload get configuration rx_offload enable|disable per_port rx_offload enable|disable per_queue Above last 2 commands should be run when the port is stopped. And can be one of "vlan_strip", "ipv4_cksum", ... Signed-off-by: Wei Dai --- app/test-pmd/cmdline.c | 374 + app/test-pmd/testpmd.c | 19 ++- app/test-pmd/testpmd.h | 1 + 3 files changed, 392 insertions(+), 2 deletions(-) diff --git a/app/test-pmd/cmdline.c b/app/test-pmd/cmdline.c index 40b31ad..ae735d0 100644 --- a/app/test-pmd/cmdline.c +++ b/app/test-pmd/cmdline.c @@ -15996,6 +15996,376 @@ cmdline_parse_inst_t cmd_ptype_mapping_update = { }, }; +/* Get Rx offloads capability */ +struct cmd_rx_offload_get_capa_result { + cmdline_fixed_string_t rx_offload; + cmdline_fixed_string_t get; + cmdline_fixed_string_t capability; + portid_t port_id; +}; + +cmdline_parse_token_string_t cmd_rx_offload_get_capa_rx_offload = + TOKEN_STRING_INITIALIZER + (struct cmd_rx_offload_get_capa_result, +rx_offload, "rx_offload"); +cmdline_parse_token_string_t cmd_rx_offload_get_capa_get = + TOKEN_STRING_INITIALIZER + (struct cmd_rx_offload_get_capa_result, +get, "get"); +cmdline_parse_token_string_t cmd_rx_offload_get_capa_capability = + TOKEN_STRING_INITIALIZER + (struct cmd_rx_offload_get_capa_result, +capability, "capability"); +cmdline_parse_token_num_t cmd_rx_offload_get_capa_port_id = + TOKEN_NUM_INITIALIZER + (struct cmd_rx_offload_get_capa_result, +port_id, UINT16); + +static void +print_rx_offloads(uint64_t offloads) +{ + uint64_t single_offload; + int begin; + int end; + int bit; + + if (offloads == 0) + return; + + begin = __builtin_ctzll(offloads); + end = sizeof(offloads) * CHAR_BIT - __builtin_clzll(offloads); + + single_offload = 1 << begin; + for (bit = begin; bit < end; bit++) { + if (offloads & single_offload) + printf(" %s", + rte_eth_dev_rx_offload_name(single_offload)); + single_offload <<= 1; + } +} + +static void +cmd_rx_offload_get_capa_parsed( + void *parsed_result, + __attribute__((unused)) struct cmdline *cl, + __attribute__((unused)) void *data) +{ + struct cmd_rx_offload_get_capa_result *res = parsed_result; + struct rte_eth_dev_info dev_info; + portid_t port_id = res->port_id; + uint64_t queue_offloads; + uint64_t port_offloads; + + rte_eth_dev_info_get(port_id, &dev_info); + queue_offloads = dev_info.rx_queue_offload_capa; + port_offloads = dev_info.rx_offload_capa ^ queue_offloads; + + printf("Rx Offloading Capabilities of port %d :\n", port_id); + printf(" Per Queue :"); + print_rx_offloads(queue_offloads); + + printf("\n"); + printf(" Per Port :"); + print_rx_offloads(port_offloads); + printf("\n\n"); +} + +cmdline_parse_inst_t cmd_rx_offload_get_capa = { + .f = cmd_rx_offload_get_capa_parsed, + .data = NULL, + .help_str = "rx_offload get capability ", + .tokens = { + (void *)&cmd_rx_offload_get_capa_rx_offload, + (void *)&cmd_rx_offload_get_capa_get, + (void *)&cmd_rx_offload_get_capa_capability, + (void *)&cmd_rx_offload_get_capa_port_id, + NULL, + } +}; + +/* Get Rx offloads configuration */ +struct cmd_rx_offload_get_configuration_result { + cmdline_fixed_string_t rx_offload; + cmdline_fixed_string_t get; + cmdline_fixed_string_t configuration; + portid_t port_id; +}; + +cmdline_parse_token_string_t cmd_rx_offload_get_configuration_rx_offload = + TOKEN_STRING_INITIALIZER + (struct cmd_rx_offload_get_configuration_result, +rx_offload, "rx_offload"); +cmdline_parse_token_string_t cmd_rx_offload_get_configuration_get = + TOKEN_STRING_INITIALIZER + (struct cmd_rx_offload_get_configuration_result, +get, "get"); +cmdline_parse_token_string_t cmd_rx_offload_get_configuration_configuration = + TOKEN_STRING_INITIALIZER + (struct cmd_rx_offload_get_configuration_result, +configuration, "configuration"); +cmdline_parse_token_num_t cmd_rx_offload_get_configuration_port_id = + TOKEN_NUM_INITIALIZER + (struct cmd_rx_offload_get_configuration_result, +port_id, UINT16); + +static void +cmd_rx_offload_get_configuration_parsed( + void *parsed_result, + __attribute__((unused)) struct cmdline *cl, + __attribute__((unused)) void *data) +{ + struct cmd_rx_offload_
[dpdk-dev] [PATCH v4 3/3] pp/testpmd: add commands to test new Tx offload API
Add following testpmd run-time commands to support test of new Tx offload API: tx_offload get capability tx_offload get configuration tx_offload enable|disable per_port tx_offload enable|disable per_queue Above last 2 commands should be run when the port is stopped. And can be one of "vlan_insert", "udp_cksum", ... Signed-off-by: Wei Dai --- app/test-pmd/cmdline.c | 379 + app/test-pmd/testpmd.c | 15 +- app/test-pmd/testpmd.h | 1 + 3 files changed, 393 insertions(+), 2 deletions(-) diff --git a/app/test-pmd/cmdline.c b/app/test-pmd/cmdline.c index ae735d0..e977910 100644 --- a/app/test-pmd/cmdline.c +++ b/app/test-pmd/cmdline.c @@ -16366,6 +16366,381 @@ cmdline_parse_inst_t cmd_config_per_queue_rx_offload = { } }; +/* Get Tx offloads capability */ +struct cmd_tx_offload_get_capa_result { + cmdline_fixed_string_t tx_offload; + cmdline_fixed_string_t get; + cmdline_fixed_string_t capability; + portid_t port_id; +}; + +cmdline_parse_token_string_t cmd_tx_offload_get_capa_tx_offload = + TOKEN_STRING_INITIALIZER + (struct cmd_tx_offload_get_capa_result, +tx_offload, "tx_offload"); +cmdline_parse_token_string_t cmd_tx_offload_get_capa_get = + TOKEN_STRING_INITIALIZER + (struct cmd_tx_offload_get_capa_result, +get, "get"); +cmdline_parse_token_string_t cmd_tx_offload_get_capa_capability = + TOKEN_STRING_INITIALIZER + (struct cmd_tx_offload_get_capa_result, +capability, "capability"); +cmdline_parse_token_num_t cmd_tx_offload_get_capa_port_id = + TOKEN_NUM_INITIALIZER + (struct cmd_tx_offload_get_capa_result, +port_id, UINT16); + +static void +print_tx_offloads(uint64_t offloads) +{ + uint64_t single_offload; + int begin; + int end; + int bit; + + if (offloads == 0) + return; + + begin = __builtin_ctzll(offloads); + end = sizeof(offloads) * CHAR_BIT - __builtin_clzll(offloads); + + single_offload = 1 << begin; + for (bit = begin; bit < end; bit++) { + if (offloads & single_offload) + printf(" %s", + rte_eth_dev_tx_offload_name(single_offload)); + single_offload <<= 1; + } +} + +static void +cmd_tx_offload_get_capa_parsed( + void *parsed_result, + __attribute__((unused)) struct cmdline *cl, + __attribute__((unused)) void *data) +{ + struct cmd_tx_offload_get_capa_result *res = parsed_result; + struct rte_eth_dev_info dev_info; + portid_t port_id = res->port_id; + uint64_t queue_offloads; + uint64_t port_offloads; + + rte_eth_dev_info_get(port_id, &dev_info); + queue_offloads = dev_info.tx_queue_offload_capa; + port_offloads = dev_info.tx_offload_capa ^ queue_offloads; + + printf("Tx Offloading Capabilities of port %d :\n", port_id); + printf(" Per Queue :"); + print_tx_offloads(queue_offloads); + printf("\n"); + printf(" Per Port :"); + print_tx_offloads(port_offloads); + printf("\n\n"); +} + +cmdline_parse_inst_t cmd_tx_offload_get_capa = { + .f = cmd_tx_offload_get_capa_parsed, + .data = NULL, + .help_str = "tx_offload get capability ", + .tokens = { + (void *)&cmd_tx_offload_get_capa_tx_offload, + (void *)&cmd_tx_offload_get_capa_get, + (void *)&cmd_tx_offload_get_capa_capability, + (void *)&cmd_tx_offload_get_capa_port_id, + NULL, + } +}; + +/* Get Tx offloads configuration */ +struct cmd_tx_offload_get_configuration_result { + cmdline_fixed_string_t tx_offload; + cmdline_fixed_string_t get; + cmdline_fixed_string_t configuration; + portid_t port_id; +}; + +cmdline_parse_token_string_t cmd_tx_offload_get_configuration_tx_offload = + TOKEN_STRING_INITIALIZER + (struct cmd_tx_offload_get_configuration_result, +tx_offload, "tx_offload"); +cmdline_parse_token_string_t cmd_tx_offload_get_configuration_get = + TOKEN_STRING_INITIALIZER + (struct cmd_tx_offload_get_configuration_result, +get, "get"); +cmdline_parse_token_string_t cmd_tx_offload_get_configuration_configuration = + TOKEN_STRING_INITIALIZER + (struct cmd_tx_offload_get_configuration_result, +configuration, "configuration"); +cmdline_parse_token_num_t cmd_tx_offload_get_configuration_port_id = + TOKEN_NUM_INITIALIZER + (struct cmd_tx_offload_get_configuration_result, +port_id, UINT16); + +static void +cmd_tx_offload_get_configuration_parsed( + void *parsed_result, + __attribute__((unused)) struct cmdline *cl, + __attribute__((unused)) void *data) +{ + struct cmd_tx_offl
[dpdk-dev] [PATCH v4 0/3] app/testpmd: add new commands to test new Tx/Rx offload
Existed testpmd commands can't support per queue offload configuration. And there are different commands to enable or disable different offloading. This patch set add following commands to support new Tx/Rx offloading API test. To get Rx offload capability of a port, please run: testpmd > rx_offload get capability To get current Rx offload per queue and per port configuration of a port, run: tesstpmd > rx_offload get configuration To enable or disable a Rx per port offloading, please run: testpmd > rx_offload enable|disable per_port vlan_strip|ipv4_cksum|... This command will set|clear the associated bit in dev->dev_conf.rxmode.offloads for rte_eth_dev_configure and tx_conf->offloads of all Rx queues for rte_eth_rx_queue_setup( ). To enable or disable a Tx per port offloading, please run: testpmd > rx_offload enable|disable per_queue vlan_strip|ipv4_cksum|... Same commands like "tx_offload ..." are also added to support new Tx offload API test. --- v4: improve testpmd command per port offload to set or clear the port configuration and the queue configuration of all queues. v3: add enum rte_eth_rx_offload_type and enum rte_eth_tx_offload_type free memory of port->rx_offloads and port->tx_offloads when testpmd is existed v2: use rte_eth_dev_rx_offload_name() and rte_eth_dev_tx_offload_name(). remove static const strings of Rx/Tx offload names. Wei Dai (3): ethdev: add enum type for loop on Rx/Tx offloads app/testpmd: add commands to test new Rx offload API pp/testpmd: add commands to test new Tx offload API app/test-pmd/cmdline.c| 753 ++ app/test-pmd/testpmd.c| 34 +- app/test-pmd/testpmd.h| 2 + lib/librte_ether/rte_ethdev.h | 44 +++ 4 files changed, 829 insertions(+), 4 deletions(-) -- 2.7.5
[dpdk-dev] [PATCH v4 1/3] ethdev: add enum type for loop on Rx/Tx offloads
This patch adds enum rte_eth_rx_offload_type and enum rte_eth_tx_offload_type. For a loop on all Rx offloads, it is convenient to begin with the first enum member ETH_RX_OFFLOAD_FIRST_FEATURE and to end at ETH_RX_OFFLOAD_TOTAL_NUM. A loop on all Tx offloads can begin with ETH_TX_OFFLOAD_FIRST_FEATURE and end at ETH_TX_OFFLOAD_TOTAL_NUM. Signed-off-by: Wei Dai --- lib/librte_ether/rte_ethdev.h | 44 +++ 1 file changed, 44 insertions(+) diff --git a/lib/librte_ether/rte_ethdev.h b/lib/librte_ether/rte_ethdev.h index 0361533..0089ea3 100644 --- a/lib/librte_ether/rte_ethdev.h +++ b/lib/librte_ether/rte_ethdev.h @@ -946,6 +946,27 @@ struct rte_eth_conf { DEV_RX_OFFLOAD_VLAN_FILTER | \ DEV_RX_OFFLOAD_VLAN_EXTEND) +enum rte_eth_rx_offload_type { + ETH_RX_OFFLOAD_FIRST_FEATURE = 0, + ETH_RX_OFFLOAD_VLAN_STRIP = ETH_RX_OFFLOAD_FIRST_FEATURE, + ETH_RX_OFFLOAD_IPV4_CKSUM, + ETH_RX_OFFLOAD_UDP_CKSUM, + ETH_RX_OFFLOAD_TCP_CKSUM, + ETH_RX_OFFLOAD_TCP_LRO, + ETH_RX_OFFLOAD_QINQ_STRIP, + ETH_RX_OFFLOAD_OUTER_IPV4_CKSUM, + ETH_RX_OFFLOAD_MACSEC_STRIP, + ETH_RX_OFFLOAD_HEADER_SPLIT, + ETH_RX_OFFLOAD_VLAN_FILTER, + ETH_RX_OFFLOAD_VLAN_EXTEND, + ETH_RX_OFFLOAD_JUMBO_FRAME, + ETH_RX_OFFLOAD_CRC_STRIP, + ETH_RX_OFFLOAD_SCATTER, + ETH_RX_OFFLOAD_TIMESTAMP, + ETH_RX_OFFLOAD_SECURITY, + ETH_RX_OFFLOAD_TOTAL_NUM +}; + /* * If new Rx offload capabilities are defined, they also must be * mentioned in rte_rx_offload_names in rte_ethdev.c file. @@ -981,6 +1002,29 @@ struct rte_eth_conf { */ #define DEV_TX_OFFLOAD_SECURITY 0x0002 +enum rte_eth_tx_offload_type { + ETH_TX_OFFLOAD_FIRST_FEATURE = 0, + ETH_TX_OFFLOAD_VLAN_INSERT = ETH_TX_OFFLOAD_FIRST_FEATURE, + ETH_TX_OFFLOAD_IPV4_CKSUM, + ETH_TX_OFFLOAD_UDP_CKSUM, + ETH_TX_OFFLOAD_TCP_CKSUM, + ETH_TX_OFFLOAD_SCTP_CKSUM, + ETH_TX_OFFLOAD_TCP_TSO, + ETH_TX_OFFLOAD_UDP_TSO, + ETH_TX_OFFLOAD_OUTER_IPV4_CKSUM, + ETH_TX_OFFLOAD_QINQ_INSERT, + ETH_TX_OFFLOAD_VXLAN_TNL_TSO, + ETH_TX_OFFLOAD_GRE_TNL_TSO, + ETH_TX_OFFLOAD_IPIP_TNL_TSO, + ETH_TX_OFFLOAD_GENEVE_TNL_TSO, + ETH_TX_OFFLOAD_MACSEC_INSERT, + ETH_TX_OFFLOAD_MT_LOCKFREE, + ETH_TX_OFFLOAD_MULTI_SEGS, + ETH_TX_OFFLOAD_MBUF_FAST_FREE, + ETH_TX_OFFLOAD_SECURITY, + ETH_TX_OFFLOAD_TOTAL_NUM +}; + /* * If new Tx offload capabilities are defined, they also must be * mentioned in rte_tx_offload_names in rte_ethdev.c file. -- 2.7.5
[dpdk-dev] [PATCH v4 3/3] app/testpmd: add commands to test new Tx offload API
Add following testpmd run-time commands to support test of new Tx offload API: tx_offload get capability tx_offload get configuration tx_offload enable|disable per_port tx_offload enable|disable per_queue Above last 2 commands should be run when the port is stopped. And can be one of "vlan_insert", "udp_cksum", ... Signed-off-by: Wei Dai --- app/test-pmd/cmdline.c | 379 + app/test-pmd/testpmd.c | 15 +- app/test-pmd/testpmd.h | 1 + 3 files changed, 393 insertions(+), 2 deletions(-) diff --git a/app/test-pmd/cmdline.c b/app/test-pmd/cmdline.c index ae735d0..e977910 100644 --- a/app/test-pmd/cmdline.c +++ b/app/test-pmd/cmdline.c @@ -16366,6 +16366,381 @@ cmdline_parse_inst_t cmd_config_per_queue_rx_offload = { } }; +/* Get Tx offloads capability */ +struct cmd_tx_offload_get_capa_result { + cmdline_fixed_string_t tx_offload; + cmdline_fixed_string_t get; + cmdline_fixed_string_t capability; + portid_t port_id; +}; + +cmdline_parse_token_string_t cmd_tx_offload_get_capa_tx_offload = + TOKEN_STRING_INITIALIZER + (struct cmd_tx_offload_get_capa_result, +tx_offload, "tx_offload"); +cmdline_parse_token_string_t cmd_tx_offload_get_capa_get = + TOKEN_STRING_INITIALIZER + (struct cmd_tx_offload_get_capa_result, +get, "get"); +cmdline_parse_token_string_t cmd_tx_offload_get_capa_capability = + TOKEN_STRING_INITIALIZER + (struct cmd_tx_offload_get_capa_result, +capability, "capability"); +cmdline_parse_token_num_t cmd_tx_offload_get_capa_port_id = + TOKEN_NUM_INITIALIZER + (struct cmd_tx_offload_get_capa_result, +port_id, UINT16); + +static void +print_tx_offloads(uint64_t offloads) +{ + uint64_t single_offload; + int begin; + int end; + int bit; + + if (offloads == 0) + return; + + begin = __builtin_ctzll(offloads); + end = sizeof(offloads) * CHAR_BIT - __builtin_clzll(offloads); + + single_offload = 1 << begin; + for (bit = begin; bit < end; bit++) { + if (offloads & single_offload) + printf(" %s", + rte_eth_dev_tx_offload_name(single_offload)); + single_offload <<= 1; + } +} + +static void +cmd_tx_offload_get_capa_parsed( + void *parsed_result, + __attribute__((unused)) struct cmdline *cl, + __attribute__((unused)) void *data) +{ + struct cmd_tx_offload_get_capa_result *res = parsed_result; + struct rte_eth_dev_info dev_info; + portid_t port_id = res->port_id; + uint64_t queue_offloads; + uint64_t port_offloads; + + rte_eth_dev_info_get(port_id, &dev_info); + queue_offloads = dev_info.tx_queue_offload_capa; + port_offloads = dev_info.tx_offload_capa ^ queue_offloads; + + printf("Tx Offloading Capabilities of port %d :\n", port_id); + printf(" Per Queue :"); + print_tx_offloads(queue_offloads); + printf("\n"); + printf(" Per Port :"); + print_tx_offloads(port_offloads); + printf("\n\n"); +} + +cmdline_parse_inst_t cmd_tx_offload_get_capa = { + .f = cmd_tx_offload_get_capa_parsed, + .data = NULL, + .help_str = "tx_offload get capability ", + .tokens = { + (void *)&cmd_tx_offload_get_capa_tx_offload, + (void *)&cmd_tx_offload_get_capa_get, + (void *)&cmd_tx_offload_get_capa_capability, + (void *)&cmd_tx_offload_get_capa_port_id, + NULL, + } +}; + +/* Get Tx offloads configuration */ +struct cmd_tx_offload_get_configuration_result { + cmdline_fixed_string_t tx_offload; + cmdline_fixed_string_t get; + cmdline_fixed_string_t configuration; + portid_t port_id; +}; + +cmdline_parse_token_string_t cmd_tx_offload_get_configuration_tx_offload = + TOKEN_STRING_INITIALIZER + (struct cmd_tx_offload_get_configuration_result, +tx_offload, "tx_offload"); +cmdline_parse_token_string_t cmd_tx_offload_get_configuration_get = + TOKEN_STRING_INITIALIZER + (struct cmd_tx_offload_get_configuration_result, +get, "get"); +cmdline_parse_token_string_t cmd_tx_offload_get_configuration_configuration = + TOKEN_STRING_INITIALIZER + (struct cmd_tx_offload_get_configuration_result, +configuration, "configuration"); +cmdline_parse_token_num_t cmd_tx_offload_get_configuration_port_id = + TOKEN_NUM_INITIALIZER + (struct cmd_tx_offload_get_configuration_result, +port_id, UINT16); + +static void +cmd_tx_offload_get_configuration_parsed( + void *parsed_result, + __attribute__((unused)) struct cmdline *cl, + __attribute__((unused)) void *data) +{ + struct cmd_tx_offl
[dpdk-dev] [PATCH v4 2/3] app/testpmd: add commands to test new Rx offload API
Add following testpmd run-time commands to support test of new Rx offload API: rx_offload get capability rx_offload get configuration rx_offload enable|disable per_port rx_offload enable|disable per_queue Above last 2 commands should be run when the port is stopped. And can be one of "vlan_strip", "ipv4_cksum", ... Signed-off-by: Wei Dai --- app/test-pmd/cmdline.c | 374 + app/test-pmd/testpmd.c | 19 ++- app/test-pmd/testpmd.h | 1 + 3 files changed, 392 insertions(+), 2 deletions(-) diff --git a/app/test-pmd/cmdline.c b/app/test-pmd/cmdline.c index 40b31ad..ae735d0 100644 --- a/app/test-pmd/cmdline.c +++ b/app/test-pmd/cmdline.c @@ -15996,6 +15996,376 @@ cmdline_parse_inst_t cmd_ptype_mapping_update = { }, }; +/* Get Rx offloads capability */ +struct cmd_rx_offload_get_capa_result { + cmdline_fixed_string_t rx_offload; + cmdline_fixed_string_t get; + cmdline_fixed_string_t capability; + portid_t port_id; +}; + +cmdline_parse_token_string_t cmd_rx_offload_get_capa_rx_offload = + TOKEN_STRING_INITIALIZER + (struct cmd_rx_offload_get_capa_result, +rx_offload, "rx_offload"); +cmdline_parse_token_string_t cmd_rx_offload_get_capa_get = + TOKEN_STRING_INITIALIZER + (struct cmd_rx_offload_get_capa_result, +get, "get"); +cmdline_parse_token_string_t cmd_rx_offload_get_capa_capability = + TOKEN_STRING_INITIALIZER + (struct cmd_rx_offload_get_capa_result, +capability, "capability"); +cmdline_parse_token_num_t cmd_rx_offload_get_capa_port_id = + TOKEN_NUM_INITIALIZER + (struct cmd_rx_offload_get_capa_result, +port_id, UINT16); + +static void +print_rx_offloads(uint64_t offloads) +{ + uint64_t single_offload; + int begin; + int end; + int bit; + + if (offloads == 0) + return; + + begin = __builtin_ctzll(offloads); + end = sizeof(offloads) * CHAR_BIT - __builtin_clzll(offloads); + + single_offload = 1 << begin; + for (bit = begin; bit < end; bit++) { + if (offloads & single_offload) + printf(" %s", + rte_eth_dev_rx_offload_name(single_offload)); + single_offload <<= 1; + } +} + +static void +cmd_rx_offload_get_capa_parsed( + void *parsed_result, + __attribute__((unused)) struct cmdline *cl, + __attribute__((unused)) void *data) +{ + struct cmd_rx_offload_get_capa_result *res = parsed_result; + struct rte_eth_dev_info dev_info; + portid_t port_id = res->port_id; + uint64_t queue_offloads; + uint64_t port_offloads; + + rte_eth_dev_info_get(port_id, &dev_info); + queue_offloads = dev_info.rx_queue_offload_capa; + port_offloads = dev_info.rx_offload_capa ^ queue_offloads; + + printf("Rx Offloading Capabilities of port %d :\n", port_id); + printf(" Per Queue :"); + print_rx_offloads(queue_offloads); + + printf("\n"); + printf(" Per Port :"); + print_rx_offloads(port_offloads); + printf("\n\n"); +} + +cmdline_parse_inst_t cmd_rx_offload_get_capa = { + .f = cmd_rx_offload_get_capa_parsed, + .data = NULL, + .help_str = "rx_offload get capability ", + .tokens = { + (void *)&cmd_rx_offload_get_capa_rx_offload, + (void *)&cmd_rx_offload_get_capa_get, + (void *)&cmd_rx_offload_get_capa_capability, + (void *)&cmd_rx_offload_get_capa_port_id, + NULL, + } +}; + +/* Get Rx offloads configuration */ +struct cmd_rx_offload_get_configuration_result { + cmdline_fixed_string_t rx_offload; + cmdline_fixed_string_t get; + cmdline_fixed_string_t configuration; + portid_t port_id; +}; + +cmdline_parse_token_string_t cmd_rx_offload_get_configuration_rx_offload = + TOKEN_STRING_INITIALIZER + (struct cmd_rx_offload_get_configuration_result, +rx_offload, "rx_offload"); +cmdline_parse_token_string_t cmd_rx_offload_get_configuration_get = + TOKEN_STRING_INITIALIZER + (struct cmd_rx_offload_get_configuration_result, +get, "get"); +cmdline_parse_token_string_t cmd_rx_offload_get_configuration_configuration = + TOKEN_STRING_INITIALIZER + (struct cmd_rx_offload_get_configuration_result, +configuration, "configuration"); +cmdline_parse_token_num_t cmd_rx_offload_get_configuration_port_id = + TOKEN_NUM_INITIALIZER + (struct cmd_rx_offload_get_configuration_result, +port_id, UINT16); + +static void +cmd_rx_offload_get_configuration_parsed( + void *parsed_result, + __attribute__((unused)) struct cmdline *cl, + __attribute__((unused)) void *data) +{ + struct cmd_rx_offload_
[dpdk-dev] [PATCH] net/avp: conforming to new ethdev offload API
This change updates the AVP driver to conform with the new ethdev offload API. As per these commits: commit cba7f53b717d ("ethdev: introduce Tx queue offloads API")' commit ce17eddefc20 ("ethdev: introduce Rx queue offloads API")' Signed-off-by: Allain Legacy --- drivers/net/avp/avp_ethdev.c | 14 +- 1 file changed, 5 insertions(+), 9 deletions(-) diff --git a/drivers/net/avp/avp_ethdev.c b/drivers/net/avp/avp_ethdev.c index dba99120f..2f3cdef1a 100644 --- a/drivers/net/avp/avp_ethdev.c +++ b/drivers/net/avp/avp_ethdev.c @@ -2074,12 +2074,6 @@ avp_dev_start(struct rte_eth_dev *eth_dev) goto unlock; } - /* disable features that we do not support */ - eth_dev->data->dev_conf.rxmode.hw_ip_checksum = 0; - eth_dev->data->dev_conf.rxmode.hw_vlan_filter = 0; - eth_dev->data->dev_conf.rxmode.hw_vlan_extend = 0; - eth_dev->data->dev_conf.rxmode.hw_strip_crc = 0; - /* update link state */ ret = avp_dev_ctrl_set_link_state(eth_dev, 1); if (ret < 0) { @@ -,10 +2216,12 @@ static int avp_vlan_offload_set(struct rte_eth_dev *eth_dev, int mask) { struct avp_dev *avp = AVP_DEV_PRIVATE_TO_HW(eth_dev->data->dev_private); + struct rte_eth_conf *dev_conf = ð_dev->data->dev_conf; + uint64_t offloads = dev_conf->rxmode.offloads; if (mask & ETH_VLAN_STRIP_MASK) { if (avp->host_features & RTE_AVP_FEATURE_VLAN_OFFLOAD) { - if (eth_dev->data->dev_conf.rxmode.hw_vlan_strip) + if (offloads & DEV_RX_OFFLOAD_VLAN_STRIP) avp->features |= RTE_AVP_FEATURE_VLAN_OFFLOAD; else avp->features &= ~RTE_AVP_FEATURE_VLAN_OFFLOAD; @@ -2235,12 +2231,12 @@ avp_vlan_offload_set(struct rte_eth_dev *eth_dev, int mask) } if (mask & ETH_VLAN_FILTER_MASK) { - if (eth_dev->data->dev_conf.rxmode.hw_vlan_filter) + if (offloads & DEV_RX_OFFLOAD_VLAN_FILTER) PMD_DRV_LOG(ERR, "VLAN filter offload not supported\n"); } if (mask & ETH_VLAN_EXTEND_MASK) { - if (eth_dev->data->dev_conf.rxmode.hw_vlan_extend) + if (offloads & DEV_RX_OFFLOAD_VLAN_EXTEND) PMD_DRV_LOG(ERR, "VLAN extend offload not supported\n"); } -- 2.12.1
Re: [dpdk-dev] [PATCH v4 1/3] ethdev: add enum type for loop on Rx/Tx offloads
Hi Daiwei: > -Original Message- > From: dev [mailto:dev-boun...@dpdk.org] On Behalf Of Wei Dai > Sent: Monday, March 19, 2018 8:41 PM > To: Lu, Wenzhuo ; Wu, Jingjing > > Cc: dev@dpdk.org; Dai, Wei > Subject: [dpdk-dev] [PATCH v4 1/3] ethdev: add enum type for loop on Rx/Tx > offloads > > This patch adds enum rte_eth_rx_offload_type and enum > rte_eth_tx_offload_type. For a loop on all Rx offloads, it is convenient to > begin with the first enum member ETH_RX_OFFLOAD_FIRST_FEATURE and to > end at ETH_RX_OFFLOAD_TOTAL_NUM. > A loop on all Tx offloads can begin with ETH_TX_OFFLOAD_FIRST_FEATURE > and end at ETH_TX_OFFLOAD_TOTAL_NUM. > > Signed-off-by: Wei Dai > --- > lib/librte_ether/rte_ethdev.h | 44 > +++ > 1 file changed, 44 insertions(+) > > diff --git a/lib/librte_ether/rte_ethdev.h b/lib/librte_ether/rte_ethdev.h > index 0361533..0089ea3 100644 > --- a/lib/librte_ether/rte_ethdev.h > +++ b/lib/librte_ether/rte_ethdev.h > @@ -946,6 +946,27 @@ struct rte_eth_conf { >DEV_RX_OFFLOAD_VLAN_FILTER | \ >DEV_RX_OFFLOAD_VLAN_EXTEND) > > +enum rte_eth_rx_offload_type { > + ETH_RX_OFFLOAD_FIRST_FEATURE = 0, > + ETH_RX_OFFLOAD_VLAN_STRIP = ETH_RX_OFFLOAD_FIRST_FEATURE, > + ETH_RX_OFFLOAD_IPV4_CKSUM, > + ETH_RX_OFFLOAD_UDP_CKSUM, > + ETH_RX_OFFLOAD_TCP_CKSUM, > + ETH_RX_OFFLOAD_TCP_LRO, > + ETH_RX_OFFLOAD_QINQ_STRIP, > + ETH_RX_OFFLOAD_OUTER_IPV4_CKSUM, > + ETH_RX_OFFLOAD_MACSEC_STRIP, > + ETH_RX_OFFLOAD_HEADER_SPLIT, > + ETH_RX_OFFLOAD_VLAN_FILTER, > + ETH_RX_OFFLOAD_VLAN_EXTEND, > + ETH_RX_OFFLOAD_JUMBO_FRAME, > + ETH_RX_OFFLOAD_CRC_STRIP, > + ETH_RX_OFFLOAD_SCATTER, > + ETH_RX_OFFLOAD_TIMESTAMP, > + ETH_RX_OFFLOAD_SECURITY, > + ETH_RX_OFFLOAD_TOTAL_NUM > +}; This looks a little duplicate with rte_rx_offload_names in rte_ethdev.c In patch 2/3 and 3/3, I think if you simply loop from offload bit 0 to bit 63 to check the name, that should still work right? then this patch could be unnecessary. Regards Qi > + > /* > * If new Rx offload capabilities are defined, they also must be > * mentioned in rte_rx_offload_names in rte_ethdev.c file. > @@ -981,6 +1002,29 @@ struct rte_eth_conf { > */ > #define DEV_TX_OFFLOAD_SECURITY 0x0002 > > +enum rte_eth_tx_offload_type { > + ETH_TX_OFFLOAD_FIRST_FEATURE = 0, > + ETH_TX_OFFLOAD_VLAN_INSERT = ETH_TX_OFFLOAD_FIRST_FEATURE, > + ETH_TX_OFFLOAD_IPV4_CKSUM, > + ETH_TX_OFFLOAD_UDP_CKSUM, > + ETH_TX_OFFLOAD_TCP_CKSUM, > + ETH_TX_OFFLOAD_SCTP_CKSUM, > + ETH_TX_OFFLOAD_TCP_TSO, > + ETH_TX_OFFLOAD_UDP_TSO, > + ETH_TX_OFFLOAD_OUTER_IPV4_CKSUM, > + ETH_TX_OFFLOAD_QINQ_INSERT, > + ETH_TX_OFFLOAD_VXLAN_TNL_TSO, > + ETH_TX_OFFLOAD_GRE_TNL_TSO, > + ETH_TX_OFFLOAD_IPIP_TNL_TSO, > + ETH_TX_OFFLOAD_GENEVE_TNL_TSO, > + ETH_TX_OFFLOAD_MACSEC_INSERT, > + ETH_TX_OFFLOAD_MT_LOCKFREE, > + ETH_TX_OFFLOAD_MULTI_SEGS, > + ETH_TX_OFFLOAD_MBUF_FAST_FREE, > + ETH_TX_OFFLOAD_SECURITY, > + ETH_TX_OFFLOAD_TOTAL_NUM > +}; > + > /* > * If new Tx offload capabilities are defined, they also must be > * mentioned in rte_tx_offload_names in rte_ethdev.c file. > -- > 2.7.5
[dpdk-dev] [PATCH] ip_frag: fix double free of chained mbufs
The first mbuf and the last mbuf to be visited in the preceeding loop are not set to NULL in the fragmentation table. This creates the possibility of a double free when the fragmentation table is later freed with rte_ip_frag_table_destroy(). Fixes: 95908f52393d ("ip_frag: free mbufs on reassembly table destroy") Signed-off-by: Allain Legacy --- lib/librte_ip_frag/rte_ipv4_reassembly.c | 4 +++- lib/librte_ip_frag/rte_ipv6_reassembly.c | 2 ++ 2 files changed, 5 insertions(+), 1 deletion(-) diff --git a/lib/librte_ip_frag/rte_ipv4_reassembly.c b/lib/librte_ip_frag/rte_ipv4_reassembly.c index 82e831ca3..42974fb8b 100644 --- a/lib/librte_ip_frag/rte_ipv4_reassembly.c +++ b/lib/librte_ip_frag/rte_ipv4_reassembly.c @@ -25,7 +25,7 @@ ipv4_frag_reassemble(struct ip_frag_pkt *fp) /*start from the last fragment. */ m = fp->frags[IP_LAST_FRAG_IDX].mb; ofs = fp->frags[IP_LAST_FRAG_IDX].ofs; - curr_idx = IP_LAST_FRAG_IDX; + curr_idx = IP_LAST_FRAG_IDX; while (ofs != first_len) { @@ -59,7 +59,9 @@ ipv4_frag_reassemble(struct ip_frag_pkt *fp) /* chain with the first fragment. */ rte_pktmbuf_adj(m, (uint16_t)(m->l2_len + m->l3_len)); rte_pktmbuf_chain(fp->frags[IP_FIRST_FRAG_IDX].mb, m); + fp->frags[curr_idx].mb = NULL; m = fp->frags[IP_FIRST_FRAG_IDX].mb; + fp->frags[IP_FIRST_FRAG_IDX].mb = NULL; /* update mbuf fields for reassembled packet. */ m->ol_flags |= PKT_TX_IP_CKSUM; diff --git a/lib/librte_ip_frag/rte_ipv6_reassembly.c b/lib/librte_ip_frag/rte_ipv6_reassembly.c index 3479fabb8..db249fe60 100644 --- a/lib/librte_ip_frag/rte_ipv6_reassembly.c +++ b/lib/librte_ip_frag/rte_ipv6_reassembly.c @@ -82,7 +82,9 @@ ipv6_frag_reassemble(struct ip_frag_pkt *fp) /* chain with the first fragment. */ rte_pktmbuf_adj(m, (uint16_t)(m->l2_len + m->l3_len)); rte_pktmbuf_chain(fp->frags[IP_FIRST_FRAG_IDX].mb, m); + fp->frags[curr_idx].mb = NULL; m = fp->frags[IP_FIRST_FRAG_IDX].mb; + fp->frags[IP_FIRST_FRAG_IDX].mb = NULL; /* update mbuf fields for reassembled packet. */ m->ol_flags |= PKT_TX_IP_CKSUM; -- 2.12.1
[dpdk-dev] [PATCH v2] ip_frag: fix double free of chained mbufs
The first mbuf and the last mbuf to be visited in the preceding loop are not set to NULL in the fragmentation table. This creates the possibility of a double free when the fragmentation table is later freed with rte_ip_frag_table_destroy(). Fixes: 95908f52393d ("ip_frag: free mbufs on reassembly table destroy") Signed-off-by: Allain Legacy --- lib/librte_ip_frag/rte_ipv4_reassembly.c | 2 ++ lib/librte_ip_frag/rte_ipv6_reassembly.c | 2 ++ 2 files changed, 4 insertions(+) diff --git a/lib/librte_ip_frag/rte_ipv4_reassembly.c b/lib/librte_ip_frag/rte_ipv4_reassembly.c index 82e831ca3..4956b99ea 100644 --- a/lib/librte_ip_frag/rte_ipv4_reassembly.c +++ b/lib/librte_ip_frag/rte_ipv4_reassembly.c @@ -59,7 +59,9 @@ ipv4_frag_reassemble(struct ip_frag_pkt *fp) /* chain with the first fragment. */ rte_pktmbuf_adj(m, (uint16_t)(m->l2_len + m->l3_len)); rte_pktmbuf_chain(fp->frags[IP_FIRST_FRAG_IDX].mb, m); + fp->frags[curr_idx].mb = NULL; m = fp->frags[IP_FIRST_FRAG_IDX].mb; + fp->frags[IP_FIRST_FRAG_IDX].mb = NULL; /* update mbuf fields for reassembled packet. */ m->ol_flags |= PKT_TX_IP_CKSUM; diff --git a/lib/librte_ip_frag/rte_ipv6_reassembly.c b/lib/librte_ip_frag/rte_ipv6_reassembly.c index 3479fabb8..db249fe60 100644 --- a/lib/librte_ip_frag/rte_ipv6_reassembly.c +++ b/lib/librte_ip_frag/rte_ipv6_reassembly.c @@ -82,7 +82,9 @@ ipv6_frag_reassemble(struct ip_frag_pkt *fp) /* chain with the first fragment. */ rte_pktmbuf_adj(m, (uint16_t)(m->l2_len + m->l3_len)); rte_pktmbuf_chain(fp->frags[IP_FIRST_FRAG_IDX].mb, m); + fp->frags[curr_idx].mb = NULL; m = fp->frags[IP_FIRST_FRAG_IDX].mb; + fp->frags[IP_FIRST_FRAG_IDX].mb = NULL; /* update mbuf fields for reassembled packet. */ m->ol_flags |= PKT_TX_IP_CKSUM; -- 2.12.1
[dpdk-dev] [PATCH] examples/l3fwd: adding event queue support
This patch set to add the support for eventdev based queue mode support to the l3fwd application. 1. Eventdev support with parallel queue 2. Eventdev support with atomic queue This patch adds - New command line parameter is added named as "dequeue-mode" which identifies dequeue method i.e. dequeue via eventdev or polling (default is polling) . If dequeue mode is via: a. eventdev: New parameters are added -e, -a, -l to cater eventdev config, adapter config and link configuration respectively. "--config" option will be invalid in this case. b. poll mode: It will work as of existing way and option for eventdev parameters(-e, -a, -l) will be invalid. - Functions are added in l3fwd_em.c and l3fwd_lpm.c for packet I/O operation The main purpose of this RFC is get comments on the approach. This is a *not tested* code. Signed-off-by: Sunil Kumar Kori --- examples/l3fwd/Makefile | 2 +- examples/l3fwd/l3fwd.h | 21 ++ examples/l3fwd/l3fwd_em.c | 100 examples/l3fwd/l3fwd_eventdev.c | 541 examples/l3fwd/l3fwd_eventdev.h | 85 +++ examples/l3fwd/l3fwd_lpm.c | 100 examples/l3fwd/main.c | 318 +++ examples/l3fwd/meson.build | 2 +- 8 files changed, 1120 insertions(+), 49 deletions(-) create mode 100644 examples/l3fwd/l3fwd_eventdev.c create mode 100644 examples/l3fwd/l3fwd_eventdev.h diff --git a/examples/l3fwd/Makefile b/examples/l3fwd/Makefile index cccdd9d..94ba537 100644 --- a/examples/l3fwd/Makefile +++ b/examples/l3fwd/Makefile @@ -5,7 +5,7 @@ APP = l3fwd # all source are stored in SRCS-y -SRCS-y := main.c l3fwd_lpm.c l3fwd_em.c +SRCS-y := main.c l3fwd_lpm.c l3fwd_em.c l3fwd_eventdev.c # Build using pkg-config variables if possible $(shell pkg-config --exists libdpdk) diff --git a/examples/l3fwd/l3fwd.h b/examples/l3fwd/l3fwd.h index c962dea..675abd1 100644 --- a/examples/l3fwd/l3fwd.h +++ b/examples/l3fwd/l3fwd.h @@ -69,6 +69,17 @@ struct lcore_conf { void *ipv6_lookup_struct; } __rte_cache_aligned; +struct l3fwd_lkp_mode { + void (*setup)(int); + int (*check_ptype)(int); + rte_rx_callback_fn cb_parse_ptype; + int (*main_loop)(void *); + void* (*get_ipv4_lookup_struct)(int); + void* (*get_ipv6_lookup_struct)(int); +}; + +extern struct l3fwd_lkp_mode l3fwd_lkp; + extern volatile bool force_quit; /* ethernet addresses of ports */ @@ -81,11 +92,16 @@ extern uint32_t enabled_port_mask; /* Used only in exact match mode. */ extern int ipv6; /**< ipv6 is false by default. */ extern uint32_t hash_entry_number; +extern int promiscuous_on; +extern int numa_on; /**< NUMA is enabled by default. */ +extern uint16_t nb_rxd; extern xmm_t val_eth[RTE_MAX_ETHPORTS]; extern struct lcore_conf lcore_conf[RTE_MAX_LCORE]; +extern struct rte_mempool *pktmbuf_pool[]; + /* Send burst of packets on an output interface */ static inline int send_burst(struct lcore_conf *qconf, uint16_t n, uint16_t port) @@ -208,5 +224,10 @@ lpm_get_ipv4_l3fwd_lookup_struct(const int socketid); void * lpm_get_ipv6_l3fwd_lookup_struct(const int socketid); +int +prepare_ptype_parser(uint16_t portid, uint16_t queueid); + +void +check_all_ports_link_status(uint16_t port_num, uint32_t port_mask); #endif /* __L3_FWD_H__ */ diff --git a/examples/l3fwd/l3fwd_em.c b/examples/l3fwd/l3fwd_em.c index 9dc3b8c..199ff75 100644 --- a/examples/l3fwd/l3fwd_em.c +++ b/examples/l3fwd/l3fwd_em.c @@ -21,12 +21,14 @@ #include #include #include +#include #include #include #include #include #include "l3fwd.h" +#include "l3fwd_eventdev.h" #if defined(RTE_ARCH_X86) || defined(RTE_MACHINE_CPUFLAG_CRC32) #define EM_HASH_CRC 1 @@ -612,6 +614,101 @@ em_cb_parse_ptype(uint16_t port __rte_unused, uint16_t queue __rte_unused, return nb_pkts; } +/* main EM processing loop for eventdev*/ +int +em_eventdev_main_loop(__attribute__((unused)) void *dummy) +{ + struct rte_event ev[MAX_PKT_BURST]; + struct rte_mbuf *pkts_burst[MAX_PKT_BURST]; + struct rte_event_port_conf event_port_conf; + unsigned int lcore_id; + uint64_t prev_tsc, diff_tsc, cur_tsc; + int i, nb_rx; + uint8_t queueid; + uint16_t portid, dequeue_len; + uint8_t event_port_id = INVALID_EVENDEV_ID; + struct lcore_conf *qconf; + const uint64_t drain_tsc = (rte_get_tsc_hz() + US_PER_S - 1) / + US_PER_S * BURST_TX_DRAIN_US; + + prev_tsc = 0; + + lcore_id = rte_lcore_id(); + qconf = &lcore_conf[lcore_id]; + + if (qconf->n_rx_queue == 0) { + RTE_LOG(INFO, L3FWD, "lcore %u has nothing to do\n", lcore_id); + return 0; + } + + RTE_LOG(INFO, L3FWD, "entering main loop on lcore %u\n", lcore_id); + + for (i = 0; i < qconf->n_rx_queue; i++) { + + portid = qconf->rx_qu
[dpdk-dev] [PATCH v2 0/3] net/mlx5: use Netlink in VF mode
When MLX5 behaves in VF mode and the hypervisor have **trusted** this VF, to be able to receive specific traffic some requests must be done to configure the NIC. There is no API currently available to do it though Verbs, but there is in Linux side using Netlink. The specific cases are: - Enable/disable promiscuous mode. - Enable/disable allmulti mode. - Add/remove mac addresses. Changes in v2: Embed Netlink socket communication inside the PMD. Nelio Laranjeiro (3): net/mlx5: use Netlink to add/remove MAC addresses net/mlx5: use Netlink to enable promisc / all multicast mode net/mlx5: add a parameter for Netlink support in VF doc/guides/nics/mlx5.rst | 11 + drivers/net/mlx5/Makefile | 1 + drivers/net/mlx5/mlx5.c| 29 ++ drivers/net/mlx5/mlx5.h| 14 + drivers/net/mlx5/mlx5_ethdev.c | 27 ++ drivers/net/mlx5/mlx5_mac.c| 25 +- drivers/net/mlx5/mlx5_nl.c | 623 + drivers/net/mlx5/mlx5_rxmode.c | 8 + 8 files changed, 736 insertions(+), 2 deletions(-) create mode 100644 drivers/net/mlx5/mlx5_nl.c -- 2.11.0
[dpdk-dev] [PATCH v2 1/3] net/mlx5: use Netlink to add/remove MAC addresses
VF devices are not able to receive traffic unless it fully requests it though Netlink. This will cause the request to be processed by the PF which will add/remove the MAC address to the VF table if the VF is trusted. Signed-off-by: Nelio Laranjeiro Acked-by: Adrien Mazarguil --- drivers/net/mlx5/Makefile | 1 + drivers/net/mlx5/mlx5.c| 22 ++ drivers/net/mlx5/mlx5.h| 13 + drivers/net/mlx5/mlx5_ethdev.c | 27 +++ drivers/net/mlx5/mlx5_mac.c| 25 +- drivers/net/mlx5/mlx5_nl.c | 529 + 6 files changed, 615 insertions(+), 2 deletions(-) create mode 100644 drivers/net/mlx5/mlx5_nl.c diff --git a/drivers/net/mlx5/Makefile b/drivers/net/mlx5/Makefile index afda4118f..5cca2ba53 100644 --- a/drivers/net/mlx5/Makefile +++ b/drivers/net/mlx5/Makefile @@ -59,6 +59,7 @@ SRCS-$(CONFIG_RTE_LIBRTE_MLX5_PMD) += mlx5_rss.c SRCS-$(CONFIG_RTE_LIBRTE_MLX5_PMD) += mlx5_mr.c SRCS-$(CONFIG_RTE_LIBRTE_MLX5_PMD) += mlx5_flow.c SRCS-$(CONFIG_RTE_LIBRTE_MLX5_PMD) += mlx5_socket.c +SRCS-$(CONFIG_RTE_LIBRTE_MLX5_PMD) += mlx5_nl.c ifeq ($(CONFIG_RTE_LIBRTE_MLX5_DLOPEN_DEPS),y) INSTALL-$(CONFIG_RTE_LIBRTE_MLX5_PMD)-lib += $(LIB_GLUE) diff --git a/drivers/net/mlx5/mlx5.c b/drivers/net/mlx5/mlx5.c index 95dd49420..5ec152ab1 100644 --- a/drivers/net/mlx5/mlx5.c +++ b/drivers/net/mlx5/mlx5.c @@ -13,6 +13,7 @@ #include #include #include +#include /* Verbs header. */ /* ISO C doesn't support unnamed structs/unions, disabling -pedantic. */ @@ -205,6 +206,10 @@ mlx5_dev_close(struct rte_eth_dev *dev) rte_free(priv->reta_idx); if (priv->primary_socket) mlx5_socket_uninit(dev); + if (priv->config.vf) + mlx5_nl_mac_addr_flush(dev); + if (priv->nl_socket >= 0) + close(priv->nl_socket); ret = mlx5_hrxq_ibv_verify(dev); if (ret) DRV_LOG(WARNING, "port %u some hash Rx queue still remain", @@ -597,6 +602,7 @@ mlx5_pci_probe(struct rte_pci_driver *pci_drv __rte_unused, int err = 0; struct ibv_context *attr_ctx = NULL; struct ibv_device_attr_ex device_attr; + unsigned int vf; unsigned int mps; unsigned int cqe_comp; unsigned int tunnel_en = 0; @@ -646,6 +652,14 @@ mlx5_pci_probe(struct rte_pci_driver *pci_drv __rte_unused, continue; DRV_LOG(INFO, "PCI information matches, using device \"%s\"", list[i]->name); + vf = ((pci_dev->id.device_id == + PCI_DEVICE_ID_MELLANOX_CONNECTX4VF) || + (pci_dev->id.device_id == + PCI_DEVICE_ID_MELLANOX_CONNECTX4LXVF) || + (pci_dev->id.device_id == + PCI_DEVICE_ID_MELLANOX_CONNECTX5VF) || + (pci_dev->id.device_id == + PCI_DEVICE_ID_MELLANOX_CONNECTX5EXVF)); attr_ctx = mlx5_glue->open_device(list[i]); rte_errno = errno; err = rte_errno; @@ -871,6 +885,7 @@ mlx5_pci_probe(struct rte_pci_driver *pci_drv __rte_unused, DRV_LOG(DEBUG, "hardware Rx end alignment padding is %ssupported", (config.hw_padding ? "" : "not ")); + config.vf = vf; config.tso = ((device_attr_ex.tso_caps.max_tso > 0) && (device_attr_ex.tso_caps.supported_qpts & (1 << IBV_QPT_RAW_PACKET))); @@ -947,6 +962,13 @@ mlx5_pci_probe(struct rte_pci_driver *pci_drv __rte_unused, eth_dev->tx_pkt_burst = removed_tx_burst; eth_dev->dev_ops = &mlx5_dev_ops; /* Register MAC address. */ + priv->nl_socket = -1; + priv->nl_sn = 0; + if (vf) { + priv->nl_socket = mlx5_nl_init(RTMGRP_LINK); + if (priv->nl_socket < 0) + priv->nl_socket = -1; + } claim_zero(mlx5_mac_addr_add(eth_dev, &mac, 0, 0)); TAILQ_INIT(&priv->flows); TAILQ_INIT(&priv->ctrl_flows); diff --git a/drivers/net/mlx5/mlx5.h b/drivers/net/mlx5/mlx5.h index faacfd9d6..6a7e9f310 100644 --- a/drivers/net/mlx5/mlx5.h +++ b/drivers/net/mlx5/mlx5.h @@ -78,6 +78,7 @@ struct mlx5_dev_config { unsigned int hw_vlan_strip:1; /* VLAN stripping is supported. */ unsigned int hw_fcs_strip:1; /* FCS stripping is supported. */ unsigned int hw_padding:1; /* End alignment padding is supported. */ + unsigned int vf:1; /* This is a VF. */ unsigned int mps:2; /* Multi-packet send supported mode. */ unsigned int tunnel_en:1; /* Whether tunnel stateless offloads are supported. */ @@ -154,6 +155,8 @@ struct priv { struct mlx5_dev_config config; /* Device configuration. */
[dpdk-dev] [PATCH v2 2/3] net/mlx5: use Netlink to enable promisc / all multicast mode
VF devices are not able to receive promisc or allmulti traffic unless it fully requests it though Netlink. This will cause the request to be processed by the PF which will handle the request and enable it. This requires the VF to be trusted by the PF. Signed-off-by: Nelio Laranjeiro Acked-by: Adrien Mazarguil --- drivers/net/mlx5/mlx5_nl.c | 94 ++ drivers/net/mlx5/mlx5_rxmode.c | 8 2 files changed, 102 insertions(+) diff --git a/drivers/net/mlx5/mlx5_nl.c b/drivers/net/mlx5/mlx5_nl.c index 21edefc0f..562d219fa 100644 --- a/drivers/net/mlx5/mlx5_nl.c +++ b/drivers/net/mlx5/mlx5_nl.c @@ -527,3 +527,97 @@ mlx5_nl_mac_addr_flush(struct rte_eth_dev *dev) mlx5_nl_mac_addr_remove(dev, m); } } + +/** + * Enable promiscuous / all multicast mode through Netlink. + * + * @param dev + * Pointer to Ethernet device structure. + * @param flags + * IFF_PROMISC for promiscuous, IFF_ALLMULTI for allmulti. + * @param enable + * Nonzero to enable, disable otherwise. + * + * @return + * 0 on success, a negative errno value otherwise and rte_errno is set. + */ +static int +mlx5_nl_device_flags(struct rte_eth_dev *dev, uint32_t flags, int enable) +{ + struct priv *priv = dev->data->dev_private; + int iface_idx = mlx5_ifindex(dev); + struct { + struct nlmsghdr hdr; + struct ifinfomsg ifi; + } req = { + .hdr = { + .nlmsg_len = NLMSG_LENGTH(sizeof(struct ifinfomsg)), + .nlmsg_type = RTM_NEWLINK, + .nlmsg_flags = NLM_F_REQUEST, + }, + .ifi = { + .ifi_flags = enable ? flags : 0, + .ifi_change = flags, + .ifi_index = iface_idx, + }, + }; + int fd; + int ret; + + assert(!(flags & ~(IFF_PROMISC | IFF_ALLMULTI))); + if (priv->nl_socket < 0) + return 0; + fd = priv->nl_socket; + ret = mlx5_nl_send(fd, &req.hdr, priv->nl_sn++); + if (ret < 0) + return ret; + return 0; +} + +/** + * Enable promiscuous mode through Netlink. + * + * @param dev + * Pointer to Ethernet device structure. + * @param enable + * Nonzero to enable, disable otherwise. + * + * @return + * 0 on success, a negative errno value otherwise and rte_errno is set. + */ +int +mlx5_nl_promisc(struct rte_eth_dev *dev, int enable) +{ + int ret = mlx5_nl_device_flags(dev, IFF_PROMISC, enable); + + if (ret) + DRV_LOG(DEBUG, + "port %u cannot %s promisc mode: Netlink error %s", + dev->data->port_id, enable ? "enable" : "disable", + strerror(rte_errno)); + return ret; +} + +/** + * Enable all multicast mode through Netlink. + * + * @param dev + * Pointer to Ethernet device structure. + * @param enable + * Nonzero to enable, disable otherwise. + * + * @return + * 0 on success, a negative errno value otherwise and rte_errno is set. + */ +int +mlx5_nl_allmulti(struct rte_eth_dev *dev, int enable) +{ + int ret = mlx5_nl_device_flags(dev, IFF_ALLMULTI, enable); + + if (ret) + DRV_LOG(DEBUG, + "port %u cannot %s allmulti mode: Netlink error %s", + dev->data->port_id, enable ? "enable" : "disable", + strerror(rte_errno)); + return ret; +} diff --git a/drivers/net/mlx5/mlx5_rxmode.c b/drivers/net/mlx5/mlx5_rxmode.c index e43a4b030..c1c0f21c7 100644 --- a/drivers/net/mlx5/mlx5_rxmode.c +++ b/drivers/net/mlx5/mlx5_rxmode.c @@ -35,6 +35,8 @@ mlx5_promiscuous_enable(struct rte_eth_dev *dev) int ret; dev->data->promiscuous = 1; + if (((struct priv *)dev->data->dev_private)->config.vf) + mlx5_nl_promisc(dev, 1); ret = mlx5_traffic_restart(dev); if (ret) DRV_LOG(ERR, "port %u cannot enable promiscuous mode: %s", @@ -53,6 +55,8 @@ mlx5_promiscuous_disable(struct rte_eth_dev *dev) int ret; dev->data->promiscuous = 0; + if (((struct priv *)dev->data->dev_private)->config.vf) + mlx5_nl_promisc(dev, 0); ret = mlx5_traffic_restart(dev); if (ret) DRV_LOG(ERR, "port %u cannot disable promiscuous mode: %s", @@ -71,6 +75,8 @@ mlx5_allmulticast_enable(struct rte_eth_dev *dev) int ret; dev->data->all_multicast = 1; + if (((struct priv *)dev->data->dev_private)->config.vf) + mlx5_nl_allmulti(dev, 1); ret = mlx5_traffic_restart(dev); if (ret) DRV_LOG(ERR, "port %u cannot enable allmulicast mode: %s", @@ -89,6 +95,8 @@ mlx5_allmulticast_disable(struct rte_eth_dev *dev) int ret; dev->data->all_multicast = 0; + if (((struct priv *)dev->data->dev_private)->config.vf) +
[dpdk-dev] [PATCH v2 3/3] net/mlx5: add a parameter for Netlink support in VF
All Netlink request the PMD will do can also be done by a iproute2 command line interface, letting the operator configure the VF behavior without having to modify the application nor reaching PMD limits (e.g. MAC address number limit). Signed-off-by: Nelio Laranjeiro Acked-by: Adrien Mazarguil --- doc/guides/nics/mlx5.rst | 11 +++ drivers/net/mlx5/mlx5.c | 9 - drivers/net/mlx5/mlx5.h | 1 + 3 files changed, 20 insertions(+), 1 deletion(-) diff --git a/doc/guides/nics/mlx5.rst b/doc/guides/nics/mlx5.rst index 46d26e4c8..4c7bf1c56 100644 --- a/doc/guides/nics/mlx5.rst +++ b/doc/guides/nics/mlx5.rst @@ -135,6 +135,8 @@ Limitations - Flows with a VXLAN Network Identifier equal (or ends to be equal) to 0 are not supported. - VXLAN TSO and checksum offloads are not supported on VM. +- VF: flow rules created on VF devices can only match traffic targeted at the + configured MAC addresses (see ``rte_eth_dev_mac_addr_add()``). Statistics -- @@ -335,6 +337,15 @@ Run-time configuration Enabled by default. +- ``vf_nl_en`` parameter [int] + + A nonzero value enables Netlink requests from the VF to add/remove MAC + addresses or/and enable/disable promiscuous/all multicast on the Netdevice. + Otherwise the relevant configuration must be run with Linux iproute2 tools. + This is a prerequisite to receive this kind of traffic. + + Enabled by default, valid only on VF devices ignored otherwise. + Prerequisites - diff --git a/drivers/net/mlx5/mlx5.c b/drivers/net/mlx5/mlx5.c index 5ec152ab1..89bada409 100644 --- a/drivers/net/mlx5/mlx5.c +++ b/drivers/net/mlx5/mlx5.c @@ -69,6 +69,9 @@ /* Device parameter to enable hardware Rx vector. */ #define MLX5_RX_VEC_EN "rx_vec_en" +/* Activate Netlink support in VF mode. */ +#define MLX5_VF_NL_EN "vf_nl_en" + #ifndef HAVE_IBV_MLX5_MOD_MPW #define MLX5DV_CONTEXT_FLAGS_MPW_ALLOWED (1 << 2) #define MLX5DV_CONTEXT_FLAGS_ENHANCED_MPW (1 << 3) @@ -412,6 +415,8 @@ mlx5_args_check(const char *key, const char *val, void *opaque) config->tx_vec_en = !!tmp; } else if (strcmp(MLX5_RX_VEC_EN, key) == 0) { config->rx_vec_en = !!tmp; + } else if (strcmp(MLX5_VF_NL_EN, key) == 0) { + config->vf_nl_en = !!tmp; } else { DRV_LOG(WARNING, "%s: unknown parameter", key); rte_errno = EINVAL; @@ -443,6 +448,7 @@ mlx5_args(struct mlx5_dev_config *config, struct rte_devargs *devargs) MLX5_TXQ_MAX_INLINE_LEN, MLX5_TX_VEC_EN, MLX5_RX_VEC_EN, + MLX5_VF_NL_EN, NULL, }; struct rte_kvargs *kvlist; @@ -748,6 +754,7 @@ mlx5_pci_probe(struct rte_pci_driver *pci_drv __rte_unused, .txq_inline = MLX5_ARG_UNSET, .txqs_inline = MLX5_ARG_UNSET, .inline_max_packet_sz = MLX5_ARG_UNSET, + .vf_nl_en = 1, }; len = snprintf(name, sizeof(name), PCI_PRI_FMT, @@ -964,7 +971,7 @@ mlx5_pci_probe(struct rte_pci_driver *pci_drv __rte_unused, /* Register MAC address. */ priv->nl_socket = -1; priv->nl_sn = 0; - if (vf) { + if (vf && config.vf_nl_en) { priv->nl_socket = mlx5_nl_init(RTMGRP_LINK); if (priv->nl_socket < 0) priv->nl_socket = -1; diff --git a/drivers/net/mlx5/mlx5.h b/drivers/net/mlx5/mlx5.h index 6a7e9f310..7cabf43cc 100644 --- a/drivers/net/mlx5/mlx5.h +++ b/drivers/net/mlx5/mlx5.h @@ -88,6 +88,7 @@ struct mlx5_dev_config { unsigned int tx_vec_en:1; /* Tx vector is enabled. */ unsigned int rx_vec_en:1; /* Rx vector is enabled. */ unsigned int mpw_hdr_dseg:1; /* Enable DSEGs in the title WQEBB. */ + unsigned int vf_nl_en:1; /* Enable Netlink requests in VF mode. */ unsigned int tso_max_payload_sz; /* Maximum TCP payload for TSO. */ unsigned int ind_table_max_size; /* Maximum indirection table size. */ int txq_inline; /* Maximum packet size for inlining. */ -- 2.11.0
[dpdk-dev] rte_eth_dev_count() returns 0 in shared library mode.
Hi Everyone, Need urgent help over here. After cloning the latest git, when i compiled the dpdk in shared library mode by modifying the common_base config file and rebuild the examples the api rte_eth_dev_count() returns 0 However in the static library mode, rte_eth_dev_count() returns 2. Can someone please let me know what iam missing over here. Here are details -- Code -- /* Check that there is an even number of ports to send/receive on. */ nb_ports = rte_eth_dev_count(); if (nb_ports < 2 || (nb_ports & 1)) rte_exit(EXIT_FAILURE, "Error: number of ports must be even\n"); #make config T=x86_64-native-linuxapp-gcc install examples --- Libraries built by changing the SHARED_LIB. config/common_base +++ b/config/common_base @@ -38,7 +38,7 @@ CONFIG_RTE_ARCH_STRICT_ALIGN=n # # Compile to share library # -CONFIG_RTE_BUILD_SHARED_LIB=n +CONFIG_RTE_BUILD_SHARED_LIB=y -- [root@localhost dpdk_new]# ./examples/skeleton/x86_64-native-linuxapp-gcc/basicfwd EAL: Detected 72 lcore(s) EAL: Multi-process socket /var/run/.rte_unix EAL: Probing VFIO support... EAL: Error - exiting with code: 1 Cause: Error: number of ports must be even -- [root@localhost dpdk_new]# ldd ./examples/skeleton/x86_64-native-linuxapp-gcc/basicfwd linux-vdso.so.1 => (0x7fffa858a000) librte_flow_classify.so.1.1 => /root/DPDK/dpdk_new/x86_64-native-linuxapp-gcc/lib/librte_flow_classify.so.1.1 (0x7f7272f5f000) librte_pipeline.so.3.1 => /root/DPDK/dpdk_new/x86_64-native-linuxapp-gcc/lib/librte_pipeline.so.3.1 (0x7f7272d58000) librte_table.so.3.1 => /root/DPDK/dpdk_new/x86_64-native-linuxapp-gcc/lib/librte_table.so.3.1 (0x7f7272b3e000) librte_port.so.3.1 => /root/DPDK/dpdk_new/x86_64-native-linuxapp-gcc/lib/librte_port.so.3.1 (0x7f727291d000) librte_pdump.so.2.1 => /root/DPDK/dpdk_new/x86_64-native-linuxapp-gcc/lib/librte_pdump.so.2.1 (0x7f7272513000) librte_distributor.so.1.1 => /root/DPDK/dpdk_new/x86_64-native-linuxapp-gcc/lib/librte_distributor.so.1.1 (0x7f727230d000) librte_ip_frag.so.1.1 => /root/DPDK/dpdk_new/x86_64-native-linuxapp-gcc/lib/librte_ip_frag.so.1.1 (0x7f7272103000) librte_gro.so.1.1 => /root/DPDK/dpdk_new/x86_64-native-linuxapp-gcc/lib/librte_gro.so.1.1 (0x7f7271efc000) librte_gso.so.1.1 => /root/DPDK/dpdk_new/x86_64-native-linuxapp-gcc/lib/librte_gso.so.1.1 (0x7f7271cf7000) --snip-- [root@localhost dpdk_new]# ./usertools/dpdk-devbind.py --status Network devices using DPDK-compatible driver :04:00.0 'I350 Gigabit Network Connection 1521' drv=igb_uio unused= :04:00.1 'I350 Gigabit Network Connection 1521' drv=igb_uio unused= Static Library Mode - [root@localhost dpdk_new]# ./examples/skeleton/x86_64-native-linuxapp-gcc/basicfwd EAL: Detected 72 lcore(s) EAL: Multi-process socket /var/run/.rte_unix EAL: Probing VFIO support... EAL: PCI device :04:00.0 on NUMA socket 0 EAL: probe driver: 8086:1521 net_e1000_igb EAL: PCI device :04:00.1 on NUMA socket 0 EAL: probe driver: 8086:1521 net_e1000_igb EAL: PCI device :04:00.2 on NUMA socket 0 EAL: probe driver: 8086:1521 net_e1000_igb EAL: PCI device :04:00.3 on NUMA socket 0 EAL: probe driver: 8086:1521 net_e1000_igb EAL: PCI device :05:00.0 on NUMA socket 0 EAL: probe driver: 8086:1572 net_i40e EAL: PCI device :05:00.1 on NUMA socket 0 EAL: probe driver: 8086:1572 net_i40e Port 0 MAC: 14 02 ec 6b 36 10 Port 1 MAC: 14 02 ec 6b 36 11 --- Git Log --- commit 09e1e8d256b0832a64af6d13bf96fcb49e1e7ded Author: Hemant Agrawal Date: Fri Feb 23 15:28:03 2018 +0530 mk: fix dependencies of dpaaX drivers This patch fixes the build dependency of various dpaaX components, when the dpaa or fslmc bus is disabled, or VFIO is disabled. Regards, Venky
[dpdk-dev] [PATCH] net/mlx5: setup RSS regardless of queue count
From: Dahir Osman In some environments it is desirable to have the NIC perform RSS normally on the packet regardless of the number of queues configured. The RSS hash result that is stored in the mbuf can then be used by the application to make decisions about how to distribute workloads to threads, secondary processes, or even virtual machines if the application is a virtual switch. This change to the mlx5 driver aligns with how other drivers in the Intel family work. Signed-off-by: Allain Legacy --- drivers/net/mlx5/mlx5_rxq.c | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/drivers/net/mlx5/mlx5_rxq.c b/drivers/net/mlx5/mlx5_rxq.c index ff58c4921..e6b05b0ad 100644 --- a/drivers/net/mlx5/mlx5_rxq.c +++ b/drivers/net/mlx5/mlx5_rxq.c @@ -1028,7 +1028,8 @@ mlx5_priv_rxq_new(struct priv *priv, uint16_t idx, uint16_t desc, tmpl->rxq.crc_present ? "disabled" : "enabled", tmpl->rxq.crc_present << 2); /* Save port ID. */ - tmpl->rxq.rss_hash = priv->rxqs_n > 1; + tmpl->rxq.rss_hash = (!!(dev->data->dev_conf.rxmode.mq_mode & +ETH_MQ_RX_RSS)); tmpl->rxq.port_id = dev->data->port_id; tmpl->priv = priv; tmpl->rxq.mp = mp; -- 2.12.1
[dpdk-dev] [PATCH v2] net/mlx4: support CRC strip toggling
Previous to this commit mlx4 CRC stripping was executed by default and there was no verbs API to disable it. Current support in MLNX_OFED_4.3-1.5.0.0 and above Signed-off-by: Ophir Munk --- v1: initial version v2: following internal reviews drivers/net/mlx4/mlx4.c | 4 drivers/net/mlx4/mlx4.h | 1 + drivers/net/mlx4/mlx4_rxq.c | 33 +++-- drivers/net/mlx4/mlx4_rxtx.c | 5 - drivers/net/mlx4/mlx4_rxtx.h | 1 + 5 files changed, 41 insertions(+), 3 deletions(-) diff --git a/drivers/net/mlx4/mlx4.c b/drivers/net/mlx4/mlx4.c index ee93daf..d4f4323 100644 --- a/drivers/net/mlx4/mlx4.c +++ b/drivers/net/mlx4/mlx4.c @@ -578,6 +578,10 @@ mlx4_pci_probe(struct rte_pci_driver *pci_drv, struct rte_pci_device *pci_dev) } DEBUG("supported RSS hash fields mask: %016" PRIx64, priv->hw_rss_sup); + priv->hw_fcs_strip = !!(device_attr_ex.raw_packet_caps & +IBV_RAW_PACKET_CAP_SCATTER_FCS); + DEBUG("FCS stripping toggling is %ssupported", + (priv->hw_fcs_strip ? "" : "not ")); /* Configure the first MAC address by default. */ if (mlx4_get_mac(priv, &mac.addr_bytes)) { ERROR("cannot get MAC address, is mlx4_en loaded?" diff --git a/drivers/net/mlx4/mlx4.h b/drivers/net/mlx4/mlx4.h index 19c8a22..3ae3ce6 100644 --- a/drivers/net/mlx4/mlx4.h +++ b/drivers/net/mlx4/mlx4.h @@ -105,6 +105,7 @@ struct priv { uint32_t isolated:1; /**< Toggle isolated mode. */ uint32_t hw_csum:1; /**< Checksum offload is supported. */ uint32_t hw_csum_l2tun:1; /**< Checksum support for L2 tunnels. */ + uint32_t hw_fcs_strip:1; /**< FCS stripping toggling is supported. */ uint64_t hw_rss_sup; /**< Supported RSS hash fields (Verbs format). */ struct rte_intr_handle intr_handle; /**< Port interrupt handle. */ struct mlx4_drop *drop; /**< Shared resources for drop flow rules. */ diff --git a/drivers/net/mlx4/mlx4_rxq.c b/drivers/net/mlx4/mlx4_rxq.c index 7a036ed..6748355 100644 --- a/drivers/net/mlx4/mlx4_rxq.c +++ b/drivers/net/mlx4/mlx4_rxq.c @@ -491,6 +491,8 @@ mlx4_rxq_attach(struct rxq *rxq) const char *msg; struct ibv_cq *cq = NULL; struct ibv_wq *wq = NULL; + unsigned int create_flags = 0; + unsigned int comp_mask = 0; volatile struct mlx4_wqe_data_seg (*wqes)[]; unsigned int i; int ret; @@ -503,6 +505,11 @@ mlx4_rxq_attach(struct rxq *rxq) msg = "CQ creation failure"; goto error; } + /* By default, FCS (CRC) is stripped by hardware. */ + if (rxq->crc_present) { + create_flags |= IBV_WQ_FLAGS_SCATTER_FCS; + comp_mask |= IBV_WQ_INIT_ATTR_FLAGS; + } wq = mlx4_glue->create_wq (priv->ctx, &(struct ibv_wq_init_attr){ @@ -511,6 +518,8 @@ mlx4_rxq_attach(struct rxq *rxq) .max_sge = sges_n, .pd = priv->pd, .cq = cq, + .comp_mask = comp_mask, + .create_flags = create_flags, }); if (!wq) { ret = errno ? errno : EINVAL; @@ -649,9 +658,10 @@ mlx4_rxq_detach(struct rxq *rxq) uint64_t mlx4_get_rx_queue_offloads(struct priv *priv) { - uint64_t offloads = DEV_RX_OFFLOAD_SCATTER | - DEV_RX_OFFLOAD_CRC_STRIP; + uint64_t offloads = DEV_RX_OFFLOAD_SCATTER; + if (priv->hw_fcs_strip) + offloads |= DEV_RX_OFFLOAD_CRC_STRIP; if (priv->hw_csum) offloads |= DEV_RX_OFFLOAD_CHECKSUM; return offloads; @@ -781,6 +791,24 @@ mlx4_rx_queue_setup(struct rte_eth_dev *dev, uint16_t idx, uint16_t desc, (void *)dev, idx); return -rte_errno; } + /* By default, FCS (CRC) is stripped by hardware. */ + unsigned char crc_present; + if (conf->offloads & DEV_RX_OFFLOAD_CRC_STRIP) { + crc_present = 0; + } else if (priv->hw_fcs_strip) { + crc_present = 1; + } else { + WARN("%p: CRC stripping has been disabled but will still" +" be performed by hardware, make sure MLNX_OFED and" +" firmware are up to date", +(void *)dev); + crc_present = 0; + } + DEBUG("%p: CRC stripping is %s, %u bytes will be subtracted from" + " incoming frames to hide it", + (void *)dev, + crc_present ? "disabled" : "enabled", + crc_present << 2); *rxq = (struct rxq){ .priv = priv, .mp = mp, @@ -794,6 +822,7 @@ mlx4_rx_queue_setup(struct rte_eth_dev *dev, uint16_t idx, uint16_t desc, .c
Re: [dpdk-dev] [PATCH v1 0/9] mempool: prepare to add bucket driver
Hi Andrew, Thank you for this nice rework. Globally, the patchset looks good to me. I'm sending some comments as reply to specific patches. On Sat, Mar 10, 2018 at 03:39:33PM +, Andrew Rybchenko wrote: > The initial patch series [1] is split into two to simplify processing. > The second series relies on this one and will add bucket mempool driver > and related ops. > > The patch series has generic enhancements suggested by Olivier. > Basically it adds driver callbacks to calculate required memory size and > to populate objects using provided memory area. It allows to remove > so-called capability flags used before to tell generic code how to > allocate and slice allocated memory into mempool objects. > Clean up which removes get_capabilities and register_memory_area is > not strictly required, but I think right thing to do. > Existing mempool drivers are updated. > > I've kept rte_mempool_populate_iova_tab() intact since it seems to > be not directly related XMEM API functions. The function rte_mempool_populate_iova_tab() (actually, it was rte_mempool_populate_phys_tab()) was introduced to support XMEM API. In my opinion, it can also be deprecated. > It breaks ABI since changes rte_mempool_ops. Also it removes > rte_mempool_ops_register_memory_area() and > rte_mempool_ops_get_capabilities() since corresponding callbacks are > removed. > > Internal global functions are not listed in map file since it is not > a part of external API. > > [1] http://dpdk.org/ml/archives/dev/2018-January/088698.html > > RFCv1 -> RFCv2: > - add driver ops to calculate required memory size and populate > mempool objects, remove extra flags which were required before > to control it > - transition of octeontx and dpaa drivers to the new callbacks > - change info API to get information from driver required to > API user to know contiguous block size > - remove get_capabilities (not required any more and may be > substituted with more in info get API) > - remove register_memory_area since it is substituted with > populate callback which can do more > - use SPDX tags > - avoid all objects affinity to single lcore > - fix bucket get_count > - deprecate XMEM API > - avoid introduction of a new function to flush cache > - fix NO_CACHE_ALIGN case in bucket mempool > > RFCv2 -> v1: > - split the series in two > - squash octeontx patches which implement calc_mem_size and populate > callbacks into the patch which removes get_capabilities since it is > the easiest way to untangle the tangle of tightly related library > functions and flags advertised by the driver > - consistently name default callbacks > - move default callbacks to dedicated file > - see detailed description in patches > > Andrew Rybchenko (7): > mempool: add op to calculate memory size to be allocated > mempool: add op to populate objects using provided memory > mempool: remove callback to get capabilities > mempool: deprecate xmem functions > mempool/octeontx: prepare to remove register memory area op > mempool/dpaa: prepare to remove register memory area op > mempool: remove callback to register memory area > > Artem V. Andreev (2): > mempool: ensure the mempool is initialized before populating > mempool: support flushing the default cache of the mempool > > doc/guides/rel_notes/deprecation.rst| 12 +- > doc/guides/rel_notes/release_18_05.rst | 32 ++- > drivers/mempool/dpaa/dpaa_mempool.c | 13 +- > drivers/mempool/octeontx/rte_mempool_octeontx.c | 64 -- > lib/librte_mempool/Makefile | 3 +- > lib/librte_mempool/meson.build | 5 +- > lib/librte_mempool/rte_mempool.c| 159 +++ > lib/librte_mempool/rte_mempool.h| 260 > +--- > lib/librte_mempool/rte_mempool_ops.c| 37 ++-- > lib/librte_mempool/rte_mempool_ops_default.c| 51 + > lib/librte_mempool/rte_mempool_version.map | 11 +- > test/test/test_mempool.c| 31 --- > 12 files changed, 437 insertions(+), 241 deletions(-) > create mode 100644 lib/librte_mempool/rte_mempool_ops_default.c > > -- > 2.7.4 >
Re: [dpdk-dev] [PATCH v1 1/9] mempool: add op to calculate memory size to be allocated
On Sat, Mar 10, 2018 at 03:39:34PM +, Andrew Rybchenko wrote: > Size of memory chunk required to populate mempool objects depends > on how objects are stored in the memory. Different mempool drivers > may have different requirements and a new operation allows to > calculate memory size in accordance with driver requirements and > advertise requirements on minimum memory chunk size and alignment > in a generic way. > > Bump ABI version since the patch breaks it. > > Signed-off-by: Andrew Rybchenko Looks good to me. Just see below for few minor comments. > --- > RFCv2 -> v1: > - move default calc_mem_size callback to rte_mempool_ops_default.c > - add ABI changes to release notes > - name default callback consistently: rte_mempool_op__default() > - bump ABI version since it is the first patch which breaks ABI > - describe default callback behaviour in details > - avoid introduction of internal function to cope with depration typo (depration) >(keep it to deprecation patch) > - move cache-line or page boundary chunk alignment to default callback > - highlight that min_chunk_size and align parameters are output only [...] > --- a/lib/librte_mempool/Makefile > +++ b/lib/librte_mempool/Makefile > @@ -11,11 +11,12 @@ LDLIBS += -lrte_eal -lrte_ring > > EXPORT_MAP := rte_mempool_version.map > > -LIBABIVER := 3 > +LIBABIVER := 4 > > # all source are stored in SRCS-y > SRCS-$(CONFIG_RTE_LIBRTE_MEMPOOL) += rte_mempool.c > SRCS-$(CONFIG_RTE_LIBRTE_MEMPOOL) += rte_mempool_ops.c > +SRCS-$(CONFIG_RTE_LIBRTE_MEMPOOL) += rte_mempool_ops_default.c > # install includes > SYMLINK-$(CONFIG_RTE_LIBRTE_MEMPOOL)-include := rte_mempool.h > > diff --git a/lib/librte_mempool/meson.build b/lib/librte_mempool/meson.build > index 7a4f3da..9e3b527 100644 > --- a/lib/librte_mempool/meson.build > +++ b/lib/librte_mempool/meson.build > @@ -1,7 +1,8 @@ > # SPDX-License-Identifier: BSD-3-Clause > # Copyright(c) 2017 Intel Corporation > > -version = 2 > -sources = files('rte_mempool.c', 'rte_mempool_ops.c') > +version = 4 > +sources = files('rte_mempool.c', 'rte_mempool_ops.c', > + 'rte_mempool_ops_default.c') > headers = files('rte_mempool.h') > deps += ['ring'] It's strange to see that meson does not have the same .so version than the legacy build system. +CC Bruce in case he wants to fix this issue separately. [...] > --- a/lib/librte_mempool/rte_mempool_version.map > +++ b/lib/librte_mempool/rte_mempool_version.map > @@ -51,3 +51,11 @@ DPDK_17.11 { > rte_mempool_populate_iova_tab; > > } DPDK_16.07; > + > +DPDK_18.05 { > + global: > + > + rte_mempool_op_calc_mem_size_default; > + > +} DPDK_17.11; > + Another minor comment. When applying the patch with git am: Applying: mempool: add op to calculate memory size to be allocated .git/rebase-apply/patch:399: new blank line at EOF. + warning: 1 line adds whitespace errors.
Re: [dpdk-dev] [PATCH v1 2/9] mempool: add op to populate objects using provided memory
On Sat, Mar 10, 2018 at 03:39:35PM +, Andrew Rybchenko wrote: > The callback allows to customize how objects are stored in the > memory chunk. Default implementation of the callback which simply > puts objects one by one is available. > > Signed-off-by: Andrew Rybchenko [...] > --- a/lib/librte_mempool/rte_mempool.c > +++ b/lib/librte_mempool/rte_mempool.c > @@ -99,7 +99,8 @@ static unsigned optimize_object_size(unsigned obj_size) > } > > static void > -mempool_add_elem(struct rte_mempool *mp, void *obj, rte_iova_t iova) > +mempool_add_elem(struct rte_mempool *mp, __rte_unused void *opaque, > + void *obj, rte_iova_t iova) > { > struct rte_mempool_objhdr *hdr; > struct rte_mempool_objtlr *tlr __rte_unused; > @@ -116,9 +117,6 @@ mempool_add_elem(struct rte_mempool *mp, void *obj, > rte_iova_t iova) > tlr = __mempool_get_trailer(obj); > tlr->cookie = RTE_MEMPOOL_TRAILER_COOKIE; > #endif > - > - /* enqueue in ring */ > - rte_mempool_ops_enqueue_bulk(mp, &obj, 1); > } > > /* call obj_cb() for each mempool element */ Before this patch, the purpose of mempool_add_elem() was to add an object into a mempool: 1- write object header and trailers 2- chain it into the list of objects 3- add it into the ring/stack/whatever (=enqueue) Now, the enqueue is done in rte_mempool_op_populate_default() or will be done in the driver. I'm not sure it's a good idea to separate 3- from 2-, because an object that is chained into the list is expected to be in the ring/stack too. This risk of mis-synchronization is also enforced by the fact that ops->populate() can be provided by the driver and mempool_add_elem() is passed as a callback pointer. It's not clear to me why rte_mempool_ops_enqueue_bulk() is removed from mempool_add_elem(). > @@ -396,16 +394,13 @@ rte_mempool_populate_iova(struct rte_mempool *mp, char > *vaddr, > else > off = RTE_PTR_ALIGN_CEIL(vaddr, RTE_CACHE_LINE_SIZE) - vaddr; > > - while (off + total_elt_sz <= len && mp->populated_size < mp->size) { > - off += mp->header_size; > - if (iova == RTE_BAD_IOVA) > - mempool_add_elem(mp, (char *)vaddr + off, > - RTE_BAD_IOVA); > - else > - mempool_add_elem(mp, (char *)vaddr + off, iova + off); > - off += mp->elt_size + mp->trailer_size; > - i++; > - } > + if (off > len) > + return -EINVAL; I think there is a memory leak here (memhdr), but it's my fault ;) I introduced a similar code in commit 84121f1971: if (i == 0) return -EINVAL; I can send a patch for it if you want.
Re: [dpdk-dev] [PATCH v1 3/9] mempool: remove callback to get capabilities
Hi, On Thu, Mar 15, 2018 at 03:44:34PM +0300, Andrew Rybchenko wrote: [...] > > > Aha, you're saying that virtual-contiguous and IOVA-contiguous > > > requirements are different things that it could be usecases where > > > virtual contiguous is important but IOVA-contiguos is not required. > > > It is perfectly fine. > > > As I understand IOVA-contiguous (physical) typically means > > > virtual-contiguous as well. Requirements to have everything > > > virtually contiguous and some blocks physically contiguous are > > > unlikely. So, it may be reduced to either virtual or physical > > > contiguous. If mempool does not care about physical contiguous at > > > all, MEMPOOL_F_NO_PHYS_CONTIG flag should be used and min_chunk_size > > > should mean virtual contiguous requirements. If mempool requires > > > physical contiguous objects, there is *no* MEMPOOL_F_NO_PHYS_CONTIG > > > flag and min_chunk_size means physical contiguous requirements. Just as a side note, from what I understood, having VA="contiguous" and IOVA="don't care" would be helpful for mbuf pools with mellanox drivers because perform better in that case.
Re: [dpdk-dev] [PATCH v1 3/9] mempool: remove callback to get capabilities
On Sat, Mar 10, 2018 at 03:39:36PM +, Andrew Rybchenko wrote: > The callback was introduced to let generic code to know octeontx > mempool driver requirements to use single physically contiguous > memory chunk to store all objects and align object address to > total object size. Now these requirements are met using a new > callbacks to calculate required memory chunk size and to populate > objects using provided memory chunk. > > These capability flags are not used anywhere else. > > Restricting capabilities to flags is not generic and likely to > be insufficient to describe mempool driver features. If required > in the future, API which returns structured information may be > added. > > Signed-off-by: Andrew Rybchenko Looks fine... > --- a/drivers/mempool/octeontx/rte_mempool_octeontx.c > +++ b/drivers/mempool/octeontx/rte_mempool_octeontx.c > @@ -126,14 +126,29 @@ octeontx_fpavf_get_count(const struct rte_mempool *mp) > return octeontx_fpa_bufpool_free_count(pool); > } > > -static int > -octeontx_fpavf_get_capabilities(const struct rte_mempool *mp, > - unsigned int *flags) > +static ssize_t > +octeontx_fpavf_calc_mem_size(const struct rte_mempool *mp, > + uint32_t obj_num, uint32_t pg_shift, > + size_t *min_chunk_size, size_t *align) > { > - RTE_SET_USED(mp); > - *flags |= (MEMPOOL_F_CAPA_PHYS_CONTIG | > - MEMPOOL_F_CAPA_BLK_ALIGNED_OBJECTS); > - return 0; > + ssize_t mem_size; > + > + /* > + * Simply need space for one more object to be able to > + * fullfil alignment requirements. > + */ ...ah, just one typo: fullfil -> fulfil or fulfill
Re: [dpdk-dev] [PATCH v1 8/9] mempool: ensure the mempool is initialized before populating
On Sat, Mar 10, 2018 at 03:39:41PM +, Andrew Rybchenko wrote: > From: "Artem V. Andreev" > > Callback to calculate required memory area size may require mempool > driver data to be already allocated and initialized. > > Signed-off-by: Artem V. Andreev > Signed-off-by: Andrew Rybchenko > --- > RFCv2 -> v1: > - rename helper function as mempool_ops_alloc_once() > > lib/librte_mempool/rte_mempool.c | 29 ++--- > 1 file changed, 22 insertions(+), 7 deletions(-) > > diff --git a/lib/librte_mempool/rte_mempool.c > b/lib/librte_mempool/rte_mempool.c > index 844d907..12085cd 100644 > --- a/lib/librte_mempool/rte_mempool.c > +++ b/lib/librte_mempool/rte_mempool.c > @@ -322,6 +322,21 @@ rte_mempool_free_memchunks(struct rte_mempool *mp) > } > } > > +static int > +mempool_ops_alloc_once(struct rte_mempool *mp) > +{ > + int ret; > + > + /* create the internal ring if not already done */ > + if ((mp->flags & MEMPOOL_F_POOL_CREATED) == 0) { > + ret = rte_mempool_ops_alloc(mp); > + if (ret != 0) > + return ret; > + mp->flags |= MEMPOOL_F_POOL_CREATED; > + } > + return 0; > +} > + > /* Add objects in the pool, using a physically contiguous memory > * zone. Return the number of objects added, or a negative value > * on error. > @@ -336,13 +351,9 @@ rte_mempool_populate_iova(struct rte_mempool *mp, char > *vaddr, > struct rte_mempool_memhdr *memhdr; > int ret; > > - /* create the internal ring if not already done */ > - if ((mp->flags & MEMPOOL_F_POOL_CREATED) == 0) { > - ret = rte_mempool_ops_alloc(mp); > - if (ret != 0) > - return ret; > - mp->flags |= MEMPOOL_F_POOL_CREATED; > - } > + ret = mempool_ops_alloc_once(mp); > + if (ret != 0) > + return ret; > > /* mempool is already populated */ > if (mp->populated_size >= mp->size) > @@ -515,6 +526,10 @@ rte_mempool_populate_default(struct rte_mempool *mp) > unsigned mz_id, n; > int ret; > > + ret = mempool_ops_alloc_once(mp); > + if (ret != 0) > + return ret; > + > /* mempool must not be populated */ > if (mp->nb_mem_chunks != 0) > return -EEXIST; Is there a reason why we need to add it in rte_mempool_populate_default() but not in rte_mempool_populate_virt() and rte_mempool_populate_iova_tab()?
Re: [dpdk-dev] [PATCH v2 23/41] mempool: add support for the new allocation methods
Hi Anatoly, Please find some comments below. On Wed, Mar 07, 2018 at 04:56:51PM +, Anatoly Burakov wrote: > If a user has specified that the zone should have contiguous memory, > use the new _contig allocation API's instead of normal ones. > Otherwise, account for the fact that unless we're in IOVA_AS_VA > mode, we cannot guarantee that the pages would be physically > contiguous, so we calculate the memzone size and alignments as if > we were getting the smallest page size available. > > Signed-off-by: Anatoly Burakov [...] > @@ -563,10 +585,46 @@ rte_mempool_populate_default(struct rte_mempool *mp) > /* update mempool capabilities */ > mp->flags |= mp_flags; > > - if (rte_eal_has_hugepages()) { > - pg_shift = 0; /* not needed, zone is physically contiguous */ > + no_contig = mp->flags & MEMPOOL_F_NO_PHYS_CONTIG; > + force_contig = mp->flags & MEMPOOL_F_CAPA_PHYS_CONTIG; > + > + /* > + * there are several considerations for page size and page shift here. I would add a little word here to describe what page size and page shift are used for: These values impact the result of rte_mempool_xmem_size() (*), which returns the amount of memory that should be allocated to store the desired number of objects. When not zero, it allocates more memory for the padding between objects, to ensure that an object does not cross a page boundary. (*) it is renamed in Andrew's patchset about mempool_ops API, but it seems the memory rework may be pushed before. > + * > + * if we don't need our mempools to have physically contiguous objects, > + * then just set page shift and page size to 0, because the user has > + * indicated that there's no need to care about anything. > + * > + * if we do need contiguous objects, there is also an option to reserve > + * the entire mempool memory as one contiguous block of memory, in > + * which case the page shift and alignment wouldn't matter as well. > + * > + * if we require contiguous objects, but not necessarily the entire > + * mempool reserved space to be contiguous, then there are two options. > + * > + * if our IO addresses are virtual, not actual physical (IOVA as VA > + * case), then no page shift needed - our memory allocation will give us > + * contiguous physical memory as far as the hardware is concerned, so > + * act as if we're getting contiguous memory. > + * > + * if our IO addresses are physical, we may get memory from bigger > + * pages, or we might get memory from smaller pages, and how much of it > + * we require depends on whether we want bigger or smaller pages. > + * However, requesting each and every memory size is too much work, so > + * what we'll do instead is walk through the page sizes available, pick > + * the smallest one and set up page shift to match that one. We will be > + * wasting some space this way, but it's much nicer than looping around > + * trying to reserve each and every page size. > + */ This comment is helpful to understand, thanks. (by the way, reading it makes me think we should rename MEMPOOL_F_*_PHYS_CONTIG as MEMPOOL_F_*_IOVA_CONTIG) > + > + if (no_contig || force_contig || rte_eal_iova_mode() == RTE_IOVA_VA) { > pg_sz = 0; > + pg_shift = 0; > align = RTE_CACHE_LINE_SIZE; > + } else if (rte_eal_has_hugepages()) { > + pg_sz = get_min_page_size(); > + pg_shift = rte_bsf32(pg_sz); > + align = pg_sz; > } else { > pg_sz = getpagesize(); > pg_shift = rte_bsf32(pg_sz); > @@ -585,23 +643,34 @@ rte_mempool_populate_default(struct rte_mempool *mp) > goto fail; > } > > - mz = rte_memzone_reserve_aligned(mz_name, size, > - mp->socket_id, mz_flags, align); > - /* not enough memory, retry with the biggest zone we have */ > - if (mz == NULL) > - mz = rte_memzone_reserve_aligned(mz_name, 0, > + if (force_contig) { > + /* > + * if contiguous memory for entire mempool memory was > + * requested, don't try reserving again if we fail. > + */ > + mz = rte_memzone_reserve_aligned_contig(mz_name, size, > + mp->socket_id, mz_flags, align); > + } else { > + mz = rte_memzone_reserve_aligned(mz_name, size, > mp->socket_id, mz_flags, align); > + /* not enough memory, retry with the biggest zone we > + * have > + */ > + if (mz == NULL) > + mz = rte_memzone_reserve_aligned(mz_name, 0, > + mp->s
Re: [dpdk-dev] [PATCH 00/41] Memory Hotplug for DPDK
Hi Anatoly, On Sat, Mar 03, 2018 at 01:45:48PM +, Anatoly Burakov wrote: > This patchset introduces dynamic memory allocation for DPDK (aka memory > hotplug). Based upon RFC submitted in December [1]. > > Dependencies (to be applied in specified order): > - IPC bugfixes patchset [2] > - IPC improvements patchset [3] > - IPC asynchronous request API patch [4] > - Function to return number of sockets [5] > > Deprecation notices relevant to this patchset: > - General outline of memory hotplug changes [6] > - EAL NUMA node count changes [7] > > The vast majority of changes are in the EAL and malloc, the external API > disruption is minimal: a new set of API's are added for contiguous memory > allocation for rte_memzone, and a few API additions in rte_memory due to > switch to memseg_lists as opposed to memsegs. Every other API change is > internal to EAL, and all of the memory allocation/freeing is handled > through rte_malloc, with no externally visible API changes. > > Quick outline of all changes done as part of this patchset: > > * Malloc heap adjusted to handle holes in address space > * Single memseg list replaced by multiple memseg lists > * VA space for hugepages is preallocated in advance > * Added alloc/free for pages happening as needed on rte_malloc/rte_free > * Added contiguous memory allocation API's for rte_memzone > * Integrated Pawel Wodkowski's patch for registering/unregistering memory >with VFIO [8] > * Callbacks for registering memory allocations > * Multiprocess support done via DPDK IPC introduced in 18.02 > > The biggest difference is a "memseg" now represents a single page (as opposed > to > being a big contiguous block of pages). As a consequence, both memzones and > malloc elements are no longer guaranteed to be physically contiguous, unless > the user asks for it at reserve time. To preserve whatever functionality that > was dependent on previous behavior, a legacy memory option is also provided, > however it is expected (or perhaps vainly hoped) to be temporary solution. > > Why multiple memseg lists instead of one? Since memseg is a single page now, > the list of memsegs will get quite big, and we need to locate pages somehow > when we allocate and free them. We could of course just walk the list and > allocate one contiguous chunk of VA space for memsegs, but this > implementation uses separate lists instead in order to speed up many > operations with memseg lists. > > For v1, the following limitations are present: > - FreeBSD does not even compile, let alone run > - No 32-bit support > - There are some minor quality-of-life improvements planned that aren't > ready yet and will be part of v2 > - VFIO support is only smoke-tested (but is expected to work), VFIO support > with secondary processes is not tested; work is ongoing to validate VFIO > for all use cases > - Dynamic mapping/unmapping memory with VFIO is not supported in sPAPR > IOMMU mode - help from sPAPR maintainers requested > > Nevertheless, this patchset should be testable under 64-bit Linux, and > should work for all use cases bar those mentioned above. > > [1] http://dpdk.org/dev/patchwork/bundle/aburakov/Memory_RFC/ > [2] http://dpdk.org/dev/patchwork/bundle/aburakov/IPC_Fixes/ > [3] http://dpdk.org/dev/patchwork/bundle/aburakov/IPC_Improvements/ > [4] http://dpdk.org/dev/patchwork/bundle/aburakov/IPC_Async_Request/ > [5] http://dpdk.org/dev/patchwork/bundle/aburakov/Num_Sockets/ > [6] http://dpdk.org/dev/patchwork/patch/34002/ > [7] http://dpdk.org/dev/patchwork/patch/33853/ > [8] http://dpdk.org/dev/patchwork/patch/24484/ I did a quick pass on your patches (unfortunately, I don't have the time to really dive in it). I have few questions/comments: - This is really a big patchset. Thank you for working on this topic. I'll try to test our application with it as soon as possible. - I see from patch 17 that it is possible that rte_malloc() expands the heap by requesting more memory to the OS? Did I understand well? Today, a good property of rte_malloc() compared to malloc() is that it won't interrupt the process (the worst case is a spinlock). This is appreciable on a dataplane core. Will it change? - It's not a big issue, but I have the feeling that the "const" statement is often forgotten in the patchset. I think it is helpful for both optimization, documentation and to detect bugs that modifies/free something that should not. I'm sending some other dummy comments as replies to patches. Thanks, Olivier
Re: [dpdk-dev] [PATCH 03/41] eal: make malloc heap a doubly-linked list
On Sat, Mar 03, 2018 at 01:45:51PM +, Anatoly Burakov wrote: > As we are preparing for dynamic memory allocation, we need to be > able to handle holes in our malloc heap, hence we're switching to > doubly linked list, and prepare infrastructure to support it. > > Since our heap is now aware where are our first and last elements, > there is no longer any need to have a dummy element at the end of > each heap, so get rid of that as well. Instead, let insert/remove/ > join/split operations handle end-of-list conditions automatically. > > Signed-off-by: Anatoly Burakov > --- > lib/librte_eal/common/include/rte_malloc_heap.h | 6 + > lib/librte_eal/common/malloc_elem.c | 200 > +++- > lib/librte_eal/common/malloc_elem.h | 14 +- > lib/librte_eal/common/malloc_heap.c | 8 +- > 4 files changed, 179 insertions(+), 49 deletions(-) > > diff --git a/lib/librte_eal/common/include/rte_malloc_heap.h > b/lib/librte_eal/common/include/rte_malloc_heap.h > index ba99ed9..9ec4b62 100644 > --- a/lib/librte_eal/common/include/rte_malloc_heap.h > +++ b/lib/librte_eal/common/include/rte_malloc_heap.h > @@ -13,12 +13,18 @@ > /* Number of free lists per heap, grouped by size. */ > #define RTE_HEAP_NUM_FREELISTS 13 > > +/* dummy definition, for pointers */ > +struct malloc_elem; > + > /** > * Structure to hold malloc heap > */ > struct malloc_heap { > rte_spinlock_t lock; > LIST_HEAD(, malloc_elem) free_head[RTE_HEAP_NUM_FREELISTS]; > + struct malloc_elem *first; > + struct malloc_elem *last; > + > unsigned alloc_count; > size_t total_size; > } __rte_cache_aligned; > diff --git a/lib/librte_eal/common/malloc_elem.c > b/lib/librte_eal/common/malloc_elem.c > index ea041e2..eb41200 100644 > --- a/lib/librte_eal/common/malloc_elem.c > +++ b/lib/librte_eal/common/malloc_elem.c > @@ -31,6 +31,7 @@ malloc_elem_init(struct malloc_elem *elem, > elem->heap = heap; > elem->ms = ms; > elem->prev = NULL; > + elem->next = NULL; > memset(&elem->free_list, 0, sizeof(elem->free_list)); > elem->state = ELEM_FREE; > elem->size = size; > @@ -39,15 +40,56 @@ malloc_elem_init(struct malloc_elem *elem, > set_trailer(elem); > } > > -/* > - * Initialize a dummy malloc_elem header for the end-of-memseg marker > - */ > void > -malloc_elem_mkend(struct malloc_elem *elem, struct malloc_elem *prev) > +malloc_elem_insert(struct malloc_elem *elem) > { > - malloc_elem_init(elem, prev->heap, prev->ms, 0); > - elem->prev = prev; > - elem->state = ELEM_BUSY; /* mark busy so its never merged */ > + struct malloc_elem *prev_elem, *next_elem; > + struct malloc_heap *heap = elem->heap; > + > + if (heap->first == NULL && heap->last == NULL) { > + /* if empty heap */ > + heap->first = elem; > + heap->last = elem; > + prev_elem = NULL; > + next_elem = NULL; > + } else if (elem < heap->first) { > + /* if lower than start */ > + prev_elem = NULL; > + next_elem = heap->first; > + heap->first = elem; > + } else if (elem > heap->last) { > + /* if higher than end */ > + prev_elem = heap->last; > + next_elem = NULL; > + heap->last = elem; > + } else { > + /* the new memory is somewhere inbetween start and end */ > + uint64_t dist_from_start, dist_from_end; > + > + dist_from_end = RTE_PTR_DIFF(heap->last, elem); > + dist_from_start = RTE_PTR_DIFF(elem, heap->first); > + > + /* check which is closer, and find closest list entries */ > + if (dist_from_start < dist_from_end) { > + prev_elem = heap->first; > + while (prev_elem->next < elem) > + prev_elem = prev_elem->next; > + next_elem = prev_elem->next; > + } else { > + next_elem = heap->last; > + while (next_elem->prev > elem) > + next_elem = next_elem->prev; > + prev_elem = next_elem->prev; > + } > + } > + > + /* insert new element */ > + elem->prev = prev_elem; > + elem->next = next_elem; > + if (prev_elem) > + prev_elem->next = elem; > + if (next_elem) > + next_elem->prev = elem; > } Would it be possible here to use a TAILQ? If yes, it could be easier to read.
Re: [dpdk-dev] [PATCH 08/41] eal: make malloc free return resulting malloc element
On Sat, Mar 03, 2018 at 01:45:56PM +, Anatoly Burakov wrote: > Signed-off-by: Anatoly Burakov > --- > lib/librte_eal/common/malloc_elem.c | 4 ++-- > lib/librte_eal/common/malloc_elem.h | 2 +- > lib/librte_eal/common/malloc_heap.c | 4 ++-- > 3 files changed, 5 insertions(+), 5 deletions(-) > > diff --git a/lib/librte_eal/common/malloc_elem.c > b/lib/librte_eal/common/malloc_elem.c > index 008f5a3..c18f050 100644 > --- a/lib/librte_eal/common/malloc_elem.c > +++ b/lib/librte_eal/common/malloc_elem.c > @@ -379,7 +379,7 @@ malloc_elem_join_adjacent_free(struct malloc_elem *elem) > * blocks either immediately before or immediately after newly freed block > * are also free, the blocks are merged together. > */ > -int > +struct malloc_elem * > malloc_elem_free(struct malloc_elem *elem) > { > void *ptr; > @@ -397,7 +397,7 @@ malloc_elem_free(struct malloc_elem *elem) > > memset(ptr, 0, data_len); > > - return 0; > + return elem; > } > > /* An explanation about why this change is needed would make sense I think. Thanks, Olivier
Re: [dpdk-dev] [PATCH 13/41] eal: replace memseg with memseg lists
On Sat, Mar 03, 2018 at 01:46:01PM +, Anatoly Burakov wrote: [...] > --- a/config/common_base > +++ b/config/common_base > @@ -61,7 +61,20 @@ CONFIG_RTE_CACHE_LINE_SIZE=64 > CONFIG_RTE_LIBRTE_EAL=y > CONFIG_RTE_MAX_LCORE=128 > CONFIG_RTE_MAX_NUMA_NODES=8 > -CONFIG_RTE_MAX_MEMSEG=256 > +CONFIG_RTE_MAX_MEMSEG_LISTS=32 > +# each memseg list will be limited to either RTE_MAX_MEMSEG_PER_LIST pages > +# or RTE_MAX_MEM_PER_LIST gigabytes worth of memory, whichever is the > smallest > +CONFIG_RTE_MAX_MEMSEG_PER_LIST=8192 > +CONFIG_RTE_MAX_MEM_PER_LIST=32 > +# a "type" is a combination of page size and NUMA node. total number of > memseg > +# lists per type will be limited to either RTE_MAX_MEMSEG_PER_TYPE pages > (split > +# over multiple lists of RTE_MAX_MEMSEG_PER_LIST pages), or > RTE_MAX_MEM_PER_TYPE > +# gigabytes of memory (split over multiple lists of RTE_MAX_MEM_PER_LIST), > +# whichever is the smallest > +CONFIG_RTE_MAX_MEMSEG_PER_TYPE=32768 > +CONFIG_RTE_MAX_MEM_PER_TYPE=128 > +# legacy mem mode only > +CONFIG_RTE_MAX_LEGACY_MEMSEG=256 Would it be possible to suffix CONFIG_RTE_MAX_MEM_PER_LIST and CONFIG_RTE_MAX_MEM_PER_TYPE with _GB? It's not that obvious that is it gigabytes. What is the impact of changing one of these values on the ABI? And what would be the impact on performance? The underlying question is: shall we increase these values to avoid changing them later?
Re: [dpdk-dev] [PATCH 14/41] eal: add support for mapping hugepages at runtime
On Sat, Mar 03, 2018 at 01:46:02PM +, Anatoly Burakov wrote: > Nothing uses this code yet. The bulk of it is copied from old > memory allocation code (linuxapp eal_memory.c). We provide an > EAL-internal API to allocate either one page or multiple pages, > guaranteeing that we'll get contiguous VA for all of the pages > that we requested. > > For single-file segments, we will use fallocate() to grow and > shrink memory segments, however fallocate() is not supported > on all kernel versions, so we will fall back to using > ftruncate() to grow the file, and disable shrinking as there's > little we can do there. This will enable vhost use cases where > having single file segments is of great value even without > support for hot-unplugging memory. > > Not supported on FreeBSD. > > Locking is done via fcntl() because that way, when it comes to > taking out write locks or unlocking on deallocation, we don't > have to keep original fd's around. Plus, using fcntl() gives us > ability to lock parts of a file, which is useful for single-file > segments. > > Signed-off-by: Anatoly Burakov Few minor typos: [...] > +static void > +resotre_numa(int *oldpolicy, struct bitmask *oldmask) restore [...] > +static off_t > +getFileSize(int fd) should it be get_file_size()? [...] > +static int > +alloc_page(struct rte_memseg *ms, void *addr, uint64_t size, int socket_id, > + struct hugepage_info *hi, unsigned int list_idx, > + unsigned int seg_idx) > +{ > + int cur_socket_id = 0; > + uint64_t map_offset; > + char path[PATH_MAX]; > + int ret = 0; > + int fd; > + > + fd = get_page_fd(path, sizeof(path), hi, list_idx, seg_idx); > + if (fd < 0) > + return -1; > + > + > + if (internal_config.single_file_segments) { > + map_offset = seg_idx * size; > + ret = resize_hugefile(fd, map_offset, size, true); > + if (ret < 1) > + goto resized; > + } else { > + map_offset = 0; > + if (ftruncate(fd, size) < 0) { > + RTE_LOG(DEBUG, EAL, "%s(): ftruncate() failed: %s\n", > + __func__, strerror(errno)); > + goto resized; > + } > + /* we've allocated a page - take out a read lock. we're using > + * fcntl() locks rather than flock() here because doing that > + * gives us one huge advantage - fcntl() locks are per-process, > + * not per-file descriptor, which means that we don't have to > + * keep the original fd's around to keep a lock on the file. > + * > + * this is useful, because when it comes to unmapping pages, we > + * will have to take out a write lock (to figure out if another > + * process still has this page mapped), and to do itwith flock() typo: itwith
Re: [dpdk-dev] [PATCH 17/41] eal: enable memory hotplug support in rte_malloc
On Sat, Mar 03, 2018 at 01:46:05PM +, Anatoly Burakov wrote: > This set of changes enables rte_malloc to allocate and free memory > as needed. The way it works is, first malloc checks if there is > enough memory already allocated to satisfy user's request. If there > isn't, we try and allocate more memory. The reverse happens with > free - we free an element, check its size (including free element > merging due to adjacency) and see if it's bigger than hugepage > size and that its start and end span a hugepage or more. Then we > remove the area from malloc heap (adjusting element lengths where > appropriate), and deallocate the page. > > For legacy mode, runtime alloc/free of pages is disabled. > > It is worth noting that memseg lists are being sorted by page size, > and that we try our best to satisfy user's request. That is, if > the user requests an element from a 2MB page memory, we will check > if we can satisfy that request from existing memory, if not we try > and allocate more 2MB pages. If that fails and user also specified > a "size is hint" flag, we then check other page sizes and try to > allocate from there. If that fails too, then, depending on flags, > we may try allocating from other sockets. In other words, we try > our best to give the user what they asked for, but going to other > sockets is last resort - first we try to allocate more memory on > the same socket. > > Signed-off-by: Anatoly Burakov [...] > @@ -123,48 +125,356 @@ find_suitable_element(struct malloc_heap *heap, size_t > size, > * scan fails. Once the new memseg is added, it re-scans and should return > * the new element after releasing the lock. > */ > -void * > -malloc_heap_alloc(struct malloc_heap *heap, > - const char *type __attribute__((unused)), size_t size, unsigned > flags, > - size_t align, size_t bound) > +static void * > +heap_alloc(struct malloc_heap *heap, const char *type __rte_unused, size_t > size, > + unsigned int flags, size_t align, size_t bound) > { > struct malloc_elem *elem; > > size = RTE_CACHE_LINE_ROUNDUP(size); > align = RTE_CACHE_LINE_ROUNDUP(align); > > - rte_spinlock_lock(&heap->lock); > - > elem = find_suitable_element(heap, size, flags, align, bound); > if (elem != NULL) { > elem = malloc_elem_alloc(elem, size, align, bound); > + > /* increase heap's count of allocated elements */ > heap->alloc_count++; > } > - rte_spinlock_unlock(&heap->lock); > > return elem == NULL ? NULL : (void *)(&elem[1]); > } The comment on top of the function says "after releasing the lock" but it seems it's not relevant anymore because the lock is removed. [...] > int > malloc_heap_free(struct malloc_elem *elem) > { > struct malloc_heap *heap; > - struct malloc_elem *ret; > + void *start, *aligned_start, *end, *aligned_end; > + size_t len, aligned_len; > + struct rte_memseg_list *msl; > + int n_pages, page_idx, max_page_idx, ret; > > if (!malloc_elem_cookies_ok(elem) || elem->state != ELEM_BUSY) > return -1; > > /* elem may be merged with previous element, so keep heap address */ > heap = elem->heap; > + msl = elem->msl; > > rte_spinlock_lock(&(heap->lock)); > > - ret = malloc_elem_free(elem); > + elem = malloc_elem_free(elem); > > - rte_spinlock_unlock(&(heap->lock)); > + /* anything after this is a bonus */ > + ret = 0; > + The fact that there was previously 2 rte_spinlock_unlock() calls looks strange to me. Is there something wrong in a previous patch?
Re: [dpdk-dev] [PATCH 18/41] test: fix malloc autotest to support memory hotplug
On Sat, Mar 03, 2018 at 01:46:06PM +, Anatoly Burakov wrote: > The test was expecting memory already being allocated on all sockets, > and thus was failing because calling rte_malloc could trigger memory > hotplug event and allocate memory where there was none before. > > Fix it to instead report availability of memory on specific sockets > by attempting to allocate a page and see if that succeeds. Technically, > this can still cause failure as memory might not be available at the > time of check, but become available by the time the test is run, but > this is a corner case not worth considering. > > Signed-off-by: Anatoly Burakov > --- > test/test/test_malloc.c | 52 > + > 1 file changed, 44 insertions(+), 8 deletions(-) > > diff --git a/test/test/test_malloc.c b/test/test/test_malloc.c > index 8484fb6..2aaf1b8 100644 > --- a/test/test/test_malloc.c > +++ b/test/test/test_malloc.c > @@ -22,6 +22,8 @@ > #include > #include > > +#include "../../lib/librte_eal/common/eal_memalloc.h" > + I guess there is no way to test without importing a private EAL function, correct? If yes, maybe it deserves a quick explanation.
[dpdk-dev] DPDK build issues with gcc-8
Tried building DPDK with gcc-8 and got several new warnings. Working on trivial patches for most of them, but the DPDK rte_table interface for hashing has a more fundamental problem which Gcc-8 discovers. The code is trying to cast an RTE hash function into a RTE table hash function with a different signature. It works because the two are mostly the same, but it really shouldn't be done that way: Warning: CC rte_table_hash_cuckoo.o lib/librte_table/rte_table_hash_cuckoo.c: In function ‘rte_table_hash_cuckoo_create’: lib/librte_table/rte_table_hash_cuckoo.c:110:16: error: cast between incompatible function types from ‘rte_table_hash_op_hash’ {aka ‘long unsigned int (*)(void *, void *, unsigned int, long unsigned int)’} to ‘uint32_t (*)(const void *, uint32_t, uint32_t)’ {aka ‘unsigned int (*)(const void *, unsigned int, unsigned int)’} [-Werror=cast-function-type] .hash_func = (rte_hash_function)(p->f_hash), ^ RTE hash: /** Type of function that can be used for calculating the hash value. */ typedef uint32_t (*rte_hash_function)(const void *key, uint32_t key_len, uint32_t init_val); RTE table hash: /** Hash function */ typedef uint64_t (*rte_table_hash_op_hash)( void *key, void *key_mask, uint32_t key_size, uint64_t seed); Just turning off the warning is not an acceptable solution. The API should really be fixed. My preference would be to eliminate rte_table_hash_op_hash typedef and keep rte_hash_function.
[dpdk-dev] [PATCH 0/2] Gcc-8 build fixes
A couple of trivial patches since GCC-8 finds more string overflows than earlier versions. Stephen Hemminger (2): rte_mempool: fix strncpy warnings rte_metrics: fix strncpy truncation warning lib/librte_mbuf/rte_mbuf_pool_ops.c | 4 ++-- lib/librte_metrics/rte_metrics.c| 7 +-- 2 files changed, 7 insertions(+), 4 deletions(-) -- 2.16.2
[dpdk-dev] [PATCH 1/2] rte_mempool: fix strncpy warnings
Gcc-8 discovers issue with platform_mempool_ops. rte_mbuf_pool_ops.c:26:3: error: ‘strncpy’ output truncated before terminating nul copying as many bytes from a string as its length [-Werror=stringop-truncation] strncpy(mz->addr, ops_name, strlen(ops_name)); The issue is that the size should be size in destination not source. Fixes: a3acc3144a76 ("mbuf: add pool ops selection functions") --- lib/librte_mbuf/rte_mbuf_pool_ops.c | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/lib/librte_mbuf/rte_mbuf_pool_ops.c b/lib/librte_mbuf/rte_mbuf_pool_ops.c index 48cc342002a5..661cf4fb401f 100644 --- a/lib/librte_mbuf/rte_mbuf_pool_ops.c +++ b/lib/librte_mbuf/rte_mbuf_pool_ops.c @@ -23,7 +23,7 @@ rte_mbuf_set_platform_mempool_ops(const char *ops_name) RTE_MEMPOOL_OPS_NAMESIZE, SOCKET_ID_ANY, 0); if (mz == NULL) return -rte_errno; - strncpy(mz->addr, ops_name, strlen(ops_name)); + strncpy(mz->addr, ops_name, sizeof(mz->addr) - 1); return 0; } else if (strcmp(mz->addr, ops_name) == 0) { return 0; @@ -62,7 +62,7 @@ rte_mbuf_set_user_mempool_ops(const char *ops_name) return -rte_errno; } - strncpy(mz->addr, ops_name, strlen(ops_name)); + strncpy(mz->addr, ops_name, sizeof(mz->addr) - 1); return 0; } -- 2.16.2
[dpdk-dev] [PATCH 2/2] rte_metrics: fix strncpy truncation warning
Fixes: librte_metrics/rte_metrics.c:218:4: error: ‘strncpy’ specified bound 64 equals destination size [-Werror=stringop-truncation] strncpy(names[idx_name].name, ^ stats->metadata[idx_name].name, ~~~ RTE_METRICS_MAX_NAME_LEN); ~ Signed-off-by: Stephen Hemminger --- lib/librte_metrics/rte_metrics.c | 7 +-- 1 file changed, 5 insertions(+), 2 deletions(-) diff --git a/lib/librte_metrics/rte_metrics.c b/lib/librte_metrics/rte_metrics.c index 556ae1ba8b4d..ae79d4ab276e 100644 --- a/lib/librte_metrics/rte_metrics.c +++ b/lib/librte_metrics/rte_metrics.c @@ -214,10 +214,13 @@ rte_metrics_get_names(struct rte_metric_name *names, rte_spinlock_unlock(&stats->lock); return return_value; } - for (idx_name = 0; idx_name < stats->cnt_stats; idx_name++) + + for (idx_name = 0; idx_name < stats->cnt_stats; idx_name++) { + memset(names[idx_name].name, 0, RTE_METRICS_MAX_NAME_LEN); strncpy(names[idx_name].name, stats->metadata[idx_name].name, - RTE_METRICS_MAX_NAME_LEN); + RTE_METRICS_MAX_NAME_LEN - 1); + } } return_value = stats->cnt_stats; rte_spinlock_unlock(&stats->lock); -- 2.16.2
Re: [dpdk-dev] [PATCH] net/mlx5: fix compilation error of rdma-core on ARM
Wednesday, March 14, 2018 7:52 PM, Yongseok Koh: > rdma-core v16 has a bug. The following compilation error occurs on ARM > hosts. > > In file included > from /auto/mswg/yskoh/git/dpdk.org/drivers/net/mlx5/mlx5_glue.h:16:0, > from /auto/mswg/yskoh/git/dpdk.org/drivers/net/mlx5/mlx5_glue.c:11: > /usr/include/infiniband/mlx5dv.h:144:2: error: unknown type name ‘off_t’ > off_t uar_mmap_offset; > ^ > > As a temporary fix, sys/types.h is included in PMD. This has been fixed in > rdma-core v17. This can be removed when all the Linux distros are shipped > with rdma-core v17 or back-ported fix. As of now, RedHat 7.5 is known to > have rdma-core v16. > > Signed-off-by: Yongseok Koh > Acked-by: Adrien Mazarguil Applied to next-net-mlx, thanks.
Re: [dpdk-dev] [PATCH v2 0/3] net/mlx5: cleanup link status
Monday, March 12, 2018 3:43 PM, Nelio Laranjeiro: > This series applies on top of [1] and cleans up the DPDK API implementation > for the link status. > > [1] > https://emea01.safelinks.protection.outlook.com/?url=https%3A%2F%2Fdp > dk.org%2Fdev%2Fpatchwork%2Fpatch%2F35653%2F&data=02%7C01%7Csha > hafs%40mellanox.com%7C53adbbb3cbac4c7036f508d5881f6948%7Ca652971c > 7d2e4d9ba6a4d149256f461b%7C0%7C0%7C636564590863924988&sdata=ic9Y > UOQLHq719ebOXxX%2FlS%2BgzWX7R7CFa7qXrOEG7xs%3D&reserved=0 > > Changes in v2: > - Removes kernel version verification, the bug it tried to detected was fixed > several commit after in the PMD. Implementation in mlx5 kernel driver is > only available since v4.9. > - Removes the alarm handler as this can be worked around by not > acknowledging > the event until the link becomes stable. > - Clean-up the API implementation by letting the application handle the > interrupt and decide by itself to start/stop the device. > > Nelio Laranjeiro (3): > net/mlx5: remove kernel version check > net/mlx5: fix link status behavior > net/mlx5: fix link status to use wait to complete Series applied to next-net-mlx, thanks.
Re: [dpdk-dev] [PATCH v4 1/3] ethdev: add enum type for loop on Rx/Tx offloads
Hi Zhang Qi Thanks for your feedback. > -Original Message- > From: Zhang, Qi Z > Sent: Monday, March 19, 2018 10:06 PM > To: Dai, Wei ; Lu, Wenzhuo ; > Wu, Jingjing > Cc: dev@dpdk.org; Dai, Wei > Subject: RE: [dpdk-dev] [PATCH v4 1/3] ethdev: add enum type for loop on > Rx/Tx offloads > > Hi Daiwei: > > > -Original Message- > > From: dev [mailto:dev-boun...@dpdk.org] On Behalf Of Wei Dai > > Sent: Monday, March 19, 2018 8:41 PM > > To: Lu, Wenzhuo ; Wu, Jingjing > > > > Cc: dev@dpdk.org; Dai, Wei > > Subject: [dpdk-dev] [PATCH v4 1/3] ethdev: add enum type for loop on > > Rx/Tx offloads > > > > This patch adds enum rte_eth_rx_offload_type and enum > > rte_eth_tx_offload_type. For a loop on all Rx offloads, it is > > convenient to begin with the first enum member > > ETH_RX_OFFLOAD_FIRST_FEATURE and to end at > ETH_RX_OFFLOAD_TOTAL_NUM. > > A loop on all Tx offloads can begin with > ETH_TX_OFFLOAD_FIRST_FEATURE > > and end at ETH_TX_OFFLOAD_TOTAL_NUM. > > > > Signed-off-by: Wei Dai > > --- > > lib/librte_ether/rte_ethdev.h | 44 > > +++ > > 1 file changed, 44 insertions(+) > > > > diff --git a/lib/librte_ether/rte_ethdev.h > > b/lib/librte_ether/rte_ethdev.h index 0361533..0089ea3 100644 > > --- a/lib/librte_ether/rte_ethdev.h > > +++ b/lib/librte_ether/rte_ethdev.h > > @@ -946,6 +946,27 @@ struct rte_eth_conf { > > DEV_RX_OFFLOAD_VLAN_FILTER | \ > > DEV_RX_OFFLOAD_VLAN_EXTEND) > > > > +enum rte_eth_rx_offload_type { > > + ETH_RX_OFFLOAD_FIRST_FEATURE = 0, > > + ETH_RX_OFFLOAD_VLAN_STRIP = ETH_RX_OFFLOAD_FIRST_FEATURE, > > + ETH_RX_OFFLOAD_IPV4_CKSUM, > > + ETH_RX_OFFLOAD_UDP_CKSUM, > > + ETH_RX_OFFLOAD_TCP_CKSUM, > > + ETH_RX_OFFLOAD_TCP_LRO, > > + ETH_RX_OFFLOAD_QINQ_STRIP, > > + ETH_RX_OFFLOAD_OUTER_IPV4_CKSUM, > > + ETH_RX_OFFLOAD_MACSEC_STRIP, > > + ETH_RX_OFFLOAD_HEADER_SPLIT, > > + ETH_RX_OFFLOAD_VLAN_FILTER, > > + ETH_RX_OFFLOAD_VLAN_EXTEND, > > + ETH_RX_OFFLOAD_JUMBO_FRAME, > > + ETH_RX_OFFLOAD_CRC_STRIP, > > + ETH_RX_OFFLOAD_SCATTER, > > + ETH_RX_OFFLOAD_TIMESTAMP, > > + ETH_RX_OFFLOAD_SECURITY, > > + ETH_RX_OFFLOAD_TOTAL_NUM > > +}; > > This looks a little duplicate with rte_rx_offload_names in rte_ethdev.c rte_rx_offload_names is a static string array defined in rte_ethdev.c, if it can be referred by other files, the "static" must be removed. > > In patch 2/3 and 3/3, I think if you simply loop from offload bit 0 to bit 63 > to > check the name, that should still work right? > then this patch could be unnecessary. Yes, using loop from bit 0 to bit 63 in patch 2/3 and 3/3 can work right. But I think this enum type is still useful for other application which need to loop on all possible offloads. I will submit this patch as a standalone one and using loop from bit 0 to bit 63 in next version of patch set. > Regards > Qi > > + > > /* > > * If new Rx offload capabilities are defined, they also must be > > * mentioned in rte_rx_offload_names in rte_ethdev.c file. > > @@ -981,6 +1002,29 @@ struct rte_eth_conf { > > */ > > #define DEV_TX_OFFLOAD_SECURITY 0x0002 > > > > +enum rte_eth_tx_offload_type { > > + ETH_TX_OFFLOAD_FIRST_FEATURE = 0, > > + ETH_TX_OFFLOAD_VLAN_INSERT = ETH_TX_OFFLOAD_FIRST_FEATURE, > > + ETH_TX_OFFLOAD_IPV4_CKSUM, > > + ETH_TX_OFFLOAD_UDP_CKSUM, > > + ETH_TX_OFFLOAD_TCP_CKSUM, > > + ETH_TX_OFFLOAD_SCTP_CKSUM, > > + ETH_TX_OFFLOAD_TCP_TSO, > > + ETH_TX_OFFLOAD_UDP_TSO, > > + ETH_TX_OFFLOAD_OUTER_IPV4_CKSUM, > > + ETH_TX_OFFLOAD_QINQ_INSERT, > > + ETH_TX_OFFLOAD_VXLAN_TNL_TSO, > > + ETH_TX_OFFLOAD_GRE_TNL_TSO, > > + ETH_TX_OFFLOAD_IPIP_TNL_TSO, > > + ETH_TX_OFFLOAD_GENEVE_TNL_TSO, > > + ETH_TX_OFFLOAD_MACSEC_INSERT, > > + ETH_TX_OFFLOAD_MT_LOCKFREE, > > + ETH_TX_OFFLOAD_MULTI_SEGS, > > + ETH_TX_OFFLOAD_MBUF_FAST_FREE, > > + ETH_TX_OFFLOAD_SECURITY, > > + ETH_TX_OFFLOAD_TOTAL_NUM > > +}; > > + > > /* > > * If new Tx offload capabilities are defined, they also must be > > * mentioned in rte_tx_offload_names in rte_ethdev.c file. > > -- > > 2.7.5
Re: [dpdk-dev] [PATCH v3] net/i40e: fix intr callback unregister by adding retry
Is it better to fix this way? Or other ideas? - /* register callback func to eal lib */ - rte_intr_callback_unregister(intr_handle, -i40e_dev_interrupt_handler, dev); - + /* unregister callback func to eal lib */ + while ((ret = rte_intr_callback_unregister(intr_handle, + i40e_dev_interrupt_handler, dev)) == -EAGAIN) + ; + if (ret < 0) + PMD_INIT_LOG(ERR, + "cannot unregister interrupt handler %d", + ret); > -Original Message- > From: Zhang, Qi Z [mailto:qi.z.zh...@intel.com] > Sent: Wednesday, March 14, 2018 9:42 PM > To: Rybalchenko, Kirill ; wangyunjian > ; dev@dpdk.org > Cc: caihe > Subject: RE: [dpdk-dev] [PATCH v3] net/i40e: fix intr callback unregister by > adding retry > > > > > -Original Message- > > From: dev [mailto:dev-boun...@dpdk.org] On Behalf Of Rybalchenko, > > Kirill > > Sent: Wednesday, March 14, 2018 8:08 PM > > To: wangyunjian ; dev@dpdk.org > > Cc: ca...@huawei.com > > Subject: Re: [dpdk-dev] [PATCH v3] net/i40e: fix intr callback > > unregister by adding retry > > > > > > > > > -Original Message- > > > From: dev [mailto:dev-boun...@dpdk.org] On Behalf Of wangyunjian > > > Sent: Wednesday 14 March 2018 12:02 > > > To: dev@dpdk.org > > > Cc: ca...@huawei.com; Yunjian Wang > > > Subject: [dpdk-dev] [PATCH v3] net/i40e: fix intr callback > > > unregister by adding retry > > > > > > From: Yunjian Wang > > > > > > The nic's interrupt source has some active callbacks, when the port > > hotplug. > > > Add a retry to give more port's a chance to uninit before returning an > error. > > > > > > Fixes: d42aaf30008b ("i40e: support port hotplug") > > > > > > Signed-off-by: Yunjian Wang > > > --- > > > v2: > > > * Fix return value check > > > --- > > > v3: > > > * Fix comment suggested by Kirill Rybalchenko > > > --- > > > drivers/net/i40e/i40e_ethdev.c | 18 +++--- > > > 1 file changed, 15 insertions(+), 3 deletions(-) > > > > > > diff --git a/drivers/net/i40e/i40e_ethdev.c > > > b/drivers/net/i40e/i40e_ethdev.c index 508b417..6bb31f6 100644 > > > --- a/drivers/net/i40e/i40e_ethdev.c > > > +++ b/drivers/net/i40e/i40e_ethdev.c > > > @@ -1533,6 +1533,7 @@ void i40e_flex_payload_reg_set_default(struct > > > i40e_hw *hw) > > > struct rte_flow *p_flow; > > > int ret; > > > uint8_t aq_fail = 0; > > > + int retries = 0; > > > > > > PMD_INIT_FUNC_TRACE(); > > > > > > @@ -1574,9 +1575,20 @@ void i40e_flex_payload_reg_set_default(struct > > > i40e_hw *hw) > > > /* disable uio intr before callback unregister */ > > > rte_intr_disable(intr_handle); > > > > > > - /* register callback func to eal lib */ > > > - rte_intr_callback_unregister(intr_handle, > > > - i40e_dev_interrupt_handler, dev); > > > + /* unregister callback func to eal lib */ > > > + do { > > > + ret = rte_intr_callback_unregister(intr_handle, > > > + i40e_dev_interrupt_handler, dev); > > > + if (ret >= 0) { > > > + break; > > > + } else if (ret != -EAGAIN) { > > > + PMD_INIT_LOG(ERR, > > > + "intr callback unregister failed: %d", > > > + ret); > > > + return ret; > > > + } > > > + i40e_msec_delay(100); > > > + } while (retries++ < 5); > > > > > > i40e_rm_ethtype_filter_list(pf); > > > i40e_rm_tunnel_filter_list(pf); > > > -- > > > 1.8.3.1 > > > > > > > Reviewed-by: Kirill Rybalchenko > > > Acked-by: Qi Zhang
[dpdk-dev] [PATCH v5 0/2] app/testpmd: add new commands to test new Tx/Rx offloads
Existed testpmd commands can't support per queue offload configuration. And there are different commands to enable or disable different offloading. This patch set add following commands to support new Tx/Rx offloading API test. To get Rx offload capability of a port, please run: testpmd > rx_offload get capability To get current Rx offload per queue and per port configuration of a port, run: tesstpmd > rx_offload get configuration To enable or disable a Rx per port offloading, please run: testpmd > rx_offload enable|disable per_port vlan_strip|ipv4_cksum|... This command will set|clear the associated bit in dev->dev_conf.rxmode.offloads for rte_eth_dev_configure and tx_conf->offloads of all Rx queues for rte_eth_rx_queue_setup( ). To enable or disable a Tx per port offloading, please run: testpmd > rx_offload enable|disable per_queue vlan_strip|ipv4_cksum|... Same commands like "tx_offload ..." are also added to support new Tx offload API test. --- v5: don't depend on enum types defined in rte_ethdev. v4: improve testpmd command per port offload to set or clear the port configuration and the queue configuration of all queues. v3: add enum rte_eth_rx_offload_type and enum rte_eth_tx_offload_type free memory of port->rx_offloads and port->tx_offloads when testpmd is existed v2: use rte_eth_dev_rx_offload_name() and rte_eth_dev_tx_offload_name(). remove static const strings of Rx/Tx offload names. Wei Dai (2): app/testpmd: add commands to test new Rx offload API app/testpmd: add commands to test new Tx offload API app/test-pmd/cmdline.c | 751 + app/test-pmd/testpmd.c | 34 ++- app/test-pmd/testpmd.h | 2 + 3 files changed, 783 insertions(+), 4 deletions(-) -- 2.7.5