> -----Original Message----- > From: David Christensen <d...@linux.vnet.ibm.com> > Sent: Friday, October 16, 2020 1:50 AM > To: dev@dpdk.org; maxime.coque...@redhat.com; Xia, Chenbo > <chenbo....@intel.com>; Wang, Zhihong <zhihong.w...@intel.com> > Cc: sta...@dpdk.org; David Christensen <d...@linux.vnet.ibm.com> > Subject: [PATCH v3] net/vhost: fix xstats wrong after clearing stats > > The PMD API allows stats and xstats values to be cleared separately. > This is a problem for the vhost PMD since some of the xstats values are > derived from existing stats values. For example: > > testpmd> show port xstats all > ... > tx_unicast_packets: 17562959 > ... > testpmd> clear port stats all > ... > show port xstats all > ... > tx_unicast_packets: 18446744073709551615 > ... > > Modify the driver so that stats and xstats values are stored, updated, > and cleared separately. > > Fixes: 4d6cf2ac93dc ("net/vhost: add extended statistics") > Cc: sta...@dpdk.org > > Signed-off-by: David Christensen <d...@linux.vnet.ibm.com> > --- > v3: > * Modified comment that unicast packets include unsent packets > * Change Cc: to sta...@dpdk.org > v2: > * Removed newly unused vq loops > * Added "fixes" message > * Renamed vhost_count_multicast_broadcast to vhost_count_xcast_packets > > drivers/net/vhost/rte_eth_vhost.c | 76 +++++++++++++++---------------- > 1 file changed, 38 insertions(+), 38 deletions(-) > > diff --git a/drivers/net/vhost/rte_eth_vhost.c > b/drivers/net/vhost/rte_eth_vhost.c > index e55278af6..886b3afe0 100644 > --- a/drivers/net/vhost/rte_eth_vhost.c > +++ b/drivers/net/vhost/rte_eth_vhost.c > @@ -73,6 +73,9 @@ enum vhost_xstats_pkts { > VHOST_BROADCAST_PKT, > VHOST_MULTICAST_PKT, > VHOST_UNICAST_PKT, > + VHOST_PKT, > + VHOST_BYTE, > + VHOST_MISSED_PKT, > VHOST_ERRORS_PKT, > VHOST_ERRORS_FRAGMENTED, > VHOST_ERRORS_JABBER, > @@ -149,11 +152,11 @@ struct vhost_xstats_name_off { > /* [rx]_is prepended to the name string here */ > static const struct vhost_xstats_name_off vhost_rxport_stat_strings[] = { > {"good_packets", > - offsetof(struct vhost_queue, stats.pkts)}, > + offsetof(struct vhost_queue, stats.xstats[VHOST_PKT])}, > {"total_bytes", > - offsetof(struct vhost_queue, stats.bytes)}, > + offsetof(struct vhost_queue, stats.xstats[VHOST_BYTE])}, > {"missed_pkts", > - offsetof(struct vhost_queue, stats.missed_pkts)}, > + offsetof(struct vhost_queue, stats.xstats[VHOST_MISSED_PKT])}, > {"broadcast_packets", > offsetof(struct vhost_queue, stats.xstats[VHOST_BROADCAST_PKT])}, > {"multicast_packets", > @@ -189,11 +192,11 @@ static const struct vhost_xstats_name_off > vhost_rxport_stat_strings[] = { > /* [tx]_ is prepended to the name string here */ > static const struct vhost_xstats_name_off vhost_txport_stat_strings[] = { > {"good_packets", > - offsetof(struct vhost_queue, stats.pkts)}, > + offsetof(struct vhost_queue, stats.xstats[VHOST_PKT])}, > {"total_bytes", > - offsetof(struct vhost_queue, stats.bytes)}, > + offsetof(struct vhost_queue, stats.xstats[VHOST_BYTE])}, > {"missed_pkts", > - offsetof(struct vhost_queue, stats.missed_pkts)}, > + offsetof(struct vhost_queue, stats.xstats[VHOST_MISSED_PKT])}, > {"broadcast_packets", > offsetof(struct vhost_queue, stats.xstats[VHOST_BROADCAST_PKT])}, > {"multicast_packets", > @@ -287,23 +290,6 @@ vhost_dev_xstats_get(struct rte_eth_dev *dev, struct > rte_eth_xstat *xstats, > if (n < nxstats) > return nxstats; > > - for (i = 0; i < dev->data->nb_rx_queues; i++) { > - vq = dev->data->rx_queues[i]; > - if (!vq) > - continue; > - vq->stats.xstats[VHOST_UNICAST_PKT] = vq->stats.pkts > - - (vq->stats.xstats[VHOST_BROADCAST_PKT] > - + vq->stats.xstats[VHOST_MULTICAST_PKT]); > - } > - for (i = 0; i < dev->data->nb_tx_queues; i++) { > - vq = dev->data->tx_queues[i]; > - if (!vq) > - continue; > - vq->stats.xstats[VHOST_UNICAST_PKT] = vq->stats.pkts > - + vq->stats.missed_pkts > - - (vq->stats.xstats[VHOST_BROADCAST_PKT] > - + vq->stats.xstats[VHOST_MULTICAST_PKT]); > - } > for (t = 0; t < VHOST_NB_XSTATS_RXPORT; t++) { > xstats[count].value = 0; > for (i = 0; i < dev->data->nb_rx_queues; i++) { > @@ -334,7 +320,7 @@ vhost_dev_xstats_get(struct rte_eth_dev *dev, struct > rte_eth_xstat *xstats, > } > > static inline void > -vhost_count_multicast_broadcast(struct vhost_queue *vq, > +vhost_count_xcast_packets(struct vhost_queue *vq, > struct rte_mbuf *mbuf) > { > struct rte_ether_addr *ea = NULL; > @@ -346,20 +332,27 @@ vhost_count_multicast_broadcast(struct vhost_queue > *vq, > pstats->xstats[VHOST_BROADCAST_PKT]++; > else > pstats->xstats[VHOST_MULTICAST_PKT]++; > + } else { > + pstats->xstats[VHOST_UNICAST_PKT]++; > } > } > > static void > -vhost_update_packet_xstats(struct vhost_queue *vq, > - struct rte_mbuf **bufs, > - uint16_t count) > +vhost_update_packet_xstats(struct vhost_queue *vq, struct rte_mbuf **bufs, > + uint16_t count, uint64_t nb_bytes, > + uint64_t nb_missed) > { > uint32_t pkt_len = 0; > uint64_t i = 0; > uint64_t index; > struct vhost_stats *pstats = &vq->stats; > > + pstats->xstats[VHOST_BYTE] += nb_bytes; > + pstats->xstats[VHOST_MISSED_PKT] += nb_missed; > + pstats->xstats[VHOST_UNICAST_PKT] += nb_missed; > + > for (i = 0; i < count ; i++) { > + pstats->xstats[VHOST_PKT]++; > pkt_len = bufs[i]->pkt_len; > if (pkt_len == 64) { > pstats->xstats[VHOST_64_PKT]++; > @@ -375,7 +368,7 @@ vhost_update_packet_xstats(struct vhost_queue *vq, > else if (pkt_len > 1522) > pstats->xstats[VHOST_1523_TO_MAX_PKT]++; > } > - vhost_count_multicast_broadcast(vq, bufs[i]); > + vhost_count_xcast_packets(vq, bufs[i]); > } > } > > @@ -385,6 +378,7 @@ eth_vhost_rx(void *q, struct rte_mbuf **bufs, uint16_t > nb_bufs) > struct vhost_queue *r = q; > uint16_t i, nb_rx = 0; > uint16_t nb_receive = nb_bufs; > + uint64_t nb_bytes = 0; > > if (unlikely(rte_atomic32_read(&r->allow_queuing) == 0)) > return 0; > @@ -419,10 +413,11 @@ eth_vhost_rx(void *q, struct rte_mbuf **bufs, > uint16_t nb_bufs) > if (r->internal->vlan_strip) > rte_vlan_strip(bufs[i]); > > - r->stats.bytes += bufs[i]->pkt_len; > + nb_bytes += bufs[i]->pkt_len; > } > > - vhost_update_packet_xstats(r, bufs, nb_rx); > + r->stats.bytes += nb_bytes; > + vhost_update_packet_xstats(r, bufs, nb_rx, nb_bytes, 0); > > out: > rte_atomic32_set(&r->while_queuing, 0); > @@ -436,6 +431,8 @@ eth_vhost_tx(void *q, struct rte_mbuf **bufs, uint16_t > nb_bufs) > struct vhost_queue *r = q; > uint16_t i, nb_tx = 0; > uint16_t nb_send = 0; > + uint64_t nb_bytes = 0; > + uint64_t nb_missed = 0; > > if (unlikely(rte_atomic32_read(&r->allow_queuing) == 0)) > return 0; > @@ -476,20 +473,23 @@ eth_vhost_tx(void *q, struct rte_mbuf **bufs, > uint16_t nb_bufs) > break; > } > > + for (i = 0; likely(i < nb_tx); i++) > + nb_bytes += bufs[i]->pkt_len; > + > + nb_missed = nb_bufs - nb_tx; > + > r->stats.pkts += nb_tx; > + r->stats.bytes += nb_bytes; > r->stats.missed_pkts += nb_bufs - nb_tx; > > - for (i = 0; likely(i < nb_tx); i++) > - r->stats.bytes += bufs[i]->pkt_len; > - > - vhost_update_packet_xstats(r, bufs, nb_tx); > + vhost_update_packet_xstats(r, bufs, nb_tx, nb_bytes, nb_missed); > > - /* According to RFC2863 page42 section ifHCOutMulticastPkts and > - * ifHCOutBroadcastPkts, the counters "multicast" and "broadcast" > - * are increased when packets are not transmitted successfully. > + /* According to RFC2863, ifHCOutUcastPkts, ifHCOutMulticastPkts and > + * ifHCOutBroadcastPkts counters are increased when packets are not > + * transmitted successfully. > */ > for (i = nb_tx; i < nb_bufs; i++) > - vhost_count_multicast_broadcast(r, bufs[i]); > + vhost_count_xcast_packets(r, bufs[i]); > > for (i = 0; likely(i < nb_tx); i++) > rte_pktmbuf_free(bufs[i]); > -- > 2.18.4
Reviewed-by: Chenbo Xia <chenbo....@intel.com>