On Wed, Jun 17, 2015 at 12:47:28PM -0300, Flavio Leitner wrote: > On Wed, Jun 17, 2015 at 12:28:23PM -0300, Flavio Leitner wrote: > > On Wed, Jun 17, 2015 at 10:54:25AM +0800, Dongjun wrote: > > > > > > > > > On 2015/6/17 1:44, Daniele Di Proietto wrote: > > > > > > > >On 16/06/2015 07:40, "Pravin Shelar" <pshe...@nicira.com> wrote: > > > > > > > >>On Mon, Jun 8, 2015 at 7:42 PM, Pravin Shelar <pshe...@nicira.com> > > > >>wrote: > > > >>>On Mon, Jun 8, 2015 at 6:13 PM, Wei li <l...@dtdream.com> wrote: > > > >>>>When tx queue is shared among CPUS,the pkts always be flush in > > > >>>>'netdev_dpdk_eth_send' > > > >>>>So it is unnecessarily for flushing in netdev_dpdk_rxq_recv > > > >>>>Otherwise tx will be accessed without locking > > > >>>> > > > >>>>Signed-off-by: Wei li <l...@dtdream.com> > > > >>>>--- > > > >>>>v1->v2: fix typo > > > >>>> > > > >>>> lib/netdev-dpdk.c | 7 +++++-- > > > >>>> 1 file changed, 5 insertions(+), 2 deletions(-) > > > >>>> > > > >>>>diff --git a/lib/netdev-dpdk.c b/lib/netdev-dpdk.c > > > >>>>index 63243d8..1e0a483 100644 > > > >>>>--- a/lib/netdev-dpdk.c > > > >>>>+++ b/lib/netdev-dpdk.c > > > >>>>@@ -892,8 +892,11 @@ netdev_dpdk_rxq_recv(struct netdev_rxq *rxq_, > > > >>>>struct dp_packet **packets, > > > >>>> int nb_rx; > > > >>>> > > > >>>> /* There is only one tx queue for this core. Do not flush other > > > >>>>- * queueus. */ > > > >>>>- if (rxq_->queue_id == rte_lcore_id()) { > > > >>>>+ * queues. > > > >>>>+ * Do not flush tx queue which is shared among CPUs > > > >>>>+ * since it is always flushed */ > > > >>>>+ if (rxq_->queue_id == rte_lcore_id() && > > > >>>>+ OVS_LIKELY(!dev->txq_needs_locking)) { > > > >>>> dpdk_queue_flush(dev, rxq_->queue_id); > > > >>>> } > > > >>>Patch looks good, But Daniele has plan to get rid of queue flushing > > > >>>logic. do you see problem with doing this? > > > >>Daniele, > > > >>I am not sure if we should fix this queue flushing logic if we are > > > >>going to remove it. So I would like to discuss it first. > > > >>You mentioned that the removing flushing logic results in performance > > > >>drop. Can you explain it? How much is performance drop and which is > > > >>the case? > > > >Hi Pravin, > > > > > > > >sorry for the late reply. I suspected that removing the queues in > > > >netdev-dpdk > > > >was going to have a performance impact in the following scenario: a batch > > > >of > > > >packets from different megaflows with the same action (output on port 1). > > > >Here's what happens: > > > > > > > >1) With the queues in netdev-dpdk. dpif-netdev calls netdev_send() for > > > >each > > > > packet. netdev_dpdk_send() just copies the pointer in the queue. > > > > Before > > > > receiving the next batch dpdk_queue_flush() call rte_eth_tx_burst() > > > > with > > > > a whole batch of packets > > > >2) Without the queues in netdev-dpdk. dpif-netdev calls netdev_send() for > > > >each > > > > packet. netdev_dpdk_send() calls rte_eth_tx_burst() for each packet. > > > > > > > >Scenario 2 could be slower because rte_eth_tx_burst() is called for each > > > >packet > > > >(instead of each batch). After some testing this turns out not to be the > > > >case. > > > >It appears that the bottleneck is not rte_eth_tx_burst(), and copying the > > > >pointers > > > >in the temporary queue makes the code slower. > > > Hi Daniele, > > > I get something else impacting the performance too. Here it is: > > > In netdev_dpdk_rxq_recv(), the flush condition is "txq->count != 0", and > > > in > > > dpdk_queue_pkts(), the flush condition is "txq->count == MAX_TX_QUEUE_LEN" > > > or "(rte_get_timer_cycles() - txq->tsc) >= DRAIN_TSC". > > > They don't match. Our thoughts is to get a batch flush, but > > > netdev_dpdk_rxq_recv() may do it earlier sometime. > > > > > > I did some local change, stealing the cycle interval from > > > dpdk_queue_pkts(). > > > I got performance improvement, about at least 10% from guest VNIC to dpdk > > > phy NIC to outside. > > > I didn't compare the performance to "Scenario 2" for "Scenario 2" > > > modification is not that easy. :( > > > > > > $ git diff > > > diff --git a/lib/netdev-dpdk.c b/lib/netdev-dpdk.c > > > index 3af1ee7..98d33c3 100644 > > > --- a/lib/netdev-dpdk.c > > > +++ b/lib/netdev-dpdk.c > > > @@ -929,10 +929,13 @@ netdev_dpdk_rxq_recv(struct netdev_rxq *rxq_, struct > > > dp_packet **packets, > > > > > > /* There is only one tx queue for this core. Do not flush other > > > * queueus. */ > > > - if (rxq_->queue_id == rte_lcore_id()) { > > > - dpdk_queue_flush(dev, rxq_->queue_id); > > > - } > > > + if (rxq_->queue_id == rte_lcore_id() && > > > OVS_LIKELY(!dev->txq_needs_locking)) { > > > + struct dpdk_tx_queue *txq = &dev->tx_q[rxq_->queue_id]; > > > > > > + if (txq->count != 0 && (rte_get_timer_cycles() - txq->tsc) >= > > > DRAIN_TSC) { > > > + dpdk_queue_flush__(dev, rxq_->queue_id); // instead of > > > dpdk_queue_flush > > > + } > > > + } > > > nb_rx = rte_eth_rx_burst(rx->port_id, rxq_->queue_id, > > > (struct rte_mbuf **) packets, > > > NETDEV_MAX_BURST);
Before patch: 3.5Mpps After patch: 3.3Mpps fbl _______________________________________________ dev mailing list dev@openvswitch.org http://openvswitch.org/mailman/listinfo/dev