Hi Wenzuo, On Mon, Jan 15, 2018 at 08:31:52AM +0000, Lu, Wenzhuo wrote: > Hi Simon, > > > > -----Original Message----- > > From: wei.guo.si...@gmail.com [mailto:wei.guo.si...@gmail.com] > > Sent: Wednesday, January 10, 2018 4:00 PM > > To: Lu, Wenzhuo <wenzhuo...@intel.com> > > Cc: dev@dpdk.org; Simon Guo <wei.guo.si...@gmail.com> > > Subject: [PATCH v4] app/testpmd: add option ring-bind-lcpu to bind Q with > > CPU > > > > From: Simon Guo <wei.guo.si...@gmail.com> > > > > Currently the rx/tx queue is allocated from the buffer pool on socket of: > > - port's socket if --port-numa-config specified > > - or ring-numa-config setting per port > > > > All the above will "bind" queue to single socket per port configuration. > > But it can actually archieve better performance if one port's queue can be > > spread across multiple NUMA nodes, and the rx/tx queue is allocated per > > lcpu socket. > > > > This patch adds a new option "--ring-bind-lcpu"(no parameter). With this, > > testpmd can utilize the PCI-e bus bandwidth on another NUMA nodes. > > > > When --port-numa-config or --ring-numa-config option is specified, this -- > > ring-bind-lcpu option will be suppressed. > > > > Test result: > > 64bytes package, running in PowerPC with Mellanox > > CX-4 card, single port(100G), with 8 cores, fw mode: > > - Without this patch: 52.5Mpps throughput > > - With this patch: 66Mpps throughput > > ~25% improvement > > > > Signed-off-by: Simon Guo <wei.guo.si...@gmail.com> > This patch is fine. But for the new parameter, please update this doc, > run_app.rst. Thanks. ah...yes. I will rework on v5.
Thanks, - Simon