Hi Florin, Thanks for your timely reply.
> > I’m doing testing with things like envoy and at 200 CPS cli is not > affected. Can be an issue related to configuration or maybe hardware. This is really caused by my environment. I did this test on the local virtual machine. When I test with a real physical environment and a Spirent tester,Low speed CPS, i.e. 1K req/s, will not affect cli. But when the CPS reaches 30000 req/s, the VPP cli command line becomes very stuck and cannot execute commands normally. My test machine is Intel (R) Xeon (R) CPU e5-2670, there are 24 cores. > > top - 14:44:52 up 3:06, 4 users, load average: 5.65, 4.96, 3.95 > Tasks: 311 total, 6 running, 305 sleeping, 0 stopped, 0 zombie > %Cpu0 : 0.0 us, 0.0 sy, 0.0 ni, 96.2 id, 0.0 wa, 0.0 hi, 3.8 si, > 0.0 st > %Cpu1 : 99.7 us, 0.3 sy, 0.0 ni, 0.0 id, 0.0 wa, 0.0 hi, 0.0 si, > 0.0 st > %Cpu2 : 99.7 us, 0.3 sy, 0.0 ni, 0.0 id, 0.0 wa, 0.0 hi, 0.0 si, > 0.0 st > %Cpu3 : 0.0 us, 0.0 sy, 0.0 ni,100.0 id, 0.0 wa, 0.0 hi, 0.0 si, > 0.0 st > %Cpu4 : 0.0 us, 0.0 sy, 0.0 ni,100.0 id, 0.0 wa, 0.0 hi, 0.0 si, > 0.0 st > %Cpu5 : 0.3 us, 0.0 sy, 0.0 ni, 99.7 id, 0.0 wa, 0.0 hi, 0.0 si, > 0.0 st > %Cpu6 : 79.1 us, 20.9 sy, 0.0 ni, 0.0 id, 0.0 wa, 0.0 hi, 0.0 si, > 0.0 st > %Cpu7 : 75.8 us, 24.2 sy, 0.0 ni, 0.0 id, 0.0 wa, 0.0 hi, 0.0 si, > 0.0 st > %Cpu8 : 75.7 us, 24.3 sy, 0.0 ni, 0.0 id, 0.0 wa, 0.0 hi, 0.0 si, > 0.0 st > %Cpu9 : 77.2 us, 22.8 sy, 0.0 ni, 0.0 id, 0.0 wa, 0.0 hi, 0.0 si, > 0.0 st > %Cpu10 : 0.0 us, 0.0 sy, 0.0 ni,100.0 id, 0.0 wa, 0.0 hi, 0.0 si, > 0.0 st > %Cpu11 : 0.0 us, 0.0 sy, 0.0 ni,100.0 id, 0.0 wa, 0.0 hi, 0.0 si, > 0.0 st > %Cpu12 : 0.0 us, 0.0 sy, 0.0 ni,100.0 id, 0.0 wa, 0.0 hi, 0.0 si, > 0.0 st > %Cpu13 : 0.0 us, 0.0 sy, 0.0 ni,100.0 id, 0.0 wa, 0.0 hi, 0.0 si, > 0.0 st > %Cpu14 : 0.0 us, 0.0 sy, 0.0 ni,100.0 id, 0.0 wa, 0.0 hi, 0.0 si, > 0.0 st > %Cpu15 : 0.7 us, 0.0 sy, 0.0 ni, 99.3 id, 0.0 wa, 0.0 hi, 0.0 si, > 0.0 st > %Cpu16 : 0.0 us, 0.0 sy, 0.0 ni,100.0 id, 0.0 wa, 0.0 hi, 0.0 si, > 0.0 st > %Cpu17 : 0.0 us, 0.0 sy, 0.0 ni,100.0 id, 0.0 wa, 0.0 hi, 0.0 si, > 0.0 st > %Cpu18 : 0.0 us, 0.0 sy, 0.0 ni,100.0 id, 0.0 wa, 0.0 hi, 0.0 si, > 0.0 st > %Cpu19 : 0.0 us, 0.0 sy, 0.0 ni,100.0 id, 0.0 wa, 0.0 hi, 0.0 si, > 0.0 st > %Cpu20 : 0.0 us, 0.0 sy, 0.0 ni,100.0 id, 0.0 wa, 0.0 hi, 0.0 si, > 0.0 st > %Cpu21 : 0.0 us, 0.0 sy, 0.0 ni,100.0 id, 0.0 wa, 0.0 hi, 0.0 si, > 0.0 st > %Cpu22 : 0.0 us, 0.0 sy, 0.0 ni,100.0 id, 0.0 wa, 0.0 hi, 0.0 si, > 0.0 st > %Cpu23 : 0.0 us, 0.0 sy, 0.0 ni,100.0 id, 0.0 wa, 0.0 hi, 0.0 si, > 0.0 st cpu1 is vpp main core,and cpu2 is vpp worker.cpu6-9 is nginx worker. > > Pressed send a bit too soon. I also tried a simple proxy scenario locally > with -c 500 and envoy and got about 490-500 CPS. Then did the same with -c > 1000 and got close to 1k req/s. > > I’m guessing performance is capped by how fast we manage to establish a > connection as the proxy will introduce a decent amount of latency and ab > needs to wait until the requests in flight are answered. > > Regards, > Florin > > >> On Dec 23, 2021, at 8:57 PM, Florin Coras < fcoras.li...@gmail.com > wrote: >> >> Hi Jie, >> >> Inline. >> >> >> >>> On Dec 23, 2021, at 6:34 PM, deng...@webray.com.cn wrote: >>> Hi Florin, >>> Thanks for your reply! My intention was to ask you on vpp-dev, but because >>> I am not familiar with vpp-dev, replay was selected as private email in >>> this conversation ( >>> https://lists.fd.io/g/vpp-dev/topic/81698902?p=%2C%2C%2C20%2C0%2C0%2C0%3A%3Arelevance%2C%2C%23hoststack%2C20%2C2%2C0%2C81698902%2Cct%3D1&ct=1 >>> ). >>> >> >> >> Have you registered with the mailing list? Maybe that’s why you’re hitting >> issues. >> >> >>> The VPP version I use is 21.10. When the RPS of HTTP is more than 200, the >>> utilization rate of the main core will be very high. I think active opens >>> is in the main thread, and each version should have this issues. >>> >> >> >> I’m doing testing with things like envoy and at 200 CPS cli is not >> affected. Can be an issue related to configuration or maybe hardware. >> >> >>> In our usage scenario (using VPP's host stack to speed up nginx reverse >>> proxy), since TCP active opens is generated by main thread, there will be >>> some problems under a large number of HTTP requests. >>> * The CPU utilization of main core is close to 100% , which will cause the >>> execution of CLI commands to be very slow. >>> >>> >> >> It’s 100% because main thread switches to polling mode. It’s not that all >> cpu is consumed just that main thread is not sleeping. >> >>> * Because active opens is in the main thread and barrier sync, the >>> scalability of multi-core performance is not very good. >>> >>> >> >> Connects do indeed rely on barrier syncs but they are paced so they should >> not starve the workers. >> > > Regards, Jie
-=-=-=-=-=-=-=-=-=-=-=- Links: You receive all messages sent to this group. View/Reply Online (#20666): https://lists.fd.io/g/vpp-dev/message/20666 Mute This Topic: https://lists.fd.io/mt/87933432/21656 Mute #hoststack:https://lists.fd.io/g/vpp-dev/mutehashtag/hoststack Mute #vpp-hoststack:https://lists.fd.io/g/vpp-dev/mutehashtag/vpp-hoststack Group Owner: vpp-dev+ow...@lists.fd.io Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub [arch...@mail-archive.com] -=-=-=-=-=-=-=-=-=-=-=-