https://bugs.dpdk.org/show_bug.cgi?id=1043
Bug ID: 1043 Summary: [dpdk-22.07]vm2vm_virtio_net_perf_cbdma/vm2vm_split_ri ng_iperf_with_tso_and_cbdma_enable: iperf test no data between 2 VMs Product: DPDK Version: 22.03 Hardware: All OS: All Status: UNCONFIRMED Severity: normal Priority: Normal Component: vhost/virtio Assignee: dev@dpdk.org Reporter: weix.l...@intel.com Target Milestone: --- [Environment] DPDK version: Use make showversion or for a non-released version: git remote -v && git show-ref --heads commit 7cac53f205ebd04d8ebd3ee6a9dd84f698d4ada3 (HEAD -> main, tag: v22.07-rc2, origin/main, origin/HEAD) Author: Thomas Monjalon <tho...@monjalon.net> Date: Mon Jun 27 04:03:44 2022 +0200 version: 22.07-rc2 Signed-off-by: Thomas Monjalon <tho...@monjalon.net> Other software versions: QEMU-7.0.0 OS: Ubuntu 22.04 LTS/Linux 5.15.45-051545-generic Compiler: gcc version 11.2.0 (Ubuntu 11.2.0-19ubuntu1) Hardware platform: Intel(R) Xeon(R) Platinum 8280M CPU @ 2.70GHz NIC hardware: N/A NIC firmware: N/A [Test Setup] Steps to reproduce List the steps to reproduce the issue. 1. Bind 2 CBDMA channels to vfio-pci: dpdk-devbind.py --force --bind=vfio-pci 0000:80:04.0 0000:80:04.1 2. Start vhost-tetpmd: x86_64-native-linuxapp-gcc/app/dpdk-testpmd -l 28-36 -n 4 -a 0000:80:04.0 -a 0000:80:04.1 --file-prefix=vhost_247798_20220628141037 --vdev 'net_vhost0,iface=vhost-net0,queues=1,tso=1,dmas=[txq0;rxq0]' --vdev 'net_vhost1,iface=vhost-net1,queues=1,tso=1,dmas=[txq0;rxq0]' --iova=va -- -i --nb-cores=2 --txd=1024 --rxd=1024 --txq=1 --rxq=1 --lcore-dma=[lcore29@0000:80:04.0,lcore30@0000:80:04.1] start 3. Start VM0: taskset -c 20,21,22,23,24,25,26,27 /home/QEMU/qemu-7.0.0/bin/qemu-system-x86_64 -name vm0 -enable-kvm -pidfile /tmp/.vm0.pid -daemonize -monitor unix:/tmp/vm0_monitor.sock,server,nowait -netdev user,id=nttsip1,hostfwd=tcp:10.239.252.220:6000-:22 -device e1000,netdev=nttsip1 -chardev socket,id=char0,path=/root/dpdk/vhost-net0 -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce,queues=1 -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:01,disable-modern=false,mrg_rxbuf=on,csum=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on -cpu host -smp 8 -m 16384 -object memory-backend-file,id=mem,size=16384M,mem-path=/mnt/huge,share=on -numa node,memdev=mem -mem-prealloc -chardev socket,path=/tmp/vm0_qga0.sock,server,nowait,id=vm0_qga0 -device virtio-serial -device virtserialport,chardev=vm0_qga0,name=org.qemu.guest_agent.0 -vnc :4 -drive file=/home/image/ubuntu2004.img 4. Start VM1: taskset -c 48,49,50,51,52,53,54,55 /home/QEMU/qemu-7.0.0/bin/qemu-system-x86_64 -name vm1 -enable-kvm -pidfile /tmp/.vm1.pid -daemonize -monitor unix:/tmp/vm1_monitor.sock,server,nowait -device e1000,netdev=nttsip1 -netdev user,id=nttsip1,hostfwd=tcp:10.239.252.220:6001-:22 -chardev socket,id=char0,path=/root/dpdk/vhost-net1 -netdev type=vhost-user,id=netdev0,chardev=char0,vhostforce,queues=1 -device virtio-net-pci,netdev=netdev0,mac=52:54:00:00:00:02,disable-modern=false,mrg_rxbuf=on,csum=on,guest_csum=on,host_tso4=on,guest_tso4=on,guest_ecn=on -cpu host -smp 8 -m 16384 -object memory-backend-file,id=mem,size=16384M,mem-path=/mnt/huge,share=on -numa node,memdev=mem -mem-prealloc -chardev socket,path=/tmp/vm1_qga0.sock,server,nowait,id=vm1_qga0 -device virtio-serial -device virtserialport,chardev=vm1_qga0,name=org.qemu.guest_agent.0 -vnc :5 -drive file=/home/image/ubuntu2004_2.img 5. SSH login VM0 and VM1 to config IP: [VM0] ssh root@10.239.252.220 -p 6000 ifconfig ens4 1.1.1.1 [VM1] ssh root@10.239.252.220 -p 6001 ifconfig ens4 1.1.1.2 6. Use iperf test tool to test between 2 VMs [VM0] iperf -s -i 1 [VM1] iperf -c 1.1.1.1 -i 1 -t 60 Show the output from the previous commands. There is no any data. [Expected Result] Explain what is the expected result in text or as an example output: root@virtiovm:~# iperf -c 1.1.1.1 -i 1 -t 60 ------------------------------------------------------------ Client connecting to 1.1.1.1, TCP port 5001 TCP window size: 85.0 KByte (default) ------------------------------------------------------------ [ 1] local 1.1.1.2 port 42240 connected with 1.1.1.1 port 5001 [ ID] Interval Transfer Bandwidth [ 1] 0.0000-1.0000 sec 2.45 GBytes 21.0 Gbits/sec [ 1] 1.0000-2.0000 sec 2.42 GBytes 20.8 Gbits/sec [ 1] 2.0000-3.0000 sec 2.39 GBytes 20.6 Gbits/sec [ 1] 3.0000-4.0000 sec 2.41 GBytes 20.7 Gbits/sec [ 1] 4.0000-5.0000 sec 2.33 GBytes 20.0 Gbits/sec [ 1] 5.0000-6.0000 sec 2.40 GBytes 20.6 Gbits/sec [ 1] 6.0000-7.0000 sec 2.39 GBytes 20.6 Gbits/sec [ 1] 7.0000-8.0000 sec 2.41 GBytes 20.7 Gbits/sec [ 1] 8.0000-9.0000 sec 2.35 GBytes 20.2 Gbits/sec [ 1] 9.0000-10.0000 sec 2.40 GBytes 20.6 Gbits/sec [ 1] 10.0000-11.0000 sec 2.37 GBytes 20.4 Gbits/sec [ 1] 11.0000-12.0000 sec 2.41 GBytes 20.7 Gbits/sec [ 1] 12.0000-13.0000 sec 2.42 GBytes 20.7 Gbits/sec [ 1] 13.0000-14.0000 sec 2.41 GBytes 20.7 Gbits/sec [ 1] 14.0000-15.0000 sec 2.40 GBytes 20.6 Gbits/sec [ 1] 15.0000-16.0000 sec 2.40 GBytes 20.6 Gbits/sec [ 1] 16.0000-17.0000 sec 2.54 GBytes 21.8 Gbits/sec [ 1] 17.0000-18.0000 sec 2.41 GBytes 20.7 Gbits/sec [ 1] 18.0000-19.0000 sec 2.39 GBytes 20.6 Gbits/sec [ 1] 19.0000-20.0000 sec 2.40 GBytes 20.6 Gbits/sec [ 1] 20.0000-21.0000 sec 2.40 GBytes 20.6 Gbits/sec [ 1] 21.0000-22.0000 sec 2.42 GBytes 20.8 Gbits/sec [ 1] 22.0000-23.0000 sec 2.51 GBytes 21.6 Gbits/sec [ 1] 23.0000-24.0000 sec 2.40 GBytes 20.6 Gbits/sec [ 1] 24.0000-25.0000 sec 2.41 GBytes 20.7 Gbits/sec [ 1] 25.0000-26.0000 sec 2.43 GBytes 20.9 Gbits/sec [ 1] 26.0000-27.0000 sec 2.41 GBytes 20.7 Gbits/sec [ 1] 27.0000-28.0000 sec 2.40 GBytes 20.6 Gbits/sec [ 1] 28.0000-29.0000 sec 2.41 GBytes 20.7 Gbits/sec [ 1] 29.0000-30.0000 sec 2.38 GBytes 20.4 Gbits/sec [ 1] 30.0000-31.0000 sec 2.38 GBytes 20.5 Gbits/sec [ 1] 31.0000-32.0000 sec 2.35 GBytes 20.2 Gbits/sec [ 1] 32.0000-33.0000 sec 2.39 GBytes 20.5 Gbits/sec [ 1] 33.0000-34.0000 sec 2.39 GBytes 20.5 Gbits/sec [ 1] 34.0000-35.0000 sec 2.41 GBytes 20.7 Gbits/sec [ 1] 35.0000-36.0000 sec 2.33 GBytes 20.0 Gbits/sec [ 1] 36.0000-37.0000 sec 2.41 GBytes 20.7 Gbits/sec [ 1] 37.0000-38.0000 sec 2.40 GBytes 20.6 Gbits/sec [ 1] 38.0000-39.0000 sec 2.41 GBytes 20.7 Gbits/sec [ 1] 39.0000-40.0000 sec 2.39 GBytes 20.5 Gbits/sec [ 1] 40.0000-41.0000 sec 2.45 GBytes 21.0 Gbits/sec [ 1] 41.0000-42.0000 sec 2.40 GBytes 20.6 Gbits/sec [ 1] 42.0000-43.0000 sec 2.40 GBytes 20.6 Gbits/sec [ 1] 43.0000-44.0000 sec 2.40 GBytes 20.7 Gbits/sec [ 1] 44.0000-45.0000 sec 2.41 GBytes 20.7 Gbits/sec [ 1] 45.0000-46.0000 sec 2.40 GBytes 20.6 Gbits/sec [ 1] 46.0000-47.0000 sec 2.38 GBytes 20.5 Gbits/sec [ 1] 47.0000-48.0000 sec 2.45 GBytes 21.0 Gbits/sec [ 1] 48.0000-49.0000 sec 2.42 GBytes 20.8 Gbits/sec [ 1] 49.0000-49.6028 sec 1.49 GBytes 21.2 Gbits/sec [ 1] 0.0000-49.6028 sec 119 GBytes 20.7 Gbits/sec root@virtiovm:~# Regression Is this issue a regression: (Y/N) Y Version the regression was introduced: Specify git id if known. Bad commit: commit 3a6ee8dafb21d7a55af59b573195a9dc18732476 Author: Maxime Coquelin <maxime.coque...@redhat.com> Date: Wed Jun 8 14:49:43 2022 +0200 net/vhost: enable compliant offloading mode This patch enables the compliant offloading flags mode by default, which prevents the Rx path to set Tx offload flags, which is illegal. A new legacy-ol-flags devarg is introduced to enable the legacy behaviour. Signed-off-by: Maxime Coquelin <maxime.coque...@redhat.com> Reviewed-by: Chenbo Xia <chenbo....@intel.com> -- You are receiving this mail because: You are the assignee for the bug.