OK, so crash is clearly happening in DPDK code so you will need somebody familiar with that driver to take a look.
— Damjan > On 07.03.2022., at 07:01, Xu, Ting <ting...@intel.com> wrote: > > Hi, Damjan > > Thanks for your help, and the backtrace from gdb is below (a file with same > content is attached for better format). I use the commit > ce4083ce48958d9d3956e8317445a5552780af1a (“dpdk: offloads cleanup”) to get > these info. The previous one commit 3b7ef512f190a506f62af53536b586b4800f66c1 > ("misc: fix the uninitialization error") does not cause the error. > > Thread 1 "vpp_main" received signal SIGSEGV, Segmentation fault. > 0x00007fff7211f958 in ice_sq_send_cmd_nolock (hw=0x0, cq=0x0, desc=0x0, > buf=0x0, buf_size=0, cd=0x0) at > ../src-dpdk/drivers/net/ice/base/ice_controlq.c:889 > 889 { > (gdb) bt > #0 0x00007fff7211f958 in ice_sq_send_cmd_nolock (hw=0x0, cq=0x0, desc=0x0, > buf=0x0, buf_size=0, cd=0x0) at > ../src-dpdk/drivers/net/ice/base/ice_controlq.c:889 > #1 0x00007fff721434f9 in ice_sq_send_cmd (hw=0x7fd2bf7f9b00, > cq=0x7fd2bf7fb5a0, desc=0x7fff6d361f40, buf=0x7fe2c025d000, buf_size=6, > cd=0x0) at ../src-dpdk/drivers/net/ice/base/ice_controlq.c:1076 > #2 0x00007fff721724bc in ice_sq_send_cmd_retry (hw=0x7fd2bf7f9b00, > cq=0x7fd2bf7fb5a0, desc=0x7fff6d361f40, buf=0x7fe2c025d000, buf_size=6, > cd=0x0) at ../src-dpdk/drivers/net/ice/base/ice_common.c:1415 > #3 0x00007fff72180687 in ice_aq_send_cmd (hw=0x7fd2bf7f9b00, > desc=0x7fff6d361f40, buf=0x7fe2c025d000, buf_size=6, cd=0x0) at > ../src-dpdk/drivers/net/ice/base/ice_common.c:1474 > #4 0x00007fff72181130 in ice_aq_alloc_free_res (hw=0x7fd2bf7f9b00, > num_entries=1, buf=0x7fe2c025d000, buf_size=6, opc=ice_aqc_opc_alloc_res, > cd=0x0) at ../src-dpdk/drivers/net/ice/base/ice_common.c:1810 > #5 0x00007fff72181255 in ice_alloc_hw_res (hw=0x7fd2bf7f9b00, type=96, > num=1, btm=false, res=0x7fff6d364452) at > ../src-dpdk/drivers/net/ice/base/ice_common.c:1840 > #6 0x00007fff72327d2c in ice_alloc_prof_id (hw=0x7fd2bf7f9b00, > blk=ICE_BLK_RSS, prof_id=0x7fff6d3644ba "5r") at > ../src-dpdk/drivers/net/ice/base/ice_flex_pipe.c:3305 > #7 0x00007fff72348519 in ice_add_prof (hw=0x7fd2bf7f9b00, blk=ICE_BLK_RSS, > id=17179875328, ptypes=0x7fe2c025ddbc "", attr=0x0, attr_cnt=0, > es=0x7fe2c025dc90, masks=0x7fe2c025dd5a) at > ../src-dpdk/drivers/net/ice/base/ice_flex_pipe.c:4980 > #8 0x00007fff72364b71 in ice_flow_add_prof_sync (hw=0x7fd2bf7f9b00, > blk=ICE_BLK_RSS, dir=ICE_FLOW_RX, prof_id=17179875328, segs=0x7fe2c025dec0, > segs_cnt=1 '\001', acts=0x0, acts_cnt=0 '\000', prof=0x7fff6d368fb8) at > ../src-dpdk/drivers/net/ice/base/ice_flow.c:2054 > #9 0x00007fff7236574a in ice_flow_add_prof (hw=0x7fd2bf7f9b00, > blk=ICE_BLK_RSS, dir=ICE_FLOW_RX, prof_id=17179875328, segs=0x7fe2c025dec0, > segs_cnt=1 '\001', acts=0x0, acts_cnt=0 '\000', prof=0x7fff6d368fb8) at > ../src-dpdk/drivers/net/ice/base/ice_flow.c:2371 > #10 0x00007fff7238bd74 in ice_add_rss_cfg_sync (hw=0x7fd2bf7f9b00, > vsi_handle=0, cfg=0x7fff6d369010) at > ../src-dpdk/drivers/net/ice/base/ice_flow.c:3884 > #11 0x00007fff7238beef in ice_add_rss_cfg (hw=0x7fd2bf7f9b00, vsi_handle=0, > cfg=0x7fff6d3690b0) at ../src-dpdk/drivers/net/ice/base/ice_flow.c:3937 > #12 0x00007fff724e6301 in ice_add_rss_cfg_wrap (pf=0x7fd2bf7fc7d0, vsi_id=0, > cfg=0x7fff6d3690b0) at ../src-dpdk/drivers/net/ice/ice_ethdev.c:2792 > #13 0x00007fff724e6457 in ice_rss_hash_set (pf=0x7fd2bf7fc7d0, rss_hf=12220) > at ../src-dpdk/drivers/net/ice/ice_ethdev.c:2834 > #14 0x00007fff724fc253 in ice_init_rss (pf=0x7fd2bf7fc7d0) at > ../src-dpdk/drivers/net/ice/ice_ethdev.c:3102 > #15 0x00007fff724fc369 in ice_dev_configure (dev=0x7fff746a0100 > <rte_eth_devices>) at ../src-dpdk/drivers/net/ice/ice_ethdev.c:3131 > #16 0x00007fff70d9c3e4 in rte_eth_dev_configure (port_id=0, nb_rx_q=8, > nb_tx_q=5, dev_conf=0x7fff6d36ecc0) at > ../src-dpdk/lib/ethdev/rte_ethdev.c:1578 > #17 0x00007fff73e10178 in dpdk_device_setup (xd=0x7fff7c8f4f00) at > /root/networking.dataplane.fdio.vpp/src/plugins/dpdk/device/common.c:156 > #18 0x00007fff73e47b84 in dpdk_lib_init (dm=0x7fff74691f58 <dpdk_main>) at > /root/networking.dataplane.fdio.vpp/src/plugins/dpdk/device/init.c:582 > #19 0x00007fff73e459f4 in dpdk_process (vm=0x7fff76800680, rt=0x7fff76e191c0, > f=0x0) at > /root/networking.dataplane.fdio.vpp/src/plugins/dpdk/device/init.c:1499 > #20 0x00007ffff6e7033d in vlib_process_bootstrap (_a=140735062407352) at > /root/networking.dataplane.fdio.vpp/src/vlib/main.c:1235 > #21 0x00007ffff6d0ebf8 in clib_calljmp () at > /root/networking.dataplane.fdio.vpp/src/vppinfra/longjmp.S:123 > #22 0x00007fff6f66f8b0 in ?? () > #23 0x00007ffff6e6fd5f in vlib_process_startup (vm=0x7fff76800680, > p=0x7fff76e191c0, f=0x0) at > /root/networking.dataplane.fdio.vpp/src/vlib/main.c:1260 > #24 0x00007ffff6e6b4fa in dispatch_process (vm=0x7fff76800680, > p=0x7fff76e191c0, f=0x0, last_time_stamp=2826656650676320) at > /root/networking.dataplane.fdio.vpp/src/vlib/main.c:1316 > #25 0x00007ffff6e6bdf5 in vlib_main_or_worker_loop (vm=0x7fff76800680, > is_main=1) at /root/networking.dataplane.fdio.vpp/src/vlib/main.c:1515 > #26 0x00007ffff6e6e45a in vlib_main_loop (vm=0x7fff76800680) at > /root/networking.dataplane.fdio.vpp/src/vlib/main.c:1728 > #27 0x00007ffff6e6e242 in vlib_main (vm=0x7fff76800680, input=0x7fff6f66ffa8) > at /root/networking.dataplane.fdio.vpp/src/vlib/main.c:2017 > #28 0x00007ffff6ed02ce in thread0 (arg=140735181489792) at > /root/networking.dataplane.fdio.vpp/src/vlib/unix/main.c:671 > #29 0x00007ffff6d0ebf8 in clib_calljmp () at > /root/networking.dataplane.fdio.vpp/src/vppinfra/longjmp.S:123 > #30 0x00007fffffffc9f0 in ?? () > #31 0x00007ffff6ecfdfe in vlib_unix_main (argc=59, argv=0x446500) at > /root/networking.dataplane.fdio.vpp/src/vlib/unix/main.c:751 > #32 0x0000000000406b23 in main (argc=59, argv=0x446500) at > /root/networking.dataplane.fdio.vpp/src/vpp/vnet/main.c:342 > > Please tell me if any more info needed > > Best Regards, > Xu Ting > >> -----Original Message----- >> From: Damjan Marion <dmar...@me.com> >> Sent: Friday, March 4, 2022 9:13 PM >> To: Xu, Ting <ting...@intel.com> >> Cc: vpp-dev@lists.fd.io >> Subject: Re: [vpp-dev] Segmentation fault when dpdk number-rx-queues > 1 >> in startup.conf >> >> >> Dear Xu Ting, >> >> Data you provided is not sufficient to help you. >> i.e. providing backtrace may help us understand where problem is. >> >> — >> Damjan >> >> >> >>> On 03.03.2022., at 08:10, Xu, Ting <ting...@intel.com> wrote: >>> >>> Hi, all >>> >>> I meet one issue that when I set dpdk rx queue number larger than 1 >> (which also enables RSS) in startup.conf, for example: >>> >>> dev default { >>> # Number of receive queues, enables RSS >>> # Default is 1 >>> num-rx-queues 2 >>> } >>> >>> When start VPP, it will meet segmentation fault, the error log is: >>> >>> …… >>> dpdk [debug ]: [0] interface dpdk_eth0 created >>> interface/rx-queue [debug ]: set_input_node: node dpdk-input for >> interface dpdk_eth0 >>> interface/rx-queue [debug ]: register: interface dpdk_eth0 queue-id 0 >> thread 1 >>> interface/rx-queue [debug ]: register: interface dpdk_eth0 queue-id 1 >> thread 2 >>> dpdk [debug ]: [0] configuring device name: 0000:d8:00.0, numa: 1, driver: >> net_ice, bus: pci >>> dpdk [debug ]: [0] Supported RX offloads: vlan-strip ipv4-cksum udp-cksum >> tcp-cksum qinq-strip >>> outer-ipv4-cksum vlan-filter vlan-extend scatter >>> timestamp keep-crc rss-hash >>> dpdk [debug ]: [0] Configured RX offloads: ipv4-cksum scatter >>> dpdk [debug ]: [0] Supported TX offloads: vlan-insert ipv4-cksum udp- >> cksum tcp-cksum sctp-cksum >>> tcp-tso outer-ipv4-cksum qinq-insert multi-segs >>> mbuf-fast-free outer-udp-cksum >>> dpdk [debug ]: [0] Configured TX offloads: ipv4-cksum udp-cksum tcp- >> cksum multi-segs >>> Segmentation fault (core dumped) >>> >>> I think I find the bad commit: >> ce4083ce48958d9d3956e8317445a5552780af1a (“dpdk: offloads cleanup”) >>> Does anyone also meet issue? Is there any solution to it? Thanks! >>> >>> Best Regards, >>> Xu Ting >>> >>> > > <backtrace.txt>
-=-=-=-=-=-=-=-=-=-=-=- Links: You receive all messages sent to this group. View/Reply Online (#20972): https://lists.fd.io/g/vpp-dev/message/20972 Mute This Topic: https://lists.fd.io/mt/89520993/21656 Group Owner: vpp-dev+ow...@lists.fd.io Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub [arch...@mail-archive.com] -=-=-=-=-=-=-=-=-=-=-=-