Yes, mpls routes and ip routes that push label are not MP safe. I fixed this in our node that programs these routes. Fix is taking the barrier lock before programming such routes and releasing it once done.
-Rajith On Wed, 13 Oct 2021 at 1:12 PM, Stanislav Zaikin <zsta...@gmail.com> wrote: > Hi Rajith, > > Did you find the root cause? I'm facing the same problem, the load_balance > element in the pool seems to be good. And event a helper from gdbinit says > that the element isn't free: > > #0 __GI_raise (sig=sig@entry=6) at ../sysdeps/unix/sysv/linux/raise.c:51 > 51 ../sysdeps/unix/sysv/linux/raise.c: No such file or directory. > (gdb) bt > #0 __GI_raise (sig=sig@entry=6) at ../sysdeps/unix/sysv/linux/raise.c:51 > #1 0x00007f7848cce921 in __GI_abort () at abort.c:79 > #2 0x00000000004079a5 in os_exit (code=1) at > /home/vpp/src/vpp/vnet/main.c:433 > #3 0x00007f784a81b6a8 in unix_signal_handler (signum=6, > si=0x7f77a86b5cb0, uc=0x7f77a86b5b80) at /home/vpp/src/vlib/unix/main.c:187 > #4 <signal handler called>; > #5 __GI_raise (sig=sig@entry=6) at ../sysdeps/unix/sysv/linux/raise.c:51 > #6 0x00007f7848cce921 in __GI_abort () at abort.c:79 > #7 0x00007f7849c66799 in os_panic () at > /home/vpp/src/vppinfra/unix-misc.c:177 > #8 0x00007f7849bacf49 in debugger () at /home/vpp/src/vppinfra/error.c:84 > #9 0x00007f7849baccc7 in _clib_error (how_to_die=2, function_name=0x0, > line_number=0, fmt=0x7f784bbc0110 "%s:%d (%s) assertion `%s' > fails") at /home/vpp/src/vppinfra/error.c:143 > #10 0x00007f784adf55b4 in load_balance_get (lbi=16) at > /home/vpp/src/vnet/dpo/load_balance.h:222 > #11 0x00007f784adf4031 in mpls_lookup_node_fn_hsw (vm=0x7f77d279af40, > node=0x7f77d615bc80, from_frame=0x7f77d4d62cc0) at > /home/vpp/src/vnet/mpls/mpls_lookup.c:197 > #12 0x00007f784a794076 in dispatch_node (vm=0x7f77d279af40, > node=0x7f77d615bc80, type=VLIB_NODE_TYPE_INTERNAL, > dispatch_state=VLIB_NODE_STATE_POLLING, frame=0x7f77d4d62cc0, > last_time_stamp=1518151280278706) at /home/vpp/src/vlib/main.c:1217 > #13 0x00007f784a7949e7 in dispatch_pending_node (vm=0x7f77d279af40, > pending_frame_index=2, last_time_stamp=1518151280278706) at > /home/vpp/src/vlib/main.c:1376 > #14 0x00007f784a78e441 in vlib_main_or_worker_loop (vm=0x7f77d279af40, > is_main=0) at /home/vpp/src/vlib/main.c:1904 > #15 0x00007f784a78d2e7 in vlib_worker_loop (vm=0x7f77d279af40) at > /home/vpp/src/vlib/main.c:2038 > #16 0x00007f784a7e595d in vlib_worker_thread_fn (arg=0x7f77c931a0c0) at > /home/vpp/src/vlib/threads.c:1868 > #17 0x00007f7849bd1214 in clib_calljmp () at > /home/vpp/src/vppinfra/longjmp.S:123 > #18 0x00007f67a37fec90 in ?? () > #19 0x00007f784a7ddb83 in vlib_worker_thread_bootstrap_fn > (arg=0x7f77c931a0c0) at /home/vpp/src/vlib/threads.c:585 > Backtrace stopped: previous frame inner to this frame (corrupt stack?) > (gdb) select 10 > (gdb) source ~/vpp/extras/gdb/gdbinit > Loading vpp functions... > Load vlLoad pe > Load pifi > Load node_name_from_index > Load vnet_buffer_opaque > Load vnet_buffer_opaque2 > Load bitmap_get > Done loading vpp functions... > (gdb) pifi load_balance_pool 16 > pool_is_free_index (load_balance_pool, 16)$12 = 0 > > On Fri, 17 Jul 2020 at 14:54, Rajith PR via lists.fd.io <rajith= > rtbrick....@lists.fd.io> wrote: > >> Hi All, >> >> The crash has occurred again in load_balance_get(). This >> time lbi=320017171. May be an invalid value??. Also, this time the flow is >> from mpls_lookup_node_fn_avx2(). >> Any pointes to fix the issue would be very helpful. >> >> Thread 10 (Thread 0x7fc1c089b700 (LWP 438)): >> #0 0x00007fc27b1a8722 in __GI___waitpid (pid=5267, >> stat_loc=stat_loc@entry=0x7fc1fefc2a18, options=options@entry=0) >> at ../sysdeps/unix/sysv/linux/waitpid.c:30 >> #1 0x00007fc27b113107 in do_system (line=<optimized out>) at >> ../sysdeps/posix/system.c:149 >> #2 0x00007fc27bc2c76b in bd_signal_handler_cb (signo=6) at >> /development/librtbrickinfra/bd/src/bd.c:770 >> #3 0x00007fc26f5670ac in rtb_bd_signal_handler (signo=6) at >> /development/libvpp/src/vlib/unix/main.c:80 >> #4 0x00007fc26f567447 in unix_signal_handler (signum=6, >> si=0x7fc1fefc31f0, uc=0x7fc1fefc30c0) >> at /development/libvpp/src/vlib/unix/main.c:180 >> #5 <signal handler called> >> #6 __GI_raise (sig=sig@entry=6) at ../sysdeps/unix/sysv/linux/raise.c:51 >> #7 0x00007fc27b1048b1 in __GI_abort () at abort.c:79 >> #8 0x00007fc270d09f86 in os_panic () at >> /development/libvpp/src/vpp/vnet/main.c:559 >> #9 0x00007fc26f1a0825 in debugger () at >> /development/libvpp/src/vppinfra/error.c:84 >> #10 0x00007fc26f1a0bf4 in _clib_error (how_to_die=2, function_name=0x0, >> line_number=0, fmt=0x7fc27055f7a8 "%s:%d (%s) assertion `%s' fails") >> at /development/libvpp/src/vppinfra/error.c:143 >> #11 0x00007fc26f95e36d in load_balance_get (lbi=320017171) at >> /development/libvpp/src/vnet/dpo/load_balance.h:222 >> #12 0x00007fc26f95fdeb in mpls_lookup_node_fn_avx2 (vm=0x7fc1feb95280, >> node=0x7fc1fe770300, from_frame=0x7fc1fea9bd80) >> at /development/libvpp/src/vnet/mpls/mpls_lookup.c:396 >> #13 0x00007fc26f4fecef in dispatch_node (vm=0x7fc1feb95280, >> node=0x7fc1fe770300, type=VLIB_NODE_TYPE_INTERNAL, >> dispatch_state=VLIB_NODE_STATE_POLLING, frame=0x7fc1fea9bd80, >> last_time_stamp=959252489496170) >> at /development/libvpp/src/vlib/main.c:1207 >> #14 0x00007fc26f4ff4aa in dispatch_pending_node (vm=0x7fc1feb95280, >> pending_frame_index=2, last_time_stamp=959252489496170) >> at /development/libvpp/src/vlib/main.c:1375 >> #15 0x00007fc26f5010ee in vlib_main_or_worker_loop (vm=0x7fc1feb95280, >> is_main=0) at /development/libvpp/src/vlib/main.c:1826 >> #16 0x00007fc26f501971 in vlib_worker_loop (vm=0x7fc1feb95280) at >> /development/libvpp/src/vlib/main.c:1934 >> #17 0x00007fc26f54069b in vlib_worker_thread_fn (arg=0x7fc1fc320b40) at >> /development/libvpp/src/vlib/threads.c:1803 >> #18 0x00007fc26f1c1600 in clib_calljmp () from >> /usr/local/lib/libvppinfra.so.1.0.1 >> #19 0x00007fc1c089aec0 in ?? () >> #20 0x00007fc26f53ac32 in vlib_worker_thread_bootstrap_fn >> (arg=0x7fc1fc320b40) at /development/libvpp/src/vlib/threads.c:573 >> >> Thanks, >> Rajith >> >> On Tue, Jul 7, 2020 at 6:28 PM Rajith PR via lists.fd.io <rajith= >> rtbrick....@lists.fd.io> wrote: >> >>> Hi Benoit, >>> >>> I have all those fixes. I had reported this issue (27407), the others i >>> found during my tests and added barrier protection in all those places. >>> This ASSERT seems to be not due to pool expansion, as what I have seen >>> mostly in those cases is an invalid pointer access causing a SIGSEGV. >>> Here its an abort triggered from *pool_elt_at_index *as the index is >>> already freed. >>> >>> Is it not possible that the main thread freed the pool entry, while the >>> worker thread was holding an index. The backtraces I have attached indicate >>> that the main thread was freeing fib entries. >>> >>> Thanks, >>> Rajith >>> >>> >>> On Tue, Jul 7, 2020 at 5:50 PM Benoit Ganne (bganne) <bga...@cisco.com> >>> wrote: >>> >>>> Hi Rajith, >>>> >>>> You are probably missing https://gerrit.fd.io/r/c/vpp/+/27407 >>>> https://gerrit.fd.io/r/c/vpp/+/27454 and maybe >>>> https://gerrit.fd.io/r/c/vpp/+/27448 >>>> >>>> Best >>>> ben >>>> >>>> > -----Original Message----- >>>> > From: vpp-dev@lists.fd.io <vpp-dev@lists.fd.io> On Behalf Of Rajith >>>> PR via >>>> > lists.fd.io >>>> > Sent: mardi 7 juillet 2020 14:11 >>>> > To: vpp-dev <vpp-dev@lists.fd.io> >>>> > Subject: [vpp-dev]: ASSERT in load_balance_get() >>>> > >>>> > Hi All, >>>> > >>>> > During our scale testing of routes we have hit an ASSERT in >>>> > load_balance_get() . From the code it looks like the lb_index(148) >>>> > referred to is already returned to the pool by the main thread >>>> causing the >>>> > ASSERT in the worker. The version in 19.08. We have two workers and a >>>> > main thread. >>>> > >>>> > Any inputs to fix the issue is highly appreciated. >>>> > >>>> > The complete bt is pasted below: >>>> > >>>> > Thread 11 (Thread 0x7f988ebe3700 (LWP 398)): >>>> > #0 0x00007f99464cef54 in vlib_worker_thread_barrier_check () at >>>> > /development/libvpp/src/vlib/threads.h:430 >>>> > #1 0x00007f99464d6b9b in vlib_main_or_worker_loop (vm=0x7f98d8ba3540, >>>> > is_main=0) at /development/libvpp/src/vlib/main.c:1744 >>>> > #2 0x00007f99464d7971 in vlib_worker_loop (vm=0x7f98d8ba3540) at >>>> > /development/libvpp/src/vlib/main.c:1934 >>>> > #3 0x00007f994651669b in vlib_worker_thread_fn (arg=0x7f98d6191a40) >>>> at >>>> > /development/libvpp/src/vlib/threads.c:1803 >>>> > #4 0x00007f9946197600 in clib_calljmp () from >>>> > /usr/local/lib/libvppinfra.so.1.0.1 >>>> > #5 0x00007f988ebe2ec0 in ?? () >>>> > #6 0x00007f9946510c32 in vlib_worker_thread_bootstrap_fn >>>> > (arg=0x7f98d6191a40) at /development/libvpp/src/vlib/threads.c:573 >>>> > Backtrace stopped: previous frame inner to this frame (corrupt stack?) >>>> > >>>> > Thread 10 (Thread 0x7f988f3e4700 (LWP 397)): >>>> > #0 0x00007f995217d722 in __GI___waitpid (pid=2595, >>>> > stat_loc=stat_loc@entry=0x7f98d8fd0118, options=options@entry=0) at >>>> > ../sysdeps/unix/sysv/linux/waitpid.c:30 >>>> > #1 0x00007f99520e8107 in do_system (line=<optimized out>) at >>>> > ../sysdeps/posix/system.c:149 >>>> > #2 0x00007f9952c016ca in bd_signal_handler_cb (signo=6) at >>>> > /development/librtbrickinfra/bd/src/bd.c:770 >>>> > #3 0x00007f994653d0ac in rtb_bd_signal_handler (signo=6) at >>>> > /development/libvpp/src/vlib/unix/main.c:80 >>>> > #4 0x00007f994653d447 in unix_signal_handler (signum=6, >>>> > si=0x7f98d8fd08f0, uc=0x7f98d8fd07c0) at >>>> > /development/libvpp/src/vlib/unix/main.c:180 >>>> > #5 <signal handler called> >>>> > #6 __GI_raise (sig=sig@entry=6) at >>>> ../sysdeps/unix/sysv/linux/raise.c:51 >>>> > #7 0x00007f99520d98b1 in __GI_abort () at abort.c:79 >>>> > #8 0x00007f9947cdec86 in os_panic () at >>>> > /development/libvpp/src/vpp/vnet/main.c:559 >>>> > #9 0x00007f9946176825 in debugger () at >>>> > /development/libvpp/src/vppinfra/error.c:84 >>>> > #10 0x00007f9946176bf4 in _clib_error (how_to_die=2, >>>> function_name=0x0, >>>> > line_number=0, fmt=0x7f99475271b8 "%s:%d (%s) assertion `%s' fails") >>>> > at /development/libvpp/src/vppinfra/error.c:143 >>>> > #11 0x00007f99468d046c in load_balance_get (lbi=148) at >>>> > /development/libvpp/src/vnet/dpo/load_balance.h:222 >>>> > #12 0x00007f99468d4d44 in ip4_local_check_src (b=0x1002535e00, >>>> > ip0=0x1002535f52, last_check=0x7f98d8fd1234, error0=0x7f98d8fd11e8 >>>> > "\016\r") >>>> > at /development/libvpp/src/vnet/ip/ip4_forward.c:1583 >>>> > #13 0x00007f99468d58e1 in ip4_local_inline (vm=0x7f98d8ba2e40, >>>> > node=0x7f98d8711f40, frame=0x7f98d9585bc0, head_of_feature_arc=1) >>>> > at /development/libvpp/src/vnet/ip/ip4_forward.c:1870 >>>> > #14 0x00007f99468d5a08 in ip4_local_node_fn_avx2 (vm=0x7f98d8ba2e40, >>>> > node=0x7f98d8711f40, frame=0x7f98d9585bc0) at >>>> > /development/libvpp/src/vnet/ip/ip4_forward.c:1889 >>>> > #15 0x00007f99464d4cef in dispatch_node (vm=0x7f98d8ba2e40, >>>> > node=0x7f98d8711f40, type=VLIB_NODE_TYPE_INTERNAL, >>>> > dispatch_state=VLIB_NODE_STATE_POLLING, >>>> > frame=0x7f98d9585bc0, last_time_stamp=531887125605324) at >>>> > /development/libvpp/src/vlib/main.c:1207 >>>> > #16 0x00007f99464d54aa in dispatch_pending_node (vm=0x7f98d8ba2e40, >>>> > pending_frame_index=3, last_time_stamp=531887125605324) at >>>> > /development/libvpp/src/vlib/main.c:1375 >>>> > #17 0x00007f99464d70ee in vlib_main_or_worker_loop (vm=0x7f98d8ba2e40, >>>> > is_main=0) at /development/libvpp/src/vlib/main.c:1826 >>>> > #18 0x00007f99464d7971 in vlib_worker_loop (vm=0x7f98d8ba2e40) at >>>> > /development/libvpp/src/vlib/main.c:1934 >>>> > #19 0x00007f994651669b in vlib_worker_thread_fn (arg=0x7f98d6191940) >>>> at >>>> > /development/libvpp/src/vlib/threads.c:1803 >>>> > #20 0x00007f9946197600 in clib_calljmp () from >>>> > /usr/local/lib/libvppinfra.so.1.0.1 >>>> > #21 0x00007f988f3e3ec0 in ?? () >>>> > #22 0x00007f9946510c32 in vlib_worker_thread_bootstrap_fn >>>> > (arg=0x7f98d6191940) at /development/libvpp/src/vlib/threads.c:573 >>>> > Backtrace stopped: previous frame inner to this frame (corrupt stack?) >>>> > >>>> > Thread 1 (Thread 0x7f995306c740 (LWP 249)): >>>> > #0 0x00007f994650dd9e in clib_time_now (c=0x7f9946776e40 >>>> > <vlib_global_main>) at /development/libvpp/src/vppinfra/time.h:217 >>>> > #1 0x00007f994650de74 in vlib_time_now (vm=0x7f9946776e40 >>>> > <vlib_global_main>) at /development/libvpp/src/vlib/main.h:268 >>>> > #2 0x00007f99465159ab in vlib_worker_thread_barrier_sync_int >>>> > (vm=0x7f9946776e40 <vlib_global_main>, >>>> > func_name=0x7f994764e2d0 <__FUNCTION__.42472> >>>> "adj_last_lock_gone") at >>>> > /development/libvpp/src/vlib/threads.c:1486 >>>> > #3 0x00007f994743fb9f in adj_last_lock_gone (adj=0x7f98d5f0fc00) at >>>> > /development/libvpp/src/vnet/adj/adj.c:256 >>>> > #4 0x00007f994744062e in adj_node_last_lock_gone >>>> (node=0x7f98d5f0fc00) at >>>> > /development/libvpp/src/vnet/adj/adj.c:546 >>>> > #5 0x00007f99473e74be in fib_node_unlock (node=0x7f98d5f0fc00) at >>>> > /development/libvpp/src/vnet/fib/fib_node.c:215 >>>> > #6 0x00007f9947440186 in adj_unlock (adj_index=37) at >>>> > /development/libvpp/src/vnet/adj/adj.c:346 >>>> > #7 0x00007f99474277ba in adj_dpo_unlock (dpo=0x7f98d8bb5288) at >>>> > /development/libvpp/src/vnet/adj/adj_nbr.c:1008 >>>> > #8 0x00007f994744b9bd in dpo_unlock (dpo=0x7f98d8bb5288) at >>>> > /development/libvpp/src/vnet/dpo/dpo.c:378 >>>> > #9 0x00007f994744b379 in dpo_copy (dst=0x7f98d896dba0, >>>> > src=0x7f98d8bb52b8) at /development/libvpp/src/vnet/dpo/dpo.c:274 >>>> > #10 0x00007f994744b2ef in dpo_reset (dpo=0x7f98d896dba0) at >>>> > /development/libvpp/src/vnet/dpo/dpo.c:239 >>>> > #11 0x00007f994745b394 in load_balance_destroy (lb=0x7f98d896db80) at >>>> > /development/libvpp/src/vnet/dpo/load_balance.c:880 >>>> > #12 0x00007f994745b777 in load_balance_unlock (dpo=0x7f98d8bb5398) at >>>> > /development/libvpp/src/vnet/dpo/load_balance.c:906 >>>> > #13 0x00007f994744b9bd in dpo_unlock (dpo=0x7f98d8bb5398) at >>>> > /development/libvpp/src/vnet/dpo/dpo.c:378 >>>> > #14 0x00007f994744b379 in dpo_copy (dst=0x7f98d88ecd58, >>>> > src=0x7f98d8bb53c8) at /development/libvpp/src/vnet/dpo/dpo.c:274 >>>> > #15 0x00007f994744b2ef in dpo_reset (dpo=0x7f98d88ecd58) at >>>> > /development/libvpp/src/vnet/dpo/dpo.c:239 >>>> > #16 0x00007f99473f5025 in fib_entry_src_action_uninstall >>>> > (fib_entry=0x7f98d88ecd30) at >>>> > /development/libvpp/src/vnet/fib/fib_entry_src.c:707 >>>> > #17 0x00007f99473ec71c in fib_entry_source_removed >>>> > (fib_entry=0x7f98d88ecd30, old_flags=FIB_ENTRY_FLAG_NONE) at >>>> > /development/libvpp/src/vnet/fib/fib_entry.c:992 >>>> > #18 0x00007f99473ec926 in fib_entry_path_remove (fib_entry_index=4159, >>>> > source=FIB_SOURCE_API, rpaths=0x7f98d9e73a00) >>>> > at /development/libvpp/src/vnet/fib/fib_entry.c:1082 >>>> > #19 0x00007f99473d8d1d in fib_table_entry_path_remove2 (fib_index=5, >>>> > prefix=0x7f98d8bb56d0, source=FIB_SOURCE_API, rpaths=0x7f98d9e73a00) >>>> > at /development/libvpp/src/vnet/fib/fib_table.c:663 >>>> > >>>> > #20 0x00007f9947dcf3a6 in rtb_vpp_route_mapping_process >>>> > (table=0x5576fd90ef3c, object=0x557701f7f07c, action=1 '\001') >>>> > at /development/libvpp/src/vpp/rtbrick/rtb_vpp_route.c:919 >>>> > #21 0x00007f9947dcf477 in rtb_vpp_route_mapping_del_cb >>>> > (table=0x5576fd90ef3c, object=0x557701f7f07c) at >>>> > /development/libvpp/src/vpp/rtbrick/rtb_vpp_route.c:947 >>>> > #22 0x00007f994f98bd9a in bds_handle_object_plugin >>>> > (table_handle=0x5576fd90ef3c, index=0x557701f7f07c, old_index=0x0, >>>> > type=BDS_PC_DEL) >>>> > at >>>> /development/librtbrickinfra/libbds/code/bds/src/bds_plugin.c:992 >>>> > #23 0x00007f994f98ab6c in bds_deferred_object_plugins >>>> > (table_handle=0x5576fd90ef3c, obj_handle=0x557701f7f07c, >>>> > obj_deleted=0x7f98d8bb58c2 "") >>>> > at >>>> /development/librtbrickinfra/libbds/code/bds/src/bds_plugin.c:458 >>>> > #24 0x00007f994f98b2af in bds_deferred_plugin_cb >>>> > (table_handle=0x5576fd90ef3c, obj_handle=0x557701f7f07c, >>>> > obj_deleted=0x7f98d8bb58c2 "") >>>> > at >>>> /development/librtbrickinfra/libbds/code/bds/src/bds_plugin.c:656 >>>> > #25 0x00007f994f97fed9 in bds_object_handle_plugin_publish >>>> > (table_handle=0x5576fd90ef3c, obj_handle=0x557701f7f07c, >>>> > old_obj_handle=0x0) >>>> > ---Type <return> to continue, or q <return> to quit--- >>>> > at >>>> /development/librtbrickinfra/libbds/code/bds/src/bds_object.c:3403 >>>> > #26 0x00007f994f983600 in bds_object_delete >>>> (obj_template=0x5576f8637c5c) >>>> > at /development/librtbrickinfra/libbds/code/bds/src/bds_object.c:4867 >>>> > #27 0x00007f9947dcd772 in rtb_vpp_route_mapping_object_add_del >>>> > (route_mapping=0x7f98d8bb5a70, action=1 '\001') >>>> > at /development/libvpp/src/vpp/rtbrick/rtb_vpp_route.c:383 >>>> > #28 0x00007f9947dd0986 in rtb_vpp_adj_adjacency_route_handle >>>> > (adj_api_out=0x7f98d8bb5ba0, action=1 '\001') at >>>> > /development/libvpp/src/vpp/rtbrick/rtb_vpp_adj.c:132 >>>> > #29 0x00007f9947dd0b93 in rtb_vpp_adj_api_out_process >>>> > (table=0x5576fdc0b13c, object=0x55770212be6c, action=1 '\001') >>>> > at /development/libvpp/src/vpp/rtbrick/rtb_vpp_adj.c:178 >>>> > #30 0x00007f9947dd0c57 in rtb_vpp_adj_api_out_chng_cb >>>> > (table=0x5576fdc0b13c, object=0x55770102780c, >>>> object_old=0x55770212be6c) >>>> > at /development/libvpp/src/vpp/rtbrick/rtb_vpp_adj.c:207 >>>> > #31 0x00007f994f98bd9a in bds_handle_object_plugin >>>> > (table_handle=0x5576fdc0b13c, index=0x55770102780c, >>>> > old_index=0x55770212be6c, type=BDS_PC_CHANGE) >>>> > at >>>> /development/librtbrickinfra/libbds/code/bds/src/bds_plugin.c:992 >>>> > #32 0x00007f994f98ad7f in bds_deferred_object_plugins >>>> > (table_handle=0x5576fdc0b13c, obj_handle=0x55770102780c, >>>> > obj_deleted=0x7f98d8bb5de2 "") >>>> > at >>>> /development/librtbrickinfra/libbds/code/bds/src/bds_plugin.c:492 >>>> > #33 0x00007f994f98b2af in bds_deferred_plugin_cb >>>> > (table_handle=0x5576fdc0b13c, obj_handle=0x55770102780c, >>>> > obj_deleted=0x7f98d8bb5de2 "") >>>> > at >>>> /development/librtbrickinfra/libbds/code/bds/src/bds_plugin.c:656 >>>> > #34 0x00007f994f97fed9 in bds_object_handle_plugin_publish >>>> > (table_handle=0x5576fdc0b13c, obj_handle=0x55770102780c, >>>> > old_obj_handle=0x55770212be6c) >>>> > at >>>> /development/librtbrickinfra/libbds/code/bds/src/bds_object.c:3403 >>>> > #35 0x00007f994f9801b9 in bds_object_handle_features >>>> > (table=0x5576fdc0b13c, index=0x55770102780c, old_index=0x55770212be6c, >>>> > add_vector=3, is_diff_index=0 '\000') >>>> > at >>>> /development/librtbrickinfra/libbds/code/bds/src/bds_object.c:3515 >>>> > #36 0x00007f994f980dc5 in bds_build_index (table=0x5576fdc0b13c, >>>> > obj_map=0x557702d6fefc, cp_offset=0x7f98d8bb5f80, >>>> cp_len=0x7f98d8bb6180, >>>> > data_offset=38, >>>> > add_bitmap=65535, is_diff_index=0 '\000') at >>>> > /development/librtbrickinfra/libbds/code/bds/src/bds_object.c:3832 >>>> > >>>> > #37 0x00007f994f9816bd in bds_create_object_index_map >>>> > (obj_template=0x5576f8083e5c, use_seq=0 '\000', skip_index=0 '\000') >>>> > at >>>> /development/librtbrickinfra/libbds/code/bds/src/bds_object.c:4024 >>>> > #38 0x00007f994f98206c in bds_object_add >>>> (object_template=0x5576f8083e5c) >>>> > at /development/librtbrickinfra/libbds/code/bds/src/bds_object.c:4267 >>>> > #39 0x00007f9945acfd12 in fib_adjacency_api_out_obj_add_del >>>> > (fib_adjacency=0x7f98d8bb6820, action=0 '\000') >>>> > at >>>> > /development/libforwarding/pi/fib/code/src/fib_adjacency_api_out.c:342 >>>> > #40 0x00007f9945acec9a in fib_adjacency_local_obj_process >>>> > (table=0x5576fdc0512c, obj_handle=0x5577018317fc, action=0 '\000') >>>> > at /development/libforwarding/pi/fib/code/src/fib_adjacency.c:90 >>>> > #41 0x00007f9945aced39 in fib_adjacency_local_obj_chng_cb >>>> > (tblHndl=0x5576fdc0512c, newObjHndl=0x5577018317fc, >>>> > oldObjHndl=0x5577016fed0c) >>>> > at /development/libforwarding/pi/fib/code/src/fib_adjacency.c:111 >>>> > #42 0x00007f994f98bd9a in bds_handle_object_plugin >>>> > (table_handle=0x5576fdc0512c, index=0x5577018317fc, >>>> > old_index=0x5577016fed0c, type=BDS_PC_CHANGE) >>>> > at >>>> /development/librtbrickinfra/libbds/code/bds/src/bds_plugin.c:992 >>>> > #43 0x00007f994f98ad7f in bds_deferred_object_plugins >>>> > (table_handle=0x5576fdc0512c, obj_handle=0x5577018317fc, >>>> > obj_deleted=0x7f98d8bb6d8f "") >>>> > at >>>> /development/librtbrickinfra/libbds/code/bds/src/bds_plugin.c:492 >>>> > #44 0x00007f994f98b2af in bds_deferred_plugin_cb >>>> > (table_handle=0x5576fdc0512c, obj_handle=0x5577018317fc, >>>> > obj_deleted=0x7f98d8bb6d8f "") >>>> > at >>>> /development/librtbrickinfra/libbds/code/bds/src/bds_plugin.c:656 >>>> > #45 0x00007f994f98b390 in bds_deferred_plugin_cb_expiry >>>> > (table_handle=0x5576fdc0512c) at >>>> > /development/librtbrickinfra/libbds/code/bds/src/bds_plugin.c:696 >>>> > #46 0x00007f995270a070 in timer_dispatch (item=0x557701811cb0, >>>> > p=QB_LOOP_HIGH) at >>>> > /development/librtbrickinfra/libqb/lib/loop_timerlist.c:56 >>>> > #47 0x00007f9952706006 in qb_loop_run_level (level=0x5576f78f5830) at >>>> > /development/librtbrickinfra/libqb/lib/loop.c:43 >>>> > #48 0x00007f9952707038 in qb_loop_run_vpp_wrapper (lp=0x5576f78f57c0) >>>> at >>>> > /development/librtbrickinfra/libqb/lib/loop.c:350 >>>> > #49 0x00007f9952717cd8 in lib_qb_service_start_event_wrapper_loop () >>>> at >>>> > /development/librtbrickinfra/libqb/lib/wrapper/lib_qb_service.c:290 >>>> > #50 0x00007f98cfb8386c in rtb_vpp_epoll_process (vm=0x7f9946776e40 >>>> > <vlib_global_main>, rt=0x7f98d8ba6000, f=0x0) >>>> > ---Type <return> to continue, or q <return> to quit--- >>>> > at /development/libvpp/src/plugins/rtbrick/rtb_node.c:81 >>>> > #51 0x00007f99464d571b in vlib_process_bootstrap (_a=140294421412800) >>>> at >>>> > /development/libvpp/src/vlib/main.c:1468 >>>> > #52 0x00007f9946197600 in clib_calljmp () from >>>> > /usr/local/lib/libvppinfra.so.1.0.1 >>>> > #53 0x00007f98d726fb90 in ?? () >>>> > #54 0x00007f99464d5820 in vlib_process_startup (vm=0x7f99464d588a >>>> > <vlib_process_resume+96>, p=0x8, f=0x7f9946776e40 <vlib_global_main>) >>>> > at /development/libvpp/src/vlib/main.c:1490 >>>> > Backtrace stopped: previous frame inner to this frame (corrupt stack?) >>>> > >>>> > >>>> > Thanks, >>>> > Rajith >>>> >>> >>> > > >> > > -- > Best regards > Stanislav Zaikin > -- NOTICE TO RECIPIENT This e-mail message and any attachments are confidential and may be privileged. If you received this e-mail in error, any review, use, dissemination, distribution, or copying of this e-mail is strictly prohibited. Please notify us immediately of the error by return e-mail and please delete this message from your system. For more information about Rtbrick, please visit us at www.rtbrick.com <http://www.rtbrick.com>
-=-=-=-=-=-=-=-=-=-=-=- Links: You receive all messages sent to this group. View/Reply Online (#20318): https://lists.fd.io/g/vpp-dev/message/20318 Mute This Topic: https://lists.fd.io/mt/75353568/21656 Group Owner: vpp-dev+ow...@lists.fd.io Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub [arch...@mail-archive.com] -=-=-=-=-=-=-=-=-=-=-=-