Hi Florin,
Hi Rajith,

It shouldn't be the pool expansion case, I have
8341f76fd1cd4351961cd8161cfed2814fc55103.
Moreover, in this case _e would be different from
&load_balance_pool[3604]. I've
found some of those expansions (in other places), in those cases a pointer
to the element has a different address.


On Thu, 14 Oct 2021 at 06:45, Rajith PR <raj...@rtbrick.com> wrote:

> HI Stanislav,
>
> My guess is you don't have the commit below.
>
> commit 8341f76fd1cd4351961cd8161cfed2814fc55103
> Author: Dave Barach <d...@barachs.net>
> Date:   Wed Jun 3 08:05:15 2020 -0400
>
>     fib: add barrier sync, pool/vector expand cases
>
>     load_balance_alloc_i(...) is not thread safe when the
>     load_balance_pool or combined counter vectors expand.
>
>     Type: fix
>
>     Signed-off-by: Dave Barach <d...@barachs.net>
>     Change-Id: I7f295ed77350d1df0434d5ff461eedafe79131de
>
> Thanks,
> Rajith
>
> On Thu, Oct 14, 2021 at 3:57 AM Florin Coras <fcoras.li...@gmail.com>
> wrote:
>
>> Hi Stanislav,
>>
>> The only thing I can think of is that main thread grows the pool, or the
>> pool’s bitmap, without a worker barrier while the worker that asserts is
>> trying to access it. Is main thread busy doing something (e.g., adding
>> routes/interfaces) when the assert happens?
>>
>> Regards,
>> Florin
>>
>> On Oct 13, 2021, at 2:52 PM, Stanislav Zaikin <zsta...@gmail.com> wrote:
>>
>> Hi Florin,
>>
>> I wasn't aware of those helper functions, thanks! But yeah, it also
>> returns 0 (sorry, but there's the trace of another crash)
>>
>> Thread 3 "vpp_wk_0" received signal SIGABRT, Aborted.
>> [Switching to Thread 0x7f9cc0f6a700 (LWP 3546)]
>> __GI_raise (sig=sig@entry=6) at ../sysdeps/unix/sysv/linux/raise.c:51
>> 51 ../sysdeps/unix/sysv/linux/raise.c: No such file or directory.
>> (gdb) bt
>> #0  __GI_raise (sig=sig@entry=6) at ../sysdeps/unix/sysv/linux/raise.c:51
>> #1  0x00007f9d61542921 in __GI_abort () at abort.c:79
>> #2  0x00007f9d624da799 in os_panic () at
>> /home/vpp/vpp/src/vppinfra/unix-misc.c:177
>> #3  0x00007f9d62420f49 in debugger () at
>> /home/vpp/vpp/src/vppinfra/error.c:84
>> #4  0x00007f9d62420cc7 in _clib_error (how_to_die=2, function_name=0x0,
>> line_number=0, fmt=0x7f9d644348d0 "%s:%d (%s) assertion `%s' fails") at
>> /home/vpp/vpp/src/vppinfra/error.c:143
>> #5  0x00007f9d636695b4 in load_balance_get (lbi=4569) at
>> /home/vpp/vpp/src/vnet/dpo/load_balance.h:222
>> #6  0x00007f9d63668247 in mpls_lookup_node_fn_hsw (vm=0x7f9ceb0138c0,
>> node=0x7f9ceee6f700, from_frame=0x7f9cef9c9240) at
>> /home/vpp/vpp/src/vnet/mpls/mpls_lookup.c:229
>> #7  0x00007f9d63008076 in dispatch_node (vm=0x7f9ceb0138c0,
>> node=0x7f9ceee6f700, type=VLIB_NODE_TYPE_INTERNAL,
>> dispatch_state=VLIB_NODE_STATE_POLLING, frame=0x7f9cef9c9240,
>> last_time_stamp=1837178878370487) at /home/vpp/vpp/src/vlib/main.c:1217
>> #8  0x00007f9d630089e7 in dispatch_pending_node (vm=0x7f9ceb0138c0,
>> pending_frame_index=2, last_time_stamp=1837178878370487) at
>> /home/vpp/vpp/src/vlib/main.c:1376
>> #9  0x00007f9d63002441 in vlib_main_or_worker_loop (vm=0x7f9ceb0138c0,
>> is_main=0) at /home/vpp/vpp/src/vlib/main.c:1904
>> #10 0x00007f9d630012e7 in vlib_worker_loop (vm=0x7f9ceb0138c0) at
>> /home/vpp/vpp/src/vlib/main.c:2038
>> #11 0x00007f9d6305995d in vlib_worker_thread_fn (arg=0x7f9ce1b88540) at
>> /home/vpp/vpp/src/vlib/threads.c:1868
>> #12 0x00007f9d62445214 in clib_calljmp () at
>> /home/vpp/vpp/src/vppinfra/longjmp.S:123
>> #13 0x00007f9cc0f69c90 in ?? ()
>> #14 0x00007f9d63051b83 in vlib_worker_thread_bootstrap_fn
>> (arg=0x7f9ce1b88540) at /home/vpp/vpp/src/vlib/threads.c:585
>> #15 0x00007f9cda360355 in eal_thread_loop (arg=0x0) at
>> ../src-dpdk/lib/librte_eal/linux/eal_thread.c:127
>> #16 0x00007f9d629246db in start_thread (arg=0x7f9cc0f6a700) at
>> pthread_create.c:463
>> #17 0x00007f9d6162371f in clone () at
>> ../sysdeps/unix/sysv/linux/x86_64/clone.S:95
>> (gdb) select 5
>> (gdb)
>> *print pifi( load_balance_pool, 4569 )$1 = 0*
>> (gdb) source ~/vpp/extras/gdb/gdbinit
>> Loading vpp functions...
>> Load vlLoad pe
>> Load pifi
>> Load node_name_from_index
>> Load vnet_buffer_opaque
>> Load vnet_buffer_opaque2
>> Load bitmap_get
>> Done loading vpp functions...
>> (gdb) pifi load_balance_pool 4569
>> pool_is_free_index (load_balance_pool, 4569)$2 = 0
>>
>> On Wed, 13 Oct 2021 at 21:55, Florin Coras <fcoras.li...@gmail.com>
>> wrote:
>>
>>> Hi Stanislav,
>>>
>>> Just to make sure the gdb macro is okay, could you run from gdb:
>>> pifi(pool, index)? The function is defined in gdb_funcs.c.
>>>
>>> Regards,
>>> Florin
>>>
>>> On Oct 13, 2021, at 11:30 AM, Stanislav Zaikin <zsta...@gmail.com>
>>> wrote:
>>>
>>> Hello folks,
>>>
>>> I'm facing a strange issue with 2 worker threads. Sometimes I get a
>>> crash either in "ip6-lookup" or "mpls-lookup" nodes. They happen with
>>> assert in the *pool_elt_at_index* macro and always inside the "
>>> *load_balance_get*" function. But the load_balance dpo looks perfectly
>>> good, I mean it still has a lock and on regular deletion (in the case when
>>> the load_balance dpo is deleted) it should be erased properly (with
>>> dpo_reset). It happens usually when the main core is executing
>>> vlib_worker_thread_barrier_sync_int(), and the other worker is executing
>>> vlib_worker_thread_barrier_check().
>>> And the strangest thing is, when I run the vpp's gdb helper for checking
>>> "pool_index_is_free" or pifi, it shows me that the index isn't free (and
>>> the macro in that case shouldn't fire).
>>>
>>> Any thoughts and inputs are appreciated.
>>>
>>> Thread 3 "vpp_wk_0" received signal SIGABRT, Aborted.
>>> [Switching to Thread 0x7fb4f2e22700 (LWP 3244)]
>>> __GI_raise (sig=sig@entry=6) at ../sysdeps/unix/sysv/linux/raise.c:51
>>> 51 ../sysdeps/unix/sysv/linux/raise.c: No such file or directory.
>>> (gdb) bt
>>> #0  __GI_raise (sig=sig@entry=6) at
>>> ../sysdeps/unix/sysv/linux/raise.c:51
>>> #1  0x00007fb5933fa921 in __GI_abort () at abort.c:79
>>> #2  0x00007fb594392799 in os_panic () at
>>> /home/vpp/vpp/src/vppinfra/unix-misc.c:177
>>> #3  0x00007fb5942d8f49 in debugger () at
>>> /home/vpp/vpp/src/vppinfra/error.c:84
>>> #4  0x00007fb5942d8cc7 in _clib_error (how_to_die=2, function_name=0x0,
>>> line_number=0, fmt=0x7fb5962ec8d0 "%s:%d (%s) assertion `%s' fails") at
>>> /home/vpp/vpp/src/vppinfra/error.c:143
>>> #5  0x00007fb5954bd694 in load_balance_get (lbi=3604) at
>>> /home/vpp/vpp/src/vnet/dpo/load_balance.h:222
>>> #6  0x00007fb5954bc070 in ip6_lookup_inline (vm=0x7fb51ceccd00,
>>> node=0x7fb520f6b700, frame=0x7fb52128e4c0) at
>>> /home/vpp/vpp/src/vnet/ip/ip6_forward.h:117
>>> #7  0x00007fb5954bbdd5 in ip6_lookup_node_fn_hsw (vm=0x7fb51ceccd00,
>>> node=0x7fb520f6b700, frame=0x7fb52128e4c0) at
>>> /home/vpp/vpp/src/vnet/ip/ip6_forward.c:736
>>> #8  0x00007fb594ec0076 in dispatch_node (vm=0x7fb51ceccd00,
>>> node=0x7fb520f6b700, type=VLIB_NODE_TYPE_INTERNAL,
>>> dispatch_state=VLIB_NODE_STATE_POLLING, frame=0x7fb52128e4c0,
>>> last_time_stamp=1808528151240447) at /home/vpp/vpp/src/vlib/main.c:1217
>>> #9  0x00007fb594ec09e7 in dispatch_pending_node (vm=0x7fb51ceccd00,
>>> pending_frame_index=5, last_time_stamp=1808528151240447) at
>>> /home/vpp/vpp/src/vlib/main.c:1376
>>> #10 0x00007fb594eba441 in vlib_main_or_worker_loop (vm=0x7fb51ceccd00,
>>> is_main=0) at /home/vpp/vpp/src/vlib/main.c:1904
>>> #11 0x00007fb594eb92e7 in vlib_worker_loop (vm=0x7fb51ceccd00) at
>>> /home/vpp/vpp/src/vlib/main.c:2038
>>> #12 0x00007fb594f1195d in vlib_worker_thread_fn (arg=0x7fb513a48100) at
>>> /home/vpp/vpp/src/vlib/threads.c:1868
>>> #13 0x00007fb5942fd214 in clib_calljmp () at
>>> /home/vpp/vpp/src/vppinfra/longjmp.S:123
>>> #14 0x00007fb4f2e21c90 in ?? ()
>>> #15 0x00007fb594f09b83 in vlib_worker_thread_bootstrap_fn
>>> (arg=0x7fb513a48100) at /home/vpp/vpp/src/vlib/threads.c:585
>>> #16 0x00007fb50c218355 in eal_thread_loop (arg=0x0) at
>>> ../src-dpdk/lib/librte_eal/linux/eal_thread.c:127
>>> #17 0x00007fb5947dc6db in start_thread (arg=0x7fb4f2e22700) at
>>> pthread_create.c:463
>>> #18 0x00007fb5934db71f in clone () at
>>> ../sysdeps/unix/sysv/linux/x86_64/clone.S:95
>>> (gdb) select 5
>>> (gdb) print _e
>>> $1 = (load_balance_t *) 0x7fb52651e580
>>> (gdb) print load_balance_pool[3604]
>>> $2 = {cacheline0 = 0x7fb52651e580 "\001", lb_n_buckets = 1,
>>> lb_n_buckets_minus_1 = 0, lb_proto = DPO_PROTO_IP6, lb_flags =
>>> LOAD_BALANCE_FLAG_NONE, lb_fib_entry_flags = (FIB_ENTRY_FLAG_CONNECTED |
>>> FIB_ENTRY_FLAG_LOCAL), lb_locks = 1, lb_map = 4294967295, lb_urpf = 4094,
>>> lb_hash_config = 31, lb_buckets = 0x0,
>>>   lb_buckets_inline = {{{{dpoi_type = DPO_RECEIVE, dpoi_proto =
>>> DPO_PROTO_IP6, dpoi_next_node = 2, dpoi_index = 2094}, as_u64 =
>>> 8993661649164}}, {{{dpoi_type = DPO_FIRST, dpoi_proto = DPO_PROTO_IP4,
>>> dpoi_next_node = 0, dpoi_index = 0}, as_u64 = 0}}, {{{dpoi_type =
>>> DPO_FIRST, dpoi_proto = DPO_PROTO_IP4,
>>>           dpoi_next_node = 0, dpoi_index = 0}, as_u64 = 0}},
>>> {{{dpoi_type = DPO_FIRST, dpoi_proto = DPO_PROTO_IP4, dpoi_next_node = 0,
>>> dpoi_index = 0}, as_u64 = 0}}}}
>>> (gdb) print &load_balance_pool[3604]
>>> $3 = (load_balance_t *) 0x7fb52651e580
>>> (gdb) source ~/vpp/extras/gdb/gdbinit
>>> Loading vpp functions...
>>> Load vlLoad pe
>>> Load pifi
>>> Load node_name_from_index
>>> Load vnet_buffer_opaque
>>> Load vnet_buffer_opaque2
>>> Load bitmap_get
>>> Done loading vpp functions...
>>> (gdb) pifi load_balance_pool 3604
>>> pool_is_free_index (load_balance_pool, 3604)$4 = 0
>>> (gdb) info threads
>>>   Id   Target Id         Frame
>>>   1    Thread 0x7fb596bd2c40 (LWP 727) "vpp_main" 0x00007fb594f1439b in
>>> clib_time_now_internal (c=0x7fb59517ccc0 <vlib_global_main>,
>>> n=1808528155236639) at /home/vpp/vpp/src/vppinfra/time.h:215
>>>   2    Thread 0x7fb4f3623700 (LWP 2976) "eal-intr-thread"
>>> 0x00007fb5934dba47 in epoll_wait (epfd=17, events=0x7fb4f3622d80,
>>> maxevents=1, timeout=-1) at ../sysdeps/unix/sysv/linux/epoll_wait.c:30
>>> * 3    Thread 0x7fb4f2e22700 (LWP 3244) "vpp_wk_0" __GI_raise
>>> (sig=sig@entry=6) at ../sysdeps/unix/sysv/linux/raise.c:51
>>>   4    Thread 0x7fb4f2621700 (LWP 3246) "vpp_wk_1" 0x00007fb594ebf897 in
>>> vlib_worker_thread_barrier_check () at /home/vpp/vpp/src/vlib/threads.h:439
>>>
>>> --
>>> Best regards
>>> Stanislav Zaikin
>>>
>>>
>>>
>>>
>>>
>>
>> --
>> Best regards
>> Stanislav Zaikin
>>
>>
>>
>> 
>>
>>
> NOTICE TO RECIPIENT This e-mail message and any attachments are
> confidential and may be privileged. If you received this e-mail in error,
> any review, use, dissemination, distribution, or copying of this e-mail is
> strictly prohibited. Please notify us immediately of the error by return
> e-mail and please delete this message from your system. For more
> information about Rtbrick, please visit us at www.rtbrick.com
>


-- 
Best regards
Stanislav Zaikin
-=-=-=-=-=-=-=-=-=-=-=-
Links: You receive all messages sent to this group.
View/Reply Online (#20327): https://lists.fd.io/g/vpp-dev/message/20327
Mute This Topic: https://lists.fd.io/mt/86295132/21656
Group Owner: vpp-dev+ow...@lists.fd.io
Unsubscribe: https://lists.fd.io/g/vpp-dev/unsub [arch...@mail-archive.com]
-=-=-=-=-=-=-=-=-=-=-=-

Reply via email to