On 27 Feb 2024, at 9:49, LIU Yulong wrote:

> Yes, that makes sense.
>
> Another question is how to distinguish the core at line of
> ovs_mutex_trylock in revalidator_sweep__ is after the free(ukey),
> since the core trace has no timestamp.

This is hard to figure out without adding a time variable (and make sure it’s 
not optimized out) in revalidator_sweep__()

As you are using OVS-DPDK, you can read the values from the pmd->ctx.now 
context, which should be close.

> This line in the function 'ukey_create__' should be the only place
> where ovs allocated the memory for ukey:
> https://github.com/openvswitch/ovs/blob/master/ofproto/ofproto-dpif-upcall.c#L1777
>
> Right?

Yes, this should be the only place.

>
> If it is true, I will update the buffer structure, and a
> counter_upcall_ukey_allocate as well.
>
> On Tue, Feb 27, 2024 at 3:34 PM Eelco Chaudron <echau...@redhat.com> wrote:
>>
>>
>>
>> On 27 Feb 2024, at 4:44, LIU Yulong wrote:
>>
>>> @Eelco, as you suggested, added such circular buffer to my local OVS:
>>> https://github.com/gotostack/ovs/commit/939d88c3c5fcdb446b01f2afa8f1e80c3929db46
>>
>> I should also add allocate logging, or else you might not know if a buffer 
>> was allocated at the same address.
>> Maybe add a bool to the record structure to indicate if it’s an allocate or 
>> free.
>>
>> //Eelco
>>
>>> gdb shows such data structure:
>>> 2232     ukey_free_buffer.index = (ukey_free_buffer.index + 1) % (1024
>>> * 1024);  // Circular buffer
>>> (gdb) p ukey_free_buffer
>>> $1 = {
>>>   records = {{
>>>       ukey_addr = 0x7f8a0d871700,
>>>       timestamp = 1709003328
>>>     }, {
>>>       ukey_addr = 0x7f8a0f969120,
>>>       timestamp = 1709003365
>>>     }, {
>>>       ukey_addr = 0x7f8a0defe190,
>>>       timestamp = 1709003393
>>>     }, {
>>>       ukey_addr = 0x7f8a0984aea0,
>>>       timestamp = 1709003452
>>>     }...},
>>>   index = 3,
>>>   mutex = {
>>>     lock = {
>>>       __data = {
>>>         __lock = 1,
>>>         __count = 0,
>>>         __owner = 45210,
>>>         __nusers = 1,
>>>         __kind = 2,
>>>         __spins = 0,
>>>         __elision = 0,
>>>         __list = {
>>>           __prev = 0x0,
>>>           __next = 0x0
>>>         }
>>>       },
>>>       __size = 
>>> "\001\000\000\000\000\000\000\000\232\260\000\000\001\000\000\000\002",
>>> '\000' <repeats 22 times>,
>>>       __align = 1
>>>     },
>>>     where = 0x55c35a347d18 "ofproto/ofproto-dpif-upcall.c:2229"
>>>   }
>>> }
>>>
>>> and counter_upcall_ukey_free is:
>>> $2 = {name = 0x5622b448f612 "upcall_ukey_free", count = 0x5622b41047f0
>>> <upcall_ukey_free_count>, total = 79785, last_total = 79785, min = {0,
>>> 0, 0, 0, 0, 55, 22681, 11703, 13877, 12750, 0, 18719}, hr = {79785,
>>>     0 <repeats 59 times>}}
>>>
>>> Let's see how this goes.
>>>
>>> Thank you.
>>>
>>> On Tue, Feb 27, 2024 at 9:05 AM LIU Yulong <liuyulong...@gmail.com> wrote:
>>>>
>>>> @Ilya, thank you, I will add that patch.
>>>>
>>>> @Eelco, thank you again, I will add a RL log to the free(ukey). Hope
>>>> we can get something useful.
>>>>
>>>>
>>>> On Mon, Feb 26, 2024 at 7:55 PM Ilya Maximets <i.maxim...@ovn.org> wrote:
>>>>>
>>>>> On 2/26/24 11:20, Eelco Chaudron wrote:
>>>>>>
>>>>>>
>>>>>> On 26 Feb 2024, at 11:10, LIU Yulong wrote:
>>>>>>
>>>>>>> Hi Eelco,
>>>>>>>
>>>>>>> Thank you for the quick response.
>>>>>>>
>>>>>>> I did not add those logs, because in order to reproduce the issue, we
>>>>>>> have to send lots of packets to the host.
>>>>>>> So there are too many ukeys created/deleted to do logging.
>>>>>>
>>>>>> Maybe a circular buffer with all alloc/free (+ 1ukey address, and 
>>>>>> timestamp), 1 or 2 Mb of memory can hold a lot.
>>>>>>
>>>>>>> And can we ensure that this [1] is the only place for ovs to free the 
>>>>>>> ukey?
>>>>>>>
>>>>>>> [1] 
>>>>>>> https://github.com/openvswitch/ovs/blob/v2.17.8/ofproto/ofproto-dpif-upcall.c#L2084
>>>>>>
>>>>>> Yes, this should be the only place, and should always be done through an 
>>>>>> RCU delayed delete.
>>>>>>
>>>>>>> For last mail, can this issue be concurrent read-and-update/delete?
>>>>>>> The revalidator_sweep__ is trying to lock the ukey->mutex, while
>>>>>>> another thread is updating the ukey->mutex to NULL and free ukey.
>>>>>>
>>>>>> This should not happen as the delete should happen by the delayed RCU 
>>>>>> delete, and if the ukey is still in the cmap after the delayed delete 
>>>>>> (quiescent state) something is wrong.
>>>>>
>>>>>
>>>>> I agree with Eelco and I don't see any abvious issues with the current
>>>>> implementation.
>>>>>
>>>>> However, the ususal suspect for RCU problems is entering quiescent state
>>>>> while iterating RCU-protected structure.  Though I'm not sure how that can
>>>>> happen in the revalidator, usually such issues are hiding somewhere way
>>>>> down the call stack.  I made a small patch that can help to be sure that
>>>>> this doesn't actually happen in your setup:
>>>>> https://github.com/igsilya/ovs/commit/8268347a159b5afa884f5b3008897878b5b520f5
>>>>>
>>>>> Could you try it?
>>>>>
>>>>> The change will log an error message and abort the process if we happen
>>>>> to enter quiescent state while iterating over the hash map.  Core dump
>>>>> will point to a problematic call.
>>>>>
>>>>> Best regards, Ilya Maximets.
>>>>>
>>>>>>
>>>>>>> LIU Yulong
>>>>>>>
>>>>>>>
>>>>>>> On Mon, Feb 26, 2024 at 5:41 PM Eelco Chaudron <echau...@redhat.com> 
>>>>>>> wrote:
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> On 26 Feb 2024, at 9:33, LIU Yulong wrote:
>>>>>>>>
>>>>>>>>> Hi,
>>>>>>>>>
>>>>>>>>> I have read the code by comparing the call stack of the core files
>>>>>>>>> carefully, and found
>>>>>>>>> a potential race condition. Please confirm whether the following 3 
>>>>>>>>> threads
>>>>>>>>> have a race condition. Just did some code trace, can such
>>>>>>>>> race condition happen?
>>>>>>>>>
>>>>>>>>> * PMD thread1 ===================================:
>>>>>>>>> -> pmd_thread_main
>>>>>>>>> -> dp_netdev_process_rxq_port
>>>>>>>>> -> dp_netdev_input
>>>>>>>>> -> dp_netdev_input__
>>>>>>>>> -> handle_packet_upcall
>>>>>>>>> -> dp_netdev_upcall
>>>>>>>>> -> upcall_cb
>>>>>>>>> -> ukey_install
>>>>>>>>> -> ukey_install__
>>>>>>>>> -> try_ukey_replace:
>>>>>>>>>             ovs_mutex_lock(&new_ukey->mutex);
>>>>>>>>> <---------- the CMAP_FOR_EACH loop in the revalidator_sweep__ run a
>>>>>>>>> bit earlier than the cmap_replace next line, so the old_ukey can be
>>>>>>>>> iterated. [1]
>>>>>>>>>             cmap_replace(&umap->cmap, &old_ukey->cmap_node,
>>>>>>>>>                          &new_ukey->cmap_node, new_ukey->hash);
>>>>>>>>>             ovsrcu_postpone(ukey_delete__, old_ukey);
>>>>>>>>> <-------- delete the ukey asynchronously. [2]
>>>>>>>>>             transition_ukey(old_ukey, UKEY_DELETED);       <--------
>>>>>>>>> transition the ukey state to UKEY_DELETED, most core files show that
>>>>>>>>> the ukey last state change was at this line. [3]
>>>>>>>>>             transition_ukey(new_ukey, UKEY_VISIBLE);
>>>>>>>>>
>>>>>>>>> [1] 
>>>>>>>>> https://github.com/openvswitch/ovs/blob/v2.17.8/ofproto/ofproto-dpif-upcall.c#L1892
>>>>>>>>> [2] 
>>>>>>>>> https://github.com/openvswitch/ovs/blob/v2.17.8/ofproto/ofproto-dpif-upcall.c#L1896
>>>>>>>>> [3] 
>>>>>>>>> https://github.com/openvswitch/ovs/blob/v2.17.8/ofproto/ofproto-dpif-upcall.c#L1897
>>>>>>>>>
>>>>>>>>> This function try_ukey_replace was called many times, because the
>>>>>>>>> `counter_upcall_ukey_replace` is not zero.
>>>>>>>>> For instance:
>>>>>>>>> {
>>>>>>>>>   name = 0x55ba9755206b "upcall_ukey_replace",
>>>>>>>>>   count = 0x55ba971c7610 <upcall_ukey_replace_count>,
>>>>>>>>>   total = 2287997,
>>>>>>>>>   last_total = 2287997,
>>>>>>>>>   min = {221, 247, 444, 278, 324, 570, 379, 464, 283, 280, 0, 427},
>>>>>>>>>   hr = {3300, 4378, 3557, 4554, 3748, 3710, 4340, 3559, 4296, 3759,
>>>>>>>>> 3522, 4136, 3660, 4428, 3802, 3652, 3880, 3375, 4806, 4221, 4158,
>>>>>>>>> 3816, 3750, 3846, 3761, 3653, 4293, 3816, 3723, 3691, 4033, 468, 4117,
>>>>>>>>> 3659, 4007, 3536,
>>>>>>>>>     3439, 4440, 3388, 4079, 3876, 3865, 4339, 3757, 3481, 4027, 3989,
>>>>>>>>> 3633, 3737, 3564, 3403, 3992, 3793, 4390, 4124, 4354, 4164, 4383,
>>>>>>>>> 4237, 3667}
>>>>>>>>> }
>>>>>>>>>
>>>>>>>>> * RCU thread2 ===================================:
>>>>>>>>> -> ovsrcu_postpone_thread
>>>>>>>>> -> ovsrcu_call_postponed
>>>>>>>>> -> ukey_delete__                                <------------ This
>>>>>>>>> function is not thead safe IMO, it has mark
>>>>>>>>> OVS_NO_THREAD_SAFETY_ANALYSIS. [4]
>>>>>>>>>
>>>>>>>>>         recirc_refs_unref(&ukey->recircs);
>>>>>>>>>         xlate_cache_delete(ukey->xcache);
>>>>>>>>>         ofpbuf_delete(ovsrcu_get(struct ofpbuf *, &ukey->actions));
>>>>>>>>>         ovs_mutex_destroy(&ukey->mutex);         <-------------- Just
>>>>>>>>> set ukey mutex to NULL. [5][6][7]
>>>>>>>>>         free(ukey);
>>>>>>>>>
>>>>>>>>> [4] 
>>>>>>>>> https://github.com/openvswitch/ovs/blob/v2.17.8/ofproto/ofproto-dpif-upcall.c#L2074
>>>>>>>>> [5] 
>>>>>>>>> https://github.com/openvswitch/ovs/blob/v2.17.8/ofproto/ofproto-dpif-upcall.c#L2083
>>>>>>>>> [6] 
>>>>>>>>> https://github.com/openvswitch/ovs/blob/v2.17.8/lib/ovs-thread.c#L131
>>>>>>>>> [7] 
>>>>>>>>> https://github.com/openvswitch/ovs/blob/v2.17.8/lib/ovs-thread.c#L124
>>>>>>>>>
>>>>>>>>> * revalidator thread3 ===================================:
>>>>>>>>>
>>>>>>>>> -> udpif_revalidator
>>>>>>>>> -> revalidator_sweep
>>>>>>>>> -> revalidator_sweep__
>>>>>>>>>
>>>>>>>>>         CMAP_FOR_EACH(ukey, cmap_node, &umap->cmap) {
>>>>>>>>>             enum ukey_state ukey_state;
>>>>>>>>>
>>>>>>>>>             if (ovs_mutex_trylock(&ukey->mutex)) {  <--------------
>>>>>>>>> Core at here, because of the NULL pointer. [8]
>>>>>>>>>                 continue;
>>>>>>>>>             }
>>>>>>>>> [8] 
>>>>>>>>> https://github.com/openvswitch/ovs/blob/v2.17.8/ofproto/ofproto-dpif-upcall.c#L2900
>>>>>>>>>
>>>>>>>>> CMIIW, if this race condition can happen, IMO, it is mostly because
>>>>>>>>> the umap is not locked during the sweep CMAP_FOR_EACH loop.
>>>>>>>>> Or some RCU protection did not work properly.
>>>>>>>>
>>>>>>>> The code looks fine to me, we are taking the lock when we update the 
>>>>>>>> cmap, which is a requirement, however iterating over the cmap as done 
>>>>>>>> above does not require a lock.
>>>>>>>>
>>>>>>>> If the RCU protection would have been broken, we would have seen a lot 
>>>>>>>> more errors. Did you have any luck adding the traces I suggested 
>>>>>>>> earlier, to see if it’s a use after free?
>>>>>>>>
>>>>>>>> Cheers,
>>>>>>>>
>>>>>>>> Eelco
>>>>>>>>
>>>>>>>>> On Wed, Feb 21, 2024 at 6:40 PM Eelco Chaudron <echau...@redhat.com> 
>>>>>>>>> wrote:
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On 21 Feb 2024, at 4:26, LIU Yulong wrote:
>>>>>>>>>>
>>>>>>>>>>> Thank you very much for your reply.
>>>>>>>>>>>
>>>>>>>>>>> The problem is not easy to reproduce, we have to wait a random long 
>>>>>>>>>>> time to see
>>>>>>>>>>> if the issue happens again. It can be more than one day or longer.
>>>>>>>>>>> OVS 2.17 with dpdk 20.11 had run to core before, so it's hard to say
>>>>>>>>>>> if it is related to DPDK.
>>>>>>>>>>> I'm running the ovs without offload to see if the issue can happen 
>>>>>>>>>>> in
>>>>>>>>>>> recent days.
>>>>>>>>>>>
>>>>>>>>>>> And again, TLDR, paste more thread call stacks.
>>>>>>>>>>> Most of the threads are in the state of sched_yield, nanosleep,
>>>>>>>>>>> epoll_wait and  poll.
>>>>>>>>>>
>>>>>>>>>> If this looks like a memory trash issue, it might be hard to figure 
>>>>>>>>>> out. Does the ukey show any kind of pattern, i.e. does the trashed 
>>>>>>>>>> data look like anything known?
>>>>>>>>>> Maybe it’s a use after free, so you could add some debugging code 
>>>>>>>>>> logging/recording all free and xmalloc of the ukey structure, to see 
>>>>>>>>>> that when it crashes it was actually allocated?
>>>>>>>>>>
>>>>>>>>>> Hope this helps you getting started.
>>>>>>>>>>
>>>>>>>>>> //Eelco
>>>>>>>>>>
>>>>>>>>>>> The following threads are in working state. So hope this can have
>>>>>>>>>>> clues for investigation.
>>>>>>>>>>>
>>>>>>>>>>> Thread 14 (Thread 0x7fd34002b700 (LWP 91928)):
>>>>>>>>>>> #0  0x00007fd344487b6d in recvmsg () at 
>>>>>>>>>>> ../sysdeps/unix/syscall-template.S:81
>>>>>>>>>>> #1  0x0000562773cb8d03 in mp_handle ()
>>>>>>>>>>> #2  0x00007fd344480e65 in start_thread (arg=0x7fd34002b700) at
>>>>>>>>>>> pthread_create.c:307
>>>>>>>>>>> #3  0x00007fd34260988d in clone () at
>>>>>>>>>>> ../sysdeps/unix/sysv/linux/x86_64/clone.S:111
>>>>>>>>>>>
>>>>>>>>>>> Thread 13 (Thread 0x7fd3359d7700 (LWP 91929)):
>>>>>>>>>>> #0  0x00007fd34448799d in accept () at 
>>>>>>>>>>> ../sysdeps/unix/syscall-template.S:81
>>>>>>>>>>> #1  0x0000562773cd8f3c in socket_listener ()
>>>>>>>>>>> #2  0x00007fd344480e65 in start_thread (arg=0x7fd3359d7700) at
>>>>>>>>>>> pthread_create.c:307
>>>>>>>>>>> #3  0x00007fd34260988d in clone () at
>>>>>>>>>>> ../sysdeps/unix/sysv/linux/x86_64/clone.S:111
>>>>>>>>>>>
>>>>>>>>>>> Thread 6 (Thread 0x7fd304663700 (LWP 91965)):
>>>>>>>>>>> #0  0x00007fd34448771d in read () at 
>>>>>>>>>>> ../sysdeps/unix/syscall-template.S:81
>>>>>>>>>>> #1  0x00007fd343b42bfb in _mlx5dv_devx_get_event () from 
>>>>>>>>>>> /lib64/libmlx5.so.1
>>>>>>>>>>> #2  0x0000562773936d86 in mlx5_vdpa_event_handle ()
>>>>>>>>>>> #3  0x00007fd344480e65 in start_thread (arg=0x7fd304663700) at
>>>>>>>>>>> pthread_create.c:307
>>>>>>>>>>> #4  0x00007fd34260988d in clone () at
>>>>>>>>>>> ../sysdeps/unix/sysv/linux/x86_64/clone.S:111
>>>>>>>>>>>
>>>>>>>>>>> Thread 2 (Thread 0x7fd305730700 (LWP 91943)):
>>>>>>>>>>> #0  ccmap_find_slot_protected (count=<synthetic pointer>,
>>>>>>>>>>> hash=hash@entry=1669671676, b=b@entry=0x7fd2f8012a80) at
>>>>>>>>>>> lib/ccmap.c:278
>>>>>>>>>>> #1  ccmap_inc_bucket_existing (b=b@entry=0x7fd2f8012a80,
>>>>>>>>>>> hash=hash@entry=1669671676, inc=inc@entry=1) at lib/ccmap.c:281
>>>>>>>>>>> #2  0x0000562773d4b015 in ccmap_try_inc
>>>>>>>>>>> (impl=impl@entry=0x7fd2f8012a40, hash=hash@entry=1669671676,
>>>>>>>>>>> inc=inc@entry=1) at lib/ccmap.c:464
>>>>>>>>>>> #3  0x0000562773d4b224 in ccmap_inc 
>>>>>>>>>>> (ccmap=ccmap@entry=0x7fd2f802a7e8,
>>>>>>>>>>> hash=1669671676) at lib/ccmap.c:485
>>>>>>>>>>> #4  0x0000562773d4975a in classifier_replace (cls=<optimized out>,
>>>>>>>>>>> rule=rule@entry=0x7fd2fac70e28, version=<optimized out>,
>>>>>>>>>>> conjs=<optimized out>, n_conjs=<optimized out>)
>>>>>>>>>>>     at lib/classifier.c:579
>>>>>>>>>>> #5  0x0000562773d49e99 in classifier_insert (cls=<optimized out>,
>>>>>>>>>>> rule=rule@entry=0x7fd2fac70e28, version=<optimized out>,
>>>>>>>>>>> conj=<optimized out>, n_conj=<optimized out>)
>>>>>>>>>>>     at lib/classifier.c:694
>>>>>>>>>>> #6  0x0000562773d00fc8 in replace_rule_start
>>>>>>>>>>> (ofproto=ofproto@entry=0x5627778cc420, ofm=ofm@entry=0x7fd3057235f0,
>>>>>>>>>>> old_rule=<optimized out>, new_rule=new_rule@entry=0x7fd2fac70e20)
>>>>>>>>>>>     at ofproto/ofproto.c:5645
>>>>>>>>>>> #7  0x0000562773d010e4 in add_flow_start (ofproto=0x5627778cc420,
>>>>>>>>>>> ofm=0x7fd3057235f0) at ofproto/ofproto.c:5256
>>>>>>>>>>> #8  0x0000562773d0122d in modify_flows_start__
>>>>>>>>>>> (ofproto=ofproto@entry=0x5627778cc420, ofm=ofm@entry=0x7fd3057235f0)
>>>>>>>>>>> at ofproto/ofproto.c:5824
>>>>>>>>>>> #9  0x0000562773d01eac in modify_flow_start_strict
>>>>>>>>>>> (ofm=0x7fd3057235f0, ofproto=0x5627778cc420) at 
>>>>>>>>>>> ofproto/ofproto.c:5953
>>>>>>>>>>> #10 ofproto_flow_mod_start (ofproto=0x5627778cc420,
>>>>>>>>>>> ofm=ofm@entry=0x7fd3057235f0) at ofproto/ofproto.c:8112
>>>>>>>>>>> #11 0x0000562773d0225a in ofproto_flow_mod_learn_start
>>>>>>>>>>> (ofm=ofm@entry=0x7fd3057235f0) at ofproto/ofproto.c:5491
>>>>>>>>>>> #12 0x0000562773d040ad in ofproto_flow_mod_learn
>>>>>>>>>>> (ofm=ofm@entry=0x7fd3057235f0, keep_ref=<optimized out>,
>>>>>>>>>>> limit=<optimized out>, 
>>>>>>>>>>> below_limitp=below_limitp@entry=0x7fd305723510)
>>>>>>>>>>>     at ofproto/ofproto.c:5576
>>>>>>>>>>> #13 0x0000562773d2641e in xlate_learn_action
>>>>>>>>>>> (ctx=ctx@entry=0x7fd305729a60, learn=learn@entry=0x562777db4618) at
>>>>>>>>>>> ofproto/ofproto-dpif-xlate.c:5547
>>>>>>>>>>> #14 0x0000562773d2aafb in do_xlate_actions (ofpacts=<optimized out>,
>>>>>>>>>>> ofpacts_len=<optimized out>, ctx=0x7fd305729a60,
>>>>>>>>>>> is_last_action=<optimized out>, group_bucket_action=<optimized out>)
>>>>>>>>>>>     at ofproto/ofproto-dpif-xlate.c:7232
>>>>>>>>>>> #15 0x0000562773d26c85 in xlate_recursively
>>>>>>>>>>> (actions_xlator=0x562773d29490 <do_xlate_actions>,
>>>>>>>>>>> is_last_action=true, deepens=false, rule=0x562777db4470,
>>>>>>>>>>> ctx=0x7fd305729a60)
>>>>>>>>>>>     at ofproto/ofproto-dpif-xlate.c:4383
>>>>>>>>>>> #16 xlate_table_action (ctx=0x7fd305729a60, in_port=<optimized out>,
>>>>>>>>>>> table_id=<optimized out>, may_packet_in=<optimized out>,
>>>>>>>>>>> honor_table_miss=<optimized out>, with_ct_orig=<optimized out>,
>>>>>>>>>>>     is_last_action=true, xlator=0x562773d29490 <do_xlate_actions>) 
>>>>>>>>>>> at
>>>>>>>>>>> ofproto/ofproto-dpif-xlate.c:4512
>>>>>>>>>>> #17 0x0000562773d2ab8d in xlate_ofpact_resubmit
>>>>>>>>>>> (resubmit=0x56277781db28, resubmit=0x56277781db28,
>>>>>>>>>>> resubmit=0x56277781db28, is_last_action=true, ctx=0x7fd305729a60)
>>>>>>>>>>>     at ofproto/ofproto-dpif-xlate.c:4823
>>>>>>>>>>> #18 do_xlate_actions (ofpacts=<optimized out>, 
>>>>>>>>>>> ofpacts_len=<optimized
>>>>>>>>>>> out>, ctx=0x7fd305729a60, is_last_action=<optimized out>,
>>>>>>>>>>> group_bucket_action=<optimized out>)
>>>>>>>>>>>     at ofproto/ofproto-dpif-xlate.c:7107
>>>>>>>>>>> #19 0x0000562773d26c85 in xlate_recursively
>>>>>>>>>>> (actions_xlator=0x562773d29490 <do_xlate_actions>,
>>>>>>>>>>> is_last_action=true, deepens=false, rule=0x562777ab9220,
>>>>>>>>>>> ctx=0x7fd305729a60)
>>>>>>>>>>>     at ofproto/ofproto-dpif-xlate.c:4383
>>>>>>>>>>> #20 xlate_table_action (ctx=0x7fd305729a60, in_port=<optimized out>,
>>>>>>>>>>> table_id=<optimized out>, may_packet_in=<optimized out>,
>>>>>>>>>>> honor_table_miss=<optimized out>, with_ct_orig=<optimized out>,
>>>>>>>>>>>     is_last_action=true, xlator=0x562773d29490 <do_xlate_actions>) 
>>>>>>>>>>> at
>>>>>>>>>>> ofproto/ofproto-dpif-xlate.c:4512
>>>>>>>>>>> #21 0x0000562773d2ab8d in xlate_ofpact_resubmit
>>>>>>>>>>> (resubmit=0x562777b9f3c8, resubmit=0x562777b9f3c8,
>>>>>>>>>>> resubmit=0x562777b9f3c8, is_last_action=true, ctx=0x7fd305729a60)
>>>>>>>>>>>     at ofproto/ofproto-dpif-xlate.c:4823
>>>>>>>>>>> #22 do_xlate_actions (ofpacts=<optimized out>, 
>>>>>>>>>>> ofpacts_len=<optimized
>>>>>>>>>>> out>, ctx=0x7fd305729a60, is_last_action=<optimized out>,
>>>>>>>>>>> group_bucket_action=<optimized out>)
>>>>>>>>>>>     at ofproto/ofproto-dpif-xlate.c:7107
>>>>>>>>>>> #23 0x0000562773d26c85 in xlate_recursively
>>>>>>>>>>> (actions_xlator=0x562773d29490 <do_xlate_actions>,
>>>>>>>>>>> is_last_action=true, deepens=false, rule=0x562777d6dc90,
>>>>>>>>>>> ctx=0x7fd305729a60)
>>>>>>>>>>>     at ofproto/ofproto-dpif-xlate.c:4383
>>>>>>>>>>> #24 xlate_table_action (ctx=0x7fd305729a60, in_port=<optimized out>,
>>>>>>>>>>> table_id=<optimized out>, may_packet_in=<optimized out>,
>>>>>>>>>>> honor_table_miss=<optimized out>, with_ct_orig=<optimized out>,
>>>>>>>>>>> ---Type <return> to continue, or q <return> to quit---
>>>>>>>>>>>     is_last_action=true, xlator=0x562773d29490 <do_xlate_actions>) 
>>>>>>>>>>> at
>>>>>>>>>>> ofproto/ofproto-dpif-xlate.c:4512
>>>>>>>>>>> #25 0x0000562773d2ab8d in xlate_ofpact_resubmit
>>>>>>>>>>> (resubmit=0x562777c13c58, resubmit=0x562777c13c58,
>>>>>>>>>>> resubmit=0x562777c13c58, is_last_action=true, ctx=0x7fd305729a60)
>>>>>>>>>>>     at ofproto/ofproto-dpif-xlate.c:4823
>>>>>>>>>>> #26 do_xlate_actions (ofpacts=<optimized out>, 
>>>>>>>>>>> ofpacts_len=<optimized
>>>>>>>>>>> out>, ctx=0x7fd305729a60, is_last_action=<optimized out>,
>>>>>>>>>>> group_bucket_action=<optimized out>)
>>>>>>>>>>>     at ofproto/ofproto-dpif-xlate.c:7107
>>>>>>>>>>> #27 0x0000562773d26c85 in xlate_recursively
>>>>>>>>>>> (actions_xlator=0x562773d29490 <do_xlate_actions>,
>>>>>>>>>>> is_last_action=true, deepens=false, rule=0x562778046000,
>>>>>>>>>>> ctx=0x7fd305729a60)
>>>>>>>>>>>     at ofproto/ofproto-dpif-xlate.c:4383
>>>>>>>>>>> #28 xlate_table_action (ctx=0x7fd305729a60, in_port=<optimized out>,
>>>>>>>>>>> table_id=<optimized out>, may_packet_in=<optimized out>,
>>>>>>>>>>> honor_table_miss=<optimized out>, with_ct_orig=<optimized out>,
>>>>>>>>>>>     is_last_action=true, xlator=0x562773d29490 <do_xlate_actions>) 
>>>>>>>>>>> at
>>>>>>>>>>> ofproto/ofproto-dpif-xlate.c:4512
>>>>>>>>>>> #29 0x0000562773d2ac1c in do_xlate_actions (ofpacts=<optimized out>,
>>>>>>>>>>> ofpacts_len=<optimized out>, ctx=0x7fd305729a60,
>>>>>>>>>>> is_last_action=<optimized out>, group_bucket_action=<optimized out>)
>>>>>>>>>>>     at ofproto/ofproto-dpif-xlate.c:7110
>>>>>>>>>>> #30 0x0000562773d26c85 in xlate_recursively
>>>>>>>>>>> (actions_xlator=0x562773d29490 <do_xlate_actions>,
>>>>>>>>>>> is_last_action=true, deepens=false, rule=0x562777ad0ac0,
>>>>>>>>>>> ctx=0x7fd305729a60)
>>>>>>>>>>>     at ofproto/ofproto-dpif-xlate.c:4383
>>>>>>>>>>> #31 xlate_table_action (ctx=0x7fd305729a60, in_port=<optimized out>,
>>>>>>>>>>> table_id=<optimized out>, may_packet_in=<optimized out>,
>>>>>>>>>>> honor_table_miss=<optimized out>, with_ct_orig=<optimized out>,
>>>>>>>>>>>     is_last_action=true, xlator=0x562773d29490 <do_xlate_actions>) 
>>>>>>>>>>> at
>>>>>>>>>>> ofproto/ofproto-dpif-xlate.c:4512
>>>>>>>>>>> #32 0x0000562773d2ac1c in do_xlate_actions (ofpacts=<optimized out>,
>>>>>>>>>>> ofpacts_len=<optimized out>, ctx=0x7fd305729a60,
>>>>>>>>>>> is_last_action=<optimized out>, group_bucket_action=<optimized out>)
>>>>>>>>>>>     at ofproto/ofproto-dpif-xlate.c:7110
>>>>>>>>>>> #33 0x0000562773d26c85 in xlate_recursively
>>>>>>>>>>> (actions_xlator=0x562773d29490 <do_xlate_actions>,
>>>>>>>>>>> is_last_action=true, deepens=false, rule=0x56277785c0d0,
>>>>>>>>>>> ctx=0x7fd305729a60)
>>>>>>>>>>>     at ofproto/ofproto-dpif-xlate.c:4383
>>>>>>>>>>> #34 xlate_table_action (ctx=0x7fd305729a60, in_port=<optimized out>,
>>>>>>>>>>> table_id=<optimized out>, may_packet_in=<optimized out>,
>>>>>>>>>>> honor_table_miss=<optimized out>, with_ct_orig=<optimized out>,
>>>>>>>>>>>     is_last_action=true, xlator=0x562773d29490 <do_xlate_actions>) 
>>>>>>>>>>> at
>>>>>>>>>>> ofproto/ofproto-dpif-xlate.c:4512
>>>>>>>>>>> #35 0x0000562773d2ac1c in do_xlate_actions (ofpacts=<optimized out>,
>>>>>>>>>>> ofpacts_len=<optimized out>, ctx=0x7fd305729a60,
>>>>>>>>>>> is_last_action=<optimized out>, group_bucket_action=<optimized out>)
>>>>>>>>>>>     at ofproto/ofproto-dpif-xlate.c:7110
>>>>>>>>>>> #36 0x0000562773d26c85 in xlate_recursively
>>>>>>>>>>> (actions_xlator=0x562773d29490 <do_xlate_actions>,
>>>>>>>>>>> is_last_action=true, deepens=false, rule=0x562777a86ce0,
>>>>>>>>>>> ctx=0x7fd305729a60)
>>>>>>>>>>>     at ofproto/ofproto-dpif-xlate.c:4383
>>>>>>>>>>> #37 xlate_table_action (ctx=0x7fd305729a60, in_port=<optimized out>,
>>>>>>>>>>> table_id=<optimized out>, may_packet_in=<optimized out>,
>>>>>>>>>>> honor_table_miss=<optimized out>, with_ct_orig=<optimized out>,
>>>>>>>>>>>     is_last_action=true, xlator=0x562773d29490 <do_xlate_actions>) 
>>>>>>>>>>> at
>>>>>>>>>>> ofproto/ofproto-dpif-xlate.c:4512
>>>>>>>>>>> #38 0x0000562773d2ac1c in do_xlate_actions (ofpacts=<optimized out>,
>>>>>>>>>>> ofpacts_len=<optimized out>, ctx=0x7fd305729a60,
>>>>>>>>>>> is_last_action=<optimized out>, group_bucket_action=<optimized out>)
>>>>>>>>>>>     at ofproto/ofproto-dpif-xlate.c:7110
>>>>>>>>>>> #39 0x0000562773d26c85 in xlate_recursively
>>>>>>>>>>> (actions_xlator=0x562773d29490 <do_xlate_actions>,
>>>>>>>>>>> is_last_action=true, deepens=false, rule=0x56277781b710,
>>>>>>>>>>> ctx=0x7fd305729a60)
>>>>>>>>>>>     at ofproto/ofproto-dpif-xlate.c:4383
>>>>>>>>>>> #40 xlate_table_action (ctx=0x7fd305729a60, in_port=<optimized out>,
>>>>>>>>>>> table_id=<optimized out>, may_packet_in=<optimized out>,
>>>>>>>>>>> honor_table_miss=<optimized out>, with_ct_orig=<optimized out>,
>>>>>>>>>>>     is_last_action=true, xlator=0x562773d29490 <do_xlate_actions>) 
>>>>>>>>>>> at
>>>>>>>>>>> ofproto/ofproto-dpif-xlate.c:4512
>>>>>>>>>>> #41 0x0000562773d2ac1c in do_xlate_actions
>>>>>>>>>>> (ofpacts=ofpacts@entry=0x562777833a38,
>>>>>>>>>>> ofpacts_len=ofpacts_len@entry=32, ctx=ctx@entry=0x7fd305729a60,
>>>>>>>>>>> is_last_action=is_last_action@entry=true,
>>>>>>>>>>>     group_bucket_action=group_bucket_action@entry=false) at
>>>>>>>>>>> ofproto/ofproto-dpif-xlate.c:7110
>>>>>>>>>>> #42 0x0000562773d30f68 in clone_xlate_actions 
>>>>>>>>>>> (actions=0x562777833a38,
>>>>>>>>>>> actions_len=32, ctx=0x7fd305729a60, is_last_action=<optimized out>,
>>>>>>>>>>> group_bucket_action=<optimized out>)
>>>>>>>>>>>     at ofproto/ofproto-dpif-xlate.c:5809
>>>>>>>>>>> #43 0x0000562773d26c85 in xlate_recursively
>>>>>>>>>>> (actions_xlator=0x562773d30d70 <clone_xlate_actions>,
>>>>>>>>>>> is_last_action=true, deepens=true, rule=0x562777ad5640,
>>>>>>>>>>> ctx=0x7fd305729a60)
>>>>>>>>>>>     at ofproto/ofproto-dpif-xlate.c:4383
>>>>>>>>>>> #44 xlate_table_action (ctx=0x7fd305729a60, in_port=<optimized out>,
>>>>>>>>>>> table_id=<optimized out>, may_packet_in=<optimized out>,
>>>>>>>>>>> honor_table_miss=<optimized out>, with_ct_orig=<optimized out>,
>>>>>>>>>>>     is_last_action=true, xlator=0x562773d30d70 
>>>>>>>>>>> <clone_xlate_actions>)
>>>>>>>>>>> at ofproto/ofproto-dpif-xlate.c:4512
>>>>>>>>>>> #45 0x0000562773d2cf20 in patch_port_output
>>>>>>>>>>> (ctx=ctx@entry=0x7fd305729a60, out_dev=0x562777ebeec0,
>>>>>>>>>>> is_last_action=is_last_action@entry=true, in_dev=0x562777f43ee0,
>>>>>>>>>>> in_dev=0x562777f43ee0)
>>>>>>>>>>>     at ofproto/ofproto-dpif-xlate.c:3890
>>>>>>>>>>> #46 0x0000562773d2d2f7 in compose_output_action__
>>>>>>>>>>> (ctx=ctx@entry=0x7fd305729a60, ofp_port=2, xr=xr@entry=0x0,
>>>>>>>>>>> check_stp=check_stp@entry=true,
>>>>>>>>>>> is_last_action=is_last_action@entry=true,
>>>>>>>>>>>     truncate=truncate@entry=false) at 
>>>>>>>>>>> ofproto/ofproto-dpif-xlate.c:4205
>>>>>>>>>>> #47 0x0000562773d2fdd0 in compose_output_action (truncate=false,
>>>>>>>>>>> is_last_action=true, xr=0x0, ofp_port=<optimized out>,
>>>>>>>>>>> ctx=0x7fd305729a60) at ofproto/ofproto-dpif-xlate.c:4360
>>>>>>>>>>> #48 xlate_output_action (ctx=ctx@entry=0x7fd305729a60, 
>>>>>>>>>>> port=<optimized
>>>>>>>>>>> out>, controller_len=<optimized out>,
>>>>>>>>>>> may_packet_in=may_packet_in@entry=true,
>>>>>>>>>>> is_last_action=is_last_action@entry=true,
>>>>>>>>>>>     truncate=truncate@entry=false,
>>>>>>>>>>> group_bucket_action=group_bucket_action@entry=false) at
>>>>>>>>>>> ofproto/ofproto-dpif-xlate.c:5305
>>>>>>>>>>> ---Type <return> to continue, or q <return> to quit---
>>>>>>>>>>> #49 0x0000562773d2972f in do_xlate_actions (ofpacts=<optimized out>,
>>>>>>>>>>> ofpacts_len=<optimized out>, ctx=0x7fd305729a60,
>>>>>>>>>>> is_last_action=<optimized out>, group_bucket_action=<optimized out>)
>>>>>>>>>>>     at ofproto/ofproto-dpif-xlate.c:6960
>>>>>>>>>>> #50 0x0000562773d26c85 in xlate_recursively
>>>>>>>>>>> (actions_xlator=0x562773d29490 <do_xlate_actions>,
>>>>>>>>>>> is_last_action=true, deepens=false, rule=0x562777ad5430,
>>>>>>>>>>> ctx=0x7fd305729a60)
>>>>>>>>>>>     at ofproto/ofproto-dpif-xlate.c:4383
>>>>>>>>>>> #51 xlate_table_action (ctx=0x7fd305729a60, in_port=<optimized out>,
>>>>>>>>>>> table_id=<optimized out>, may_packet_in=<optimized out>,
>>>>>>>>>>> honor_table_miss=<optimized out>, with_ct_orig=<optimized out>,
>>>>>>>>>>>     is_last_action=true, xlator=0x562773d29490 <do_xlate_actions>) 
>>>>>>>>>>> at
>>>>>>>>>>> ofproto/ofproto-dpif-xlate.c:4512
>>>>>>>>>>> #52 0x0000562773d2ac1c in do_xlate_actions (ofpacts=<optimized out>,
>>>>>>>>>>> ofpacts_len=<optimized out>, ctx=0x7fd305729a60,
>>>>>>>>>>> is_last_action=<optimized out>, group_bucket_action=<optimized out>)
>>>>>>>>>>>     at ofproto/ofproto-dpif-xlate.c:7110
>>>>>>>>>>> #53 0x0000562773d26c85 in xlate_recursively
>>>>>>>>>>> (actions_xlator=0x562773d29490 <do_xlate_actions>,
>>>>>>>>>>> is_last_action=true, deepens=false, rule=0x562777ac2fb0,
>>>>>>>>>>> ctx=0x7fd305729a60)
>>>>>>>>>>>     at ofproto/ofproto-dpif-xlate.c:4383
>>>>>>>>>>> #54 xlate_table_action (ctx=0x7fd305729a60, in_port=<optimized out>,
>>>>>>>>>>> table_id=<optimized out>, may_packet_in=<optimized out>,
>>>>>>>>>>> honor_table_miss=<optimized out>, with_ct_orig=<optimized out>,
>>>>>>>>>>>     is_last_action=true, xlator=0x562773d29490 <do_xlate_actions>) 
>>>>>>>>>>> at
>>>>>>>>>>> ofproto/ofproto-dpif-xlate.c:4512
>>>>>>>>>>> #55 0x0000562773d2ac1c in do_xlate_actions
>>>>>>>>>>> (ofpacts=ofpacts@entry=0x56277797b318,
>>>>>>>>>>> ofpacts_len=ofpacts_len@entry=8, ctx=ctx@entry=0x7fd305729a60,
>>>>>>>>>>> is_last_action=is_last_action@entry=true,
>>>>>>>>>>>     group_bucket_action=group_bucket_action@entry=false) at
>>>>>>>>>>> ofproto/ofproto-dpif-xlate.c:7110
>>>>>>>>>>> #56 0x0000562773d330d6 in xlate_actions 
>>>>>>>>>>> (xin=xin@entry=0x7fd30572a920,
>>>>>>>>>>> xout=xout@entry=0x7fd30572ad38) at ofproto/ofproto-dpif-xlate.c:7924
>>>>>>>>>>> #57 0x0000562773d2241b in upcall_xlate (wc=0x7fd30572bfe0,
>>>>>>>>>>> odp_actions=0x7fd30572b7b0, upcall=0x7fd30572acd0,
>>>>>>>>>>> udpif=0x562777850cf0) at ofproto/ofproto-dpif-upcall.c:1340
>>>>>>>>>>> #58 process_upcall (udpif=udpif@entry=0x562777850cf0,
>>>>>>>>>>> upcall=upcall@entry=0x7fd30572acd0,
>>>>>>>>>>> odp_actions=odp_actions@entry=0x7fd30572b7b0,
>>>>>>>>>>> wc=wc@entry=0x7fd30572bfe0)
>>>>>>>>>>>     at ofproto/ofproto-dpif-upcall.c:1602
>>>>>>>>>>> #59 0x0000562773d22c39 in upcall_cb (packet=<optimized out>,
>>>>>>>>>>> flow=0x7fd30572bd40, ufid=<optimized out>, pmd_id=<optimized out>,
>>>>>>>>>>> type=<optimized out>, userdata=<optimized out>,
>>>>>>>>>>>     actions=0x7fd30572b7b0, wc=0x7fd30572bfe0,
>>>>>>>>>>> put_actions=0x7fd30572b7f0, aux=0x562777850cf0) at
>>>>>>>>>>> ofproto/ofproto-dpif-upcall.c:1461
>>>>>>>>>>> #60 0x0000562773d52198 in dp_netdev_upcall
>>>>>>>>>>> (pmd=pmd@entry=0x7fd305731010, packet_=packet_@entry=0x21152c180,
>>>>>>>>>>> flow=flow@entry=0x7fd30572bd40, wc=wc@entry=0x7fd30572bfe0,
>>>>>>>>>>>     ufid=ufid@entry=0x7fd30572b790, type=type@entry=DPIF_UC_MISS,
>>>>>>>>>>> userdata=userdata@entry=0x0, actions=actions@entry=0x7fd30572b7b0,
>>>>>>>>>>> put_actions=put_actions@entry=0x7fd30572b7f0)
>>>>>>>>>>>     at lib/dpif-netdev.c:9141
>>>>>>>>>>> #61 0x0000562773d66f13 in handle_packet_upcall
>>>>>>>>>>> (put_actions=0x7fd30572b7f0, actions=0x7fd30572b7b0,
>>>>>>>>>>> key=0x7fd30572cc40, packet=0x21152c180, pmd=0x7fd305731010) at
>>>>>>>>>>> lib/dpif-netdev.c:11303
>>>>>>>>>>> #62 fast_path_processing (pmd=pmd@entry=0x7fd305731010,
>>>>>>>>>>> packets_=packets_@entry=0x7fd30572d0c0,
>>>>>>>>>>> keys=keys@entry=0x7fd30572cc30,
>>>>>>>>>>> flow_map=flow_map@entry=0x7fd30572cae0,
>>>>>>>>>>>     index_map=index_map@entry=0x7fd30572cad0 "", in_port=<optimized
>>>>>>>>>>> out>) at lib/dpif-netdev.c:11426
>>>>>>>>>>> #63 0x0000562773d680c1 in dp_netdev_input__ (pmd=<optimized out>,
>>>>>>>>>>> packets=<optimized out>, md_is_valid=md_is_valid@entry=false,
>>>>>>>>>>> port_no=<optimized out>) at lib/dpif-netdev.c:11520
>>>>>>>>>>> #64 0x0000562773d6a15d in dp_netdev_input (pmd=<optimized out>,
>>>>>>>>>>> packets=<optimized out>, port_no=<optimized out>) at
>>>>>>>>>>> lib/dpif-netdev.c:11558
>>>>>>>>>>> #65 0x0000562773d6a2cf in dp_netdev_process_rxq_port
>>>>>>>>>>> (pmd=pmd@entry=0x7fd305731010, rxq=0x56277796a5e0, port_no=3) at
>>>>>>>>>>> lib/dpif-netdev.c:6660
>>>>>>>>>>> #66 0x0000562773d6a759 in pmd_thread_main (f_=<optimized out>) at
>>>>>>>>>>> lib/dpif-netdev.c:8267
>>>>>>>>>>> #67 0x0000562773df805f in ovsthread_wrapper (aux_=<optimized out>) 
>>>>>>>>>>> at
>>>>>>>>>>> lib/ovs-thread.c:422
>>>>>>>>>>> #68 0x00007fd344480e65 in start_thread (arg=0x7fd305730700) at
>>>>>>>>>>> pthread_create.c:307
>>>>>>>>>>> #69 0x00007fd34260988d in clone () at
>>>>>>>>>>> ../sysdeps/unix/sysv/linux/x86_64/clone.S:111
>>>>>>>>>>>
>>>>>>>>>>> Thread 1 (Thread 0x7fd334307700 (LWP 91942)):
>>>>>>>>>>> #0  0x00007fd342541337 in __GI_raise (sig=sig@entry=6) at
>>>>>>>>>>> ../nptl/sysdeps/unix/sysv/linux/raise.c:55
>>>>>>>>>>> #1  0x00007fd342542a28 in __GI_abort () at abort.c:90
>>>>>>>>>>> #2  0x0000562773e2c4ee in ovs_abort_valist (err_no=<optimized out>,
>>>>>>>>>>> format=<optimized out>, args=args@entry=0x7fd334302340) at
>>>>>>>>>>> lib/util.c:499
>>>>>>>>>>> #3  0x0000562773e2c584 in ovs_abort (err_no=err_no@entry=0,
>>>>>>>>>>> format=format@entry=0x5627740cad18 "%s: %s() passed uninitialized
>>>>>>>>>>> ovs_mutex") at lib/util.c:491
>>>>>>>>>>> #4  0x0000562773df72e1 in ovs_mutex_lock_at
>>>>>>>>>>> (l_=l_@entry=0x7fd2f907df68, where=where@entry=0x5627740a7fc0
>>>>>>>>>>> "ofproto/ofproto-dpif-upcall.c:2214") at lib/ovs-thread.c:75
>>>>>>>>>>> #5  0x0000562773d1e98d in ukey_delete 
>>>>>>>>>>> (umap=umap@entry=0x562777853400,
>>>>>>>>>>> ukey=ukey@entry=0x7fd2f907df20) at 
>>>>>>>>>>> ofproto/ofproto-dpif-upcall.c:2214
>>>>>>>>>>> #6  0x0000562773d202da in revalidator_sweep__
>>>>>>>>>>> (revalidator=revalidator@entry=0x562777897b00,
>>>>>>>>>>> purge=purge@entry=false) at ofproto/ofproto-dpif-upcall.c:3048
>>>>>>>>>>> #7  0x0000562773d241a6 in revalidator_sweep
>>>>>>>>>>> (revalidator=0x562777897b00) at ofproto/ofproto-dpif-upcall.c:3072
>>>>>>>>>>> #8  udpif_revalidator (arg=0x562777897b00) at 
>>>>>>>>>>> ofproto/ofproto-dpif-upcall.c:1086
>>>>>>>>>>> #9  0x0000562773df805f in ovsthread_wrapper (aux_=<optimized out>) 
>>>>>>>>>>> at
>>>>>>>>>>> lib/ovs-thread.c:422
>>>>>>>>>>> #10 0x00007fd344480e65 in start_thread (arg=0x7fd334307700) at
>>>>>>>>>>> pthread_create.c:307
>>>>>>>>>>> #11 0x00007fd34260988d in clone () at
>>>>>>>>>>> ../sysdeps/unix/sysv/linux/x86_64/clone.S:111
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Thanks
>>>>>>>>>>> LIU Yulong
>>>>>>>>>>>
>>>>>>>>>>> On Mon, Feb 19, 2024 at 8:12 PM Eelco Chaudron 
>>>>>>>>>>> <echau...@redhat.com> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> On 19 Feb 2024, at 13:09, Ilya Maximets wrote:
>>>>>>>>>>>>
>>>>>>>>>>>>> On 2/19/24 11:14, Eelco Chaudron wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On 19 Feb 2024, at 10:34, LIU Yulong wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Hi OVS experts,
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Our ovs-vswitchd runs to core at the 
>>>>>>>>>>>>>>> ovs_mutex_trylock(&ukey->mutex) in the
>>>>>>>>>>>>>>> function revalidator_sweep__.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> I've sent the mail before but have no response.
>>>>>>>>>>>>>>> https://mail.openvswitch.org/pipermail/ovs-discuss/2023-August/052604.html
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> So I'm trying to send this mail again. And I may apologize in 
>>>>>>>>>>>>>>> advance because
>>>>>>>>>>>>>>> I would like to post as much useful information as possible to 
>>>>>>>>>>>>>>> help identify
>>>>>>>>>>>>>>> potential issues. So this mail will have a really long text.
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> Compared to the mail 2023-August/052604.html, we upgrade the 
>>>>>>>>>>>>>>> OVS to 2.17.8
>>>>>>>>>>>>>>> and DPDK to 22.11 to pray for good luck that maybe the 
>>>>>>>>>>>>>>> community has potential
>>>>>>>>>>>>>>> fixes for this issue. But unfortunately, the ovs-vswitchd still 
>>>>>>>>>>>>>>> runs to core.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> As you mentioned it looks like some memory corruption, which I 
>>>>>>>>>>>>>> have not seen before.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Have you tried this without rte offload? This is the only 
>>>>>>>>>>>>>> feature I never used.
>>>>>>>>>>>>>> There is a 2.17.9 with DPDK 22.11.6 you could try.
>>>>>>>>>>>>>
>>>>>>>>>>>>> OVS 2.17 is not supposed to work with DPDK 22.11, it's supposed 
>>>>>>>>>>>>> to work with 21.11.
>>>>>>>>>>>>> See the compatibility table here:
>>>>>>>>>>>>> https://docs.openvswitch.org/en/latest/faq/releases/
>>>>>>>>>>>>>
>>>>>>>>>>>>> Though it's hard to tell if DPDK version is anyhow related to the 
>>>>>>>>>>>>> issue.
>>>>>>>>>>>>
>>>>>>>>>>>> My mistake, I was supposed to type 21.11.6 :( But yes if they are 
>>>>>>>>>>>> using 22.11, that could also be the problem. I would suggest using 
>>>>>>>>>>>> the supported version and see if the problem goes away.
>>>>>>>>>>>>
>>>>>>>>>>>> //Eelco
>>>>>>>>>>>>
>>>>>>>>>>>>> Best regards, Ilya Maximets.
>>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>
>>>>>>
>>>>>
>>


_______________________________________________
discuss mailing list
disc...@openvswitch.org
https://mail.openvswitch.org/mailman/listinfo/ovs-discuss

Reply via email to