Hi Ilia,

Thanks for all the additional information!

On 3/2/25 8:55 PM, support via discuss wrote:
> Probably i found something interesting here.
> top -H -p $(pidof ovs-vswitchd)
>    PID USER      PR  NI    VIRT    RES    SHR S  %CPU  %MEM     TIME+
> COMMAND                    
>  673931 root      20   0 8392700 462408   7040 R  99.9   0.1   1216:19
> ovs-vswitchd                         
>  673941 root      20   0 8392700 462408   7040 S   6.7   0.1   4:18.00
> handler10                                
>  673982 root      20   0 8392700 462408   7040 S   6.7   0.1   0:49.26
> handler51                                
>  674009 root      20   0 8392700 462408   7040 S   6.7   0.1  73:17.17
> handler78                               
>  673932 root      20   0 8392700 462408   7040 S   0.0   0.1   2:07.06
> handler1                                  
>  673933 root      20   0 8392700 462408   7040 S   0.0   0.1   0:34.96
> handler2                                  
>  673934 root      20   0 8392700 462408   7040 S   0.0   0.1   3:57.85
> handler3                                  
>  673935 root      20   0 8392700 462408   7040 S   0.0   0.1   0:41.66
> handler4                                  
>  673936 root      20   0 8392700 462408   7040 S   0.0   0.1  18:46.97
> urcu5                                      
>  673937 root      20   0 8392700 462408   7040 S   0.0   0.1   4:04.14
> handler6                                  
>  673938 root      20   0 8392700 462408   7040 S   0.0   0.1   0:44.03
> handler7                                  
>  673939 root      20   0 8392700 462408   7040 S   0.0   0.1   4:20.97
> handler8
> 
> I guess t​​his could be a real issue of lost DHCP packets.
> Next are top lines of perf record
>    1.19%  ovs-vswitchd    libc.so.6          [.] realloc               
>                                                                
>    0.64%  revalidator90   ovs-vswitchd       [.] 0x00000000000797de     
>                                              
>    0.59%  ovs-vswitchd    ovs-vswitchd       [.] 0x00000000000487de     
>                                              
>    0.58%  ovs-vswitchd    ovs-vswitchd       [.] 0x00000000000797de     
>                                              
>    0.58%  ovs-vswitchd    [kernel.kallsyms]  [k] available_idle_cpu     
>                                                  
>    0.57%  ovs-vswitchd    [kernel.kallsyms]  [k] tun_net_xmit           
>                                                     
>    0.55%  ovs-vswitchd    ovs-vswitchd       [.] 0x00000000000a561a     
>                                              
>    0.51%  ovs-vswitchd    libc.so.6          [.] 0x0000000000191747     
>                                                  
>    0.50%  ovs-vswitchd    [kernel.kallsyms]  [k] task_h_load           
>                                                      
>    0.48%  ovs-vswitchd    [kernel.kallsyms]  [k] llist_add_batch       
>                                                      
>    0.48%  revalidator90   ovs-vswitchd       [.] 0x00000000000797ca     
>                                               
>    0.47%  revalidator96   ovs-vswitchd       [.] 0x00000000000797de     
>                                              
>    0.45%  revalidator90   ovs-vswitchd       [.] 0x00000000000487de     
>                                              
>    0.44%  revalidator93   ovs-vswitchd       [.] 0x00000000000487de     
>                                              
>    0.43%  revalidator93   ovs-vswitchd       [.] 0x00000000000797de     
>                                              
>    0.40%  ovs-vswitchd    ovs-vswitchd       [.] 0x00000000000797ca     
>                                              
>    0.40%  revalidator112  ovs-vswitchd       [.] 0x00000000000797de     
>                                             
>    0.39%  ovs-vswitchd    [kernel.kallsyms]  [k] __dev_queue_xmit
> 
> I've definitely saw bug reports regarding libc and openvswitch.
> 
>  
>> I'm back to digging into this more deep.
>>     PID USER      PR  NI    VIRT    RES    SHR S  %CPU  %MEM     TIME+
>> COMMAND                    
>>    4446 root      20   0  395500  88276   6336 S   1.0   0.0  13:07.66
>> ovn_pinctrl0                                
>>    3773 root      20   0  395500  88276   6336 S   0.0   0.0   5:19.43
>> ovn-controller                              
>>    4449 root      20   0  395500  88276   6336 S   0.0   0.0   0:00.00
>> urcu1                                           
>>    4452 root      20   0  395500  88276   6336 S   0.0   0.0   0:00.07
>> ovn_statctrl2                                
>>    5832 root      20   0  395500  88276   6336 S   0.0   0.0   0:04.07
>> stopwatch3
>>
>> Interesting thing that ovn-controller as well as pinctrl doesn't seems
>> starvating for CPU as it is usage is likely about 1%.
>> Host CPU is also feeling a way good
>> %Cpu(s): 29.5 us, 20.0 sy,  0.0 ni, 50.0 id,  0.0 wa,  0.0 hi,  0.6
>> si,  0.0 st
>>
>> Despite that, ovs-vswitchd seems consuming way more CPU
>>
>>    PID USER      PR  NI    VIRT    RES    SHR S  %CPU  %MEM     TIME+
>> COMMAND                     
>>  139416 root      20   0 8378828 452288   8448 R  98.7   0.1 990:55.85
>> ovs-vswitchd                        
>>  139509 root      20   0 8378828 452288   8448 S  10.7   0.1  82:49.98
>> revalidator90                         
>>  139522 root      20   0 8378828 452288   8448 R  10.7   0.1  32:01.96
>> revalidator103                       
>>  139519 root      20   0 8378828 452288   8448 R  10.3   0.1  54:28.10
>> revalidator100                       
>>  139516 root      20   0 8378828 452288   8448 S  10.0   0.1  32:34.74
>> revalidator97                         
>>  139526 root      20   0 8378828 452288   8448 R  10.0   0.1  49:16.24
>> revalidator107                       
>>  139531 root      20   0 8378828 452288   8448 R  10.0   0.1  40:45.87
>> revalidator112                      
>>  139523 root      20   0 8378828 452288   8448 S   9.7   0.1  49:39.00
>> revalidator104                        
>>  139512 root      20   0 8378828 452288   8448 S   7.7   0.1  49:37.77
>> revalidator93                          
>>  139497 root      20   0 8378828 452288   8448 S   7.3   0.1  67:31.93
>> handler78                              
>>  139513 root      20   0 8378828 452288   8448 S   7.3   0.1  36:44.18
>> revalidator94                          
>>  139525 root      20   0 8378828 452288   8448 S   6.3   0.1  36:08.10
>> revalidator106                        
>>  139520 root      20   0 8378828 452288   8448 S   5.7   0.1  49:48.41
>> revalidator101                        
>>  139521 root      20   0 8378828 452288   8448 R   5.7   0.1  36:22.60
>> revalidator102                        
>>  139510 root      20   0 8378828 452288   8448 R   5.0   0.1  48:53.38
>> revalidator91                          
>>  139511 root      20   0 8378828 452288   8448 R   5.0   0.1  54:09.25
>> revalidator92
>>
>> More logs:
>> 15:44:45.190923 IP 0.0.0.0.68 > 255.255.255.255.67: BOOTP/DHCP,
>> Request from fa:16:3e:50:22:c4, length 300
>> 15:44:46.446755 IP 0.0.0.0.68 > 255.255.255.255.67: BOOTP/DHCP,
>> Request from fa:16:3e:50:22:c4, length 300
>> 15:44:47.491209 IP 0.0.0.0.68 > 255.255.255.255.67: BOOTP/DHCP,
>> Request from fa:16:3e:50:22:c4, length 300
>> 15:44:47.869273 IP 0.0.0.0.68 > 255.255.255.255.67: BOOTP/DHCP,
>> Request from fa:16:3e:50:22:c4, length 300
>> 15:44:48.297140 IP 0.0.0.0.68 > 255.255.255.255.67: BOOTP/DHCP,
>> Request from fa:16:3e:50:22:c4, length 300
>> 15:44:51.190709 IP 0.0.0.0.68 > 255.255.255.255.67: BOOTP/DHCP,
>> Request from fa:16:3e:50:22:c4, length 300
>>
>> Manually tried to use dhclient to obtain IP but no success. There is
>> logs that ovn-controller sent DHCPOFFER but no packet arrived to VM
>> interface.
>>> Switched back to configuration with no neutron-dhcp-agent. I have synced
>>> ovn db with neutron-ovn-db-sync-util.
>>>
>>> There is coverage/show result of node where i'm seeing this issue right
>>> now (current network bandwidth is about 300Mbps).
>>> # ovs-appctl coverage/show
>>> Event coverage, avg rate over last: 5 seconds, last minute, last hour, 
>>> hash=982b7b40:
>>> nln_changed                0.0/sec     0.000/sec 0.0197/sec   total: 8568
>>> route_table_dump           0.0/sec     0.000/sec 0.0072/sec   total: 3953
>>> netlink_sent             1158.4/sec  1109.700/sec 1018.6758/sec   total:
>>> 840427925
>>> netlink_recv_jumbo       144.2/sec   140.883/sec 212.7489/sec   total:
>>> 127752548
>>> netlink_received         1258.4/sec  1229.533/sec 1192.0311/sec   total:
>>> 1072129295
>>> netdev_set_ethtool         0.0/sec     0.000/sec 0.0006/sec   total: 361
>>> netdev_get_ethtool         0.0/sec     0.000/sec 0.0039/sec   total: 2808
>>> netdev_get_ifindex         0.0/sec     0.000/sec 0.3358/sec   total:
>>> 14218
>>> netdev_set_policing        0.0/sec     0.000/sec 0.0028/sec   total: 1685
>>> vconn_sent               596.2/sec   576.967/sec 585.4778/sec   total:
>>> 643013987
>>> vconn_received           320.0/sec   310.883/sec 307.5108/sec   total:
>>> 360905481
>>> util_xalloc              937354.8/sec 912615.150/sec 873946.4494/sec  
>>> total: 1068314545738
>>> unixctl_replied            0.0/sec     0.033/sec 0.0331/sec   total:
>>> 33146
>>> unixctl_received           0.0/sec     0.033/sec 0.0331/sec   total:
>>> 33146
>>> long_poll_interval         0.0/sec     0.000/sec 0.0000/sec   total: 1
>>> stream_open                0.0/sec     0.000/sec 0.0000/sec   total: 2
>>> pstream_open               0.0/sec     0.000/sec 0.0000/sec   total: 5
>>> seq_change               3283.6/sec  3255.167/sec 3039.7792/sec   total:
>>> 3088392125
>>> rconn_sent               596.2/sec   576.967/sec 585.4767/sec   total:
>>> 643013979
>>> rconn_queued             596.2/sec   576.967/sec 585.5044/sec   total:
>>> 643014079
>>> rconn_overflow             0.0/sec     0.000/sec 2.0272/sec   total:
>>> 16543

This means we're tail dropping messages that should be sent to the
controller because we're reaching the queue size limit (100 by default
IIUC).

>>> rconn_discarded            0.0/sec     0.000/sec 0.0278/sec   total: 100

This indicates that a controller connection (ovn-controller has a few)
has disconnected while there weere messages still queued on it to be
sent to the controller.

>>> poll_zero_timeout        275.2/sec   247.600/sec 161.9242/sec   total:
>>> 136657138
>>> poll_create_node         4749.6/sec  4417.217/sec 4062.3500/sec   total:
>>> 4007153255
>>> txn_success                0.2/sec     0.183/sec 0.1828/sec   total:
>>> 184805
>>> txn_incomplete             0.4/sec     0.367/sec 0.4856/sec   total:
>>> 372203
>>> txn_unchanged              0.0/sec     0.000/sec 0.1169/sec   total: 7146
>>> netdev_get_stats          77.6/sec    71.133/sec 70.5939/sec   total:
>>> 73986502
>>> mac_learning_moved         0.0/sec     0.000/sec 0.0000/sec   total: 425
>>> mac_learning_expired       0.6/sec     0.217/sec 0.4439/sec   total:
>>> 409542
>>> mac_learning_learned       0.2/sec     0.233/sec 0.4403/sec   total:
>>> 410444
>>> hmap_expand              298.0/sec   265.833/sec 329.0514/sec   total:
>>> 291663955
>>> hmap_pathological         10.4/sec     6.050/sec 9.1406/sec   total:
>>> 8189560
>>> hindex_expand              0.0/sec     0.000/sec 0.0000/sec   total: 12
>>> hindex_pathological        0.0/sec     0.000/sec 0.0000/sec   total: 148
>>> miniflow_malloc            0.0/sec     0.000/sec 21.7272/sec   total:
>>> 3395676
>>> flow_extract             843.4/sec   801.333/sec 727.9828/sec   total:
>>> 755385765
>>> dpif_port_del              0.0/sec     0.000/sec 0.0006/sec   total: 379
>>> dpif_port_add              0.0/sec     0.000/sec 0.0006/sec   total: 367
>>> dpif_flow_put_error        6.0/sec     7.633/sec 5.1642/sec   total:
>>> 4420389
>>> dpif_flow_put            232.2/sec   208.867/sec 126.8286/sec   total:
>>> 98257942
>>> dpif_flow_get              0.0/sec     0.000/sec 0.0000/sec   total: 386
>>> dpif_flow_flush            0.0/sec     0.000/sec 0.0000/sec   total: 1
>>> dpif_flow_del            182.2/sec   203.100/sec 118.5047/sec   total:
>>> 93717525
>>> dpif_execute_with_help     0.0/sec     0.000/sec 0.0017/sec   total: 1649
>>> dpif_execute_error         0.0/sec     0.000/sec 0.0006/sec   total: 398
>>> dpif_execute             571.8/sec   537.350/sec 439.8069/sec   total:
>>> 472799472
>>> cmap_shrink                0.8/sec     0.733/sec 1.4025/sec   total:
>>> 1180907
>>> cmap_expand                1.0/sec     0.583/sec 1.5000/sec   total:
>>> 1182882
>>> ccmap_shrink               0.0/sec     0.000/sec 0.1786/sec   total: 747
>>> ccmap_expand               0.0/sec     0.000/sec 0.1200/sec   total: 774
>>> connmgr_async_unsent       0.0/sec     0.000/sec 0.0372/sec   total: 139
>>> xlate_actions            4171.0/sec  3071.683/sec 3133.7436/sec   total:
>>> 3107954003
>>> upcall_ukey_replace        0.0/sec     0.117/sec 0.1739/sec   total:
>>> 155715
>>> upcall_ukey_contention     0.0/sec     0.000/sec 0.0003/sec   total: 157
>>> upcall_flow_limit_grew     0.0/sec     0.550/sec 0.2328/sec   total:
>>> 56359
>>> ukey_replace_contention    0.0/sec     0.017/sec 0.0008/sec   total: 847
>>> revalidate_missed_dp_flow   0.0/sec     0.117/sec 0.0472/sec   total:
>>> 106876
>>> handler_duplicate_upcall  14.0/sec    12.817/sec 11.5303/sec   total:
>>> 9317449
>>> dumped_new_flow            0.0/sec     0.000/sec 0.0000/sec   total: 842
>>> dumped_duplicate_flow      0.0/sec     0.000/sec 0.0003/sec   total:
>>> 37029
>>> rev_mac_learning           0.8/sec     0.417/sec 0.6419/sec   total:
>>> 708403
>>> rev_flow_table             0.0/sec     0.000/sec 0.0033/sec   total: 4003
>>> rev_port_toggled           0.0/sec     0.000/sec 0.0000/sec   total: 119
>>> rev_reconfigure            0.0/sec     0.000/sec 0.0008/sec   total: 691
>>> ofproto_update_port        0.0/sec     0.000/sec 0.0039/sec   total: 2430
>>> ofproto_recv_openflow    320.0/sec   310.883/sec 307.5097/sec   total:
>>> 360905473
>>> ofproto_packet_out       319.6/sec   310.633/sec 299.3358/sec   total:
>>> 360555564
>>> ofproto_flush              0.0/sec     0.000/sec 0.0000/sec   total: 1
>>> bridge_reconfigure         0.0/sec     0.000/sec 0.1125/sec   total: 5108
>>> 100 events never hit
>>>
>>> > Do you see any warning/error logs in ovs-vswitchd.log?
>>> I see no noticable errors except usual warn logs like invalid packet.
>>>
>>> > Just double checking, you're saying that with low numbers of instances
>>> > you don't have this issue?
>>> Exactly. With relatively low instances count there is no visible issue.
>>>
>>> > I wonder if you're not potentially hitting the maximum resubmit limit
>>> > (4096 resubmits) due to DHCPOFFERS being sent as broadcast. There
>>> > should be warning logs indicating that in ovs-vswitchd.
>>>
>>> There is no error or warn logs mentioning resubmit, but a lot of such
>>> logs:
>>>
>>> 2025-03-01T11:27:16.473Z|19884389|vconn|DBG|unix#33063: received:
>>> OFPT_PACKET_OUT (OF1.5) (xid=0xe3465): in_port=CONTROLLER
>>> actions=set_field:0x300->reg0,set_field:0x9->reg11,set_field:0xa-
>>> >reg12,set_field:0xe0->reg13,set_field:0x1->reg14,set_field:0x291-
>>> >reg15,set_field:0x2->metadata,set_field:0-
>>> >pkt_mark,set_field:0x9002->reg6,set_field:0x8000-
>>> >reg15,resubmit(CONTROLLER,43)
>>> data_len=60
>>> 2025-03-01T11:27:16.972Z|19884399|vconn|DBG|unix#33063: received:
>>> OFPT_PACKET_OUT (OF1.5) (xid=0xe358a): in_port=CONTROLLER
>>> actions=set_field:0x300->reg0,set_field:0x9->reg11,set_field:0xa-
>>> >reg12,set_field:0x93->reg13,set_field:0x1->reg14,set_field:0xdd-
>>> >reg15,set_field:0x2->metadata,set_field:0-
>>> >pkt_mark,set_field:0x9001->reg6,set_field:0x8000-
>>> >reg15,resubmit(CONTROLLER,43)
>>> data_len=60
>>> 2025-03-01T11:27:17.873Z|19884417|vconn|DBG|unix#33063: received:
>>> OFPT_PACKET_OUT (OF1.5) (xid=0xe3765): in_port=CONTROLLER
>>> actions=set_field:0x300->reg0,set_field:0x9->reg11,set_field:0xa-
>>> >reg12,set_field:0x93->reg13,set_field:0x1->reg14,set_field:0xdd-
>>> >reg15,set_field:0x2->metadata,set_field:0-
>>> >pkt_mark,set_field:0x9001->reg6,set_field:0x8000-
>>> >reg15,resubmit(CONTROLLER,43)
>>> data_len=60
>>> 2025-03-01T11:27:17.972Z|19884419|vconn|DBG|unix#33063: received:
>>> OFPT_PACKET_OUT (OF1.5) (xid=0xe37ba): in_port=CONTROLLER
>>> actions=set_field:0x300->reg0,set_field:0x9->reg11,set_field:0xa-
>>> >reg12,set_field:0xe0->reg13,set_field:0x1->reg14,set_field:0x291-
>>> >reg15,set_field:0x2->metadata,set_field:0-
>>> >pkt_mark,set_field:0x9002->reg6,set_field:0x8000-
>>> >reg15,resubmit(CONTROLLER,43)
>>> data_len=60

These are due to the fact that OVN's multicast groups have been split
since commit
https://github.com/ovn-org/ovn/commit/325c7b203d8bfd12bc1285ad11390c1a55cd6717

A similar problem has been reported/discussed in:
https://mail.openvswitch.org/pipermail/ovs-discuss/2025-February/053455.html

To confirm that it's ARP requests (broadcast) that's causing problems in
your case, we'd need some more logs from you:

ovn-appctl vlog/set pinctrl:dbg

We're interested in logs that look like:
2025-02-12T20:35:41.490Z|10791|pinctrl(ovn_pinctrl0)|DBG|pinctrl
received  packet-in | opcode=...

I'm guessing your topology might be similar (a "large" public logical
switch with quite a few connected logical routers).  If so, a potential
solution discussed there was to set
"other_config:broadcast-arps-to-all-routers=false" on the public logical
switch.  Would you be able to check if changing that setting helps in
your case too please?

Thanks,
Dumitru

>>>
>>> I've checked presence of error/warn logs at ovs-vswitchd using next
>>> command
>>> # cat /var/log/kolla/openvswitch/ovs-vswitchd.log | grep -i "resubmit" |
>>> grep -iE "warn|err
>>>
>>> Also i have non-deployed region running 2024.2 OpenStack and did some
>>> dumb tests like creating 1500 VMs with about 500 instances per
>>> hypervisor and didn't see this issue. However, i didn't use L2 network
>>> in this case, only virtual network like geneve/vxlan. I will try today
>>> to create sample L2 vlan with mac learning enabled at the border router
>>> and report back. I think this bug will not occur if there is no a lot of
>>> packets coming through ovs, handling DHCP with no issues...
>>>
>>> 24.02.2025 16:00, Dumitru Ceara пишет:
>>> > On 2/20/25 11:13 PM, Ilia Baikov via discuss wrote:
>>> >> Hi
>>> >>
>>> > Hi Ilia,
>>> >
>>> > + Ilya Maximets for more ovs insights.
>>> >
>>> >> Over a lot of observes and testing various number of possible
>>> fixes came
>>> >> with nothing...
>>> >>
>>> >> Yes, there are messages that contains NXT_RESUME substring.
>>> >>
>>> >> ovn-controller:
>>> >> 2025-02-18T21:05:51.771Z|1041129|vconn(ovn_pinctrl0)|DBG|unix:/
>>> var/run/
>>> >> openvswitch/br-int.mgmt: sent (Success): NXT_RESUME (OF1.5) (xid=0x0):
>>> >> cookie=0x404b44bc total_len=348
>>> >>
>>> reg0=0x308,reg11=0x9,reg12=0xa,reg13=0x13c,reg14=0x4a8,metadata=0x2,in_port=312
>>>  (via action) data_len=348 (unbuffered)
>>> >> 2025-02-18T21:08:23.729Z|1044255|vconn(ovn_pinctrl0)|DBG|unix:/
>>> var/run/
>>> >> openvswitch/br-int.mgmt: sent (Success): NXT_RESUME (OF1.5) (xid=0x0):
>>> >> cookie=0xffe9d519 total_len=348
>>> >>
>>> reg0=0x308,reg11=0x9,reg12=0xa,reg13=0x112,reg14=0x2dd,metadata=0x2,in_port=363
>>>  (via action) data_len=348 (unbuffered)
>>> >> 2025-02-18T21:10:33.865Z|1046917|vconn(ovn_pinctrl0)|DBG|unix:/
>>> var/run/
>>> >> openvswitch/br-int.mgmt: sent (Success): NXT_RESUME (OF1.5) (xid=0x0):
>>> >> cookie=0x5bc96bd6 total_len=348
>>> >>
>>> reg0=0x28b,reg11=0x9,reg12=0xa,reg13=0x57,reg14=0x423,metadata=0x2,in_port=83
>>>  (via action) data_len=348 (unbuffered)
>>> >> 2025-02-18T21:11:06.733Z|1047596|vconn(ovn_pinctrl0)|DBG|unix:/
>>> var/run/
>>> >> openvswitch/br-int.mgmt: sent (Success): NXT_RESUME (OF1.5) (xid=0x0):
>>> >> cookie=0x5bc96bd6 total_len=348
>>> >>
>>> reg0=0x28b,reg11=0x9,reg12=0xa,reg13=0x5f,reg14=0x413,metadata=0x2,in_port=91
>>>  (via action) data_len=348 (unbuffered)
>>> >> 2025-02-18T21:12:16.984Z|1048995|vconn(ovn_pinctrl0)|DBG|unix:/
>>> var/run/
>>> >> openvswitch/br-int.mgmt: sent (Success): NXT_RESUME (OF1.5) (xid=0x0):
>>> >> cookie=0xfded3de0 total_len=348
>>> >>
>>> reg0=0x308,reg11=0x9,reg12=0xa,reg13=0xc5,reg14=0x354,metadata=0x2,in_port=373
>>>  (via action) data_len=348 (unbuffered)
>>> >> 2025-02-18T21:13:03.396Z|1049934|vconn(ovn_pinctrl0)|DBG|unix:/
>>> var/run/
>>> >> openvswitch/br-int.mgmt: sent (Success): NXT_RESUME (OF1.5) (xid=0x0):
>>> >> cookie=0xfded3de0 total_len=348
>>> >>
>>> reg0=0x308,reg11=0x9,reg12=0xa,reg13=0xc5,reg14=0x354,metadata=0x2,in_port=373
>>>  (via action) data_len=348 (unbuffered)
>>> >> 2025-02-18T21:19:48.641Z|1058227|vconn(ovn_pinctrl0)|DBG|unix:/
>>> var/run/
>>> >> openvswitch/br-int.mgmt: sent (Success): NXT_RESUME (OF1.5) (xid=0x0):
>>> >> cookie=0xbbfea837 total_len=348
>>> >>
>>> reg0=0x308,reg11=0x9,reg12=0xa,reg13=0x132,reg14=0x434,metadata=0x2,in_port=302
>>>  (via action) data_len=348 (unbuffered)
>>> >> 2025-02-18T21:19:50.069Z|1058255|vconn(ovn_pinctrl0)|DBG|unix:/
>>> var/run/
>>> >> openvswitch/br-int.mgmt: sent (Success): NXT_RESUME (OF1.5) (xid=0x0):
>>> >> cookie=0x5bc96bd6 total_len=348
>>> >>
>>> reg0=0x28b,reg11=0x9,reg12=0xa,reg13=0x6f,reg14=0x148,metadata=0x2,in_port=107
>>>  (via action) data_len=348 (unbuffered)
>>> >> 2025-02-18T21:21:38.224Z|1060474|vconn(ovn_pinctrl0)|DBG|unix:/
>>> var/run/
>>> >> openvswitch/br-int.mgmt: sent (Success): NXT_RESUME (OF1.5) (xid=0x0):
>>> >> cookie=0x5bc96bd6 total_len=348
>>> >>
>>> reg0=0x28b,reg11=0x9,reg12=0xa,reg13=0xdb,reg14=0x336,metadata=0x2,in_port=215
>>>  (via action) data_len=348 (unbuffered)
>>> >> 2025-02-18T21:21:42.735Z|1060568|vconn(ovn_pinctrl0)|DBG|unix:/
>>> var/run/
>>> >> openvswitch/br-int.mgmt: sent (Success): NXT_RESUME (OF1.5) (xid=0x0):
>>> >> cookie=0xffe9d519 total_len=348
>>> >>
>>> reg0=0x308,reg11=0x9,reg12=0xa,reg13=0x112,reg14=0x2dd,metadata=0x2,in_port=363
>>>  (via action) data_len=348 (unbuffered)
>>> >>
>>> >> ovs-vswitchd
>>> >> 2025-02-18T20:55:45.673Z|1086321|vconn|DBG|unix#0: received:
>>> NXT_RESUME
>>> >> (OF1.5) (xid=0x0): cookie=0x58482a66 total_len=348
>>> >>
>>> reg0=0x308,reg11=0x9,reg12=0xa,reg13=0x151,reg14=0x4b1,metadata=0x2,in_port=375
>>>  (via action) data_len=348 (unbuffered)
>>> >> 2025-02-18T21:01:05.872Z|1092877|vconn|DBG|unix#0: received:
>>> NXT_RESUME
>>> >> (OF1.5) (xid=0x0): cookie=0x983e6272 total_len=348
>>> >>
>>> reg0=0x308,reg11=0x9,reg12=0xa,reg13=0x15f,reg14=0x4d3,metadata=0x2,in_port=368
>>>  (via action) data_len=348 (unbuffered)
>>> >> 2025-02-18T21:01:23.672Z|1093239|vconn|DBG|unix#0: received:
>>> NXT_RESUME
>>> >> (OF1.5) (xid=0x0): cookie=0xfded3de0 total_len=348
>>> >>
>>> reg0=0x308,reg11=0x9,reg12=0xa,reg13=0xc5,reg14=0x354,metadata=0x2,in_port=373
>>>  (via action) data_len=348 (unbuffered)
>>> >> 2025-02-18T21:04:03.272Z|1096513|vconn|DBG|unix#0: received:
>>> NXT_RESUME
>>> >> (OF1.5) (xid=0x0): cookie=0xfded3de0 total_len=348
>>> >>
>>> reg0=0x308,reg11=0x9,reg12=0xa,reg13=0xc5,reg14=0x354,metadata=0x2,in_port=373
>>>  (via action) data_len=348 (unbuffered)
>>> >> 2025-02-18T21:07:36.773Z|1100887|vconn|DBG|unix#0: received:
>>> NXT_RESUME
>>> >> (OF1.5) (xid=0x0): cookie=0xfded3de0 total_len=348
>>> >>
>>> reg0=0x308,reg11=0x9,reg12=0xa,reg13=0xc5,reg14=0x354,metadata=0x2,in_port=373
>>>  (via action) data_len=348 (unbuffered)
>>> >> 2025-02-18T21:07:51.472Z|1101187|vconn|DBG|unix#0: received:
>>> NXT_RESUME
>>> >> (OF1.5) (xid=0x0): cookie=0xa6d16298 total_len=348
>>> >>
>>> reg0=0x308,reg11=0x9,reg12=0xa,reg13=0x4b,reg14=0x1bb,metadata=0x2,in_port=349
>>>  (via action) data_len=348 (unbuffered)
>>> >> 2025-02-18T21:08:08.673Z|1101537|vconn|DBG|unix#0: received:
>>> NXT_RESUME
>>> >> (OF1.5) (xid=0x0): cookie=0x8b831d17 total_len=348
>>> >>
>>> reg0=0x308,reg11=0x9,reg12=0xa,reg13=0x26,reg14=0x123,metadata=0x2,in_port=34
>>>  (via action) data_len=348 (unbuffered)
>>> >> 2025-02-18T21:09:09.173Z|1102776|vconn|DBG|unix#0: received:
>>> NXT_RESUME
>>> >> (OF1.5) (xid=0x0): cookie=0xbf424a23 total_len=348
>>> >>
>>> reg0=0x308,reg11=0x9,reg12=0xa,reg13=0x85,reg14=0x2d9,metadata=0x2,in_port=371
>>>  (via action) data_len=348 (unbuffered)
>>> >> 2025-02-18T21:13:03.574Z|1108046|vconn|DBG|unix#0: received:
>>> NXT_RESUME
>>> >> (OF1.5) (xid=0x0): cookie=0xfded3de0 total_len=348
>>> >>
>>> reg0=0x308,reg11=0x9,reg12=0xa,reg13=0xc5,reg14=0x354,metadata=0x2,in_port=373
>>>  (via action) data_len=348 (unbuffered)
>>> >> 2025-02-18T21:17:18.772Z|1113273|vconn|DBG|unix#0: received:
>>> NXT_RESUME
>>> >> (OF1.5) (xid=0x0): cookie=0x5bc96bd6 total_len=348
>>> >>
>>> reg0=0x28b,reg11=0x9,reg12=0xa,reg13=0x7c,reg14=0x241,metadata=0x2,in_port=120
>>>  (via action) data_len=348 (unbuffered)
>>> >> 2025-02-18T21:18:51.673Z|1115178|vconn|DBG|unix#0: received:
>>> NXT_RESUME
>>> >> (OF1.5) (xid=0x0): cookie=0xfded3de0 total_len=348
>>> >>
>>> reg0=0x308,reg11=0x9,reg12=0xa,reg13=0xc5,reg14=0x354,metadata=0x2,in_port=373
>>>  (via action) data_len=348 (unbuffered)
>>> >> 2025-02-18T21:19:36.872Z|1116105|vconn|DBG|unix#0: received:
>>> NXT_RESUME
>>> >> (OF1.5) (xid=0x0): cookie=0x5bc96bd6 total_len=348
>>> >>
>>> reg0=0x28b,reg11=0x9,reg12=0xa,reg13=0x87,reg14=0x4b,metadata=0x2,in_port=131
>>>  (via action) data_len=348 (unbuffered)
>>> >> 2025-02-18T21:19:52.172Z|1116417|vconn|DBG|unix#0: received:
>>> NXT_RESUME
>>> >> (OF1.5) (xid=0x0): cookie=0xfded3de0 total_len=348
>>> >>
>>> reg0=0x308,reg11=0x9,reg12=0xa,reg13=0xc5,reg14=0x354,metadata=0x2,in_port=373
>>>  (via action) data_len=348 (unbuffered)
>>> >> 2025-02-18T21:20:24.572Z|1117078|vconn|DBG|unix#0: received:
>>> NXT_RESUME
>>> >> (OF1.5) (xid=0x0): cookie=0x5bc96bd6 total_len=348
>>> >>
>>> reg0=0x28b,reg11=0x9,reg12=0xa,reg13=0x8b,reg14=0x2fb,metadata=0x2,in_port=135
>>>  (via action) data_len=348 (unbuffered)
>>> >> 2025-02-18T21:21:31.672Z|1118459|vconn|DBG|unix#0: received:
>>> NXT_RESUME
>>> >> (OF1.5) (xid=0x0): cookie=0x5bc96bd6 total_len=348
>>> >>
>>> reg0=0x28b,reg11=0x9,reg12=0xa,reg13=0x93,reg14=0xdd,metadata=0x2,in_port=143
>>>  (via action) data_len=348 (unbuffered)
>>> >>
>>> >> The interesting logs is also at ovn-controller logs
>>> >> 2025-02-18T21:21:10.152Z|1059900|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:a6:57:54 13.37.13.68
>>> >> 2025-02-18T21:21:10.560Z|1059909|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:43:d4:9a 13.38.13.15
>>> >> 2025-02-18T21:21:14.178Z|1059983|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:cb:77:8d 13.37.13.72
>>> >> 2025-02-18T21:21:16.785Z|1060037|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:cc:d3:66 13.38.13.62
>>> >> 2025-02-18T21:21:19.661Z|1060097|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:7e:1b:a8 13.37.13.173
>>> >> 2025-02-18T21:21:22.582Z|1060157|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:cc:d3:66 13.38.13.62
>>> >> 2025-02-18T21:21:25.903Z|1060225|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:f6:ae:8b 13.37.13.11
>>> >> 2025-02-18T21:21:28.643Z|1060281|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:b0:98:e0 13.37.13.101
>>> >> 2025-02-18T21:21:32.687Z|1060361|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:82:6e:c3 13.37.13.183
>>> >> 2025-02-18T21:21:34.517Z|1060398|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:cb:08:c7 13.37.13.55
>>> >> 2025-02-18T21:21:38.673Z|1060484|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:f9:5b:25 13.37.13.98
>>> >> 2025-02-18T21:21:40.782Z|1060528|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:cc:d3:66 13.38.13.62
>>> >> 2025-02-18T21:21:44.097Z|1060596|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:95:0c:5e 13.37.13.197
>>> >> 2025-02-18T21:21:46.484Z|1060646|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:cc:d3:66 13.38.13.62
>>> >> 2025-02-18T21:21:50.203Z|1060722|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:92:f8:ba 13.37.13.124
>>> >> 2025-02-18T21:21:52.764Z|1060776|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:ce:b8:1e 13.37.13.126
>>> >> 2025-02-18T21:21:55.503Z|1060832|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:8d:88:f6 13.38.13.63
>>> >> 2025-02-18T21:21:59.520Z|1060914|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:8f:20:53 13.37.13.63
>>> >> 2025-02-18T21:22:01.620Z|1060958|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:8a:7a:47 13.37.13.150
>>> >> 2025-02-18T21:22:04.418Z|1061015|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:70:7e:5f 13.37.13.216
>>> >> 2025-02-18T21:22:07.477Z|1061079|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:2a:7a:ad 13.37.13.50
>>> >> 2025-02-18T21:22:11.719Z|1061165|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:cc:d3:66 13.38.13.62
>>> >> 2025-02-18T21:22:13.798Z|1061208|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:a6:57:54 13.37.13.68
>>> >> 2025-02-18T21:22:16.612Z|1061266|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:81:5f:99 13.37.13.14
>>> >> 2025-02-18T21:22:19.830Z|1061334|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:cc:d3:66 13.38.13.62
>>> >> 2025-02-18T21:22:22.460Z|1061388|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:cc:d3:66 13.38.13.62
>>> >> 2025-02-18T21:22:26.001Z|1061460|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:cc:d3:66 13.38.13.62
>>> >> 2025-02-18T21:22:28.383Z|1061506|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:cc:d3:66 13.38.13.62
>>> >> 2025-02-18T21:22:31.594Z|1061572|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:58:35:35 13.37.13.64
>>> >> 2025-02-18T21:22:35.249Z|1061648|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:cc:d3:66 13.38.13.62
>>> >> 2025-02-18T21:22:37.578Z|1061695|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:82:6e:c3 13.37.13.183
>>> >> 2025-02-18T21:22:40.514Z|1061755|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:cc:d3:66 13.38.13.62
>>> >> 2025-02-18T21:22:43.373Z|1061815|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:58:72:c4 13.37.13.65
>>> >> 2025-02-18T21:22:46.557Z|1061881|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:02:5e:ac 13.37.13.111
>>> >> 2025-02-18T21:22:50.020Z|1061951|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:cc:d3:66 13.38.13.62
>>> >> 2025-02-18T21:22:53.511Z|1062023|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:5f:fd:6f 13.37.13.15
>>> >> 2025-02-18T21:22:55.717Z|1062069|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:cc:d3:66 13.38.13.62
>>> >> 2025-02-18T21:22:58.812Z|1062133|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:8d:88:f6 13.38.13.63
>>> >> 2025-02-18T21:23:01.450Z|1062189|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:9a:da:47 13.37.13.225
>>> >> 2025-02-18T21:23:05.545Z|1062273|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:8a:7a:47 13.37.13.150
>>> >> 2025-02-18T21:23:08.512Z|1062332|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:70:7e:5f 13.37.13.216
>>> >> 2025-02-18T21:23:10.980Z|1062384|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:02:ff:8a 13.37.13.233
>>> >> 2025-02-18T21:23:14.405Z|1062454|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:cc:d3:66 13.38.13.62
>>> >> 2025-02-18T21:23:17.172Z|1062511|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:2e:73:8f 13.37.13.186
>>> >> 2025-02-18T21:23:19.435Z|1062559|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:39:8d:2f 13.37.13.153
>>> >> 2025-02-18T21:23:23.049Z|1062633|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:cc:d3:66 13.38.13.62
>>> >> 2025-02-18T21:23:25.653Z|1062687|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:d4:36:c6 13.37.13.60
>>> >> 2025-02-18T21:23:28.768Z|1062749|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:f1:64:3a 13.37.13.125
>>> >> 2025-02-18T21:23:33.340Z|1062843|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:56:c4:04 13.37.13.219
>>> >> 2025-02-18T21:23:34.400Z|1062865|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:6a:7b:a4 13.37.13.121
>>> >> 2025-02-18T21:23:39.485Z|1062968|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:cc:d3:66 13.38.13.62
>>> >> 2025-02-18T21:23:41.393Z|1063007|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:82:6e:c3 13.37.13.183
>>> >> 2025-02-18T21:23:44.410Z|1063069|pinctrl(ovn_pinctrl0)|INFO|DHCPOFFER
>>> >> fa:16:3e:52:ef:94 13.37.13.182
>>> >>
>>> >> IPs are sanitized, anyway 13.37.13 belongs to subnet #1 and
>>> 13.38.13 to
>>> >> subnet #2. Observing large regression across all infra, like
>>> almost 50%
>>> >> overall network bandwith loss over time.
>>> >>
>>> >> Any thoughts guys?
>>> >>
>>> > Do you see any warning/error logs in ovs-vswitchd.log?
>>> >
>>> >> Context: we run a lot of small instances with 1 x fixed IP with little
>>> >> amount of external traffic, about 300 per hypervisor, sometimes more,
>>> >> sometimes less anyway didn't see anything similar with plain ovs and
>>> >> neutron-openvswitch-agent (openstack).
>>> >> UPD: Once i manually cleared dhcpv4_options column in all ports seems
>>> >> like neutron-dhcp-agent (dnsmasq based) started to send dhcpoffers and
>>> >> most of instances back to normal. However, didn't manage what is the
>>> >> reason to ovn-controller not send DHCPOFFER to VIF. Seems like a big
>>> >> point of regression starting from N number of instances or
>>> something else.
>>> >>
>>> > Just double checking, you're saying that with low numbers of instances
>>> > you don't have this issue?
>>> >
>>> > I wonder if you're not potentially hitting the maximum resubmit limit
>>> > (4096 resubmits) due to DHCPOFFERS being sent as broadcast. There
>>> > should be warning logs indicating that in ovs-vswitchd.
>>> >
>>> > It might also help if you could share the output of:
>>> >
>>> > ovs-appctl coverage/show
>>> >
>>> >>> 10.02.2025 17:36, Numan Siddique:
>>> >>>>
>>> >>>> On Sat, Feb 8, 2025 at 1:27 PM support via discuss <ovs-
>>> >>>> disc...@openvswitch.org> wrote:
>>> >>>>
>>> >>>>     I'm observing for a long time next bug: ovn-controller doesn't
>>> >>>>     send DHCPOFFER but it logs that it does. However this VIF
>>> doesn't
>>> >>>>     get it at all.
>>> >>>>     But once ovn-controller restarted it successfully seems like it
>>> >>>>     actually send DHCPOFFER and immediately receives DHCPACK
>>> from VIF.
>>> >>>>
>>> >>>>
>>> >>>> When you see this issue,  can you enable vconn and rcon debug
>>> logs in
>>> >>>> both ovn-controller and ovs-vswitchd ?
>>> >>>>
>>> >>>> ovn-appctl vlog/set rconn:dbg
>>> >>>> ovn-appctl vlog/set vconn:dbg
>>> >>>>
>>> >>>> ovs-appctl vlog/set rconn:dbg
>>> >>>> ovs-appctl vlog/set rconn:dbg
>>> >>>>
>>> >>>> If ovn-controller replies with a DHCPOFFER, you should see the
>>> >>>> NXT_RESUME related logs in both ovn-controller and ovs-vswitchd.
>>> >>>>
>>> >>>> This can help you narrow down the issue.
>>> >>>>
>>> >>>> Thanks
>>> >>>> Numan
>>> >>>>
>>> >>>>     |2025-02-08T14:27:57.462Z|00053|pinctrl(ovn_pinctrl0)|INFO|
>>> DHCPOFFER
>>> >>>>     fa:16:3e:80:67:62 13.37.0.5
>>> >>>>     2025-02-08T14:28:01.746Z|00054|pinctrl(ovn_pinctrl0)|INFO|
>>> DHCPOFFER
>>> >>>>     fa:16:3e:80:67:62 13.37.0.5
>>> >>>>     2025-02-08T14:28:05.582Z|00055|pinctrl(ovn_pinctrl0)|INFO|
>>> DHCPOFFER
>>> >>>>     fa:16:3e:80:67:62 13.37.0.5
>>> >>>>     2025-02-08T14:28:13.053Z|00056|pinctrl(ovn_pinctrl0)|INFO|
>>> DHCPOFFER
>>> >>>>     fa:16:3e:80:67:62 13.37.0.5
>>> >>>>     2025-02-08T14:28:21.972Z|00057|pinctrl(ovn_pinctrl0)|INFO|
>>> DHCPOFFER
>>> >>>>     fa:16:3e:84:8e:7f 13.37.0.37
>>> >>>>     2025-02-08T14:28:25.109Z|00058|pinctrl(ovn_pinctrl0)|INFO|
>>> DHCPOFFER
>>> >>>>     fa:16:3e:84:8e:7f 13.37.0.37
>>> >>>>     2025-02-08T14:28:29.362Z|00059|pinctrl(ovn_pinctrl0)|INFO|
>>> DHCPOFFER
>>> >>>>     fa:16:3e:84:8e:7f 13.37.0.37
>>> >>>>     2025-02-08T14:28:30.003Z|00060|pinctrl(ovn_pinctrl0)|INFO|
>>> DHCPOFFER
>>> >>>>     fa:16:3e:80:67:62 13.37.0.5
>>> >>>>     2025-02-08T14:28:37.820Z|00061|pinctrl(ovn_pinctrl0)|INFO|
>>> DHCPOFFER
>>> >>>>     fa:16:3e:84:8e:7f 13.37.0.37
>>> >>>>     2025-02-08T14:28:38.619Z|00002|fatal_signal(ovn_statctrl2)|WARN|
>>> >>>> terminating
>>> >>>>     with signal 15 (Terminated) # ovn_controller is restarted by
>>> >>>>     docker restart ovn_controller
>>> >>>>     2025-02-08T14:28:50.639Z|00001|vlog|INFO|opened log file
>>> >>>>     /var/log/kolla/openvswitch/ovn-controller.log
>>> >>>>     2025-02-08T14:28:50.640Z|00002|vif_plug_representor|INFO|unable
>>> >>>>     to set udev receive buffer size: Operation not permitted
>>> >>>>     2025-02-08T14:28:50.641Z|00003|reconnect|INFO|tcp:127.0.0.1:6640
>>> >>>>     <http://127.0.0.1:6640>: connecting...
>>> >>>>     2025-02-08T14:28:50.641Z|00004|reconnect|INFO|tcp:127.0.0.1:6640
>>> >>>>     <http://127.0.0.1:6640>: connected
>>> >>>>     2025-02-08T14:28:50.738Z|00005|main|INFO|OVN internal version is
>>> >>>>     : [24.09.0-20.37.0-73.6]
>>> >>>>     2025-02-08T14:28:50.738Z|00006|main|INFO|OVS IDL reconnected,
>>> >>>>     force recompute.
>>> >>>>     2025-02-08T14:28:50.738Z|00007|reconnect|INFO|tcp:10.10.9.5:6642
>>> >>>>     <http://10.10.9.5:6642>: connecting...
>>> >>>>     2025-02-08T14:28:50.738Z|00008|main|INFO|OVNSB IDL reconnected,
>>> >>>>     force recompute.
>>> >>>>     2025-02-08T14:28:50.739Z|00009|ovn_util|INFO|statctrl:
>>> connecting
>>> >>>>     to switch: "unix:/var/run/openvswitch/br-int.mgmt"
>>> >>>>     2025-02-08T14:28:50.739Z|00010|rconn|INFO|unix:/var/run/
>>> >>>> openvswitch/br-int.mgmt:
>>> >>>>     connecting...
>>> >>>>     2025-02-08T14:28:50.739Z|00011|ovn_util|INFO|pinctrl: connecting
>>> >>>>     to switch: "unix:/var/run/openvswitch/br-int.mgmt"
>>> >>>>     2025-02-08T14:28:50.739Z|00012|rconn|INFO|unix:/var/run/
>>> >>>> openvswitch/br-int.mgmt:
>>> >>>>     connecting...
>>> >>>>     2025-02-08T14:28:50.739Z|00013|reconnect|INFO|tcp:10.10.9.5:6642
>>> >>>>     <http://10.10.9.5:6642>: connected
>>> >>>>     2025-02-08T14:28:50.739Z|00001|rconn(ovn_statctrl2)|INFO|unix:/
>>> >>>> var/run/openvswitch/br-int.mgmt:
>>> >>>>     connected
>>> >>>>     2025-02-08T14:28:50.743Z|00001|rconn(ovn_pinctrl0)|INFO|unix:/
>>> >>>> var/run/openvswitch/br-int.mgmt:
>>> >>>>     connected 2025-02-08T14:28:50.756Z|00014|ovn_util|INFO|features:
>>> >>>>     connecting to switch: "unix:/var/run/openvswitch/br-int.mgmt"
>>> >>>>     2025-02-08T14:28:50.757Z|00015|rconn|INFO|unix:/var/run/
>>> >>>> openvswitch/br-int.mgmt:
>>> >>>>     connecting... 2025-02-08T14:28:50.757Z|00016|features|INFO|OVS
>>> >>>>     Feature: ct_zero_snat, state: supported
>>> >>>>     2025-02-08T14:28:50.757Z|00017|features|INFO|OVS Feature:
>>> >>>>     ct_flush, state: supported
>>> >>>>     2025-02-08T14:28:50.757Z|00018|features|INFO|OVS Feature:
>>> >>>>     dp_hash_l4_sym_support, state: supported
>>> >>>>     2025-02-08T14:28:50.757Z|00019|main|INFO|OVS feature set
>>> changed,
>>> >>>>     force recompute.
>>> >>>>     2025-02-08T14:28:50.757Z|00020|ovn_util|INFO|ofctrl: connecting
>>> >>>>     to switch: "unix:/var/run/openvswitch/br-int.mgmt"
>>> >>>>     2025-02-08T14:28:50.757Z|00021|rconn|INFO|unix:/var/run/
>>> >>>> openvswitch/br-int.mgmt:
>>> >>>>     connecting...
>>> >>>>     2025-02-08T14:28:50.758Z|00022|rconn|INFO|unix:/var/run/
>>> >>>> openvswitch/br-int.mgmt:
>>> >>>>     connected 2025-02-08T14:28:50.758Z|00023|main|INFO|OVS OpenFlow
>>> >>>>     connection reconnected,force recompute.
>>> >>>>     2025-02-08T14:28:50.759Z|00024|rconn|INFO|unix:/var/run/
>>> >>>> openvswitch/br-int.mgmt:
>>> >>>>     connected 2025-02-08T14:28:50.760Z|00025|features|INFO|OVS
>>> >>>>     Feature: meter_support, state: supported
>>> >>>>     2025-02-08T14:28:50.760Z|00026|features|INFO|OVS Feature:
>>> >>>>     group_support, state: supported
>>> >>>>     2025-02-08T14:28:50.760Z|00027|features|INFO|OVS Feature:
>>> >>>>     sample_action_with_registers, state: supported
>>> >>>>     2025-02-08T14:28:50.760Z|00028|main|INFO|OVS feature set
>>> changed,
>>> >>>>     force recompute.
>>> >>>>     2025-02-08T14:28:51.483Z|00029|binding|INFO|Releasing lport
>>> >>>>     978a923d-1076-484b-a1d5-ff5827ae69ec from this chassis
>>> >>>>     (sb_readonly=0)
>>> >>>>     2025-02-08T14:28:53.658Z|00030|poll_loop|INFO|wakeup due to
>>> >>>>     [POLLIN] on fd 30 (10.10.9.3:60902
>>> >>>>     <http://10.10.9.3:60902><->10.10.9.5:6642
>>> >>>>     <http://10.10.9.5:6642>) at lib/stream-fd.c:157 (78% CPU usage)
>>> >>>>     2025-02-08T14:28:54.726Z|00002|pinctrl(ovn_pinctrl0)|INFO|
>>> DHCPOFFER
>>> >>>>     fa:16:3e:84:8e:7f 13.37.0.37
>>> >>>>     2025-02-08T14:28:59.134Z|00003|pinctrl(ovn_pinctrl0)|INFO|
>>> DHCPACK
>>> >>>>     fa:16:3e:84:8e:7f 13.37.0.37
>>> >>>>     2025-02-08T14:29:01.499Z|00004|pinctrl(ovn_pinctrl0)|INFO|
>>> DHCPOFFER
>>> >>>>     fa:16:3e:80:67:62 13.37.0.5
>>> >>>>     2025-02-08T14:29:03.245Z|00005|pinctrl(ovn_pinctrl0)|INFO|
>>> DHCPACK
>>> >>>>     fa:16:3e:84:8e:7f 13.37.0.37
>>> >>>>     2025-02-08T14:29:03.455Z|00031|memory|INFO|98744 kB peak
>>> resident
>>> >>>>     set size after 12.8 seconds
>>> >>>>     2025-02-08T14:29:03.455Z|00032|memory|INFO|idl-cells-
>>> >>>> OVN_Southbound:168023
>>> >>>>     idl-cells-Open_vSwitch:36874 lflow-cache-entries-cache-expr:6494
>>> >>>>     lflow-cache-entries-cache-matches:1413 lflow-cache-size-KB:20726
>>> >>>>     local_datapath_usage-KB:1 ofctrl_desired_flow_usage-KB:6342
>>> >>>>     ofctrl_installed_flow_usage-KB:4695 ofctrl_sb_flow_ref_usage-
>>> >>>> KB:2355 |
>>> >>>>
>>> >>>>     Any suggestions?
>>> >>>>
>>> >>>>     Environment:
>>> >>>>     OpenStack 2024.2
>>> >>>>     Kernel 6.8.0-50-generic #51
>>> >>>>     <https://github.com/ovn-org/ovn/issues/51>~22.04.1-Ubuntu SMP
>>> >>>>     PREEMPT_DYNAMIC
>>> >>>>
>>> >>>>     |ovs-vsctl (Open vSwitch) 3.4.0 DB Schema 8.7.0 |
>>> >>>>
>>> >>>>     |ovs-ofctl (Open vSwitch) 3.4.0 OpenFlow versions 0x1:0x6 |
>>> >>>>
>>> >>>>     |# ovn-nbctl --version ovn-nbctl 24.09.0 Open vSwitch Library
>>> >>>>     3.4.0 DB Schema 7.6.0 |
>>> >>>>
>>> >>>>     |# ovn-sbctl --version ovn-sbctl 24.09.0 Open vSwitch Library
>>> >>>>     3.4.0 DB Schema 20.37.0|
>>> >>>>
>>> > Thanks,
>>> > Dumitru
>>> >
>>
>>
>>  
> 
> 
>  
> 
> _______________________________________________
> discuss mailing list
> disc...@openvswitch.org
> https://mail.openvswitch.org/mailman/listinfo/ovs-discuss

_______________________________________________
discuss mailing list
disc...@openvswitch.org
https://mail.openvswitch.org/mailman/listinfo/ovs-discuss

Reply via email to