It took a while for my servers to come back on the network this time. I think it's due to ovirt continuing to try to migrate the VMs around like I requested. The 3 servers' names are "swm-01, swm-02 and swm-03". Eventually (about 2-3 minutes ago) they all came back online.
So I disabled and stopped the lldpad service. Nope. Started some more migrations and swm-02 and swm-03 disappeared again. No ping, SSH hung, same as before - almost as soon as the migration started. If you wall have any ideas what switch-level setting might be enabled, let me know, cause I'm stumped. I can add it to the ticket that's requesting the port configurations. I've already added the port numbers and switch name that I got from CDP. Thanks again, I really appreciate the help! cecjr On Fri, Aug 23, 2019 at 3:28 PM Dominik Holler <[email protected]> wrote: > > > > On Fri, Aug 23, 2019 at 9:19 PM Dominik Holler <[email protected]> wrote: >> >> >> >> On Fri, Aug 23, 2019 at 8:03 PM Curtis E. Combs Jr. <[email protected]> >> wrote: >>> >>> This little cluster isn't in production or anything like that yet. >>> >>> So, I went ahead and used your ethtool commands to disable pause >>> frames on both interfaces of each server. I then, chose a few VMs to >>> migrate around at random. >>> >>> swm-02 and swm-03 both went out again. Unreachable. Can't ping, can't >>> ssh, and the SSH session that I had open was unresponsive. >>> >>> Any other ideas? >>> >> >> Sorry, no. Looks like two different NICs with different drivers and frimware >> goes down together. >> This is a strong indication that the root cause is related to the switch. >> Maybe you can get some information about the switch config by >> 'lldptool get-tlv -n -i em1' >> > > Another guess: > After the optional 'lldptool get-tlv -n -i em1' > 'systemctl stop lldpad' > another try to migrate. > > >> >> >>> >>> On Fri, Aug 23, 2019 at 1:50 PM Dominik Holler <[email protected]> wrote: >>> > >>> > >>> > >>> > On Fri, Aug 23, 2019 at 6:45 PM Curtis E. Combs Jr. <[email protected]> >>> > wrote: >>> >> >>> >> Unfortunately, I can't check on the switch. Trust me, I've tried. >>> >> These servers are in a Co-Lo and I've put 5 tickets in asking about >>> >> the port configuration. They just get ignored - but that's par for the >>> >> coarse for IT here. Only about 2 out of 10 of our tickets get any >>> >> response and usually the response doesn't help. Then the system they >>> >> use auto-closes the ticket. That was why I was suspecting STP before. >>> >> >>> >> I can do ethtool. I do have root on these servers, though. Are you >>> >> trying to get me to turn off link-speed auto-negotiation? Would you >>> >> like me to try that? >>> >> >>> > >>> > It is just a suspicion, that the reason is pause frames. >>> > Let's start on a NIC which is not used for ovirtmgmt, I guess em1. >>> > Does 'ethtool -S em1 | grep pause' show something? >>> > Does 'ethtool em1 | grep pause' indicates support for pause? >>> > The current config is shown by 'ethtool -a em1'. >>> > '-A autoneg' "Specifies whether pause autonegotiation should be enabled." >>> > according to ethtool doc. >>> > Assuming flow control is enabled by default, I would try to disable it >>> > via >>> > 'ethtool -A em1 autoneg off rx off tx off' >>> > and check if it is applied via >>> > 'ethtool -a em1' >>> > and check if the behavior under load changes. >>> > >>> > >>> > >>> >> >>> >> On Fri, Aug 23, 2019 at 12:24 PM Dominik Holler <[email protected]> >>> >> wrote: >>> >> > >>> >> > >>> >> > >>> >> > On Fri, Aug 23, 2019 at 5:49 PM Curtis E. Combs Jr. >>> >> > <[email protected]> wrote: >>> >> >> >>> >> >> Sure! Right now, I only have a 500gb partition on each node shared >>> >> >> over NFS, added as storage domains. This is on each node - so, >>> >> >> currently 3. >>> >> >> >>> >> >> How can the storage cause a node to drop out? >>> >> >> >>> >> > >>> >> > Thanks, I got it. >>> >> > All three links go down on load, which causes NFS to fail. >>> >> > >>> >> > Can you check in the switch port configuration if there is some kind >>> >> > of Ethernet flow control enabled? >>> >> > Can you try to modify the behavior by changing the settings of your >>> >> > host interfaces, e.g. >>> >> > >>> >> > ethtool -A em1 autoneg off rx off tx off >>> >> > >>> >> > or >>> >> > ethtool -A em1 autoneg on rx on tx on >>> >> > ? >>> >> > >>> >> > >>> >> > >>> >> > >>> >> >> >>> >> >> On Fri, Aug 23, 2019, 11:46 AM Dominik Holler <[email protected]> >>> >> >> wrote: >>> >> >>> >>> >> >>> >>> >> >>> >>> >> >>> On Fri, Aug 23, 2019 at 5:41 PM Curtis E. Combs Jr. >>> >> >>> <[email protected]> wrote: >>> >> >>>> >>> >> >>>> Also, if it helps, the hosts will sit there, quietly, for hours or >>> >> >>>> days before anything happens. They're up and working just fine. But >>> >> >>>> then, when I manually migrate a VM from one host to another, they >>> >> >>>> become completely inaccessible. >>> >> >>>> >>> >> >>> >>> >> >>> Can you share some details about your storage? >>> >> >>> Maybe there is a feature used during live migration, which triggers >>> >> >>> the issue. >>> >> >>> >>> >> >>> >>> >> >>>> >>> >> >>>> These are vanilla-as-possible CentOS7 nodes. Very basic ovirt >>> >> >>>> install >>> >> >>>> and configuration. >>> >> >>>> >>> >> >>>> On Fri, Aug 23, 2019 at 11:33 AM Curtis E. Combs Jr. >>> >> >>>> <[email protected]> wrote: >>> >> >>>> > >>> >> >>>> > Hey Dominik, >>> >> >>>> > >>> >> >>>> > Thanks for helping. I really want to try to use ovirt. >>> >> >>>> > >>> >> >>>> > When these events happen, I cannot even SSH to the nodes due to >>> >> >>>> > the >>> >> >>>> > link being down. After a little while, the hosts come back... >>> >> >>>> > >>> >> >>>> > On Fri, Aug 23, 2019 at 11:30 AM Dominik Holler >>> >> >>>> > <[email protected]> wrote: >>> >> >>>> > > >>> >> >>>> > > Is you storage connected via NFS? >>> >> >>>> > > Can you manually access the storage on the host? >>> >> >>>> > > >>> >> >>>> > > >>> >> >>>> > > On Fri, Aug 23, 2019 at 5:19 PM Curtis E. Combs Jr. >>> >> >>>> > > <[email protected]> wrote: >>> >> >>>> > >> >>> >> >>>> > >> Sorry to dead bump this, but I'm beginning to suspect that >>> >> >>>> > >> maybe it's >>> >> >>>> > >> not STP that's the problem. >>> >> >>>> > >> >>> >> >>>> > >> 2 of my hosts just went down when a few VMs tried to migrate. >>> >> >>>> > >> >>> >> >>>> > >> Do any of you have any idea what might be going on here? I >>> >> >>>> > >> don't even >>> >> >>>> > >> know where to start. I'm going to include the dmesg in case it >>> >> >>>> > >> helps. >>> >> >>>> > >> This happens on both of the hosts whenever any migration >>> >> >>>> > >> attempts to start. >>> >> >>>> > >> >>> >> >>>> > >> >>> >> >>>> > >> >>> >> >>>> > >> >>> >> >>>> > >> >>> >> >>>> > >> >>> >> >>>> > >> >>> >> >>>> > >> >>> >> >>>> > >> >>> >> >>>> > >> [68099.245833] bnx2 0000:01:00.0 em1: NIC Copper Link is Down >>> >> >>>> > >> [68099.246055] internal: port 1(em1) entered disabled state >>> >> >>>> > >> [68184.177343] ixgbe 0000:03:00.0 p1p1: NIC Link is Down >>> >> >>>> > >> [68184.177789] ovirtmgmt: port 1(p1p1) entered disabled state >>> >> >>>> > >> [68184.177856] ovirtmgmt: topology change detected, propagating >>> >> >>>> > >> [68277.078671] INFO: task qemu-kvm:8888 blocked for more than >>> >> >>>> > >> 120 seconds. >>> >> >>>> > >> [68277.078700] "echo 0 > >>> >> >>>> > >> /proc/sys/kernel/hung_task_timeout_secs" >>> >> >>>> > >> disables this message. >>> >> >>>> > >> [68277.078723] qemu-kvm D ffff9db40c359040 0 8888 >>> >> >>>> > >> 1 0x000001a0 >>> >> >>>> > >> [68277.078727] Call Trace: >>> >> >>>> > >> [68277.078738] [<ffffffff978fd2ac>] ? >>> >> >>>> > >> avc_has_perm_flags+0xdc/0x1c0 >>> >> >>>> > >> [68277.078743] [<ffffffff97d69f19>] schedule+0x29/0x70 >>> >> >>>> > >> [68277.078746] [<ffffffff9785f3d9>] inode_dio_wait+0xd9/0x100 >>> >> >>>> > >> [68277.078751] [<ffffffff976c4010>] ? >>> >> >>>> > >> wake_bit_function+0x40/0x40 >>> >> >>>> > >> [68277.078765] [<ffffffffc09d6dd6>] nfs_getattr+0x1b6/0x250 >>> >> >>>> > >> [nfs] >>> >> >>>> > >> [68277.078768] [<ffffffff97848109>] vfs_getattr+0x49/0x80 >>> >> >>>> > >> [68277.078769] [<ffffffff97848185>] vfs_fstat+0x45/0x80 >>> >> >>>> > >> [68277.078771] [<ffffffff978486f4>] SYSC_newfstat+0x24/0x60 >>> >> >>>> > >> [68277.078774] [<ffffffff97d76d21>] ? >>> >> >>>> > >> system_call_after_swapgs+0xae/0x146 >>> >> >>>> > >> [68277.078778] [<ffffffff97739f34>] ? >>> >> >>>> > >> __audit_syscall_entry+0xb4/0x110 >>> >> >>>> > >> [68277.078782] [<ffffffff9763aaeb>] ? >>> >> >>>> > >> syscall_trace_enter+0x16b/0x220 >>> >> >>>> > >> [68277.078784] [<ffffffff97848ace>] SyS_newfstat+0xe/0x10 >>> >> >>>> > >> [68277.078786] [<ffffffff97d7706b>] tracesys+0xa3/0xc9 >>> >> >>>> > >> [68397.072384] INFO: task qemu-kvm:8888 blocked for more than >>> >> >>>> > >> 120 seconds. >>> >> >>>> > >> [68397.072413] "echo 0 > >>> >> >>>> > >> /proc/sys/kernel/hung_task_timeout_secs" >>> >> >>>> > >> disables this message. >>> >> >>>> > >> [68397.072436] qemu-kvm D ffff9db40c359040 0 8888 >>> >> >>>> > >> 1 0x000001a0 >>> >> >>>> > >> [68397.072439] Call Trace: >>> >> >>>> > >> [68397.072453] [<ffffffff978fd2ac>] ? >>> >> >>>> > >> avc_has_perm_flags+0xdc/0x1c0 >>> >> >>>> > >> [68397.072458] [<ffffffff97d69f19>] schedule+0x29/0x70 >>> >> >>>> > >> [68397.072462] [<ffffffff9785f3d9>] inode_dio_wait+0xd9/0x100 >>> >> >>>> > >> [68397.072467] [<ffffffff976c4010>] ? >>> >> >>>> > >> wake_bit_function+0x40/0x40 >>> >> >>>> > >> [68397.072480] [<ffffffffc09d6dd6>] nfs_getattr+0x1b6/0x250 >>> >> >>>> > >> [nfs] >>> >> >>>> > >> [68397.072485] [<ffffffff97848109>] vfs_getattr+0x49/0x80 >>> >> >>>> > >> [68397.072486] [<ffffffff97848185>] vfs_fstat+0x45/0x80 >>> >> >>>> > >> [68397.072488] [<ffffffff978486f4>] SYSC_newfstat+0x24/0x60 >>> >> >>>> > >> [68397.072491] [<ffffffff97d76d21>] ? >>> >> >>>> > >> system_call_after_swapgs+0xae/0x146 >>> >> >>>> > >> [68397.072495] [<ffffffff97739f34>] ? >>> >> >>>> > >> __audit_syscall_entry+0xb4/0x110 >>> >> >>>> > >> [68397.072498] [<ffffffff9763aaeb>] ? >>> >> >>>> > >> syscall_trace_enter+0x16b/0x220 >>> >> >>>> > >> [68397.072500] [<ffffffff97848ace>] SyS_newfstat+0xe/0x10 >>> >> >>>> > >> [68397.072502] [<ffffffff97d7706b>] tracesys+0xa3/0xc9 >>> >> >>>> > >> [68401.573141] bnx2 0000:01:00.0 em1: NIC Copper Link is Up, >>> >> >>>> > >> 1000 Mbps >>> >> >>>> > >> full duplex >>> >> >>>> > >> >>> >> >>>> > >> [68401.573247] internal: port 1(em1) entered blocking state >>> >> >>>> > >> [68401.573255] internal: port 1(em1) entered listening state >>> >> >>>> > >> [68403.576985] internal: port 1(em1) entered learning state >>> >> >>>> > >> [68405.580907] internal: port 1(em1) entered forwarding state >>> >> >>>> > >> [68405.580916] internal: topology change detected, propagating >>> >> >>>> > >> [68469.565589] nfs: server swm-01.hpc.moffitt.org not >>> >> >>>> > >> responding, timed out >>> >> >>>> > >> [68469.565840] nfs: server swm-01.hpc.moffitt.org not >>> >> >>>> > >> responding, timed out >>> >> >>>> > >> [68487.193932] ixgbe 0000:03:00.0 p1p1: NIC Link is Up 10 >>> >> >>>> > >> Gbps, Flow >>> >> >>>> > >> Control: RX/TX >>> >> >>>> > >> [68487.194105] ovirtmgmt: port 1(p1p1) entered blocking state >>> >> >>>> > >> [68487.194114] ovirtmgmt: port 1(p1p1) entered listening state >>> >> >>>> > >> [68489.196508] ovirtmgmt: port 1(p1p1) entered learning state >>> >> >>>> > >> [68491.200400] ovirtmgmt: port 1(p1p1) entered forwarding state >>> >> >>>> > >> [68491.200405] ovirtmgmt: topology change detected, sending >>> >> >>>> > >> tcn bpdu >>> >> >>>> > >> [68493.672423] NFS: nfs4_reclaim_open_state: Lock reclaim >>> >> >>>> > >> failed! >>> >> >>>> > >> [68494.777996] NFSD: client 10.15.28.22 testing state ID with >>> >> >>>> > >> incorrect client ID >>> >> >>>> > >> [68494.778580] NFSD: client 10.15.28.22 testing state ID with >>> >> >>>> > >> incorrect client ID >>> >> >>>> > >> >>> >> >>>> > >> >>> >> >>>> > >> On Thu, Aug 22, 2019 at 2:53 PM Curtis E. Combs Jr. >>> >> >>>> > >> <[email protected]> wrote: >>> >> >>>> > >> > >>> >> >>>> > >> > Thanks, I'm just going to revert back to bridges. >>> >> >>>> > >> > >>> >> >>>> > >> > On Thu, Aug 22, 2019 at 11:50 AM Dominik Holler >>> >> >>>> > >> > <[email protected]> wrote: >>> >> >>>> > >> > > >>> >> >>>> > >> > > >>> >> >>>> > >> > > >>> >> >>>> > >> > > On Thu, Aug 22, 2019 at 3:06 PM Curtis E. Combs Jr. >>> >> >>>> > >> > > <[email protected]> wrote: >>> >> >>>> > >> > >> >>> >> >>>> > >> > >> Seems like the STP options are so common and necessary >>> >> >>>> > >> > >> that it would >>> >> >>>> > >> > >> be a priority over seldom-used bridge_opts. I know what >>> >> >>>> > >> > >> STP is and I'm >>> >> >>>> > >> > >> not even a networking guy - never even heard of half of >>> >> >>>> > >> > >> the >>> >> >>>> > >> > >> bridge_opts that have switches in the UI. >>> >> >>>> > >> > >> >>> >> >>>> > >> > >> Anyway. I wanted to try the openvswitches, so I >>> >> >>>> > >> > >> reinstalled all of my >>> >> >>>> > >> > >> nodes and used "openvswitch (Technology Preview)" as the >>> >> >>>> > >> > >> engine-setup >>> >> >>>> > >> > >> option for the first host. I made a new Cluster for my >>> >> >>>> > >> > >> nodes, added >>> >> >>>> > >> > >> them all to the new cluster, created a new "logical >>> >> >>>> > >> > >> network" for the >>> >> >>>> > >> > >> internal network and attached it to the internal network >>> >> >>>> > >> > >> ports. >>> >> >>>> > >> > >> >>> >> >>>> > >> > >> Now, when I go to create a new VM, I don't even have >>> >> >>>> > >> > >> either the >>> >> >>>> > >> > >> ovirtmgmt switch OR the internal switch as an option. The >>> >> >>>> > >> > >> drop-down is >>> >> >>>> > >> > >> empy as if I don't have any vnic-profiles. >>> >> >>>> > >> > >> >>> >> >>>> > >> > > >>> >> >>>> > >> > > openvswitch clusters are limited to ovn networks. >>> >> >>>> > >> > > You can create one like described in >>> >> >>>> > >> > > https://www.ovirt.org/documentation/admin-guide/chap-External_Providers.html#connecting-an-ovn-network-to-a-physical-network >>> >> >>>> > >> > > >>> >> >>>> > >> > > >>> >> >>>> > >> > >> >>> >> >>>> > >> > >> On Thu, Aug 22, 2019 at 7:34 AM Tony Pearce >>> >> >>>> > >> > >> <[email protected]> wrote: >>> >> >>>> > >> > >> > >>> >> >>>> > >> > >> > Hi Dominik, would you mind sharing the use case for stp >>> >> >>>> > >> > >> > via API Only? I am keen to know this. >>> >> >>>> > >> > >> > Thanks >>> >> >>>> > >> > >> > >>> >> >>>> > >> > >> > >>> >> >>>> > >> > >> > On Thu., 22 Aug. 2019, 19:24 Dominik Holler, >>> >> >>>> > >> > >> > <[email protected]> wrote: >>> >> >>>> > >> > >> >> >>> >> >>>> > >> > >> >> >>> >> >>>> > >> > >> >> >>> >> >>>> > >> > >> >> On Thu, Aug 22, 2019 at 1:08 PM Miguel Duarte de Mora >>> >> >>>> > >> > >> >> Barroso <[email protected]> wrote: >>> >> >>>> > >> > >> >>> >>> >> >>>> > >> > >> >>> On Sat, Aug 17, 2019 at 11:27 AM >>> >> >>>> > >> > >> >>> <[email protected]> wrote: >>> >> >>>> > >> > >> >>> > >>> >> >>>> > >> > >> >>> > Hello. I have been trying to figure out an issue >>> >> >>>> > >> > >> >>> > for a very long time. >>> >> >>>> > >> > >> >>> > That issue relates to the ethernet and 10gb fc >>> >> >>>> > >> > >> >>> > links that I have on my >>> >> >>>> > >> > >> >>> > cluster being disabled any time a migration occurs. >>> >> >>>> > >> > >> >>> > >>> >> >>>> > >> > >> >>> > I believe this is because I need to have STP turned >>> >> >>>> > >> > >> >>> > on in order to >>> >> >>>> > >> > >> >>> > participate with the switch. However, there does >>> >> >>>> > >> > >> >>> > not seem to be any >>> >> >>>> > >> > >> >>> > way to tell oVirt to stop turning it off! Very >>> >> >>>> > >> > >> >>> > frustrating. >>> >> >>>> > >> > >> >>> > >>> >> >>>> > >> > >> >>> > After entering a cronjob that enables stp on all >>> >> >>>> > >> > >> >>> > bridges every 1 >>> >> >>>> > >> > >> >>> > minute, the migration issue disappears.... >>> >> >>>> > >> > >> >>> > >>> >> >>>> > >> > >> >>> > Is there any way at all to do without this cronjob >>> >> >>>> > >> > >> >>> > and set STP to be >>> >> >>>> > >> > >> >>> > ON without having to resort to such a silly >>> >> >>>> > >> > >> >>> > solution? >>> >> >>>> > >> > >> >>> >>> >> >>>> > >> > >> >>> Vdsm exposes a per bridge STP knob that you can use >>> >> >>>> > >> > >> >>> for this. By >>> >> >>>> > >> > >> >>> default it is set to false, which is probably why you >>> >> >>>> > >> > >> >>> had to use this >>> >> >>>> > >> > >> >>> shenanigan. >>> >> >>>> > >> > >> >>> >>> >> >>>> > >> > >> >>> You can, for instance: >>> >> >>>> > >> > >> >>> >>> >> >>>> > >> > >> >>> # show present state >>> >> >>>> > >> > >> >>> [vagrant@vdsm ~]$ ip a >>> >> >>>> > >> > >> >>> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue >>> >> >>>> > >> > >> >>> state UNKNOWN >>> >> >>>> > >> > >> >>> group default qlen 1000 >>> >> >>>> > >> > >> >>> link/loopback 00:00:00:00:00:00 brd >>> >> >>>> > >> > >> >>> 00:00:00:00:00:00 >>> >> >>>> > >> > >> >>> inet 127.0.0.1/8 scope host lo >>> >> >>>> > >> > >> >>> valid_lft forever preferred_lft forever >>> >> >>>> > >> > >> >>> 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 >>> >> >>>> > >> > >> >>> qdisc pfifo_fast >>> >> >>>> > >> > >> >>> state UP group default qlen 1000 >>> >> >>>> > >> > >> >>> link/ether 52:54:00:41:fb:37 brd ff:ff:ff:ff:ff:ff >>> >> >>>> > >> > >> >>> 3: eth1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 >>> >> >>>> > >> > >> >>> qdisc pfifo_fast >>> >> >>>> > >> > >> >>> state UP group default qlen 1000 >>> >> >>>> > >> > >> >>> link/ether 52:54:00:83:5b:6f brd ff:ff:ff:ff:ff:ff >>> >> >>>> > >> > >> >>> inet 192.168.50.50/24 brd 192.168.50.255 scope >>> >> >>>> > >> > >> >>> global noprefixroute eth1 >>> >> >>>> > >> > >> >>> valid_lft forever preferred_lft forever >>> >> >>>> > >> > >> >>> inet6 fe80::5054:ff:fe83:5b6f/64 scope link >>> >> >>>> > >> > >> >>> valid_lft forever preferred_lft forever >>> >> >>>> > >> > >> >>> 19: ;vdsmdummy;: <BROADCAST,MULTICAST> mtu 1500 qdisc >>> >> >>>> > >> > >> >>> noop state DOWN >>> >> >>>> > >> > >> >>> group default qlen 1000 >>> >> >>>> > >> > >> >>> link/ether 8e:5c:2e:87:fa:0b brd ff:ff:ff:ff:ff:ff >>> >> >>>> > >> > >> >>> >>> >> >>>> > >> > >> >>> # show example bridge configuration - you're looking >>> >> >>>> > >> > >> >>> for the STP knob here. >>> >> >>>> > >> > >> >>> [root@vdsm ~]$ cat bridged_net_with_stp >>> >> >>>> > >> > >> >>> { >>> >> >>>> > >> > >> >>> "bondings": {}, >>> >> >>>> > >> > >> >>> "networks": { >>> >> >>>> > >> > >> >>> "test-network": { >>> >> >>>> > >> > >> >>> "nic": "eth0", >>> >> >>>> > >> > >> >>> "switch": "legacy", >>> >> >>>> > >> > >> >>> "bridged": true, >>> >> >>>> > >> > >> >>> "stp": true >>> >> >>>> > >> > >> >>> } >>> >> >>>> > >> > >> >>> }, >>> >> >>>> > >> > >> >>> "options": { >>> >> >>>> > >> > >> >>> "connectivityCheck": false >>> >> >>>> > >> > >> >>> } >>> >> >>>> > >> > >> >>> } >>> >> >>>> > >> > >> >>> >>> >> >>>> > >> > >> >>> # issue setup networks command: >>> >> >>>> > >> > >> >>> [root@vdsm ~]$ vdsm-client -f bridged_net_with_stp >>> >> >>>> > >> > >> >>> Host setupNetworks >>> >> >>>> > >> > >> >>> { >>> >> >>>> > >> > >> >>> "code": 0, >>> >> >>>> > >> > >> >>> "message": "Done" >>> >> >>>> > >> > >> >>> } >>> >> >>>> > >> > >> >>> >>> >> >>>> > >> > >> >>> # show bridges >>> >> >>>> > >> > >> >>> [root@vdsm ~]$ brctl show >>> >> >>>> > >> > >> >>> bridge name bridge id STP enabled interfaces >>> >> >>>> > >> > >> >>> ;vdsmdummy; 8000.000000000000 no >>> >> >>>> > >> > >> >>> test-network 8000.52540041fb37 yes eth0 >>> >> >>>> > >> > >> >>> >>> >> >>>> > >> > >> >>> # show final state >>> >> >>>> > >> > >> >>> [root@vdsm ~]$ ip a >>> >> >>>> > >> > >> >>> 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue >>> >> >>>> > >> > >> >>> state UNKNOWN >>> >> >>>> > >> > >> >>> group default qlen 1000 >>> >> >>>> > >> > >> >>> link/loopback 00:00:00:00:00:00 brd >>> >> >>>> > >> > >> >>> 00:00:00:00:00:00 >>> >> >>>> > >> > >> >>> inet 127.0.0.1/8 scope host lo >>> >> >>>> > >> > >> >>> valid_lft forever preferred_lft forever >>> >> >>>> > >> > >> >>> 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 >>> >> >>>> > >> > >> >>> qdisc pfifo_fast >>> >> >>>> > >> > >> >>> master test-network state UP group default qlen 1000 >>> >> >>>> > >> > >> >>> link/ether 52:54:00:41:fb:37 brd ff:ff:ff:ff:ff:ff >>> >> >>>> > >> > >> >>> 3: eth1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 >>> >> >>>> > >> > >> >>> qdisc pfifo_fast >>> >> >>>> > >> > >> >>> state UP group default qlen 1000 >>> >> >>>> > >> > >> >>> link/ether 52:54:00:83:5b:6f brd ff:ff:ff:ff:ff:ff >>> >> >>>> > >> > >> >>> inet 192.168.50.50/24 brd 192.168.50.255 scope >>> >> >>>> > >> > >> >>> global noprefixroute eth1 >>> >> >>>> > >> > >> >>> valid_lft forever preferred_lft forever >>> >> >>>> > >> > >> >>> inet6 fe80::5054:ff:fe83:5b6f/64 scope link >>> >> >>>> > >> > >> >>> valid_lft forever preferred_lft forever >>> >> >>>> > >> > >> >>> 19: ;vdsmdummy;: <BROADCAST,MULTICAST> mtu 1500 qdisc >>> >> >>>> > >> > >> >>> noop state DOWN >>> >> >>>> > >> > >> >>> group default qlen 1000 >>> >> >>>> > >> > >> >>> link/ether 8e:5c:2e:87:fa:0b brd ff:ff:ff:ff:ff:ff >>> >> >>>> > >> > >> >>> 432: test-network: <BROADCAST,MULTICAST,UP,LOWER_UP> >>> >> >>>> > >> > >> >>> mtu 1500 qdisc >>> >> >>>> > >> > >> >>> noqueue state UP group default qlen 1000 >>> >> >>>> > >> > >> >>> link/ether 52:54:00:41:fb:37 brd ff:ff:ff:ff:ff:ff >>> >> >>>> > >> > >> >>> >>> >> >>>> > >> > >> >>> I don't think this STP parameter is exposed via >>> >> >>>> > >> > >> >>> engine UI; @Dominik >>> >> >>>> > >> > >> >>> Holler , could you confirm ? What are our plans for >>> >> >>>> > >> > >> >>> it ? >>> >> >>>> > >> > >> >>> >>> >> >>>> > >> > >> >> >>> >> >>>> > >> > >> >> STP is only available via REST-API, see >>> >> >>>> > >> > >> >> http://ovirt.github.io/ovirt-engine-api-model/4.3/#types/network >>> >> >>>> > >> > >> >> please find an example how to enable STP in >>> >> >>>> > >> > >> >> https://gist.github.com/dominikholler/4e70c9ef9929d93b6807f56d43a70b95 >>> >> >>>> > >> > >> >> >>> >> >>>> > >> > >> >> We have no plans to add STP to the web ui, >>> >> >>>> > >> > >> >> but new feature requests are always welcome on >>> >> >>>> > >> > >> >> https://bugzilla.redhat.com/enter_bug.cgi?product=ovirt-engine >>> >> >>>> > >> > >> >> >>> >> >>>> > >> > >> >> >>> >> >>>> > >> > >> >>> >>> >> >>>> > >> > >> >>> > >>> >> >>>> > >> > >> >>> > Here are some details about my systems, if you need >>> >> >>>> > >> > >> >>> > it. >>> >> >>>> > >> > >> >>> > >>> >> >>>> > >> > >> >>> > >>> >> >>>> > >> > >> >>> > selinux is disabled. >>> >> >>>> > >> > >> >>> > >>> >> >>>> > >> > >> >>> > >>> >> >>>> > >> > >> >>> > >>> >> >>>> > >> > >> >>> > >>> >> >>>> > >> > >> >>> > >>> >> >>>> > >> > >> >>> > >>> >> >>>> > >> > >> >>> > >>> >> >>>> > >> > >> >>> > >>> >> >>>> > >> > >> >>> > >>> >> >>>> > >> > >> >>> > [root@swm-02 ~]# rpm -qa | grep ovirt >>> >> >>>> > >> > >> >>> > ovirt-imageio-common-1.5.1-0.el7.x86_64 >>> >> >>>> > >> > >> >>> > ovirt-release43-4.3.5.2-1.el7.noarch >>> >> >>>> > >> > >> >>> > ovirt-imageio-daemon-1.5.1-0.el7.noarch >>> >> >>>> > >> > >> >>> > ovirt-vmconsole-host-1.0.7-2.el7.noarch >>> >> >>>> > >> > >> >>> > ovirt-hosted-engine-setup-2.3.11-1.el7.noarch >>> >> >>>> > >> > >> >>> > ovirt-ansible-hosted-engine-setup-1.0.26-1.el7.noarch >>> >> >>>> > >> > >> >>> > python2-ovirt-host-deploy-1.8.0-1.el7.noarch >>> >> >>>> > >> > >> >>> > ovirt-ansible-engine-setup-1.1.9-1.el7.noarch >>> >> >>>> > >> > >> >>> > python2-ovirt-setup-lib-1.2.0-1.el7.noarch >>> >> >>>> > >> > >> >>> > cockpit-machines-ovirt-195.1-1.el7.noarch >>> >> >>>> > >> > >> >>> > ovirt-hosted-engine-ha-2.3.3-1.el7.noarch >>> >> >>>> > >> > >> >>> > ovirt-vmconsole-1.0.7-2.el7.noarch >>> >> >>>> > >> > >> >>> > cockpit-ovirt-dashboard-0.13.5-1.el7.noarch >>> >> >>>> > >> > >> >>> > ovirt-provider-ovn-driver-1.2.22-1.el7.noarch >>> >> >>>> > >> > >> >>> > ovirt-host-deploy-common-1.8.0-1.el7.noarch >>> >> >>>> > >> > >> >>> > ovirt-host-4.3.4-1.el7.x86_64 >>> >> >>>> > >> > >> >>> > python-ovirt-engine-sdk4-4.3.2-2.el7.x86_64 >>> >> >>>> > >> > >> >>> > ovirt-host-dependencies-4.3.4-1.el7.x86_64 >>> >> >>>> > >> > >> >>> > ovirt-ansible-repositories-1.1.5-1.el7.noarch >>> >> >>>> > >> > >> >>> > [root@swm-02 ~]# cat /etc/redhat-release >>> >> >>>> > >> > >> >>> > CentOS Linux release 7.6.1810 (Core) >>> >> >>>> > >> > >> >>> > [root@swm-02 ~]# uname -r >>> >> >>>> > >> > >> >>> > 3.10.0-957.27.2.el7.x86_64 >>> >> >>>> > >> > >> >>> > You have new mail in /var/spool/mail/root >>> >> >>>> > >> > >> >>> > [root@swm-02 ~]# ip a >>> >> >>>> > >> > >> >>> > 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc >>> >> >>>> > >> > >> >>> > noqueue state UNKNOWN >>> >> >>>> > >> > >> >>> > group default qlen 1000 >>> >> >>>> > >> > >> >>> > link/loopback 00:00:00:00:00:00 brd >>> >> >>>> > >> > >> >>> > 00:00:00:00:00:00 >>> >> >>>> > >> > >> >>> > inet 127.0.0.1/8 scope host lo >>> >> >>>> > >> > >> >>> > valid_lft forever preferred_lft forever >>> >> >>>> > >> > >> >>> > 2: em1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 >>> >> >>>> > >> > >> >>> > qdisc mq master >>> >> >>>> > >> > >> >>> > test state UP group default qlen 1000 >>> >> >>>> > >> > >> >>> > link/ether d4:ae:52:8d:50:48 brd >>> >> >>>> > >> > >> >>> > ff:ff:ff:ff:ff:ff >>> >> >>>> > >> > >> >>> > 3: em2: <BROADCAST,MULTICAST> mtu 1500 qdisc noop >>> >> >>>> > >> > >> >>> > state DOWN group >>> >> >>>> > >> > >> >>> > default qlen 1000 >>> >> >>>> > >> > >> >>> > link/ether d4:ae:52:8d:50:49 brd >>> >> >>>> > >> > >> >>> > ff:ff:ff:ff:ff:ff >>> >> >>>> > >> > >> >>> > 4: p1p1: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 >>> >> >>>> > >> > >> >>> > qdisc mq master >>> >> >>>> > >> > >> >>> > ovirtmgmt state UP group default qlen 1000 >>> >> >>>> > >> > >> >>> > link/ether 90:e2:ba:1e:14:80 brd >>> >> >>>> > >> > >> >>> > ff:ff:ff:ff:ff:ff >>> >> >>>> > >> > >> >>> > 5: p1p2: <BROADCAST,MULTICAST> mtu 1500 qdisc noop >>> >> >>>> > >> > >> >>> > state DOWN group >>> >> >>>> > >> > >> >>> > default qlen 1000 >>> >> >>>> > >> > >> >>> > link/ether 90:e2:ba:1e:14:81 brd >>> >> >>>> > >> > >> >>> > ff:ff:ff:ff:ff:ff >>> >> >>>> > >> > >> >>> > 6: ovs-system: <BROADCAST,MULTICAST> mtu 1500 qdisc >>> >> >>>> > >> > >> >>> > noop state DOWN >>> >> >>>> > >> > >> >>> > group default qlen 1000 >>> >> >>>> > >> > >> >>> > link/ether a2:b8:d6:e8:b3:d8 brd >>> >> >>>> > >> > >> >>> > ff:ff:ff:ff:ff:ff >>> >> >>>> > >> > >> >>> > 7: br-int: <BROADCAST,MULTICAST> mtu 1500 qdisc >>> >> >>>> > >> > >> >>> > noop state DOWN group >>> >> >>>> > >> > >> >>> > default qlen 1000 >>> >> >>>> > >> > >> >>> > link/ether 96:a0:c1:4a:45:4b brd >>> >> >>>> > >> > >> >>> > ff:ff:ff:ff:ff:ff >>> >> >>>> > >> > >> >>> > 25: test: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu >>> >> >>>> > >> > >> >>> > 1500 qdisc noqueue >>> >> >>>> > >> > >> >>> > state UP group default qlen 1000 >>> >> >>>> > >> > >> >>> > link/ether d4:ae:52:8d:50:48 brd >>> >> >>>> > >> > >> >>> > ff:ff:ff:ff:ff:ff >>> >> >>>> > >> > >> >>> > inet 10.15.11.21/24 brd 10.15.11.255 scope >>> >> >>>> > >> > >> >>> > global test >>> >> >>>> > >> > >> >>> > valid_lft forever preferred_lft forever >>> >> >>>> > >> > >> >>> > 26: ovirtmgmt: <BROADCAST,MULTICAST,UP,LOWER_UP> >>> >> >>>> > >> > >> >>> > mtu 1500 qdisc >>> >> >>>> > >> > >> >>> > noqueue state UP group default qlen 1000 >>> >> >>>> > >> > >> >>> > link/ether 90:e2:ba:1e:14:80 brd >>> >> >>>> > >> > >> >>> > ff:ff:ff:ff:ff:ff >>> >> >>>> > >> > >> >>> > inet 10.15.28.31/24 brd 10.15.28.255 scope >>> >> >>>> > >> > >> >>> > global ovirtmgmt >>> >> >>>> > >> > >> >>> > valid_lft forever preferred_lft forever >>> >> >>>> > >> > >> >>> > 27: ;vdsmdummy;: <BROADCAST,MULTICAST> mtu 1500 >>> >> >>>> > >> > >> >>> > qdisc noop state DOWN >>> >> >>>> > >> > >> >>> > group default qlen 1000 >>> >> >>>> > >> > >> >>> > link/ether 62:e5:e5:07:99:eb brd >>> >> >>>> > >> > >> >>> > ff:ff:ff:ff:ff:ff >>> >> >>>> > >> > >> >>> > 29: vnet0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu >>> >> >>>> > >> > >> >>> > 1500 qdisc mq master >>> >> >>>> > >> > >> >>> > ovirtmgmt state UNKNOWN group default qlen 1000 >>> >> >>>> > >> > >> >>> > link/ether fe:6f:9c:95:00:02 brd >>> >> >>>> > >> > >> >>> > ff:ff:ff:ff:ff:ff >>> >> >>>> > >> > >> >>> > [root@swm-02 ~]# free -m >>> >> >>>> > >> > >> >>> > total used free >>> >> >>>> > >> > >> >>> > shared buff/cache available >>> >> >>>> > >> > >> >>> > Mem: 64413 1873 61804 >>> >> >>>> > >> > >> >>> > 9 735 62062 >>> >> >>>> > >> > >> >>> > Swap: 16383 0 16383 >>> >> >>>> > >> > >> >>> > [root@swm-02 ~]# free -h >>> >> >>>> > >> > >> >>> > total used free >>> >> >>>> > >> > >> >>> > shared buff/cache available >>> >> >>>> > >> > >> >>> > Mem: 62G 1.8G 60G >>> >> >>>> > >> > >> >>> > 9.5M 735M 60G >>> >> >>>> > >> > >> >>> > Swap: 15G 0B 15G >>> >> >>>> > >> > >> >>> > [root@swm-02 ~]# ls >>> >> >>>> > >> > >> >>> > ls lsb_release lshw >>> >> >>>> > >> > >> >>> > lslocks >>> >> >>>> > >> > >> >>> > lsmod lspci >>> >> >>>> > >> > >> >>> > lssubsys >>> >> >>>> > >> > >> >>> > lsusb.py >>> >> >>>> > >> > >> >>> > lsattr lscgroup lsinitrd >>> >> >>>> > >> > >> >>> > lslogins >>> >> >>>> > >> > >> >>> > lsns lss16toppm >>> >> >>>> > >> > >> >>> > lstopo-no-graphics >>> >> >>>> > >> > >> >>> > lsblk lscpu lsipc >>> >> >>>> > >> > >> >>> > lsmem >>> >> >>>> > >> > >> >>> > lsof lsscsi >>> >> >>>> > >> > >> >>> > lsusb >>> >> >>>> > >> > >> >>> > [root@swm-02 ~]# lscpu >>> >> >>>> > >> > >> >>> > Architecture: x86_64 >>> >> >>>> > >> > >> >>> > CPU op-mode(s): 32-bit, 64-bit >>> >> >>>> > >> > >> >>> > Byte Order: Little Endian >>> >> >>>> > >> > >> >>> > CPU(s): 16 >>> >> >>>> > >> > >> >>> > On-line CPU(s) list: 0-15 >>> >> >>>> > >> > >> >>> > Thread(s) per core: 2 >>> >> >>>> > >> > >> >>> > Core(s) per socket: 4 >>> >> >>>> > >> > >> >>> > Socket(s): 2 >>> >> >>>> > >> > >> >>> > NUMA node(s): 2 >>> >> >>>> > >> > >> >>> > Vendor ID: GenuineIntel >>> >> >>>> > >> > >> >>> > CPU family: 6 >>> >> >>>> > >> > >> >>> > Model: 44 >>> >> >>>> > >> > >> >>> > Model name: Intel(R) Xeon(R) CPU >>> >> >>>> > >> > >> >>> > X5672 @ 3.20GHz >>> >> >>>> > >> > >> >>> > Stepping: 2 >>> >> >>>> > >> > >> >>> > CPU MHz: 3192.064 >>> >> >>>> > >> > >> >>> > BogoMIPS: 6384.12 >>> >> >>>> > >> > >> >>> > Virtualization: VT-x >>> >> >>>> > >> > >> >>> > L1d cache: 32K >>> >> >>>> > >> > >> >>> > L1i cache: 32K >>> >> >>>> > >> > >> >>> > L2 cache: 256K >>> >> >>>> > >> > >> >>> > L3 cache: 12288K >>> >> >>>> > >> > >> >>> > NUMA node0 CPU(s): 0,2,4,6,8,10,12,14 >>> >> >>>> > >> > >> >>> > NUMA node1 CPU(s): 1,3,5,7,9,11,13,15 >>> >> >>>> > >> > >> >>> > Flags: fpu vme de pse tsc msr pae >>> >> >>>> > >> > >> >>> > mce cx8 apic sep >>> >> >>>> > >> > >> >>> > mtrr pge mca cmov pat pse36 clflush dts acpi mmx >>> >> >>>> > >> > >> >>> > fxsr sse sse2 ss ht >>> >> >>>> > >> > >> >>> > tm pbe syscall nx pdpe1gb rdtscp lm constant_tsc >>> >> >>>> > >> > >> >>> > arch_perfmon pebs bts >>> >> >>>> > >> > >> >>> > rep_good nopl xtopology nonstop_tsc aperfmperf >>> >> >>>> > >> > >> >>> > eagerfpu pni pclmulqdq >>> >> >>>> > >> > >> >>> > dtes64 monitor ds_cpl vmx smx est tm2 ssse3 cx16 >>> >> >>>> > >> > >> >>> > xtpr pdcm pcid dca >>> >> >>>> > >> > >> >>> > sse4_1 sse4_2 popcnt aes lahf_lm ssbd ibrs ibpb >>> >> >>>> > >> > >> >>> > stibp tpr_shadow vnmi >>> >> >>>> > >> > >> >>> > flexpriority ept vpid dtherm ida arat spec_ctrl >>> >> >>>> > >> > >> >>> > intel_stibp flush_l1d >>> >> >>>> > >> > >> >>> > [root@swm-02 ~]# >>> >> >>>> > >> > >> >>> > _______________________________________________ >>> >> >>>> > >> > >> >>> > Users mailing list -- [email protected] >>> >> >>>> > >> > >> >>> > To unsubscribe send an email to >>> >> >>>> > >> > >> >>> > [email protected] >>> >> >>>> > >> > >> >>> > Privacy Statement: >>> >> >>>> > >> > >> >>> > https://www.ovirt.org/site/privacy-policy/ >>> >> >>>> > >> > >> >>> > oVirt Code of Conduct: >>> >> >>>> > >> > >> >>> > https://www.ovirt.org/community/about/community-guidelines/ >>> >> >>>> > >> > >> >>> > List Archives: >>> >> >>>> > >> > >> >>> > https://lists.ovirt.org/archives/list/[email protected]/message/MTMZ5MF4CF2VR2D25VVPDNFN2IKE24AR/ >>> >> >>>> > >> > >> >> >>> >> >>>> > >> > >> >> _______________________________________________ >>> >> >>>> > >> > >> >> Users mailing list -- [email protected] >>> >> >>>> > >> > >> >> To unsubscribe send an email to [email protected] >>> >> >>>> > >> > >> >> Privacy Statement: >>> >> >>>> > >> > >> >> https://www.ovirt.org/site/privacy-policy/ >>> >> >>>> > >> > >> >> oVirt Code of Conduct: >>> >> >>>> > >> > >> >> https://www.ovirt.org/community/about/community-guidelines/ >>> >> >>>> > >> > >> >> List Archives: >>> >> >>>> > >> > >> >> https://lists.ovirt.org/archives/list/[email protected]/message/QBA7NYKAJNREIV6TN42VCW4IN3CX4VFG/ _______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected] Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/[email protected]/message/4HFIHDUPSP7BGT4MLRLEP6MXI4EVHQUI/

