On 05/10/2018 17:43, Karli Sjöberg wrote: > > > Den 5 okt. 2018 18:31 skrev Ben Webber <[email protected]>: > > > On 05/10/2018 17:02, Karli Sjöberg wrote: > > > > Den 5 okt. 2018 15:37 skrev Ben Webber <[email protected]> > <mailto:[email protected]>: > > Hi Miguel, > > Thanks for getting back to me so quickly! The pastebin is here: > > https://pastebin.com/xNJWiymw > > Yes, bond1 and bond2 are 802.3ad bonds and bond0 is an > active-backup bond of bond1 and bond2 > > > Ooh, that is really cool, I had no idea that was even possible! I've > gotta ask though: why? Is it to have switch redundancy without having to > shell out the extra price of having the switches stackable? > > /K > > Yes, that is exactly why, and that the servers I have contain 2 x 4 port > gigabit nics in each of them. At work we've been consolidating some of our > systems and I got a bit of old free kit which I'm aiming to use to set up a > home lab based on ceph (with the new iscsi gateway) and ovirt. The swtiches I > got are fairly basic and don't stack, but I wanted to achieve redundancy > throughout so this is the reason for trying to do it with a bond of bonds. > > > Yeah, for your use-case that makes total sense. And kudos for playing with > Ceph, we're thinking of going that route with iSCSI GW, but for VMWare > instead. > > For large deployments though, I think it's still going to be about as > expensive as you'd need to double the amount of ports needed, don't you agree? > > /K > Yes, it would probably make sense to use stacking 10 gigabit switches I would think for a larger deployment. If I were buying new hardware to do it for work, I'd most likely do that. > > > > Thanks > > Ben > > On 05/10/2018 14:09, Miguel Duarte de Mora Barroso wrote: > > On Thu, Oct 4, 2018 at 11:49 PM, Ben Webber wrote: > >> Hi, > >> > >> I'm trying to set up ovirt using the hosted-engine --deploy > command on CentOS7, but am encountering an error. I am running a slightly > unusual network configuration. I have two fairly basic non stacked gigabit > switches with port channels connecting the two switches together. I have a > lacp bond from the host consisting of 4 ports to each switch (bond1 and > bond2). I have then created an active-backup bond (bond0) using the two lacp > bonds as slaves in the hope to create ha at the switch layer using my basic > switches. There is then a VLAN (101) on bond0. > >> > >> This network configuration runs fine on the host, however, > when run, after a short while, the hosted-engine --deploy command outputs the > following error: > >> > >> ... > >> > >> [ INFO ] TASK [Force host-deploy in offline mode] > >> [ INFO ] ok: [localhost] > >> [ INFO ] TASK [Add host] > >> [ INFO ] changed: [localhost] > >> [ INFO ] TASK [Wait for the host to be up] > >> [ INFO ] ok: [localhost] > >> [ INFO ] TASK [Check host status] > >> [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, > "msg": "The host has been set in non_operational status, please check engine > logs, fix accordingly and re-deploy.\n"} > >> > >> ... > >> > >> > >> Looking in /var/log/ovirt-engine/engine.log on the machine > created, I can see the following errors logged: > >> > >> ... > >> > >> 2018-10-04 21:51:30,116+01 INFO > [org.ovirt.engine.core.vdsbroker.vdsbroker.HostSetupNetworksVDSCommand] > (EE-ManagedThreadFactory-engine-Thread-1) [59fb360a] START, > HostSetupNetworksVDSCommand(HostName = ov1.test.local, > HostSetupNetworksVdsCommandParameters:{hostId='7440c9b9-e530-4341-a317-d3a9041dc777', > vds='Host[ov1.test.local,7440c9b9-e530-4341-a317-d3a9041dc777]', > rollbackOnFailure='true', connectivityTimeout='120', > networks='[HostNetwork:{defaultRoute='true', bonding='true', > networkName='ovirtmgmt', vdsmName='ovirtmgmt', nicName='bond0', vlan='101', > vmNetwork='true', stp='false', properties='null', > ipv4BootProtocol='STATIC_IP', ipv4Address='192.168.1.11', > ipv4Netmask='255.255.255.0', ipv4Gateway='192.168.1.1', > ipv6BootProtocol='AUTOCONF', ipv6Address='null', ipv6Prefix='null', > ipv6Gateway='null', nameServers='null'}]', removedNetworks='[]', bonds='[]', > removedBonds='[]', clusterSwitchType='LEGACY'}), log id: 4f0c7eaa > >> 2018-10-04 21:51:30,121+01 INFO > [org.ovirt.engine.core.vdsbroker.vdsbroker.HostSetupNetworksVDSCommand] > (EE-ManagedThreadFactory-engine-Thread-1) [59fb360a] FINISH, > HostSetupNetworksVDSCommand, log id: 4f0c7eaa > >> 2018-10-04 21:51:30,645+01 ERROR > [org.ovirt.engine.core.vdsbroker.vdsbroker.HostSetupNetworksVDSCommand] > (EE-ManagedThreadFactory-engine-Thread-1) [59fb360a] Failed in > 'HostSetupNetworksVDS' method > >> 2018-10-04 21:51:30,687+01 ERROR > [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] > (EE-ManagedThreadFactory-engine-Thread-1) [59fb360a] EVENT_ID: > VDS_BROKER_COMMAND_FAILURE(10,802), VDSM ov1.test.local command > HostSetupNetworksVDS failed: Unknown nics in: ['bond1', 'bond2'] > >> 2018-10-04 21:51:30,688+01 ERROR > [org.ovirt.engine.core.vdsbroker.vdsbroker.HostSetupNetworksVDSCommand] > (EE-ManagedThreadFactory-engine-Thread-1) [59fb360a] Error: > VDSGenericException: VDSErrorException: Failed to HostSetupNetworksVDS, error > = Unknown nics in: ['bond1', 'bond2'], code = 23 > >> > >> ... > >> > >> > >> It looks like when HostSetupNetworksVDS is run, it is checking > that the slave interfaces to the bonds are physical network devices and being > as the slaves of bond0 are bond1 and bond2, rather than physical devices, it > then throws the error Unknown nics in: ['bond1', 'bond2']. > >> > >> Is there anything I can do or any configuration that I can put > anywhere to make it work with this "stacked bond" configuration or does ovirt > just not work when bonds are set up like this? > > Forwarding to Simone, who is an ovirt-hosted-engine-setup > expert. > > > > Please get us a pastebin with the output of 'ansible-playbook > -vvv -i > > localhost, > /usr/share/ovirt-hosted-engine-setup/ansible/get_network_interfaces.yml' > > on your engine node. > > > > One thing I want to make sure: your bond1 and bond2 > configurations are > > IEEE 802.3ad bonds, please confirm. > > > >> Thanks in advance, > >> > >> Ben > >> > >> > >> > >> > >> _______________________________________________ > >> Users mailing list -- [email protected] <mailto:[email protected]> > >> To unsubscribe send an email to [email protected] > <mailto:[email protected]> > >> Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > >> oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > >> List Archives: > https://lists.ovirt.org/archives/list/[email protected]/message/XPHQTPUINKZBSZVUDP2G66UPA5OJL3J7/ > > _______________________________________________ > > Users mailing list -- [email protected] <mailto:[email protected]> > > To unsubscribe send an email to [email protected] > <mailto:[email protected]> > > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > > List Archives: > https://lists.ovirt.org/archives/list/[email protected]/message/RG2SGSFO6XILGKPZH4RLGGEK66NDHPWF/ > _______________________________________________ > Users mailing list -- [email protected] <mailto:[email protected]> > To unsubscribe send an email to [email protected] > <mailto:[email protected]> > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > List Archives: > https://lists.ovirt.org/archives/list/[email protected]/message/6Q25LC74AYHMCREW7CNWZ4JNCADQD2LN/ > > >
_______________________________________________ Users mailing list -- [email protected] To unsubscribe send an email to [email protected] Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/[email protected]/message/7EL6IED4SNBPVWOKSLPZXQCRP75L6U5L/

