That is correct. The ib0 in all of the HV nodes are accessing iSCSI and NFS over that IB link successfully.
What we are trying to do now is create a network that utilizes the second IB port (ib1) on the cards for some of the virtual machines that live inside the environment. > On Nov 16, 2016, at 1:40 PM, Markus Stockhausen <stockhau...@collogia.de> > wrote: > > Hi, > > we are running Infiniband on the NFS storage network only. Did I get > it aight that this works or do you already have issues there? > > Best regards. > > Markus > > Web: www.collogia.de > > ________________________________________ > Von: users-boun...@ovirt.org [users-boun...@ovirt.org]" im Auftrag von > "cl...@theboggios.com [cl...@theboggios.com] > Gesendet: Mittwoch, 16. November 2016 20:10 > An: users@ovirt.org > Betreff: [ovirt-users] Adding Infiniband VM Network Fails > > Good Day; > > I am trying to add an infiniband VM network to the hosts on my ovirt > deployment, and the network configuration on the hosts fails to save. > The network bridge is added successfully, but applying the bridge to the > IB1 nic fails with little information other than it failed. > > My system: > > 6 HV nodes running CentOS 7 and OV version 4 > 1 Dedicated engine running CentOS 7 and engine version 4 in 3.6 mode. > > The HV nodes all have Mellanox IB cards, dual port. Port 0 is for iSCSI > and NFS connectivity and runs fine. Port 1 is for VM usage of the 10Gb > network. > > Have any of you had any dealings with this ? > > > _______________________________________________ > Users mailing list > Users@ovirt.org > http://lists.ovirt.org/mailman/listinfo/users > = > <InterScan_Disclaimer.txt> _______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users