That is correct. The ib0 in all of the HV nodes are accessing iSCSI and NFS 
over that IB link successfully.

What we are trying to do now is create a network that utilizes the second IB 
port (ib1) on the cards for some of the virtual machines that live inside the 
environment.

> On Nov 16, 2016, at 1:40 PM, Markus Stockhausen <stockhau...@collogia.de> 
> wrote:
> 
> Hi,
> 
> we are running Infiniband on the NFS storage network only. Did I get
> it aight that this works or do you already have issues there?
> 
> Best regards.
> 
> Markus
> 
> Web: www.collogia.de
> 
> ________________________________________
> Von: users-boun...@ovirt.org [users-boun...@ovirt.org]&quot; im Auftrag von 
> &quot;cl...@theboggios.com [cl...@theboggios.com]
> Gesendet: Mittwoch, 16. November 2016 20:10
> An: users@ovirt.org
> Betreff: [ovirt-users] Adding Infiniband VM Network Fails
> 
> Good Day;
> 
> I am trying to add an infiniband VM network to the hosts on my ovirt
> deployment, and the network configuration on the hosts fails to save.
> The network bridge is added successfully, but applying the bridge to the
> IB1 nic fails with little information other than it failed.
> 
> My system:
> 
> 6 HV nodes running CentOS 7 and OV version 4
> 1 Dedicated engine running CentOS 7 and engine version 4 in 3.6 mode.
> 
> The HV nodes all have Mellanox IB cards, dual port. Port 0 is for iSCSI
> and NFS connectivity and runs fine. Port 1 is for VM usage of the 10Gb
> network.
> 
> Have any of you had any dealings with this ?
> 
> 
> _______________________________________________
> Users mailing list
> Users@ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
> =
> <InterScan_Disclaimer.txt>

_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users

Reply via email to