Hello Howard,

I am able to run my Open MPI job to completion over TCP as you suggested for a 
sanity/configuration double check.  I also am able to complete the job using 
the RoCE fabric if I swap the breakout cable with 2 regular RoCE cables.  I am 
willing to test some custom builds to help iron out this problem.  Thank you 
again for your time and effort.

 

Brendan 

 

From: users [mailto:users-boun...@lists.open-mpi.org] On Behalf Of Howard 
Pritchard
Sent: Friday, February 03, 2017 12:53 PM
To: Open MPI Users <users@lists.open-mpi.org>
Subject: Re: [OMPI users] Open MPI over RoCE using breakout cable and switch

 

Hello Brendan,

 

Sorry for the delay in responding.  I've been on travel the past two weeks.

 

I traced through the debug output you sent.  It provided enough information

to show that for some reason, when using the breakout cable, Open MPI

is unable to complete initialization it needs to use the openib BTL.  It

correctly detects that the first port is not available, but for port 1, it

still fails to initialize.

 

To debug this further, I'd need to provide you with a custom Open MPI

to try that would have more debug output in the suspect area.

 

If you'd like to go this route let me know and I'll build a one of library

to try to debug this problem.

 

One thing to do just as a sanity check is to try tcp:

 

mpirun --mca btl tcp,self,sm ....

 

with the breakout cable.  If that doesn't work, then I think there may

be some network setup problem that needs to be resolved first before

trying custom Open MPI tarballs.

 

Thanks,

 

Howard

 

 

 

 

2017-02-01 15:08 GMT-07:00 Brendan Myers <brendan.my...@soft-forge.com 
<mailto:brendan.my...@soft-forge.com> >:

Hello Howard,

I was wondering if you have been able to look at this issue at all, or if 
anyone has any ideas on what to try next.

 

Thank you,

Brendan

 

From: users [mailto:users-boun...@lists.open-mpi.org 
<mailto:users-boun...@lists.open-mpi.org> ] On Behalf Of Brendan Myers
Sent: Tuesday, January 24, 2017 11:11 AM


To: 'Open MPI Users' <users@lists.open-mpi.org 
<mailto:users@lists.open-mpi.org> >
Subject: Re: [OMPI users] Open MPI over RoCE using breakout cable and switch

 

Hello Howard,

Here is the error output after building with debug enabled.  These CX4 Mellanox 
cards view each port as a separate device and I am using port 1 on the card 
which is device mlx5_0. 

 

Thank you,

Brendan

 

From: users [mailto:users-boun...@lists.open-mpi.org] On Behalf Of Howard 
Pritchard
Sent: Tuesday, January 24, 2017 8:21 AM
To: Open MPI Users <users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >
Subject: Re: [OMPI users] Open MPI over RoCE using breakout cable and switch

 

Hello Brendan,

 

This helps some, but looks like we need more debug output.

 

Could you build a debug version of Open MPI by adding --enable-debug

to the config options and rerun the test with the breakout cable setup

and keeping the --mca btl_base_verbose 100 command line option?

 

Thanks

 

Howard

 

 

2017-01-23 8:23 GMT-07:00 Brendan Myers <brendan.my...@soft-forge.com 
<mailto:brendan.my...@soft-forge.com> >:

Hello Howard,

Thank you for looking into this. Attached is the output you requested.  Also, I 
am using Open MPI 2.0.1.

 

Thank you,

Brendan

 

From: users [mailto:users-boun...@lists.open-mpi.org 
<mailto:users-boun...@lists.open-mpi.org> ] On Behalf Of Howard Pritchard
Sent: Friday, January 20, 2017 6:35 PM
To: Open MPI Users <users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >
Subject: Re: [OMPI users] Open MPI over RoCE using breakout cable and switch

 

Hi Brendan

 

I doubt this kind of config has gotten any testing with OMPI.  Could you rerun 
with

 

--mca btl_base_verbose 100

 

added to the command line and post the output to the list?

 

Howard

 

 

Brendan Myers <brendan.my...@soft-forge.com 
<mailto:brendan.my...@soft-forge.com> > schrieb am Fr. 20. Jan. 2017 um 15:04:

Hello,

I am attempting to get Open MPI to run over 2 nodes using a switch and a single 
breakout cable with this design:

(100GbE)QSFP <----> 2x (50GbE)QSFP       

 

Hardware Layout:

Breakout cable module A connects to switch (100GbE)

Breakout cable module B1 connects to node 1 RoCE NIC (50GbE)

Breakout cable module B2 connects to node 2 RoCE NIC (50GbE)

Switch is Mellanox SN 2700 100GbE RoCE switch

 

*         I  am able to pass RDMA traffic between the nodes with perftest 
(ib_write_bw) when using the breakout cable as the IC from both nodes to the 
switch.

*         When attempting to run a job using the breakout cable as the IC Open 
MPI aborts with failure to initialize open fabrics device errors.

*         If I replace the breakout cable with 2 standard QSFP cables the Open 
MPI job will complete correctly.  

 

 

This is the command I use, it works unless I attempt a run with the breakout 
cable used as IC:

mpirun --mca btl openib,self,sm --mca btl_openib_receive_queues 
P,65536,120,64,32 --mca btl_openib_cpc_include rdmacm  -hostfile mpi-hosts-ce 
/usr/local/bin/IMB-MPI1

 

If anyone has any idea as to why using a breakout cable is causing my jobs to 
fail please let me know.

 

Thank you,

 

Brendan T. W. Myers

brendan.my...@soft-forge.com <mailto:brendan.my...@soft-forge.com> 

Software Forge Inc

 

_______________________________________________

users mailing list

users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> 

https://rfd.newmexicoconsortium.org/mailman/listinfo/users


_______________________________________________
users mailing list
users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> 
https://rfd.newmexicoconsortium.org/mailman/listinfo/users

 


_______________________________________________
users mailing list
users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> 
https://rfd.newmexicoconsortium.org/mailman/listinfo/users

 

_______________________________________________
users mailing list
users@lists.open-mpi.org
https://rfd.newmexicoconsortium.org/mailman/listinfo/users

Reply via email to