Jackson, i am surprised with the MTU value ... IIRC, MTU for ethernet jumbo frame is 9000, not 9001.
can you run tracepath on both boxes (to check which mtu is used) ? then, can you try to set MTU=1500 on both boxes (warning, get ready to lose the connection) and try again with openmpi and intelmpi ? then, can you increase mtu to 6000 and then 9000 and see how things evolve ? also, did you configure OpenMPI with IPv6 support ? Cheers, Gilles On Wed, Mar 9, 2016 at 9:48 AM, Jackson, Gary L. <gary.jack...@jhuapl.edu> wrote: > Nope, just one ethernet interface: > > $ ifconfig > eth0 Link encap:Ethernet HWaddr 0E:47:0E:0B:59:27 > inet addr:xxx.xxx.xxx.xxx Bcast:xxx.xxx.xxx.xxx > Mask:255.255.252.0 > inet6 addr: fe80::c47:eff:fe0b:5927/64 Scope:Link > UP BROADCAST RUNNING MULTICAST MTU:9001 Metric:1 > RX packets:16962 errors:0 dropped:0 overruns:0 frame:0 > TX packets:11564 errors:0 dropped:0 overruns:0 carrier:0 > collisions:0 txqueuelen:1000 > RX bytes:28613867 (27.2 MiB) TX bytes:1092650 (1.0 MiB) > > lo Link encap:Local Loopback > inet addr:127.0.0.1 Mask:255.0.0.0 > inet6 addr: ::1/128 Scope:Host > UP LOOPBACK RUNNING MTU:65536 Metric:1 > RX packets:68 errors:0 dropped:0 overruns:0 frame:0 > TX packets:68 errors:0 dropped:0 overruns:0 carrier:0 > collisions:0 txqueuelen:0 > RX bytes:6647 (6.4 KiB) TX bytes:6647 (6.4 KiB) > > > -- > Gary Jackson > > > > > From: users <users-boun...@open-mpi.org> on behalf of Gilles Gouaillardet > <gilles.gouaillar...@gmail.com> > Reply-To: Open MPI Users <us...@open-mpi.org> > Date: Tuesday, March 8, 2016 at 9:39 AM > To: Open MPI Users <us...@open-mpi.org> > Subject: Re: [OMPI users] Poor performance on Amazon EC2 with TCP > > > Jason, > > how many Ethernet interfaces are there ? > if several, can you try again with one only > mpirun --mca btl_tcp_if_include eth0 ... > > Cheers, > > Gilles > > On Tuesday, March 8, 2016, Jackson, Gary L. <gary.jack...@jhuapl.edu> > wrote: > > > I've built OpenMPI 1.10.1 on Amazon EC2. Using NetPIPE, I'm seeing about > half the performance for MPI over TCP as I do with raw TCP. Before I start > digging in to this more deeply, does anyone know what might cause that? > > For what it's worth, I see the same issues with MPICH, but I do not see it > with Intel MPI. > > -- > Gary Jackson > > _______________________________________________ > users mailing list > us...@open-mpi.org > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > Link to this post: > http://www.open-mpi.org/community/lists/users/2016/03/28663.php