Thanks George,
I am selecting Ethernet device (em1) in mpirun script.

Here is ifconfig output:
em1       Link encap:Ethernet  HWaddr E0:DB:55:FD:38:46
          inet addr:10.30.10.121  Bcast:10.30.255.255  Mask:255.255.0.0
          inet6 addr: fe80::e2db:55ff:fefd:3846/64 Scope:Link
          UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
          RX packets:1537270190 errors:0 dropped:0 overruns:0 frame:0
          TX packets:136123598 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 txqueuelen:1000
          RX bytes:309333740659 (288.0 GiB)  TX bytes:143480101212 (133.6
GiB)
          Memory:91820000-91840000

Ifconfig uses the ioctl access method to get the full address information,
which limits hardware addresses to 8 bytes.
Because Infiniband address has 20 bytes, only the first 8 bytes are
displayed correctly.
Ifconfig is obsolete! For replacement check ip.
ib0       Link encap:InfiniBand  HWaddr
80:00:00:03:FE:80:00:00:00:00:00:00:00:00:00:00:00:00:00:00
          inet addr:10.32.10.121  Bcast:10.32.255.255  Mask:255.255.0.0
          inet6 addr: fe80::211:7500:70:6ab4/64 Scope:Link
          UP BROADCAST RUNNING MULTICAST  MTU:2044  Metric:1
          RX packets:33621 errors:0 dropped:0 overruns:0 frame:0
          TX packets:365 errors:0 dropped:5 overruns:0 carrier:0
          collisions:0 txqueuelen:256
          RX bytes:1882728 (1.7 MiB)  TX bytes:21920 (21.4 KiB)

lo        Link encap:Local Loopback
          inet addr:127.0.0.1  Mask:255.0.0.0
          inet6 addr: ::1/128 Scope:Host
          UP LOOPBACK RUNNING  MTU:16436  Metric:1
          RX packets:66889 errors:0 dropped:0 overruns:0 frame:0
          TX packets:66889 errors:0 dropped:0 overruns:0 carrier:0
          collisions:0 txqueuelen:0
          RX bytes:19005445 (18.1 MiB)  TX bytes:19005445 (18.1 MiB)






> Date: Wed, 10 Sep 2014 00:06:51 +0900
> From: George Bosilca <bosi...@icl.utk.edu>
> To: Open MPI Users <us...@open-mpi.org>
> Subject: Re: [OMPI users] Forcing OpenMPI to use Ethernet interconnect
>         instead of InfiniBand
>
> Look at your ifconfig output and select the Ethernet device (instead of the
> IPoIB one). Traditionally the name lack any fanciness, most distributions
> using eth0 as a default.
>
>   George.
>
>
> On Tue, Sep 9, 2014 at 11:24 PM, Muhammad Ansar Javed <
> muhammad.an...@seecs.edu.pk> wrote:
>
> > Hi,
> >
> > I am currently conducting some testing on system with Gigabit and
> > InfiniBand interconnects. Both Latency and Bandwidth benchmarks are doing
> > well as expected on InfiniBand interconnects but Ethernet interconnect is
> > achieving very high performance from expectations. Ethernet and
> InfiniBand
> > both are achieving equivalent performance.
> >
> > For some reason, it looks like openmpi (v1.8.1) is using the InfiniBand
> > interconnect rather than the Gigabit or TCP communication is being
> emulated
> > to use InifiniBand interconnect.
> >
> > Here are Latency and Bandwidth benchmark results.
> > #---------------------------------------------------
> > # Benchmarking PingPong
> > # processes = 2
> > # map-by node
> > #---------------------------------------------------
> >
> > Hello, world.  I am 1 on node124
> > Hello, world.  I am 0 on node123
> > Size Latency (usec) Bandwidth (Mbps)
> > 1    1.65    4.62
> > 2    1.67    9.16
> > 4    1.66    18.43
> > 8    1.66    36.74
> > 16    1.85    66.00
> > 32    1.83    133.28
> > 64    1.83    266.36
> > 128    1.88    519.10
> > 256    1.99    982.29
> > 512    2.23    1752.37
> > 1024    2.58    3026.98
> > 2048    3.32    4710.76
> >
> > I read some of the FAQs and noted that OpenMPI prefers the faster
> > available interconnect. In an effort to force it to use the gigabit
> > interconnect I ran it as follows
> >
> > 1. mpirun -np 2 -machinefile machines -map-by node --mca btl tcp --mca
> > btl_tcp_if_include em1 ./latency.ompi
> > 2. mpirun -np 2 -machinefile machines -map-by node --mca btl tcp,self,sm
> > --mca btl_tcp_if_include em1 ./latency.ompi
> > 3. mpirun -np 2 -machinefile machines -map-by node --mca btl ^openib
> --mca
> > btl_tcp_if_include em1 ./latency.ompi
> > 4. mpirun -np 2 -machinefile machines -map-by node --mca btl ^openib
> > ./latency.ompi
> >
> > None of them resulted in a significantly different benchmark output.
> >
> > I am using OpenMPI by loading module on clustered environment and don't
> > have admin access. It is configured for both TCP and OpenIB (confirmed
> from
> > ompi_info). After trying all above mentioned methods without success I
> > installed OpenMPI v1.8.2 in my home directory and disable openib with
> > following configuration options
> >
> > --disable-openib-control-hdr-padding --disable-openib-dynamic-sl
> > --disable-openib-connectx-xrc --disable-openib-udcm
> > --disable-openib-rdmacm  --disable-btl-openib-malloc-alignment
> > --disable-io-romio --without-openib --without-verbs
> >
> > Now, openib is not enabled (confirmed from ompi_info script) and there is
> > no "openib.so" file in $prefix/lib/openmpi directory as well. Still,
> above
> > mentioned mpirun commands are getting the same latency and bandwidth as
> > that of InfiniBand.
> >
> > I tried mpirun in verbose mode with following command and here is the
> > output
> >
> > Command:
> > mpirun -np 2 -machinefile machines -map-by node --mca btl tcp --mca
> > btl_base_verbose 30 --mca btl_tcp_if_include em1 ./latency.ompi
> >
> > Output:
> > [node123.prv.sciama.cluster:88310] mca: base: components_register:
> > registering btl components
> > [node123.prv.sciama.cluster:88310] mca: base: components_register: found
> > loaded component tcp
> > [node123.prv.sciama.cluster:88310] mca: base: components_register:
> > component tcp register function successful
> > [node123.prv.sciama.cluster:88310] mca: base: components_open: opening
> btl
> > components
> > [node123.prv.sciama.cluster:88310] mca: base: components_open: found
> > loaded component tcp
> > [node123.prv.sciama.cluster:88310] mca: base: components_open: component
> > tcp open function successful
> > [node124.prv.sciama.cluster:90465] mca: base: components_register:
> > registering btl components
> > [node124.prv.sciama.cluster:90465] mca: base: components_register: found
> > loaded component tcp
> > [node124.prv.sciama.cluster:90465] mca: base: components_register:
> > component tcp register function successful
> > [node124.prv.sciama.cluster:90465] mca: base: components_open: opening
> btl
> > components
> > [node124.prv.sciama.cluster:90465] mca: base: components_open: found
> > loaded component tcp
> > [node124.prv.sciama.cluster:90465] mca: base: components_open: component
> > tcp open function successful
> > Hello, world.  I am 1 on node124
> > Hello, world.  I am 0 on node123
> > Size Latency(usec) Bandwidth(Mbps)
> > 1    4.18    1.83
> > 2    3.66    4.17
> > 4    4.08    7.48
> > 8    3.12    19.57
> > 16    3.83    31.84
> > 32    3.40    71.84
> > 64    4.10    118.97
> > 128    3.89    251.19
> > 256    4.22    462.77
> > 512    2.95    1325.71
> > 1024    2.63    2969.49
> > 2048    3.38    4628.29
> > [node123.prv.sciama.cluster:88310] mca: base: close: component tcp closed
> > [node123.prv.sciama.cluster:88310] mca: base: close: unloading component
> > tcp
> > [node124.prv.sciama.cluster:90465] mca: base: close: component tcp closed
> > [node124.prv.sciama.cluster:90465] mca: base: close: unloading component
> > tcp
> >
> > Moreover, same benchmark applications using MPICH are working fine on
> > Ethernet and achieving expected Latency and Bandwidth.
> >
> > How can this be fixed?
> >
> > Thanks for help,
> >
> > --Ansar
>




-- 
Regards

Ansar Javed
HPC Lab
SEECS NUST
Contact: +92 334 438 9394
Skype: ansar.javed.859
Email: muhammad.an...@seecs.edu.pk

Reply via email to