Re: [OMPI users] Performance difference on OpenMPI, IntelMPI and ScaliMPI

2009-08-06 Thread Torgny Faxen
Pasha, se attached file. I have traced how MPI_IPROBE is called and also managed to significantly reduce the number of calls to MPI_IPROBE. Unfortunately this only resulted in the program spending time in other routines. Basically the code runs through a number of timesteps and after each time

Re: [OMPI users] Performance difference on OpenMPI, IntelMPI and ScaliMPI

2009-08-05 Thread Pavel Shamis (Pasha)
However, setting: -mca btl_openib_eager_limit 65536 gave a 15% improvement so OpenMPI is now down to 326 (from previous 376 seconds). Still a lot more than ScaliMPI with 214 seconds. Can you please run ibv_devinfo on one of compute nodes ? It is interesting to know what kind of IB HW you have

Re: [OMPI users] Performance difference on OpenMPI, IntelMPI and ScaliMPI

2009-08-05 Thread Torgny Faxen
Ralph, I can't get "opal_paffinity_alone" to work (see below). However, there is a "mpi_affinity_alone" that I tried without any improvement. However, setting: -mca btl_openib_eager_limit 65536 gave a 15% improvement so OpenMPI is now down to 326 (from previous 376 seconds). Still a lot more t

Re: [OMPI users] Performance difference on OpenMPI, IntelMPI and ScaliMPI

2009-08-05 Thread Ralph Castain
Okay, one problem is fairly clear. As Terry indicated, you have to tell us to bind or else you lose a lot of performace. Set -mca opal_paffinity_alone 1 on your cmd line and it should make a significant difference. On Wed, Aug 5, 2009 at 8:10 AM, Torgny Faxen wrote: > Ralph, > I am running thro

Re: [OMPI users] Performance difference on OpenMPI, IntelMPI and ScaliMPI

2009-08-05 Thread Torgny Faxen
Ralph, I am running through a locally provided wrapper but it translates to: /software/mpi/openmpi/1.3b2/i101017/bin/mpirun -np 144 -npernode 8 -mca mpi_show_mca_params env,file /nobac kup/rossby11/faxen/RCO_scobi/src_161.openmpi/rco2.24pe a) Upgrade.. This will take some time, it will have to

Re: [OMPI users] Performance difference on OpenMPI, IntelMPI and ScaliMPI

2009-08-05 Thread Terry Dontje
A comment to the below. I meant the 2x performance was for shared memory communications. --td Message: 3 Date: Wed, 05 Aug 2009 09:55:42 -0400 From: Terry Dontje Subject: Re: [OMPI users] Performance difference on OpenMPI, IntelMPI and ScaliMPI To: us...@open-mpi.org Message-ID

Re: [OMPI users] Performance difference on OpenMPI, IntelMPI and ScaliMPI

2009-08-05 Thread Pavel Shamis (Pasha)
If the above doesn't improve anything the next question is do you know what the sizes of the messages are? For very small messages I believe Scali shows a 2x better performance than Intel and OMPI (I think this is due to a fastpath optimization). I remember that mvapich was faster that sca

Re: [OMPI users] Performance difference on OpenMPI, IntelMPI and ScaliMPI

2009-08-05 Thread Terry Dontje
is due to a fastpath optimization). --td Message: 1 Date: Wed, 05 Aug 2009 15:15:52 +0200 From: Torgny Faxen Subject: Re: [OMPI users] Performance difference on OpenMPI, IntelMPI and ScaliMPI To: pa...@dev.mellanox.co.il, Open MPI Users Message-ID: <4a798608.5030...@nsc.liu.se> Co

Re: [OMPI users] Performance difference on OpenMPI, IntelMPI and ScaliMPI

2009-08-05 Thread Pavel Shamis (Pasha)
Torgny, We have one know issue in openib btl that it related to IPROBE - https://svn.open-mpi.org/trac/ompi/ticket/1362 Theoretical it maybe source cause of the performance degradation, but for me the performance difference sounds too big. * Do you know what is typical message size for this ap

Re: [OMPI users] Performance difference on OpenMPI, IntelMPI and ScaliMPI

2009-08-05 Thread Ralph Castain
Could you send us the mpirun cmd line? I wonder if you are missing some options that could help. Also, you might: (a) upgrade to 1.3.3 - it looks like you are using some kind of pre-release version (b) add -mca mpi_show_mca_params env,file - this will cause rank=0 to output what mca params it see

Re: [OMPI users] Performance difference on OpenMPI, IntelMPI and ScaliMPI

2009-08-05 Thread Torgny Faxen
Pasha, no collectives are being used. A simple grep in the code reveals the following MPI functions being used: MPI_Init MPI_wtime MPI_COMM_RANK MPI_COMM_SIZE MPI_BUFFER_ATTACH MPI_BSEND MPI_PACK MPI_UNPACK MPI_PROBE MPI_GET_COUNT MPI_RECV MPI_IPROBE MPI_FINALIZE where MPI_IPROBE is the clear wi

Re: [OMPI users] Performance difference on OpenMPI, IntelMPI and ScaliMPI

2009-08-05 Thread Pavel Shamis (Pasha)
Do you know if the application use some collective operations ? Thanks Pasha Torgny Faxen wrote: Hello, we are seeing a large difference in performance for some applications depending on what MPI is being used. Attached are performance numbers and oprofile output (first 30 lines) from one