Hi,
I am not an HPL expert, but this might help.

1.   rankfile mapper is avaliale only from Open MPI 1.3 version, if you are
using Open MPI 1.2.8 try -mca mpi_paffinity_alone 1
2.   if you are using Open MPI 1.3 you dont have to use mpi_leave_pinned 1 ,
since it's a default value

Lenny.

On Thu, Jul 2, 2009 at 4:47 PM, Swamy Kandadai <sw...@us.ibm.com> wrote:

>  Jeff:
>
> I am running on a 2.66 GHz Nehalem node. On this node, the turbo mode and
> hyperthreading are enabled.
> When I run LINPACK with Intel MPI, I get 82.68 GFlops without much trouble.
>
> When I ran with OpenMPI (I have OpenMPI 1.2.8 but my colleague was using
> 1.3.2). I was using the same MKL libraries both with OpenMPI and
> Intel MPI. But with OpenMPI, the best I got so far is 80.22 GFlops and I
> could never achieve close to what I am getting with Intel MPI.
> Here are muy options with OpenMPI:
>
> mpirun -n 8 --machinefile hf --mca rmaps_rank_file_path rankfile --mca
> coll_sm_info_num_procs 8 --mca btl self,sm -mca mpi_leave_pinned 1
> ./xhpl_ompi
>
> Here is my rankfile:
>
> at rankfile
> rank 0=i02n05 slot=0
> rank 1=i02n05 slot=1
> rank 2=i02n05 slot=2
> rank 3=i02n05 slot=3
> rank 4=i02n05 slot=4
> rank 5=i02n05 slot=5
> rank 6=i02n05 slot=6
> rank 7=i02n05 slot=7
>
> In this case the physical cores are 0-7 while the additional logical
> processors with hyperthreading are 8-15.
> With "top" command, I could see all the 8 tasks are running on 8 different
> physical cores. I did not see
> 2 MPI tasks running on the same physical core. Also, the program is not
> paging as the problem size
> fits in the meory.
>
> Do you have any ideas how I can improve the performance so that it matches
> with Intel MPI performance?
> Any suggestions will be greatly appreciated.
>
> Thanks
> Swamy Kandadai
>
>
> Dr. Swamy N. Kandadai
> IBM Senior Certified Executive IT Specialist
> STG WW Modular Systems Benchmark Center
> STG WW HPC and BI CoC Benchmark Center
> Phone:( 845) 433 -8429 (8-293) Fax:(845)432-9789
> sw...@us.ibm.com
> http://w3.ibm.com/sales/systems/benchmarks
>
>
>
>
>
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
>

Reply via email to