According to https://svn.open-mpi.org/trac/ompi/milestone/Open%20MPI%201.3 very soon, but you can download trunk version http://www.open-mpi.org/svn/ and check if it works for you.
how can you check mapping CPUs by OS , my cat /proc/cpuinfo shows very little info # cat /proc/cpuinfo processor : 0 cpu : Cell Broadband Engine, altivec supported clock : 3200.000000MHz revision : 48.0 (pvr 0070 3000) processor : 1 cpu : Cell Broadband Engine, altivec supported clock : 3200.000000MHz revision : 48.0 (pvr 0070 3000) processor : 2 cpu : Cell Broadband Engine, altivec supported clock : 3200.000000MHz revision : 48.0 (pvr 0070 3000) processor : 3 cpu : Cell Broadband Engine, altivec supported clock : 3200.000000MHz revision : 48.0 (pvr 0070 3000) timebase : 26666666 platform : Cell machine : CHRP IBM,0793-1RZ On Thu, Oct 23, 2008 at 3:00 PM, Mi Yan <mi...@us.ibm.com> wrote: > Hi, Lenny, > > So rank file map will be supported in OpenMPI 1.3? I'm using OpenMPI1.2.6 > and did not find parameter "rmaps_rank_file_". > Do you have idea when OpenMPI 1.3 will be available? OpenMPI 1.3 has quite > a few features I'm looking for. > > Thanks, > Mi > [image: Inactive hide details for "Lenny Verkhovsky" > <lenny.verkhov...@gmail.com>]"Lenny Verkhovsky" < > lenny.verkhov...@gmail.com> > > > > *"Lenny Verkhovsky" <lenny.verkhov...@gmail.com>* > Sent by: users-boun...@open-mpi.org > > 10/23/2008 05:48 AM Please respond to > Open MPI Users <us...@open-mpi.org> > > > To > > "Open MPI Users" <us...@open-mpi.org> > cc > > > Subject > > Re: [OMPI users] Working with a CellBlade cluster > > Hi, > > > If I understand you correctly the most suitable way to do it is by > paffinity that we have in Open MPI 1.3 and the trank. > how ever usually OS is distributing processes evenly between sockets by it > self. > > There still no formal FAQ due to a multiple reasons but you can read how to > use it in the attached scratch ( there were few name changings of the > params, so check with ompi_info ) > > shared memory is used between processes that share same machine, and openib > is used between different machines ( hostnames ), no special mca params are > needed. > > Best Regards > Lenny, > > > > On Sun, Oct 19, 2008 at 10:32 AM, Gilbert Grosdidier <* > gro...@mail.cern.ch* <gro...@mail.cern.ch>> wrote: > > Working with a CellBlade cluster (QS22), the requirement is to have one > instance of the executable running on each socket of the blade (there > are 2 > sockets). The application is of the 'domain decomposition' type, and > each > instance is required to often send/receive data with both the remote > blades and > the neighbor socket. > > Question is : which specification must be used for the mca btl > component > to force 1) shmem type messages when communicating with this neighbor > socket, > while 2) using openib to communicate with the remote blades ? > Is '-mca btl sm,openib,self' suitable for this ? > > Also, which debug flags could be used to crosscheck that the messages > are > _actually_ going thru the right channel for a given channel, please ? > > We are currently using OpenMPI 1.2.5 shipped with RHEL5.2 (ppc64). > Which version do you think is currently the most optimised for these > processors and problem type ? Should we go towards OpenMPI 1.2.8 > instead ? > Or even try some OpenMPI 1.3 nightly build ? > > Thanks in advance for your help, Gilbert. > > _______________________________________________ > users mailing list* > **us...@open-mpi.org* <us...@open-mpi.org>* > > **http://www.open-mpi.org/mailman/listinfo.cgi/users*<http://www.open-mpi.org/mailman/listinfo.cgi/users> > > *(See attached file: RANKS_FAQ.doc)* > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users >