Hi,

Here is what I put in my PBS script to allocate only single node (I want
to use 16 MPI processes in a single node).

#PBS -l mppwidth=16
#PBS -l mppnppn=16
#PBS -l mppdepth=1

Here is the output from aprun (aprun -n 16 -N 16).
Process 2 of 16 is on nid00017
Process 5 of 16 is on nid00017
Process 8 of 16 is on nid00017
Process 12 of 16 is on nid00017
Process 4 of 16 is on nid00017
Process 14 of 16 is on nid00017
Process 0 of 16 is on nid00017
Process 1 of 16 is on nid00017
Process 3 of 16 is on nid00017
Process 13 of 16 is on nid00017
Process 9 of 16 is on nid00017
Process 6 of 16 is on nid00017
Process 11 of 16 is on nid00017
Process 10 of 16 is on nid00017
Process 7 of 16 is on nid00017
Process 15 of 16 is on nid00017



I am guessing that the CrayXE6 here is different from the others in
production (it is 1 cabinet configuration for code development) and I am
afraid mpirun/mpiexec does wrong instantiation of aprun command. Do I have
to edit the script in contrib?


Thanks,
Keita

On 9/3/13 2:51 PM, "Ralph Castain" <r...@open-mpi.org> wrote:

>Interesting - and do you have an allocation? If so, what was it - i.e.,
>can you check the allocation envar to see if you have 16 slots?
>
>
>On Sep 3, 2013, at 1:38 PM, "Teranishi, Keita" <knte...@sandia.gov> wrote:
>
>> It is what I got.
>> 
>> 
>>-------------------------------------------------------------------------
>>-
>> There are not enough slots available in the system to satisfy the 16
>>slots
>> that were requested by the application:
>>  /home/knteran/test-openmpi/cpi
>> 
>> Either request fewer slots for your application, or make more slots
>> available
>> for use.
>> 
>>-------------------------------------------------------------------------
>>-
>> 
>> Thanks,
>> Keita
>> 
>> 
>> 
>> On 9/3/13 1:26 PM, "Ralph Castain" <r...@open-mpi.org> wrote:
>> 
>>> How does it fail?
>>> 
>>> On Sep 3, 2013, at 1:19 PM, "Teranishi, Keita" <knte...@sandia.gov>
>>>wrote:
>>> 
>>>> Nathan,
>>>> 
>>>> Thanks for the help.  I can run a job using openmpi, assigning a
>>>>signle
>>>> process per node.  However, I have been failing to run a job using
>>>> multiple MPI ranks in a single node.  In other words, "mpiexec
>>>> --bind-to-core --npernode 16 --n 16 ./test" never works (apron -n 16
>>>> works
>>>> fine).  DO you have any thought about it?
>>>> 
>>>> Thanks,
>>>> ---------------------------------------------
>>>> Keita Teranishi
>>>> R&D Principal Staff Member
>>>> Scalable Modeling and Analysis Systems
>>>> Sandia National Laboratories
>>>> Livermore, CA 94551
>>>> 
>>>> 
>>>> 
>>>> 
>>>> On 8/30/13 8:49 AM, "Hjelm, Nathan T" <hje...@lanl.gov> wrote:
>>>> 
>>>>> Replace install_path to where you want Open MPI installed.
>>>>> 
>>>>> ./configure --prefix=install_path
>>>>> --with-platform=contrib/platform/lanl/cray_xe6/optimized-luster
>>>>> make
>>>>> make install
>>>>> 
>>>>> To use Open MPI just set the PATH and LD_LIBRARY_PATH:
>>>>> 
>>>>> PATH=install_path/bin:$PATH
>>>>> LD_LIBRARY_PATH=install_path/lib:$LD_LIBRARY_PATH
>>>>> 
>>>>> You can then use mpicc, mpicxx, mpif90, etc to compile and either
>>>>> mpirun
>>>>> or aprun to run. If you are running at scale I would recommend
>>>>>against
>>>>> using aprun for now. I also recommend you change your programming
>>>>> environment to either PrgEnv-gnu or PrgEnv-intel. The PGI compiler
>>>>>can
>>>>> be
>>>>> a PIA. It is possible to build with the Cray compiler but it takes
>>>>> patching the config.guess and changing some autoconf stuff.
>>>>> 
>>>>> -Nathan
>>>>> 
>>>>> Please excuse the horrible Outlook-style quoting.
>>>>> ________________________________________
>>>>> From: users [users-boun...@open-mpi.org] on behalf of Teranishi,
>>>>>Keita
>>>>> [knte...@sandia.gov]
>>>>> Sent: Thursday, August 29, 2013 8:01 PM
>>>>> To: Open MPI Users
>>>>> Subject: Re: [OMPI users] [EXTERNAL] Re: What version of PMI (Cray
>>>>>XE6)
>>>>> is working for OpenMPI-1.6.5?
>>>>> 
>>>>> Thanks for the info.  Is it still possible to build by myself?  What
>>>>>is
>>>>> the procedure other than configure script?
>>>>> 
>>>>> 
>>>>> 
>>>>> 
>>>>> 
>>>>> On 8/23/13 2:37 PM, "Nathan Hjelm" <hje...@lanl.gov> wrote:
>>>>> 
>>>>>> On Fri, Aug 23, 2013 at 09:14:25PM +0000, Teranishi, Keita wrote:
>>>>>>>  Hi,
>>>>>>>  I am trying to install OpenMPI 1.6.5 on Cray XE6 and very curious
>>>>>>> with the
>>>>>>>  current support of PMI.  In the previous discussions, there was a
>>>>>>> comment
>>>>>>>  on the version of PMI (it works with 2.1.4, but fails with 3.0).
>>>>>>> Our
>>>>>> 
>>>>>> Open MPI 1.6.5 does not have support for the XE-6. Use 1.7.2
>>>>>>instead.
>>>>>> 
>>>>>>>  machine has PMI2.1.4 and PMI4.0 (default).  Which version do you
>>>>>> 
>>>>>> There was a regression in PMI 3.x.x that still exists in 4.0.x that
>>>>>> causes a warning to be printed on every rank when using mpirun. We
>>>>>>are
>>>>>> working with Cray to resolve the issue. For now use 2.1.4. See the
>>>>>> platform files in contrib/platform/lanl/cray_xe6. The platform files
>>>>>> you
>>>>>> would want to use are debug-lustre or optimized-lusre.
>>>>>> 
>>>>>> BTW, 1.7.2 is installed on Cielo and Cielito. Just run:
>>>>>> 
>>>>>> module swap PrgEnv-pgi PrgEnv-gnu (PrgEnv-intel also works)
>>>>>> module unload cray-mpich2 xt-libsci
>>>>>> module load openmpi/1.7.2
>>>>>> 
>>>>>> 
>>>>>> -Nathan Hjelm
>>>>>> Open MPI Team, HPC-3, LANL
>>>>>> _______________________________________________
>>>>>> users mailing list
>>>>>> us...@open-mpi.org
>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>> 
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>> 
>>>> _______________________________________________
>>>> users mailing list
>>>> us...@open-mpi.org
>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>> 
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>
>_______________________________________________
>users mailing list
>us...@open-mpi.org
>http://www.open-mpi.org/mailman/listinfo.cgi/users

Reply via email to