Nathan,

It is close to Cielo and use resource manager under
/opt/cray/xe-sysroot/4.1.40/usr.

Currently Loaded Modulefiles:
  1) modules/3.2.6.7                         17)
csa/3.0.0-1_2.0401.37452.4.50.gem
  2) craype-network-gemini                   18)
job/1.5.5-0.1_2.0401.35380.1.10.gem
  3) cray-mpich2/5.6.4                       19)
xpmem/0.1-2.0401.36790.4.3.gem
  4) atp/1.6.3                               20)
gni-headers/2.1-1.0401.5675.4.4.gem
  5) xe-sysroot/4.1.40                       21)
dmapp/3.2.1-1.0401.5983.4.5.gem
  6) switch/1.0-1.0401.36779.2.72.gem        22)
pmi/2.1.4-1.0000.8596.8.9.gem
  7) shared-root/1.0-1.0401.37253.3.50.gem   23)
ugni/4.0-1.0401.5928.9.5.gem
  8) pdsh/2.26-1.0401.37449.1.1.gem          24)
udreg/2.3.2-1.0401.5929.3.3.gem
  9) nodehealth/5.0-1.0401.38460.12.18.gem   25) xt-libsci/12.0.00
 10) lbcd/2.1-1.0401.35360.1.2.gem           26) xt-totalview/8.12.0
 11) hosts/1.0-1.0401.35364.1.115.gem        27) totalview-support/1.1.4
 12) configuration/1.0-1.0401.35391.1.2.gem  28) gcc/4.7.2
 13) ccm/2.2.0-1.0401.37254.2.142            29) xt-asyncpe/5.22
 14) audit/1.0.0-1.0401.37969.2.32.gem       30) eswrap/1.0.8
 15) rca/1.0.0-2.0401.38656.2.2.gem          31) craype-mc8
 16) dvs/1.8.6_0.9.0-1.0401.1401.1.120       32) PrgEnv-gnu/4.1.40


Thanks,
Keita



On 9/3/13 3:42 PM, "Nathan Hjelm" <hje...@lanl.gov> wrote:

>Hmm, what CLE release is your development cluster running? It is the value
>after PrgEnv. Ex. on Cielito we have 4.1.40.
>
>32) PrgEnv-gnu/4.1.40
>
>We have not yet fully tested Open MPI on CLE 5.x.x.
>
>-Nathan Hjelm
>HPC-3, LANL
>
>On Tue, Sep 03, 2013 at 10:33:57PM +0000, Teranishi, Keita wrote:
>> Hi,
>> 
>> Here is what I put in my PBS script to allocate only single node (I want
>> to use 16 MPI processes in a single node).
>> 
>> #PBS -l mppwidth=16
>> #PBS -l mppnppn=16
>> #PBS -l mppdepth=1
>> 
>> Here is the output from aprun (aprun -n 16 -N 16).
>> Process 2 of 16 is on nid00017
>> Process 5 of 16 is on nid00017
>> Process 8 of 16 is on nid00017
>> Process 12 of 16 is on nid00017
>> Process 4 of 16 is on nid00017
>> Process 14 of 16 is on nid00017
>> Process 0 of 16 is on nid00017
>> Process 1 of 16 is on nid00017
>> Process 3 of 16 is on nid00017
>> Process 13 of 16 is on nid00017
>> Process 9 of 16 is on nid00017
>> Process 6 of 16 is on nid00017
>> Process 11 of 16 is on nid00017
>> Process 10 of 16 is on nid00017
>> Process 7 of 16 is on nid00017
>> Process 15 of 16 is on nid00017
>> 
>> 
>> 
>> I am guessing that the CrayXE6 here is different from the others in
>> production (it is 1 cabinet configuration for code development) and I am
>> afraid mpirun/mpiexec does wrong instantiation of aprun command. Do I
>>have
>> to edit the script in contrib?
>> 
>> 
>> Thanks,
>> Keita
>> 
>> On 9/3/13 2:51 PM, "Ralph Castain" <r...@open-mpi.org> wrote:
>> 
>> >Interesting - and do you have an allocation? If so, what was it - i.e.,
>> >can you check the allocation envar to see if you have 16 slots?
>> >
>> >
>> >On Sep 3, 2013, at 1:38 PM, "Teranishi, Keita" <knte...@sandia.gov>
>>wrote:
>> >
>> >> It is what I got.
>> >> 
>> >> 
>> 
>>>>-----------------------------------------------------------------------
>>>>--
>> >>-
>> >> There are not enough slots available in the system to satisfy the 16
>> >>slots
>> >> that were requested by the application:
>> >>  /home/knteran/test-openmpi/cpi
>> >> 
>> >> Either request fewer slots for your application, or make more slots
>> >> available
>> >> for use.
>> >> 
>> 
>>>>-----------------------------------------------------------------------
>>>>--
>> >>-
>> >> 
>> >> Thanks,
>> >> Keita
>> >> 
>> >> 
>> >> 
>> >> On 9/3/13 1:26 PM, "Ralph Castain" <r...@open-mpi.org> wrote:
>> >> 
>> >>> How does it fail?
>> >>> 
>> >>> On Sep 3, 2013, at 1:19 PM, "Teranishi, Keita" <knte...@sandia.gov>
>> >>>wrote:
>> >>> 
>> >>>> Nathan,
>> >>>> 
>> >>>> Thanks for the help.  I can run a job using openmpi, assigning a
>> >>>>signle
>> >>>> process per node.  However, I have been failing to run a job using
>> >>>> multiple MPI ranks in a single node.  In other words, "mpiexec
>> >>>> --bind-to-core --npernode 16 --n 16 ./test" never works (apron -n
>>16
>> >>>> works
>> >>>> fine).  DO you have any thought about it?
>> >>>> 
>> >>>> Thanks,
>> >>>> ---------------------------------------------
>> >>>> Keita Teranishi
>> >>>> R&D Principal Staff Member
>> >>>> Scalable Modeling and Analysis Systems
>> >>>> Sandia National Laboratories
>> >>>> Livermore, CA 94551
>> >>>> 
>> >>>> 
>> >>>> 
>> >>>> 
>> >>>> On 8/30/13 8:49 AM, "Hjelm, Nathan T" <hje...@lanl.gov> wrote:
>> >>>> 
>> >>>>> Replace install_path to where you want Open MPI installed.
>> >>>>> 
>> >>>>> ./configure --prefix=install_path
>> >>>>> --with-platform=contrib/platform/lanl/cray_xe6/optimized-luster
>> >>>>> make
>> >>>>> make install
>> >>>>> 
>> >>>>> To use Open MPI just set the PATH and LD_LIBRARY_PATH:
>> >>>>> 
>> >>>>> PATH=install_path/bin:$PATH
>> >>>>> LD_LIBRARY_PATH=install_path/lib:$LD_LIBRARY_PATH
>> >>>>> 
>> >>>>> You can then use mpicc, mpicxx, mpif90, etc to compile and either
>> >>>>> mpirun
>> >>>>> or aprun to run. If you are running at scale I would recommend
>> >>>>>against
>> >>>>> using aprun for now. I also recommend you change your programming
>> >>>>> environment to either PrgEnv-gnu or PrgEnv-intel. The PGI compiler
>> >>>>>can
>> >>>>> be
>> >>>>> a PIA. It is possible to build with the Cray compiler but it takes
>> >>>>> patching the config.guess and changing some autoconf stuff.
>> >>>>> 
>> >>>>> -Nathan
>> >>>>> 
>> >>>>> Please excuse the horrible Outlook-style quoting.
>> >>>>> ________________________________________
>> >>>>> From: users [users-boun...@open-mpi.org] on behalf of Teranishi,
>> >>>>>Keita
>> >>>>> [knte...@sandia.gov]
>> >>>>> Sent: Thursday, August 29, 2013 8:01 PM
>> >>>>> To: Open MPI Users
>> >>>>> Subject: Re: [OMPI users] [EXTERNAL] Re: What version of PMI (Cray
>> >>>>>XE6)
>> >>>>> is working for OpenMPI-1.6.5?
>> >>>>> 
>> >>>>> Thanks for the info.  Is it still possible to build by myself?
>>What
>> >>>>>is
>> >>>>> the procedure other than configure script?
>> >>>>> 
>> >>>>> 
>> >>>>> 
>> >>>>> 
>> >>>>> 
>> >>>>> On 8/23/13 2:37 PM, "Nathan Hjelm" <hje...@lanl.gov> wrote:
>> >>>>> 
>> >>>>>> On Fri, Aug 23, 2013 at 09:14:25PM +0000, Teranishi, Keita wrote:
>> >>>>>>>  Hi,
>> >>>>>>>  I am trying to install OpenMPI 1.6.5 on Cray XE6 and very
>>curious
>> >>>>>>> with the
>> >>>>>>>  current support of PMI.  In the previous discussions, there
>>was a
>> >>>>>>> comment
>> >>>>>>>  on the version of PMI (it works with 2.1.4, but fails with
>>3.0).
>> >>>>>>> Our
>> >>>>>> 
>> >>>>>> Open MPI 1.6.5 does not have support for the XE-6. Use 1.7.2
>> >>>>>>instead.
>> >>>>>> 
>> >>>>>>>  machine has PMI2.1.4 and PMI4.0 (default).  Which version do
>>you
>> >>>>>> 
>> >>>>>> There was a regression in PMI 3.x.x that still exists in 4.0.x
>>that
>> >>>>>> causes a warning to be printed on every rank when using mpirun.
>>We
>> >>>>>>are
>> >>>>>> working with Cray to resolve the issue. For now use 2.1.4. See
>>the
>> >>>>>> platform files in contrib/platform/lanl/cray_xe6. The platform
>>files
>> >>>>>> you
>> >>>>>> would want to use are debug-lustre or optimized-lusre.
>> >>>>>> 
>> >>>>>> BTW, 1.7.2 is installed on Cielo and Cielito. Just run:
>> >>>>>> 
>> >>>>>> module swap PrgEnv-pgi PrgEnv-gnu (PrgEnv-intel also works)
>> >>>>>> module unload cray-mpich2 xt-libsci
>> >>>>>> module load openmpi/1.7.2
>> >>>>>> 
>> >>>>>> 
>> >>>>>> -Nathan Hjelm
>> >>>>>> Open MPI Team, HPC-3, LANL
>> >>>>>> _______________________________________________
>> >>>>>> users mailing list
>> >>>>>> us...@open-mpi.org
>> >>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> >>>>> 
>> >>>>> _______________________________________________
>> >>>>> users mailing list
>> >>>>> us...@open-mpi.org
>> >>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> >>>>> _______________________________________________
>> >>>>> users mailing list
>> >>>>> us...@open-mpi.org
>> >>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> >>>> 
>> >>>> _______________________________________________
>> >>>> users mailing list
>> >>>> us...@open-mpi.org
>> >>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> >>> 
>> >>> _______________________________________________
>> >>> users mailing list
>> >>> us...@open-mpi.org
>> >>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> >> 
>> >> _______________________________________________
>> >> users mailing list
>> >> us...@open-mpi.org
>> >> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> >
>> >_______________________________________________
>> >users mailing list
>> >us...@open-mpi.org
>> >http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>_______________________________________________
>users mailing list
>us...@open-mpi.org
>http://www.open-mpi.org/mailman/listinfo.cgi/users

Reply via email to