Nathan, It is close to Cielo and use resource manager under /opt/cray/xe-sysroot/4.1.40/usr.
Currently Loaded Modulefiles: 1) modules/3.2.6.7 17) csa/3.0.0-1_2.0401.37452.4.50.gem 2) craype-network-gemini 18) job/1.5.5-0.1_2.0401.35380.1.10.gem 3) cray-mpich2/5.6.4 19) xpmem/0.1-2.0401.36790.4.3.gem 4) atp/1.6.3 20) gni-headers/2.1-1.0401.5675.4.4.gem 5) xe-sysroot/4.1.40 21) dmapp/3.2.1-1.0401.5983.4.5.gem 6) switch/1.0-1.0401.36779.2.72.gem 22) pmi/2.1.4-1.0000.8596.8.9.gem 7) shared-root/1.0-1.0401.37253.3.50.gem 23) ugni/4.0-1.0401.5928.9.5.gem 8) pdsh/2.26-1.0401.37449.1.1.gem 24) udreg/2.3.2-1.0401.5929.3.3.gem 9) nodehealth/5.0-1.0401.38460.12.18.gem 25) xt-libsci/12.0.00 10) lbcd/2.1-1.0401.35360.1.2.gem 26) xt-totalview/8.12.0 11) hosts/1.0-1.0401.35364.1.115.gem 27) totalview-support/1.1.4 12) configuration/1.0-1.0401.35391.1.2.gem 28) gcc/4.7.2 13) ccm/2.2.0-1.0401.37254.2.142 29) xt-asyncpe/5.22 14) audit/1.0.0-1.0401.37969.2.32.gem 30) eswrap/1.0.8 15) rca/1.0.0-2.0401.38656.2.2.gem 31) craype-mc8 16) dvs/1.8.6_0.9.0-1.0401.1401.1.120 32) PrgEnv-gnu/4.1.40 Thanks, Keita On 9/3/13 3:42 PM, "Nathan Hjelm" <hje...@lanl.gov> wrote: >Hmm, what CLE release is your development cluster running? It is the value >after PrgEnv. Ex. on Cielito we have 4.1.40. > >32) PrgEnv-gnu/4.1.40 > >We have not yet fully tested Open MPI on CLE 5.x.x. > >-Nathan Hjelm >HPC-3, LANL > >On Tue, Sep 03, 2013 at 10:33:57PM +0000, Teranishi, Keita wrote: >> Hi, >> >> Here is what I put in my PBS script to allocate only single node (I want >> to use 16 MPI processes in a single node). >> >> #PBS -l mppwidth=16 >> #PBS -l mppnppn=16 >> #PBS -l mppdepth=1 >> >> Here is the output from aprun (aprun -n 16 -N 16). >> Process 2 of 16 is on nid00017 >> Process 5 of 16 is on nid00017 >> Process 8 of 16 is on nid00017 >> Process 12 of 16 is on nid00017 >> Process 4 of 16 is on nid00017 >> Process 14 of 16 is on nid00017 >> Process 0 of 16 is on nid00017 >> Process 1 of 16 is on nid00017 >> Process 3 of 16 is on nid00017 >> Process 13 of 16 is on nid00017 >> Process 9 of 16 is on nid00017 >> Process 6 of 16 is on nid00017 >> Process 11 of 16 is on nid00017 >> Process 10 of 16 is on nid00017 >> Process 7 of 16 is on nid00017 >> Process 15 of 16 is on nid00017 >> >> >> >> I am guessing that the CrayXE6 here is different from the others in >> production (it is 1 cabinet configuration for code development) and I am >> afraid mpirun/mpiexec does wrong instantiation of aprun command. Do I >>have >> to edit the script in contrib? >> >> >> Thanks, >> Keita >> >> On 9/3/13 2:51 PM, "Ralph Castain" <r...@open-mpi.org> wrote: >> >> >Interesting - and do you have an allocation? If so, what was it - i.e., >> >can you check the allocation envar to see if you have 16 slots? >> > >> > >> >On Sep 3, 2013, at 1:38 PM, "Teranishi, Keita" <knte...@sandia.gov> >>wrote: >> > >> >> It is what I got. >> >> >> >> >> >>>>----------------------------------------------------------------------- >>>>-- >> >>- >> >> There are not enough slots available in the system to satisfy the 16 >> >>slots >> >> that were requested by the application: >> >> /home/knteran/test-openmpi/cpi >> >> >> >> Either request fewer slots for your application, or make more slots >> >> available >> >> for use. >> >> >> >>>>----------------------------------------------------------------------- >>>>-- >> >>- >> >> >> >> Thanks, >> >> Keita >> >> >> >> >> >> >> >> On 9/3/13 1:26 PM, "Ralph Castain" <r...@open-mpi.org> wrote: >> >> >> >>> How does it fail? >> >>> >> >>> On Sep 3, 2013, at 1:19 PM, "Teranishi, Keita" <knte...@sandia.gov> >> >>>wrote: >> >>> >> >>>> Nathan, >> >>>> >> >>>> Thanks for the help. I can run a job using openmpi, assigning a >> >>>>signle >> >>>> process per node. However, I have been failing to run a job using >> >>>> multiple MPI ranks in a single node. In other words, "mpiexec >> >>>> --bind-to-core --npernode 16 --n 16 ./test" never works (apron -n >>16 >> >>>> works >> >>>> fine). DO you have any thought about it? >> >>>> >> >>>> Thanks, >> >>>> --------------------------------------------- >> >>>> Keita Teranishi >> >>>> R&D Principal Staff Member >> >>>> Scalable Modeling and Analysis Systems >> >>>> Sandia National Laboratories >> >>>> Livermore, CA 94551 >> >>>> >> >>>> >> >>>> >> >>>> >> >>>> On 8/30/13 8:49 AM, "Hjelm, Nathan T" <hje...@lanl.gov> wrote: >> >>>> >> >>>>> Replace install_path to where you want Open MPI installed. >> >>>>> >> >>>>> ./configure --prefix=install_path >> >>>>> --with-platform=contrib/platform/lanl/cray_xe6/optimized-luster >> >>>>> make >> >>>>> make install >> >>>>> >> >>>>> To use Open MPI just set the PATH and LD_LIBRARY_PATH: >> >>>>> >> >>>>> PATH=install_path/bin:$PATH >> >>>>> LD_LIBRARY_PATH=install_path/lib:$LD_LIBRARY_PATH >> >>>>> >> >>>>> You can then use mpicc, mpicxx, mpif90, etc to compile and either >> >>>>> mpirun >> >>>>> or aprun to run. If you are running at scale I would recommend >> >>>>>against >> >>>>> using aprun for now. I also recommend you change your programming >> >>>>> environment to either PrgEnv-gnu or PrgEnv-intel. The PGI compiler >> >>>>>can >> >>>>> be >> >>>>> a PIA. It is possible to build with the Cray compiler but it takes >> >>>>> patching the config.guess and changing some autoconf stuff. >> >>>>> >> >>>>> -Nathan >> >>>>> >> >>>>> Please excuse the horrible Outlook-style quoting. >> >>>>> ________________________________________ >> >>>>> From: users [users-boun...@open-mpi.org] on behalf of Teranishi, >> >>>>>Keita >> >>>>> [knte...@sandia.gov] >> >>>>> Sent: Thursday, August 29, 2013 8:01 PM >> >>>>> To: Open MPI Users >> >>>>> Subject: Re: [OMPI users] [EXTERNAL] Re: What version of PMI (Cray >> >>>>>XE6) >> >>>>> is working for OpenMPI-1.6.5? >> >>>>> >> >>>>> Thanks for the info. Is it still possible to build by myself? >>What >> >>>>>is >> >>>>> the procedure other than configure script? >> >>>>> >> >>>>> >> >>>>> >> >>>>> >> >>>>> >> >>>>> On 8/23/13 2:37 PM, "Nathan Hjelm" <hje...@lanl.gov> wrote: >> >>>>> >> >>>>>> On Fri, Aug 23, 2013 at 09:14:25PM +0000, Teranishi, Keita wrote: >> >>>>>>> Hi, >> >>>>>>> I am trying to install OpenMPI 1.6.5 on Cray XE6 and very >>curious >> >>>>>>> with the >> >>>>>>> current support of PMI. In the previous discussions, there >>was a >> >>>>>>> comment >> >>>>>>> on the version of PMI (it works with 2.1.4, but fails with >>3.0). >> >>>>>>> Our >> >>>>>> >> >>>>>> Open MPI 1.6.5 does not have support for the XE-6. Use 1.7.2 >> >>>>>>instead. >> >>>>>> >> >>>>>>> machine has PMI2.1.4 and PMI4.0 (default). Which version do >>you >> >>>>>> >> >>>>>> There was a regression in PMI 3.x.x that still exists in 4.0.x >>that >> >>>>>> causes a warning to be printed on every rank when using mpirun. >>We >> >>>>>>are >> >>>>>> working with Cray to resolve the issue. For now use 2.1.4. See >>the >> >>>>>> platform files in contrib/platform/lanl/cray_xe6. The platform >>files >> >>>>>> you >> >>>>>> would want to use are debug-lustre or optimized-lusre. >> >>>>>> >> >>>>>> BTW, 1.7.2 is installed on Cielo and Cielito. Just run: >> >>>>>> >> >>>>>> module swap PrgEnv-pgi PrgEnv-gnu (PrgEnv-intel also works) >> >>>>>> module unload cray-mpich2 xt-libsci >> >>>>>> module load openmpi/1.7.2 >> >>>>>> >> >>>>>> >> >>>>>> -Nathan Hjelm >> >>>>>> Open MPI Team, HPC-3, LANL >> >>>>>> _______________________________________________ >> >>>>>> users mailing list >> >>>>>> us...@open-mpi.org >> >>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >>>>> >> >>>>> _______________________________________________ >> >>>>> users mailing list >> >>>>> us...@open-mpi.org >> >>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >>>>> _______________________________________________ >> >>>>> users mailing list >> >>>>> us...@open-mpi.org >> >>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >>>> >> >>>> _______________________________________________ >> >>>> users mailing list >> >>>> us...@open-mpi.org >> >>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >>> >> >>> _______________________________________________ >> >>> users mailing list >> >>> us...@open-mpi.org >> >>> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> >> >> _______________________________________________ >> >> users mailing list >> >> us...@open-mpi.org >> >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> > >> >_______________________________________________ >> >users mailing list >> >us...@open-mpi.org >> >http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users >_______________________________________________ >users mailing list >us...@open-mpi.org >http://www.open-mpi.org/mailman/listinfo.cgi/users