Hi Gus, OpenMPI was not built with tm support.
The submission/execution hosts does not have any of the PBS environment variable set PBS_O_WORKDIR, $PBS_NODEFILE. How i can make set it regards Govind On 9 June 2010 18:45, Gus Correa <g...@ldeo.columbia.edu> wrote: > Hi Govind > > Besides what Ralph said, make sure your OpenMPI was > built with Torque ("tm") support. > > Suggestion: > Do: > > ompi_info --all | grep tm > > It should show lines like these: > > MCA ras: tm (MCA v2.0, API v2.0, Component v1.4.2) > MCA plm: tm (MCA v2.0, API v2.0, Component v1.4.2) > ... > > *** > > If your OpenMPI doesn't have torque support, > you may need to add the nodes list to your mpirun command. > > Suggestion: > > /usr/lib64/openmpi/1.4-gcc/bin/mpirun -hostfile $PBS_NODEFILE -np 4 ./hello > > *** > > Also, assuming your OpenMPI has torque support: > > Did you request 4 nodes from torque? > > If you don't request the nodes and processors, > torque will give you the default values > (which may be one processor and one node). > > Suggestion: > > A script like this (adjusted to your site), tcsh style here, > say, called run_my_pbs_job.tcsh: > > ********* > > #! /bin/tcsh > #PBS -l nodes=4:ppn=1 > #PBS -q default@your.torque.server > #PBS -N myjob > cd $PBS_O_WORKDIR > /usr/lib64/openmpi/1.4-gcc/bin/mpirun -np 4 ./hello > > ********* > > Then do: > qsub run_my_pbs_job.tcsh > > ** > > You can get more information about the PBS syntax using "man qsub". > > ** > > I hope this helps, > Gus Correa > --------------------------------------------------------------------- > Gustavo Correa > Lamont-Doherty Earth Observatory - Columbia University > Palisades, NY, 10964-8000 - USA > --------------------------------------------------------------------- > > Ralph Castain wrote: > >> >> On Jun 9, 2010, at 10:00 AM, Govind Songara wrote: >> >> Thanks Ralph after giving full path of hello it runs. >>> But it run only on one rank >>> Hello World! from process 0 out of 1 on node56.beowulf.cluster >>> >> >> Just to check things out, I would do: >> >> mpirun --display-allocation --display-map -np 4 .... >> >> That should show you the allocation and where OMPI is putting the procs. >> >> there also a error >>> >cat my-script.sh.e43 >>> stty: standard input: Invalid argument >>> >> >> Not really sure here - must be an error in the script itself. >> >> >>> >>> >>> On 9 June 2010 16:46, Ralph Castain <r...@open-mpi.org <mailto: >>> r...@open-mpi.org>> wrote: >>> >>> You need to include the path to "hello" unless it sits in your >>> PATH environment! >>> >>> On Jun 9, 2010, at 9:37 AM, Govind wrote: >>> >>> >>>> #!/bin/sh >>>> /usr/lib64/openmpi/1.4-gcc/bin/mpirun hello >>>> >>>> >>>> On 9 June 2010 16:21, David Zhang <solarbik...@gmail.com >>>> <mailto:solarbik...@gmail.com>> wrote: >>>> >>>> what does your my-script.sh looks like? >>>> >>>> On Wed, Jun 9, 2010 at 8:17 AM, Govind <govind.r...@gmail.com >>>> <mailto:govind.r...@gmail.com>> wrote: >>>> >>>> Hi, >>>> >>>> I have installed following openMPI packge on worker node >>>> from repo >>>> openmpi-libs-1.4-4.el5.x86_64 >>>> openmpi-1.4-4.el5.x86_64 >>>> mpitests-openmpi-3.0-2.el5.x86_64 >>>> mpi-selector-1.0.2-1.el5.noarch >>>> >>>> torque-client-2.3.6-2cri.el5.x86_64 >>>> torque-2.3.6-2cri.el5.x86_64 >>>> torque-mom-2.3.6-2cri.el5.x86_64 >>>> >>>> >>>> Having some problem on running MPI jobs with torque >>>> qsub -q long -l nodes=4 my-script.sh >>>> 42.pbs1 <http://42.pbs1.pp.rhul.ac.uk/> >>>> >>>> >>>> cat my-script.sh.e41 >>>> stty: standard input: Invalid argument >>>> >>>> -------------------------------------------------------------------------- >>>> mpirun was unable to launch the specified application as >>>> it could not find an executable: >>>> >>>> Executable: hello >>>> Node: node56.beowulf.cluster >>>> >>>> while attempting to start process rank 0. >>>> ================================== >>>> >>>> I could run the binary directly on the node without any >>>> problem. >>>> mpiexec -n 4 hello >>>> Hello World! from process 2 out of 4 on >>>> node56.beowulf.cluster >>>> Hello World! from process 0 out of 4 on >>>> node56.beowulf.cluster >>>> Hello World! from process 3 out of 4 on >>>> node56.beowulf.cluster >>>> Hello World! from process 1 out of 4 on >>>> node56.beowulf.cluster >>>> >>>> Could you please advise, if I missing anything here. >>>> >>>> >>>> Regards >>>> Govind >>>> >>>> _______________________________________________ >>>> users mailing list >>>> us...@open-mpi.org <mailto:us...@open-mpi.org> >>>> >>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>> >>>> >>>> >>>> >>>> -- David Zhang >>>> University of California, San Diego >>>> >>>> _______________________________________________ >>>> users mailing list >>>> us...@open-mpi.org <mailto:us...@open-mpi.org> >>>> >>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>> >>>> >>>> _______________________________________________ >>>> users mailing list >>>> us...@open-mpi.org <mailto:us...@open-mpi.org> >>>> >>>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>>> >>> >>> >>> _______________________________________________ >>> users mailing list >>> us...@open-mpi.org <mailto:us...@open-mpi.org> >>> >>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>> >>> >>> _______________________________________________ >>> users mailing list >>> us...@open-mpi.org <mailto:us...@open-mpi.org> >>> >>> http://www.open-mpi.org/mailman/listinfo.cgi/users >>> >> >> >> ------------------------------------------------------------------------ >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users >> > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users >