Hi Govind
Govind Songara wrote:
> Hi Gus,
>
> OpenMPI was not built with tm support.
>
I suspected that.
Reading your postings, it seems to be an OpenMPI rpm from
a Linux distribution, which I would guess are generic,
and have no specific support for any resource manager like Torque.
> The submission/execution hosts does not have any of the
> PBS environment variable set
> PBS_O_WORKDIR, $PBS_NODEFILE.
> How i can make set it
>
> regards
> Govind
>
I sent you a suggestion in my previous message!
Here it is again:
> If your OpenMPI doesn't have torque support,
> you may need to add the nodes list to your mpirun command.
>
> Suggestion:
>
> /usr/lib64/openmpi/1.4-gcc/bin/mpirun -hostfile $PBS_NODEFILE -np 4
> ./hello
>
I hope this helps.
Gus Correa
---------------------------------------------------------------------
Gustavo Correa
Lamont-Doherty Earth Observatory - Columbia University
Palisades, NY, 10964-8000 - USA
---------------------------------------------------------------------
On 9 June 2010 18:45, Gus Correa <g...@ldeo.columbia.edu
<mailto:g...@ldeo.columbia.edu>> wrote:
Hi Govind
Besides what Ralph said, make sure your OpenMPI was
built with Torque ("tm") support.
Suggestion:
Do:
ompi_info --all | grep tm
It should show lines like these:
MCA ras: tm (MCA v2.0, API v2.0, Component v1.4.2)
MCA plm: tm (MCA v2.0, API v2.0, Component v1.4.2)
...
***
If your OpenMPI doesn't have torque support,
you may need to add the nodes list to your mpirun command.
Suggestion:
/usr/lib64/openmpi/1.4-gcc/bin/mpirun -hostfile $PBS_NODEFILE -np 4
./hello
***
Also, assuming your OpenMPI has torque support:
Did you request 4 nodes from torque?
If you don't request the nodes and processors,
torque will give you the default values
(which may be one processor and one node).
Suggestion:
A script like this (adjusted to your site), tcsh style here,
say, called run_my_pbs_job.tcsh:
*********
#! /bin/tcsh
#PBS -l nodes=4:ppn=1
#PBS -q default@your.torque.server
#PBS -N myjob
cd $PBS_O_WORKDIR
/usr/lib64/openmpi/1.4-gcc/bin/mpirun -np 4 ./hello
*********
Then do:
qsub run_my_pbs_job.tcsh
**
You can get more information about the PBS syntax using "man qsub".
**
I hope this helps,
Gus Correa
---------------------------------------------------------------------
Gustavo Correa
Lamont-Doherty Earth Observatory - Columbia University
Palisades, NY, 10964-8000 - USA
---------------------------------------------------------------------
Ralph Castain wrote:
On Jun 9, 2010, at 10:00 AM, Govind Songara wrote:
Thanks Ralph after giving full path of hello it runs.
But it run only on one rank
Hello World! from process 0 out of 1 on node56.beowulf.cluster
Just to check things out, I would do:
mpirun --display-allocation --display-map -np 4 ....
That should show you the allocation and where OMPI is putting
the procs.
there also a error
>cat my-script.sh.e43
stty: standard input: Invalid argument
Not really sure here - must be an error in the script itself.
On 9 June 2010 16:46, Ralph Castain <r...@open-mpi.org
<mailto:r...@open-mpi.org> <mailto:r...@open-mpi.org
<mailto:r...@open-mpi.org>>> wrote:
You need to include the path to "hello" unless it sits in
your
PATH environment!
On Jun 9, 2010, at 9:37 AM, Govind wrote:
#!/bin/sh
/usr/lib64/openmpi/1.4-gcc/bin/mpirun hello
On 9 June 2010 16:21, David Zhang
<solarbik...@gmail.com <mailto:solarbik...@gmail.com>
<mailto:solarbik...@gmail.com
<mailto:solarbik...@gmail.com>>> wrote:
what does your my-script.sh looks like?
On Wed, Jun 9, 2010 at 8:17 AM, Govind
<govind.r...@gmail.com <mailto:govind.r...@gmail.com>
<mailto:govind.r...@gmail.com
<mailto:govind.r...@gmail.com>>> wrote:
Hi,
I have installed following openMPI packge on
worker node
from repo
openmpi-libs-1.4-4.el5.x86_64
openmpi-1.4-4.el5.x86_64
mpitests-openmpi-3.0-2.el5.x86_64
mpi-selector-1.0.2-1.el5.noarch
torque-client-2.3.6-2cri.el5.x86_64
torque-2.3.6-2cri.el5.x86_64
torque-mom-2.3.6-2cri.el5.x86_64
Having some problem on running MPI jobs with
torque
qsub -q long -l nodes=4 my-script.sh
42.pbs1 <http://42.pbs1.pp.rhul.ac.uk/>
cat my-script.sh.e41
stty: standard input: Invalid argument
--------------------------------------------------------------------------
mpirun was unable to launch the specified
application as
it could not find an executable:
Executable: hello
Node: node56.beowulf.cluster
while attempting to start process rank 0.
==================================
I could run the binary directly on the node
without any
problem.
mpiexec -n 4 hello
Hello World! from process 2 out of 4 on
node56.beowulf.cluster
Hello World! from process 0 out of 4 on
node56.beowulf.cluster
Hello World! from process 3 out of 4 on
node56.beowulf.cluster
Hello World! from process 1 out of 4 on
node56.beowulf.cluster
Could you please advise, if I missing
anything here.
Regards
Govind
_______________________________________________
users mailing list
us...@open-mpi.org
<mailto:us...@open-mpi.org> <mailto:us...@open-mpi.org
<mailto:us...@open-mpi.org>>
http://www.open-mpi.org/mailman/listinfo.cgi/users
-- David Zhang
University of California, San Diego
_______________________________________________
users mailing list
us...@open-mpi.org <mailto:us...@open-mpi.org>
<mailto:us...@open-mpi.org <mailto:us...@open-mpi.org>>
http://www.open-mpi.org/mailman/listinfo.cgi/users
_______________________________________________
users mailing list
us...@open-mpi.org <mailto:us...@open-mpi.org>
<mailto:us...@open-mpi.org <mailto:us...@open-mpi.org>>
http://www.open-mpi.org/mailman/listinfo.cgi/users
_______________________________________________
users mailing list
us...@open-mpi.org <mailto:us...@open-mpi.org>
<mailto:us...@open-mpi.org <mailto:us...@open-mpi.org>>
http://www.open-mpi.org/mailman/listinfo.cgi/users
_______________________________________________
users mailing list
us...@open-mpi.org <mailto:us...@open-mpi.org>
<mailto:us...@open-mpi.org <mailto:us...@open-mpi.org>>
http://www.open-mpi.org/mailman/listinfo.cgi/users
------------------------------------------------------------------------
_______________________________________________
users mailing list
us...@open-mpi.org <mailto:us...@open-mpi.org>
http://www.open-mpi.org/mailman/listinfo.cgi/users
_______________________________________________
users mailing list
us...@open-mpi.org <mailto:us...@open-mpi.org>
http://www.open-mpi.org/mailman/listinfo.cgi/users
------------------------------------------------------------------------
_______________________________________________
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users