You need to compile your cpi.c to get an executable.  This is not an MPI issue. 
 :-)

Also, mpdboot is part of a different MPI implementation named MPICH; you don't 
need to run mpdboot with Open MPI.  If you have further questions about MPICH, 
you'll need to ping them on their mailing list -- we aren't able to answer 
MPICH questions here, sorry!

(background: MPI = a book.  It's a specification.  There's a bunch of different 
implementations of that specification available; Open MPI is one [great] one.  
:-)  MPICH is another.  There are also others.)


On Apr 20, 2011, at 10:24 AM, mohd naseem wrote:

> folloeing error shows
> 
> [mpiuser@f2 programrun]$ mpiexec -np 4 ./cpi.c
> problem with execution of ./cpi.c  on  f2:  [Errno 13] Permission denied 
> problem with execution of ./cpi.c  on  f2:  [Errno 13] Permission denied 
> [mpiuser@f2 programrun]$ mpdboot -n 2 -v
> totalnum=2  numhosts=1
> there are not enough hosts on which to start all processes
> 
> 
> 
> On Wed, Apr 20, 2011 at 7:51 PM, mohd naseem <naseemshak...@gmail.com> wrote:
> sir i m still not able to trace all the hosts
> following error shows
> 
> 
> 
> [mpiuser@f2 programrun]$ mpiexec -np 4 ./cpi.c
> problem with execution of ./cpi.c  on  f2:  [Errno 13] Permission denied 
> problem with execution of ./cpi.c  on  f2:  [Errno 13] Permission denied 
> 
> 
> 
> On Tue, Apr 19, 2011 at 8:25 PM, Ralph Castain <r...@open-mpi.org> wrote:
> You have to tell mpiexec what nodes you want to use for your application. 
> This is typically done either on the command line or in a file. For now, you 
> could just do this:
> 
> mpiexec -host node1,node2,node3 -np N ./my_app
> 
> where node1,node2,node3,...  are the names or IP addresses of the nodes you 
> want to run on, and N is the number of total processes you want executed.
> 
> 
> On Apr 19, 2011, at 8:47 AM, mohd naseem wrote:
> 
>> 
>> sorry sir,
>> 
>> i am unable to understand what u are saying ? becoz i am a new user of mpi.
>> 
>> please tell me details about it and command also
>> 
>> thanks
>> 
>> 
>>  
>> On Tue, Apr 19, 2011 at 2:32 PM, Reuti <re...@staff.uni-marburg.de> wrote:
>> Good, then please supply a hostfile with the names of the machines you want 
>> to run for a particular run and give it as option to `mpiexec`. See options 
>> -np and -machinefile.
>> 
>> -- Reuti
>> 
>> 
>> Am 19.04.2011 um 06:38 schrieb mohd naseem:
>> 
>> > sir
>> > when i give mpiexec hostname command.
>> > it only give one hostname. rest are not shown.
>> >
>> >
>> >
>> >
>> >
>> >
>> > On Mon, Apr 18, 2011 at 7:46 PM, Reuti <re...@staff.uni-marburg.de> wrote:
>> > Am 18.04.2011 um 15:40 schrieb chenjie gu:
>> >
>> > > I am a green hand on Openmpi, I have the following Openmpi structure, 
>> > > however it has problem when running across multiple nodes.
>> > > I am trying to build a Bewolf Cluster between 6 nodes of our serve (HP 
>> > > Proliant G460 G7), I have installed the Openmpi on one node (assuming at 
>> > > /mirror),
>> > > ./configure --prefix=/mirror/openmpi CC=icc CXX=icpc F77=ifort FC=ifort
>> > > make all install
>> > >
>> > > using NFS, the directory of /mirror was successfully exported to the 
>> > > rest of 5 nodes. Now as I test the Openmpi, it runs very well on a 
>> > > single node,
>> > > however it hangs across multiple nodes.
>> > >
>> > > Now one possible reason as I know is that Openmpi uses TCP to exchange 
>> > > data between different nodes, so I am worried about
>> > > whether there are firewalls between each nodes, which can be factory 
>> > > integrated at somewhere(switch/NIC). Could anyone give me some
>> > > information on this point?
>> >
>> > It's not only about MPI communcation. Before you need some means to allow 
>> > the startup of the local orte daemons on each machine by passphraseless 
>> > ssh-keys or better hostbased authentication 
>> > http://arc.liv.ac.uk/SGE/howto/hostbased-ssh.html , or enable `rsh` on the 
>> > machines and tell Open MPI to use it. Is:
>> >
>> > mpiexec hostname
>> >
>> > giving you a list of the involved machines?
>> >
>> > -- Reuti
>> >
>> >
>> > > Thanks a lot,
>> > > Regards,
>> > > ArchyGU
>> > > Nanyang Technological University
>> > > _______________________________________________
>> > > users mailing list
>> > > us...@open-mpi.org
>> > > http://www.open-mpi.org/mailman/listinfo.cgi/users
>> >
>> >
>> > _______________________________________________
>> > users mailing list
>> > us...@open-mpi.org
>> > http://www.open-mpi.org/mailman/listinfo.cgi/users
>> >
>> > _______________________________________________
>> > users mailing list
>> > us...@open-mpi.org
>> > http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users


-- 
Jeff Squyres
jsquy...@cisco.com
For corporate legal information go to:
http://www.cisco.com/web/about/doing_business/legal/cri/


Reply via email to