What version of OMPI are you running? We stopped supporting bproc after the 1.2 
series, though you could always launch via ssh.

On Dec 12, 2012, at 10:25 PM, Ng Shi Wei <nsw_1...@hotmail.com> wrote:

> Dear all,
> 
> I am new in Linux and clustering. I am setting up a Beowulf Cluster using 
> several PCs according to this guide 
> http://www.tldp.org/HOWTO/html_single/Beowulf-HOWTO/.
> 
> I have setup and configure accordingly except for NFS part. Because I am not 
> requiring it for my application. I have set my ssh to login each other 
> without password. I started with 2 nodes 1st. I can compile and run in my 
> headnode using openmpi. But when I try to run my MPI application across 
> nodes, there is nothing displaying. It just like hanging there.
> 
> Headnode: master
> client: slave4
> 
> The command I used to mpirun across nodes is as below:
> Code:
> mpirun -np 4 --host slave4 output
> Since I not using NFS, so I installed OpenMPI in every nodes with same 
> locations. 
> 
> I wondering I missed out any configurations or not.
> 
> Hope someone can help me out of this problem.
> 
> Thanks in advance.
> 
> Best Regards,
> Shi Wei
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users

Reply via email to