What di you have in the "hosts" file? We don't have a native integration with 
Condor, so you'll have to specify the hosts and number of slots on each, as 
Reuti explained. You'll also need to check that your sys admin allows you to 
ssh without password to each host.


On Mar 28, 2012, at 10:02 AM, Hameed Alzahrani wrote:

> Hi,
> 
> Thanks that works fine when I submit hello program but when I tried to 
> benchmark the system it look like it does not do anything  
> mpirun -np 8 --hostfile hosts xhpl
> 
> Regards,
> 
> 
> > From: re...@staff.uni-marburg.de
> > Date: Wed, 28 Mar 2012 17:40:07 +0200
> > To: us...@open-mpi.org
> > Subject: Re: [OMPI users] Can not run a parallel job on all the nodes in 
> > the        cluster
> > 
> > Am 28.03.2012 um 17:35 schrieb Hameed Alzahrani:
> > 
> > > Hi,
> > > 
> > > Is there a specific name or location for the hostfile because I could not 
> > > figure how to specify the number of processors for each machine in the 
> > > command line.
> > 
> > No, just specify the name (or path) to it with:
> > 
> > --hostfile foobar
> > 
> > -- Reuti
> > 
> > 
> > > Regards, 
> > > 
> > > > From: re...@staff.uni-marburg.de
> > > > Date: Wed, 28 Mar 2012 17:21:39 +0200
> > > > To: us...@open-mpi.org
> > > > Subject: Re: [OMPI users] Can not run a parallel job on all the nodes 
> > > > in the    cluster
> > > > 
> > > > Hi,
> > > > 
> > > > Am 28.03.2012 um 16:55 schrieb Hameed Alzahrani:
> > > > 
> > > > > I ran hello program which return the host name when I run it using 
> > > > > mpirun -np 8 hello
> > > > > all the 8 answer returned from the same machine
> > > > > when I run it using 
> > > > > mpirun -np 8 --host host1,host2,host3 hello
> > > > > I got answers from all the machines but it is not from all processors 
> > > > > because I have 8 processors host1=4, host2=2, host3=2 the answer was 
> > > > > 3 from host1, 3 from host2 and 2 from host3.
> > > > 
> > > > If you want to specify the number of slots you can put it in a hostfile 
> > > > (otherwise a round robin assignment is just used). I'm not aware that 
> > > > it can be specified on the command line with different values for each 
> > > > machine:
> > > > 
> > > > host1 slots=4
> > > > host2 slots=2
> > > > host3 slots=2
> > > > 
> > > > -- Reuti
> > > > 
> > > > > 
> > > > > Regards,
> > > > > 
> > > > > > From: re...@staff.uni-marburg.de
> > > > > > Date: Wed, 28 Mar 2012 16:42:21 +0200
> > > > > > To: us...@open-mpi.org
> > > > > > Subject: Re: [OMPI users] Can not run a parallel job on all the 
> > > > > > nodes in the        cluster
> > > > > > 
> > > > > > Hi,
> > > > > > 
> > > > > > Am 28.03.2012 um 16:30 schrieb Hameed Alzahrani:
> > > > > > 
> > > > > > > Hi,
> > > > > > > 
> > > > > > > I mean the node that I run mpirun command from. I use condor as a 
> > > > > > > scheduler but I need to benchmark the cluster either from condor 
> > > > > > > or directly from open MPI.
> > > > > > 
> > > > > > I can't say anything regarding the Condor integration of Open MPI, 
> > > > > > but starting it directly by mpirun and supplying a valid number of 
> > > > > > ranks and hostfile should start some processes on other machines as 
> > > > > > requested. Can you run a plain mpihello first and output rank and 
> > > > > > hostname? Do you have ssh access to all the machines in questions? 
> > > > > > You have a shared home directory with the applications?
> > > > > > 
> > > > > > -- Reuti
> > > > > > 
> > > > > > 
> > > > > > > when I ran mpirun from a machine and checking the memory status 
> > > > > > > for the three machines that I have it appear that the memory 
> > > > > > > usage increased just in the same machine.
> > > > > > > 
> > > > > > > Regards,
> > > > > > > 
> > > > > > > > From: re...@staff.uni-marburg.de
> > > > > > > > Date: Wed, 28 Mar 2012 15:12:17 +0200
> > > > > > > > To: us...@open-mpi.org
> > > > > > > > Subject: Re: [OMPI users] Can not run a parallel job on all the 
> > > > > > > > nodes in the    cluster
> > > > > > > > 
> > > > > > > > Hi,
> > > > > > > > 
> > > > > > > > Am 27.03.2012 um 23:46 schrieb Hameed Alzahrani:
> > > > > > > > 
> > > > > > > > > When I run any parallel job I get the answer just from the 
> > > > > > > > > submitting node
> > > > > > > > 
> > > > > > > > what do you mean by submitting node: you use a queuing system - 
> > > > > > > > which one?
> > > > > > > > 
> > > > > > > > -- Reuti
> > > > > > > > 
> > > > > > > > 
> > > > > > > > > even when I tried to benchmark the cluster using LINPACK but 
> > > > > > > > > it look like the job just working on the submitting node is 
> > > > > > > > > there a way to make openMPI send the job equally to all the 
> > > > > > > > > nodes depending on the number of processor in the current 
> > > > > > > > > mode even if I specify that the job should use 8 processor it 
> > > > > > > > > look like openMPI use the submitting node 4 processors 
> > > > > > > > > instead of using the other processors. I tried also --host 
> > > > > > > > > but it does not work correctly in benchmarking the cluster so 
> > > > > > > > > does any one use openMPI in benchmarking a cluster or does 
> > > > > > > > > any one knows how to make openMPI divids the parallel job 
> > > > > > > > > equally to every processor on the cluster.
> > > > > > > > > 
> > > > > > > > > Regards, 
> > > > > > > > > _______________________________________________
> > > > > > > > > users mailing list
> > > > > > > > > us...@open-mpi.org
> > > > > > > > > http://www.open-mpi.org/mailman/listinfo.cgi/users
> > > > > > > > 
> > > > > > > > 
> > > > > > > > _______________________________________________
> > > > > > > > users mailing list
> > > > > > > > us...@open-mpi.org
> > > > > > > > http://www.open-mpi.org/mailman/listinfo.cgi/users
> > > > > > > _______________________________________________
> > > > > > > users mailing list
> > > > > > > us...@open-mpi.org
> > > > > > > http://www.open-mpi.org/mailman/listinfo.cgi/users
> > > > > > 
> > > > > > 
> > > > > > _______________________________________________
> > > > > > users mailing list
> > > > > > us...@open-mpi.org
> > > > > > http://www.open-mpi.org/mailman/listinfo.cgi/users
> > > > > _______________________________________________
> > > > > users mailing list
> > > > > us...@open-mpi.org
> > > > > http://www.open-mpi.org/mailman/listinfo.cgi/users
> > > > 
> > > > 
> > > > _______________________________________________
> > > > users mailing list
> > > > us...@open-mpi.org
> > > > http://www.open-mpi.org/mailman/listinfo.cgi/users
> > > _______________________________________________
> > > users mailing list
> > > us...@open-mpi.org
> > > http://www.open-mpi.org/mailman/listinfo.cgi/users
> > 
> > 
> > _______________________________________________
> > users mailing list
> > us...@open-mpi.org
> > http://www.open-mpi.org/mailman/listinfo.cgi/users
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users

Reply via email to