Hi, Is there a specific name or location for the hostfile because I could not figure how to specify the number of processors for each machine in the command line.
Regards, > From: re...@staff.uni-marburg.de > Date: Wed, 28 Mar 2012 17:21:39 +0200 > To: us...@open-mpi.org > Subject: Re: [OMPI users] Can not run a parallel job on all the nodes in the > cluster > > Hi, > > Am 28.03.2012 um 16:55 schrieb Hameed Alzahrani: > > > I ran hello program which return the host name when I run it using > > mpirun -np 8 hello > > all the 8 answer returned from the same machine > > when I run it using > > mpirun -np 8 --host host1,host2,host3 hello > > I got answers from all the machines but it is not from all processors > > because I have 8 processors host1=4, host2=2, host3=2 the answer was 3 from > > host1, 3 from host2 and 2 from host3. > > If you want to specify the number of slots you can put it in a hostfile > (otherwise a round robin assignment is just used). I'm not aware that it can > be specified on the command line with different values for each machine: > > host1 slots=4 > host2 slots=2 > host3 slots=2 > > -- Reuti > > > > > Regards, > > > > > From: re...@staff.uni-marburg.de > > > Date: Wed, 28 Mar 2012 16:42:21 +0200 > > > To: us...@open-mpi.org > > > Subject: Re: [OMPI users] Can not run a parallel job on all the nodes in > > > the cluster > > > > > > Hi, > > > > > > Am 28.03.2012 um 16:30 schrieb Hameed Alzahrani: > > > > > > > Hi, > > > > > > > > I mean the node that I run mpirun command from. I use condor as a > > > > scheduler but I need to benchmark the cluster either from condor or > > > > directly from open MPI. > > > > > > I can't say anything regarding the Condor integration of Open MPI, but > > > starting it directly by mpirun and supplying a valid number of ranks and > > > hostfile should start some processes on other machines as requested. Can > > > you run a plain mpihello first and output rank and hostname? Do you have > > > ssh access to all the machines in questions? You have a shared home > > > directory with the applications? > > > > > > -- Reuti > > > > > > > > > > when I ran mpirun from a machine and checking the memory status for the > > > > three machines that I have it appear that the memory usage increased > > > > just in the same machine. > > > > > > > > Regards, > > > > > > > > > From: re...@staff.uni-marburg.de > > > > > Date: Wed, 28 Mar 2012 15:12:17 +0200 > > > > > To: us...@open-mpi.org > > > > > Subject: Re: [OMPI users] Can not run a parallel job on all the nodes > > > > > in the cluster > > > > > > > > > > Hi, > > > > > > > > > > Am 27.03.2012 um 23:46 schrieb Hameed Alzahrani: > > > > > > > > > > > When I run any parallel job I get the answer just from the > > > > > > submitting node > > > > > > > > > > what do you mean by submitting node: you use a queuing system - which > > > > > one? > > > > > > > > > > -- Reuti > > > > > > > > > > > > > > > > even when I tried to benchmark the cluster using LINPACK but it > > > > > > look like the job just working on the submitting node is there a > > > > > > way to make openMPI send the job equally to all the nodes depending > > > > > > on the number of processor in the current mode even if I specify > > > > > > that the job should use 8 processor it look like openMPI use the > > > > > > submitting node 4 processors instead of using the other processors. > > > > > > I tried also --host but it does not work correctly in benchmarking > > > > > > the cluster so does any one use openMPI in benchmarking a cluster > > > > > > or does any one knows how to make openMPI divids the parallel job > > > > > > equally to every processor on the cluster. > > > > > > > > > > > > Regards, > > > > > > _______________________________________________ > > > > > > users mailing list > > > > > > us...@open-mpi.org > > > > > > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > > > > > > > > > > > > > _______________________________________________ > > > > > users mailing list > > > > > us...@open-mpi.org > > > > > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > > _______________________________________________ > > > > users mailing list > > > > us...@open-mpi.org > > > > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > > > > > > > _______________________________________________ > > > users mailing list > > > us...@open-mpi.org > > > http://www.open-mpi.org/mailman/listinfo.cgi/users > > _______________________________________________ > > users mailing list > > us...@open-mpi.org > > http://www.open-mpi.org/mailman/listinfo.cgi/users > > > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users