Hmmm...yes, I see what you mean. Try using "-default-hostfile mpimachines" 
instead of -hostfiles. I think that's the problem - at least, it fixed it for 
me.


On Apr 17, 2013, at 7:08 AM, "Welder, Wallace T         PWR" 
<wallace.wel...@pwr.utc.com> wrote:

> I’ve recently tried running OpenMPI v1.6.4 on multiple nodes and have noticed 
> a change in behavior that I don’t understand.  In OpenMPI version 1.4.x, 
> 1.5.x and 1.6.1, I could run a job spanning two nodes as shown below.  The 
> procedure results in 8 processes running on the first node and 8 on the 
> second node.
> 
> mpirun –hostfile mpimachines –n 1 host.exe : -n 15 node.exe
> 
> 
> where the file mpimachines looks  like:
> 
> node1 slots=8
> 
> node2 slots=8
> 
> In OpenMPI v1.6.2 and v1.6.4 (haven’t tried v1.6.3) when I try to run the 
> same way, all the processes start on node1 and none start on node2.  I’ve 
> noticed there are now runtime flags -bynode and –byslot, but  I haven’t had 
> any success with those.  I’ve also tried changing the mpimachines file to 
> look like:
> 
> node1 slots=8 max-slots=8
> 
> node2 slots=8 max-slots=8
> 
> When I tried this, I got a runtime error saying there were not enough slots 
> in the system to satisfy the 15 slots that were requested by the application 
> node.exe.  I think there is a hint in here about my problem, but I haven’t 
> been able to figure out what it is yet.
> 
> Can anyone let me know how the process has changed with these newer versions 
> of OpenMPI?
> 
> Thanks,
> 
> Wallace
> 
> 
> 
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users

Reply via email to