What version are you using?

On Aug 16, 2011, at 3:19 AM, Simone Pellegrini wrote:

> Dear all,
> I am developing a system to manage MPI tasks on top of MPI. The architecture 
> is rather simple, I have a set of scheduler processes which takes care to 
> manage the resources of a node. The idea is to have 1 (or more) of those 
> scheduler allocated on each node of a cluster and then create new MPI 
> processes (on demand) as computation is needed. Allocation of processes is 
> done using MPI_Spawn.
> 
> The system now works fine on a single node by allocating the main scheduler 
> using the following mpi command:
> mpirun --np 1 ./scheduler ...
> 
> Now when I scale to multiple nodes problems with default MPI behaviour 
> starts. For example lets assume I have 2 nodes with 8 cpu cores each. I 
> therefore set up a machine file in the following way:
> 
> s01 slots=1
> s02 slots=1
> 
> and start the node schedulers in the following way:
> mpirun --np 2 --machinefile machinefile ./scheduler ...
> 
> This allocates the processes correctly, now the problem starts when I invoke 
> MPI_Spawn. basically MPI spawn also uses the informations from the 
> machinefile and if 4 MPI processes are spawned 2 are allocated in s01 and 2 
> on s02. What I want is to allocate the processes always in the same node.
> 
> I tried to do this by specifying an MPI_Info object which is then passed to 
> the MPI_Spawn routine. I tried to set the "host" property to the hostname of 
> the machine where the scheduler is running but this didn't help.
> 
> Unfortunately there is very little documentation on this.
> 
> Thanks for the help,
> Simone
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users


Reply via email to