That should work, then. When you set the "host" property, did you give the same 
name as was in your machine file?

Debug options that might help:

-mca plm_base_verbose 5 -mca rmaps_base_verbose 5

You'll need to configure --enable-debug to get the output, but that should help 
tell us what is happening.


On Aug 16, 2011, at 5:09 AM, Simone Pellegrini wrote:

> On 08/16/2011 12:30 PM, Ralph Castain wrote:
>> What version are you using?
> 
> OpenMPI 1.4.3
> 
>> 
>> 
>> On Aug 16, 2011, at 3:19 AM, Simone Pellegrini wrote:
>> 
>>> Dear all,
>>> I am developing a system to manage MPI tasks on top of MPI. The 
>>> architecture is rather simple, I have a set of scheduler processes which 
>>> takes care to manage the resources of a node. The idea is to have 1 (or 
>>> more) of those scheduler allocated on each node of a cluster and then 
>>> create new MPI processes (on demand) as computation is needed. Allocation 
>>> of processes is done using MPI_Spawn.
>>> 
>>> The system now works fine on a single node by allocating the main scheduler 
>>> using the following mpi command:
>>> mpirun --np 1 ./scheduler ...
>>> 
>>> Now when I scale to multiple nodes problems with default MPI behaviour 
>>> starts. For example lets assume I have 2 nodes with 8 cpu cores each. I 
>>> therefore set up a machine file in the following way:
>>> 
>>> s01 slots=1
>>> s02 slots=1
>>> 
>>> and start the node schedulers in the following way:
>>> mpirun --np 2 --machinefile machinefile ./scheduler ...
>>> 
>>> This allocates the processes correctly, now the problem starts when I 
>>> invoke MPI_Spawn. basically MPI spawn also uses the informations from the 
>>> machinefile and if 4 MPI processes are spawned 2 are allocated in s01 and 2 
>>> on s02. What I want is to allocate the processes always in the same node.
>>> 
>>> I tried to do this by specifying an MPI_Info object which is then passed to 
>>> the MPI_Spawn routine. I tried to set the "host" property to the hostname 
>>> of the machine where the scheduler is running but this didn't help.
>>> 
>>> Unfortunately there is very little documentation on this.
>>> 
>>> Thanks for the help,
>>> Simone
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 


Reply via email to