Ralph,

I figured I should of run an mpi program ...here's what it does (seems
to be by-X-slot style):
---
$ /apps/x86_64/system/mpiexec-0.82/bin/mpiexec -npernode 2 mpi_hello
Hello, I am node an41 with rank 0
Hello, I am node an41 with rank 1
Hello, I am node an39 with rank 4
Hello, I am node an40 with rank 2
Hello, I am node an38 with rank 6
Hello, I am node an39 with rank 5
Hello, I am node an38 with rank 7
Hello, I am node an40 with rank 3
---

What you describe makes sense to me.  Thanks!

> -----Original Message-----
> From: users-boun...@open-mpi.org 
> [mailto:users-boun...@open-mpi.org] On Behalf Of Ralph Castain
> Sent: Monday, December 11, 2006 10:27 PM
> To: Open MPI Users
> Subject: Re: [OMPI users] Pernode request
> 
> Hi Chris
> 
> I have also implemented --npernode N now as well - it is in 
> the trunk as of
> r12826.
> 
> The testing you show below using mpiexec really doesn't tell 
> us the story -
> we need to know the rank of the various processes (and unfortunately,
> hostname just tells us the host). There is no way to tell the 
> rank from just
> the order in which the host names are printed to the screen. 
> I have a test
> program in our distribution (see 
> orte/test/mpi/hello_hostname.c) that will
> output both the rank and hostname - it would give us the 
> required info.
> 
> Regardless, I think it would make sense to provide the flexibility you
> describe. What if we selected this "by-X-slot" style by using 
> the --npernode
> option, and allowing the user to combine it with the existing 
> "--byslot"
> option? This would still launch N procs/node, but with the 
> ranking done by
> slot. If the user doesn't specify "byslot", then we default 
> to assigning
> ranks by node.
> 
> Make sense? If so, I can probably have that going before the holiday.
> 
> Ralph
> 
> 
> 
> On 12/11/06 7:51 PM, "Maestas, Christopher Daniel" 
> <cdma...@sandia.gov>
> wrote:
> 
> > Hello Ralph,
> > 
> > This is great news!  Thanks for doing this.  I will try and 
> get around
> > to it soon before the holiday break.
> > 
> > The allocation scheme always seems to get to me.  From what 
> you describe
> > that is how I would have seen it.  As I've gotten to know 
> osc mpiexec
> > through the years I think they like to do a first fit 
> approach, but now
> > that I test it I think the feature needs more testing or 
> I'm not testing
> > appropriately :-)
> > ---
> > $ /apps/x86_64/system/mpiexec-0.82/bin/mpiexec -comm=none 
> -npernode 2
> > grep HOSTNAME /etc/sysconfig/network
> > HOSTNAME="an56"
> > HOSTNAME="an56"
> > HOSTNAME="an55"
> > HOSTNAME="an53"
> > HOSTNAME="an54"
> > HOSTNAME="an55"
> > HOSTNAME="an53"
> > HOSTNAME="an54"
> > ---
> > 
> > I guess I would wonder if it would be possible to switch 
> from the method
> > what you suggest and also allow a "by-X-slot" style of 
> launch where you
> > would see for npernode = X and N nodes:
> >  proc1 - node1
> >  proc2 - node1
> >  ...
> >  proc(X*1) - node1
> > ...
> >  proc(X+1) - node2
> >  proc(X+2) - node2
> > ...
> >  proc(X*2) - node2
> > ...
> >  proc(N*X-(X-0)) - nodeN
> >  proc(N*X-(X-1)) - nodeN
> >  ...
> >  proc(X*N-1) - nodeN
> >  proc(X*N) - nodeN
> > 
> > I think that's how to best describe it.  Basically you load 
> until there
> > are X processes on each node before moving to the next.  
> This may prove
> > to be more challenging, and I can understand if it would 
> not be deemed
> > "worthy." :-)
> > 
> > -cdm 
> > 
> >> -----Original Message-----
> >> From: users-boun...@open-mpi.org
> >> [mailto:users-boun...@open-mpi.org] On Behalf Of Ralph Castain
> >> Sent: Monday, December 11, 2006 5:41 PM
> >> To: Open MPI Users
> >> Subject: Re: [OMPI users] Pernode request
> >> 
> >> Hi Chris
> >> 
> >> Okay, we have modified the pernode behavior as you requested
> >> (on the trunk
> >> as of r12821)- give it a shot and see if that does it. I have
> >> not yet added
> >> the npernode option, but hope to get that soon.
> >> 
> >> I have a question for you about the npernode option. I am
> >> assuming that you
> >> want n procs/node, but that you want it mapped by NODE. For
> >> example, proc 0
> >> goes on the first node, proc 1 goes on the second node, etc.
> >> until I get one
> >> on each node; then I wrap back to the beginning and do this
> >> again until I
> >> get the specified number of procs on each node.
> >> 
> >> Correct?
> >> Ralph
> >> 
> >>> Ralph,
> >> 
> >>> I agree with what you stated in points 1-4. That is what we
> >> are looking
> >>> for.
> >>> I understand your point now about the non-MPI users too. :-)
> >>> 
> >>> Thanks,
> >>> -cdm
> >> 
> >>>> -----Original Message-----
> >>>> From: users-bounces_at_[hidden]
> >> [mailto:users-bounces_at_[hidden]] On
> >>>> Behalf Of Ralph Castain
> >>>> Sent: Wednesday, November 29, 2006 8:01 AM
> >>>> To: Open MPI Users
> >>>> Subject: Re: [OMPI users] Pernode request
> >>>> 
> >>>> Hi Chris
> >>>> 
> >>>> Thanks for the patience and the clarification - much 
> appreciated. In
> >>>> fact, I have someone that needs to learn more about the
> >> code base, so I
> >>>> think I will assign this to him. At the least, he will have
> >> to learn a
> >>>> lot more about the mapper!
> >>>> 
> >>>> I have no problem with modifying the pernode behavior to
> >> deal with the
> >>>> case of someone specifying -np as you describe. It would be
> >> relatively
> >>>> easy to check. As I understand it, you want the behavior to be:
> >>>> 
> >>>> 1. if no -np is specified, launch one proc/node across
> >> entire allocation
> >>>> 
> >>>> 2. if -np n is specified AND n is less than the number 
> of allocated
> >>>> nodes, then launch one proc/node up to the specified
> >> number. Of course,
> >>>> this is identical to just doing -np n -bynode, but that's
> >> immaterial.
> >>>> ;-)
> >>>> 
> >>>> 3. if -np n is specified AND n is greater than the number
> >> of allocated
> >>>> nodes, error message and exit
> >>>> 
> >>>> 4. add a -npernode n option that launches n procs/node,
> >> subject to the
> >>>> same tests above.
> >>>> 
> >>>> Can you confirm?
> >>>> 
> >>>> Finally, I think you misunderstood my comment about the MPI
> >> folks. Our
> >>>> non-MPI users couldn't care less about commonality of 
> command line
> >>>> arguments across MPI implementations. Hence, I leave issues
> >> in that area
> >>>> to the MPI members of the team - they are the ones that
> >> decide how to
> >>>> deal with the myriad of different option syntaxes in the
> >> MPI community.
> >>>> 
> >>>> Gives me too much of a headache! :-)
> >>>> 
> >>>> Ralph 
> >> 
> >> 
> >> _______________________________________________
> >> users mailing list
> >> us...@open-mpi.org
> >> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >> 
> > 
> > 
> > _______________________________________________
> > users mailing list
> > us...@open-mpi.org
> > http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 


Reply via email to