You need to tell mpirun that your system doesn't have homogeneous nodes: --hetero-nodes Nodes in cluster may differ in topology, so send the topology back from each node [Default = false]
On Aug 22, 2013, at 2:48 PM, Noah Knowles <nknow...@usgs.gov> wrote: > Hi, newb here, so sorry if this is a dumb question but I haven't found an > answer. I am running OpenMPI 1.7.2 on a small Rocks 6.1, Bladecenter H > cluster. I am using the bind-to-socket option on nodes with different numbers > of cores per socket. For the sample output below, compute-0-2 has two 6-core > sockets and compute-0-3 has two 8-core sockets. > > [1,4]<stderr>:[compute-0-2.local:03268] MCW rank 4 bound to socket 0[core > 0[hwt 0]], socket 0[core 1[hwt 0]], socket 0[core 2[hwt 0]], socket 0[core > 3[hwt 0]], socket 0[core 4[hwt 0]], socket 0[core 5[hwt 0]]: > [B/B/B/B/B/B][./././././.] > [1,5]<stderr>:[compute-0-2.local:03268] MCW rank 5 bound to socket 1[core > 6[hwt 0]], socket 1[core 7[hwt 0]], socket 1[core 8[hwt 0]], socket 1[core > 9[hwt 0]], socket 1[core 10[hwt 0]], socket 1[core 11[hwt 0]]: > [./././././.][B/B/B/B/B/B] > [1,6]<stderr>:[compute-0-3.local:03816] MCW rank 6 bound to socket 0[core > 0[hwt 0]], socket 0[core 1[hwt 0]], socket 0[core 2[hwt 0]], socket 0[core > 3[hwt 0]], socket 0[core 4[hwt 0]], socket 0[core 5[hwt 0]]: > [B/B/B/B/B/B/./.][./././././././.] > [1,7]<stderr>:[compute-0-3.local:03816] MCW rank 7 bound to socket 0[core > 6[hwt 0]], socket 0[core 7[hwt 0]], socket 1[core 8[hwt 0]], socket 1[core > 9[hwt 0]], socket 1[core 10[hwt 0]], socket 1[core 11[hwt 0]]: > [././././././B/B][B/B/B/B/./././.] > > Is this behavior intended? Is there any way to cause bind-to-socket to use > all cores on a socket for the 6-core AND the 8-core nodes? Or at least to > have that last binding not spread across cores on two sockets? > I've tried a rankfile too, but had errors-- that should probably be a > separate thread though. > > Thanks, > Noah > _______________________________________________ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users