Woohoo!!! Great to hear!

If you want the best performance, then yes - you want the NUMA packages 
installed. Reason: without them, we can’t bind the memory to the same location 
as the process, and so you will likely see degraded performance.

Relates to our other email thread earlier today.


> On Jul 24, 2015, at 5:39 PM, Lane, William <william.l...@cshs.org> wrote:
> 
> Ralph,
> 
> I re-compiled my test code and ran it on almost our entire cluster (excepting
> the new hyperthread capable blade chassis), anyway, the code ran successfully:
> 
> mpirun -np 185 -display-devel-map --prefix /hpc/apps/mpi/openmpi/1.8.7/ 
> --hostfile hostfile-noslots --mca btl ^sm --hetero-nodes --bind-to core 
> /hpc/home/lanew/mpi/openmpi/ProcessColors3  >> out.txt 2>&1
> 
> I've attached the out.txt file in case you're interested in looking at it.
> 
> One question I have is WRT the warning:
> 
> --------------------------------------------------------------------------
> WARNING: a request was made to bind a process. While the system
> supports binding the process itself, at least one node does NOT
> support binding memory to the process location.
> 
>   Node:  csclprd3-6-0
> 
> This usually is due to not having the required NUMA support installed
> on the node. In some Linux distributions, the required support is
> contained in the libnumactl and libnumactl-devel packages.
> This is a warning only; your job will continue, though performance may be 
> degraded.
> --------------------------------------------------------------------------
> 
> Is it worthwhile to make sure the proper NUMA packages are
> installed on compute nodes?
> 
> Thank you for all your help w/this Ralph, now I can move on and
> get the Linpack benchmark running.
> 
> -Bill L.
> 
> 
> From: users [users-boun...@open-mpi.org <mailto:users-boun...@open-mpi.org>] 
> on behalf of Lane, William [william.l...@cshs.org 
> <mailto:william.l...@cshs.org>]
> Sent: Thursday, July 16, 2015 5:11 PM
> To: Open MPI Users
> Subject: Re: [OMPI users] OpenMPI 1.8.6, CentOS 6.3, too many slots = crash
> 
> Will do Ralph. I'll re-compile my test code against OpenMPI 1.8.7 tomorrow
> and run it again.
> 
> There are apparently now four types of systems in our cluster compute
> ring:
> two racks of older blades (HS21's), 
> a rack of newer blades (HS22's),
> IBM x3550-M3's
> and IBM x3550-M4's. 
> 
> The only systems that don't support hyperthreading are the older blades 
> (the HS21's).
> 
> The submit nodes to the cluster are Sun Sunfire X2100's (which are old
> dual core AMD Opterons), but I never use them in my compute ring.
> 
> -Bill
> 
> From: users [users-boun...@open-mpi.org <mailto:users-boun...@open-mpi.org>] 
> on behalf of Ralph Castain [r...@open-mpi.org <mailto:r...@open-mpi.org>]
> Sent: Wednesday, July 15, 2015 10:08 PM
> To: Open MPI Users
> Subject: Re: [OMPI users] OpenMPI 1.8.6, CentOS 6.3, too many slots = crash
> 
> Stable 1.8.7 has been released - please let me know if the problem is 
> resolved.
> 
> 
>> On Jul 15, 2015, at 9:53 PM, Lane, William <william.l...@cshs.org 
>> <mailto:william.l...@cshs.org>> wrote:
>> 
>> Ralph,
>> 
>> I'd rather wait for the stable release of 1.8.7, but I'm willing to give
>> it a try if my supervisor is.
>> 
>> -Bill L.
>> 
>> From: users [users-boun...@open-mpi.org <mailto:users-boun...@open-mpi.org>] 
>> on behalf of Ralph Castain [r...@open-mpi.org <mailto:r...@open-mpi.org>]
>> Sent: Tuesday, July 14, 2015 12:47 PM
>> To: Open MPI Users
>> Subject: Re: [OMPI users] OpenMPI 1.8.6, CentOS 6.3, too many slots = crash
>> 
>> Can you give it a try? I’m skeptical, but it might work. The rc is out on 
>> the web site:
>> 
>> http://www.open-mpi.org/software/ompi/v1.8/ 
>> <http://www.open-mpi.org/software/ompi/v1.8/>
>> 
>> 
>>> On Jul 14, 2015, at 11:17 AM, Lane, William <william.l...@cshs.org 
>>> <mailto:william.l...@cshs.org>> wrote:
>>> 
>>> Ralph,
>>> 
>>> Do you think the 1.8.7 release will solve the problems w/our
>>> heterogeneous cluster?
>>> 
>>> Bill L.
>>> 
>>> From: users [users-boun...@open-mpi.org 
>>> <mailto:users-boun...@open-mpi.org>] on behalf of Ralph Castain 
>>> [r...@open-mpi.org <mailto:r...@open-mpi.org>]
>>> Sent: Tuesday, July 07, 2015 8:59 PM
>>> To: Open MPI Users
>>> Subject: Re: [OMPI users] OpenMPI 1.8.6, CentOS 6.3, too many slots = crash
>>> 
>>> No need for the lstopo data anymore, Bill - I was able to recreate the 
>>> situation using some very nice hwloc functions plus your prior 
>>> descriptions. I'm not totally confident that this fix will resolve the 
>>> problem but it will clear out at least one problem.
>>> 
>>> We'll just have to see what happens and attack it next.
>>> Ralph
>>> 
>>> 
>>> On Tue, Jul 7, 2015 at 8:07 PM, Lane, William <william.l...@cshs.org 
>>> <mailto:william.l...@cshs.org>> wrote:
>>> I'm sorry I haven't been able to get the lstopo information for
>>> all the nodes, but I had to get the latest version of hwloc installed
>>> first. They've even added in some more modern blades that also
>>> support hyperthreading, ugh. They've also been doing some memory
>>> upgrades as well.
>>> 
>>>  I'm trying to get a Bash script running on the cluster via qsub
>>> that will run lstopo and output the host information to a file located
>>> in my $HOME directory but it hasn't been working (there are 60 nodes
>>> in the heterogeneous cluster that needs to have OpenMPI running).
>>> 
>>> I will try to get the lstopo information by the end of the week.
>>> 
>>> I'd be willing to do most anything to get these OpenMPI issues
>>> resolved. I'd even wash your cars for you!
>>> 
>>> -Bill L.
>>> ________________________________________
>>> From: users [users-boun...@open-mpi.org 
>>> <mailto:users-boun...@open-mpi.org>] on behalf of Ralph Castain 
>>> [r...@open-mpi.org <mailto:r...@open-mpi.org>]
>>> Sent: Tuesday, July 07, 2015 1:36 PM
>>> To: Open MPI Users
>>> Subject: Re: [OMPI users] OpenMPI 1.8.6, CentOS 6.3, too many slots = crash
>>> 
>>> I may have finally tracked this down. At least, I can now get the correct 
>>> devel map to come out, and found a memory corruption issue that only 
>>> impacted hetero operations. I can’t know if this is the root cause of the 
>>> problem Bill is seeing, however, as I have no way of actually running the 
>>> job.
>>> 
>>> I pushed this into the master and will bring it back to 1.8.7 as well as 
>>> 1.10.
>>> 
>>> Bill - would you be able/willing to give it a try there? It would be nice 
>>> to confirm this actually fixed the problem.
>>> 
>>> 
>>> > On Jun 29, 2015, at 1:58 PM, Jeff Squyres (jsquyres) <jsquy...@cisco.com 
>>> > <mailto:jsquy...@cisco.com>> wrote:
>>> >
>>> > lstopo will tell you -- if there is more than one "PU" (hwloc terminology 
>>> > for "processing unit") per core, then hyper threading is enabled.  If 
>>> > there's only one PU per core, then hyper threading is disabled.
>>> >
>>> >
>>> >> On Jun 29, 2015, at 4:42 PM, Lane, William <william.l...@cshs.org 
>>> >> <mailto:william.l...@cshs.org>> wrote:
>>> >>
>>> >> Would the output of dmidecode -t processor and/or lstopo tell me 
>>> >> conclusively
>>> >> if hyperthreading is enabled or not? Hyperthreading is supposed to be 
>>> >> enabled
>>> >> for all the IBM x3550 M3 and M4 nodes, but I'm not sure if it actually 
>>> >> is and I
>>> >> don't have access to the BIOS settings.
>>> >>
>>> >> -Bill L.
>>> >>
>>> >> From: users [users-boun...@open-mpi.org 
>>> >> <mailto:users-boun...@open-mpi.org>] on behalf of Ralph Castain 
>>> >> [r...@open-mpi.org <mailto:r...@open-mpi.org>]
>>> >> Sent: Saturday, June 27, 2015 7:21 PM
>>> >> To: Open MPI Users
>>> >> Subject: Re: [OMPI users] OpenMPI 1.8.6, CentOS 6.3, too many slots = 
>>> >> crash
>>> >>
>>> >> Bill - this is such a jumbled collection of machines that I’m having 
>>> >> trouble figuring out what I should replicate. I can create something 
>>> >> artificial here so I can try to debug this, but I need to know exactly 
>>> >> what I’m up against - can you tell me:
>>> >>
>>> >> * the architecture of each type - how many sockets, how many 
>>> >> cores/socket, HT on or off. If two nodes have the same physical setup 
>>> >> but one has HT on and the other off, then please consider those two 
>>> >> different types
>>> >>
>>> >> * how many nodes of each type
>>> >>
>>> >> Looking at your map output, it looks like the map is being done 
>>> >> correctly, but somehow the binding locale isn’t getting set in some 
>>> >> cases. You latest error output would seem out-of-step with your prior 
>>> >> reports, so something else may be going on there. As I said earlier, 
>>> >> this is the most hetero environment we’ve seen, and so there may be some 
>>> >> code paths your hitting that haven’t been well exercised before.
>>> >>
>>> >>
>>> >>
>>> >>
>>> >>> On Jun 26, 2015, at 5:22 PM, Lane, William <william.l...@cshs.org 
>>> >>> <mailto:william.l...@cshs.org>> wrote:
>>> >>>
>>> >>> Well, I managed to get a successful mpirun @ a slot count of 132 using 
>>> >>> --mca btl ^sm,
>>> >>> however when I increased the slot count to 160, mpirun crashed without 
>>> >>> any error
>>> >>> output:
>>> >>>
>>> >>> mpirun -np 160 -display-devel-map --prefix /hpc/apps/mpi/openmpi/1.8.6/ 
>>> >>> --hostfile hostfile-noslots --mca btl ^sm --hetero-nodes --bind-to core 
>>> >>> /hpc/home/lanew/mpi/openmpi/ProcessColors3 >> out.txt 2>&1
>>> >>>
>>> >>> --------------------------------------------------------------------------
>>> >>> WARNING: a request was made to bind a process. While the system
>>> >>> supports binding the process itself, at least one node does NOT
>>> >>> support binding memory to the process location.
>>> >>>
>>> >>>  Node:  csclprd3-6-1
>>> >>>
>>> >>> This usually is due to not having the required NUMA support installed
>>> >>> on the node. In some Linux distributions, the required support is
>>> >>> contained in the libnumactl and libnumactl-devel packages.
>>> >>> This is a warning only; your job will continue, though performance may 
>>> >>> be degraded.
>>> >>> --------------------------------------------------------------------------
>>> >>> --------------------------------------------------------------------------
>>> >>> A request was made to bind to that would result in binding more
>>> >>> processes than cpus on a resource:
>>> >>>
>>> >>>   Bind to:     CORE
>>> >>>   Node:        csclprd3-6-1
>>> >>>   #processes:  2
>>> >>>   #cpus:       1
>>> >>>
>>> >>> You can override this protection by adding the "overload-allowed"
>>> >>> option to your binding directive.
>>> >>> --------------------------------------------------------------------------
>>> >>>
>>> >>> But csclprd3-6-1 (a blade) does have 2 CPU's on 2 separate sockets w/2 
>>> >>> cores apiece as shown in my dmidecode output:
>>> >>>
>>> >>>    csclprd3-6-1 ~]# dmidecode -t processor
>>> >>>    # dmidecode 2.11
>>> >>>    SMBIOS 2.4 present.
>>> >>>
>>> >>>    Handle 0x0008, DMI type 4, 32 bytes
>>> >>>    Processor Information
>>> >>>            Socket Designation: Socket 1 CPU 1
>>> >>>            Type: Central Processor
>>> >>>            Family: Xeon
>>> >>>            Manufacturer: GenuineIntel
>>> >>>            ID: F6 06 00 00 01 03 00 00
>>> >>>            Signature: Type 0, Family 6, Model 15, Stepping 6
>>> >>>            Flags:
>>> >>>                    FPU (Floating-point unit on-chip)
>>> >>>                    CX8 (CMPXCHG8 instruction supported)
>>> >>>                    APIC (On-chip APIC hardware supported)
>>> >>>            Version: Intel Xeon
>>> >>>            Voltage: 2.9 V
>>> >>>            External Clock: 333 MHz
>>> >>>            Max Speed: 4000 MHz
>>> >>>            Current Speed: 3000 MHz
>>> >>>            Status: Populated, Enabled
>>> >>>            Upgrade: ZIF Socket
>>> >>>            L1 Cache Handle: 0x0004
>>> >>>            L2 Cache Handle: 0x0005
>>> >>>            L3 Cache Handle: Not Provided
>>> >>>
>>> >>>    Handle 0x0009, DMI type 4, 32 bytes
>>> >>>    Processor Information
>>> >>>            Socket Designation: Socket 2 CPU 2
>>> >>>            Type: Central Processor
>>> >>>            Family: Xeon
>>> >>>            Manufacturer: GenuineIntel
>>> >>>            ID: F6 06 00 00 01 03 00 00
>>> >>>            Signature: Type 0, Family 6, Model 15, Stepping 6
>>> >>>            Flags:
>>> >>>                    FPU (Floating-point unit on-chip)
>>> >>>                    CX8 (CMPXCHG8 instruction supported)
>>> >>>                    APIC (On-chip APIC hardware supported)
>>> >>>            Version: Intel Xeon
>>> >>>            Voltage: 2.9 V
>>> >>>            External Clock: 333 MHz
>>> >>>            Max Speed: 4000 MHz
>>> >>>            Current Speed: 3000 MHz
>>> >>>            Status: Populated, Enabled
>>> >>>            Upgrade: ZIF Socket
>>> >>>            L1 Cache Handle: 0x0006
>>> >>>            L2 Cache Handle: 0x0007
>>> >>>            L3 Cache Handle: Not Provided
>>> >>>
>>> >>>    csclprd3-6-1 ~]# lstopo
>>> >>>    Machine (16GB)
>>> >>>      Socket L#0 + L2 L#0 (4096KB)
>>> >>>        L1d L#0 (32KB) + L1i L#0 (32KB) + Core L#0 + PU L#0 (P#0)
>>> >>>        L1d L#1 (32KB) + L1i L#1 (32KB) + Core L#1 + PU L#1 (P#2)
>>> >>>      Socket L#1 + L2 L#1 (4096KB)
>>> >>>        L1d L#2 (32KB) + L1i L#2 (32KB) + Core L#2 + PU L#2 (P#1)
>>> >>>        L1d L#3 (32KB) + L1i L#3 (32KB) + Core L#3 + PU L#3 (P#3)
>>> >>>
>>> >>> csclprd3-0-1 information (which looks correct as this particular x3550 
>>> >>> should
>>> >>> have one socket populated (of two) with a 6 core Xeon (or 12 cores 
>>> >>> w/hyperthreading
>>> >>> turned on):
>>> >>>
>>> >>>    csclprd3-0-1 ~]# lstopo
>>> >>>    Machine (71GB)
>>> >>>      Socket L#0 + L3 L#0 (12MB)
>>> >>>        L2 L#0 (256KB) + L1d L#0 (32KB) + L1i L#0 (32KB) + Core L#0 + PU 
>>> >>> L#0 (P#0)
>>> >>>        L2 L#1 (256KB) + L1d L#1 (32KB) + L1i L#1 (32KB) + Core L#1 + PU 
>>> >>> L#1 (P#1)
>>> >>>        L2 L#2 (256KB) + L1d L#2 (32KB) + L1i L#2 (32KB) + Core L#2 + PU 
>>> >>> L#2 (P#2)
>>> >>>        L2 L#3 (256KB) + L1d L#3 (32KB) + L1i L#3 (32KB) + Core L#3 + PU 
>>> >>> L#3 (P#3)
>>> >>>        L2 L#4 (256KB) + L1d L#4 (32KB) + L1i L#4 (32KB) + Core L#4 + PU 
>>> >>> L#4 (P#4)
>>> >>>        L2 L#5 (256KB) + L1d L#5 (32KB) + L1i L#5 (32KB) + Core L#5 + PU 
>>> >>> L#5 (P#5)
>>> >>>
>>> >>>    csclprd3-0-1 ~]# dmidecode -t processor
>>> >>>    # dmidecode 2.11
>>> >>>    # SMBIOS entry point at 0x7f6be000
>>> >>>    SMBIOS 2.5 present.
>>> >>>
>>> >>>    Handle 0x0001, DMI type 4, 40 bytes
>>> >>>    Processor Information
>>> >>>            Socket Designation: Node 1 Socket 1
>>> >>>            Type: Central Processor
>>> >>>            Family: Xeon MP
>>> >>>            Manufacturer: Intel(R) Corporation
>>> >>>            ID: C2 06 02 00 FF FB EB BF
>>> >>>            Signature: Type 0, Family 6, Model 44, Stepping 2
>>> >>>            Flags:
>>> >>>                    FPU (Floating-point unit on-chip)
>>> >>>                    VME (Virtual mode extension)
>>> >>>                    DE (Debugging extension)
>>> >>>                    PSE (Page size extension)
>>> >>>                    TSC (Time stamp counter)
>>> >>>                    MSR (Model specific registers)
>>> >>>                    PAE (Physical address extension)
>>> >>>                    MCE (Machine check exception)
>>> >>>                    CX8 (CMPXCHG8 instruction supported)
>>> >>>                    APIC (On-chip APIC hardware supported)
>>> >>>                    SEP (Fast system call)
>>> >>>                    MTRR (Memory type range registers)
>>> >>>                    PGE (Page global enable)
>>> >>>                    MCA (Machine check architecture)
>>> >>>                    CMOV (Conditional move instruction supported)
>>> >>>                    PAT (Page attribute table)
>>> >>>                    PSE-36 (36-bit page size extension)
>>> >>>                    CLFSH (CLFLUSH instruction supported)
>>> >>>                    DS (Debug store)
>>> >>>                    ACPI (ACPI supported)
>>> >>>                    MMX (MMX technology supported)
>>> >>>                    FXSR (FXSAVE and FXSTOR instructions supported)
>>> >>>                    SSE (Streaming SIMD extensions)
>>> >>>                    SSE2 (Streaming SIMD extensions 2)
>>> >>>                    SS (Self-snoop)
>>> >>>                    HTT (Multi-threading)
>>> >>>                    TM (Thermal monitor supported)
>>> >>>                    PBE (Pending break enabled)
>>> >>>            Version: Intel(R) Xeon(R) CPU           E5645  @ 2.40GHz
>>> >>>            Voltage: 1.2 V
>>> >>>            External Clock: 5866 MHz
>>> >>>            Max Speed: 4400 MHz
>>> >>>            Current Speed: 2400 MHz
>>> >>>            Status: Populated, Enabled
>>> >>>            Upgrade: ZIF Socket
>>> >>>            L1 Cache Handle: 0x0002
>>> >>>            L2 Cache Handle: 0x0003
>>> >>>            L3 Cache Handle: 0x0004
>>> >>>            Serial Number: Not Specified
>>> >>>            Asset Tag: Not Specified
>>> >>>            Part Number: Not Specified
>>> >>>            Core Count: 6
>>> >>>            Core Enabled: 6
>>> >>>            Thread Count: 6
>>> >>>            Characteristics:
>>> >>>                    64-bit capable
>>> >>>
>>> >>>    Handle 0x005A, DMI type 4, 40 bytes
>>> >>>    Processor Information
>>> >>>            Socket Designation: Node 1 Socket 2
>>> >>>            Type: Central Processor
>>> >>>            Family: Xeon MP
>>> >>>            Manufacturer: Not Specified
>>> >>>            ID: 00 00 00 00 00 00 00 00
>>> >>>            Signature: Type 0, Family 0, Model 0, Stepping 0
>>> >>>            Flags: None
>>> >>>            Version: Not Specified
>>> >>>            Voltage: 1.2 V
>>> >>>            External Clock: 5866 MHz
>>> >>>            Max Speed: 4400 MHz
>>> >>>            Current Speed: Unknown
>>> >>>            Status: Unpopulated
>>> >>>            Upgrade: ZIF Socket
>>> >>>            L1 Cache Handle: Not Provided
>>> >>>            L2 Cache Handle: Not Provided
>>> >>>            L3 Cache Handle: Not Provided
>>> >>>            Serial Number: Not Specified
>>> >>>            Asset Tag: Not Specified
>>> >>>            Part Number: Not Specified
>>> >>>            Characteristics: None
>>> >>>
>>> >>>
>>> >>> From: users [users-boun...@open-mpi.org 
>>> >>> <mailto:users-boun...@open-mpi.org>] on behalf of Ralph Castain 
>>> >>> [r...@open-mpi.org <mailto:r...@open-mpi.org>]
>>> >>> Sent: Wednesday, June 24, 2015 6:06 AM
>>> >>> To: Open MPI Users
>>> >>> Subject: Re: [OMPI users] OpenMPI 1.8.6, CentOS 6.3, too many slots = 
>>> >>> crash
>>> >>>
>>> >>> I think trying with --mca btl ^sm makes a lot of sense and may solve 
>>> >>> the problem. I also noted that we are having trouble with the topology 
>>> >>> of several of the nodes - seeing only one socket, non-HT where you say 
>>> >>> we should see two sockets and HT-enabled. In those cases, the locality 
>>> >>> is "unknown" - given that those procs are on remote nodes from the one 
>>> >>> being impacted, I don't think it should cause a problem. However, it 
>>> >>> isn't correct, and that raises flags.
>>> >>>
>>> >>> My best guess of the root cause of that error is either we are getting 
>>> >>> bad topology info on those nodes, or we have a bug that is mishandling 
>>> >>> this scenario. It would probably be good to get this error fixed to 
>>> >>> ensure it isn't the source of the eventual crash, even though I'm not 
>>> >>> sure they are related.
>>> >>>
>>> >>> Bill: Can we examine one of the problem nodes? Let's pick csclprd3-0-1 
>>> >>> (or take another one from your list - just look for one where 
>>> >>> "locality" is reported as "unknown" for the procs in the output map). 
>>> >>> Can you run lstopo on that node and send us the output? In the above 
>>> >>> map, it is reporting a single socket with 6 cores, non-HT. Is that what 
>>> >>> lstopo shows when run on the node? Is it what you expected?
>>> >>>
>>> >>>
>>> >>> On Wed, Jun 24, 2015 at 4:07 AM, Gilles Gouaillardet 
>>> >>> <gilles.gouaillar...@gmail.com <mailto:gilles.gouaillar...@gmail.com>> 
>>> >>> wrote:
>>> >>> Bill,
>>> >>>
>>> >>> were you able to get a core file and analyze the stack with gdb ?
>>> >>>
>>> >>> I suspect the error occurs in mca_btl_sm_add_procs but this is just my 
>>> >>> best guess.
>>> >>> if this is correct, can you check the value of 
>>> >>> mca_btl_sm_component.num_smp_procs ?
>>> >>>
>>> >>> as a workaround, can you try
>>> >>> mpirun --mca btl ^sm ...
>>> >>>
>>> >>> I do not see how I can tackle the root cause without being able to 
>>> >>> reproduce the issue :-(
>>> >>>
>>> >>> can you try to reproduce the issue with the smallest hostfile, and then 
>>> >>> run lstopo on all the nodes ?
>>> >>> btw, you are not mixing 32 bits and 64 bits OS, are you ?
>>> >>>
>>> >>> Cheers,
>>> >>>
>>> >>> Gilles
>>> >>>
>>> >>>
>>> >>>
>>> >>>
>>> >>> mca_btl_sm_add_procs(
>>> >>>
>>> >>>
>>> >>>
>>> >>> int
>>> >>>
>>> >>> mca_btl_sm_add_procs
>>> >>> (
>>> >>> On Wednesday, June 24, 2015, Lane, William <william.l...@cshs.org 
>>> >>> <mailto:william.l...@cshs.org>> wrote:
>>> >>> Gilles,
>>> >>>
>>> >>> All the blades only have two core Xeons (without hyperthreading) 
>>> >>> populating both their sockets. All
>>> >>> the x3550 nodes have hyperthreading capable Xeons and Sandybridge 
>>> >>> server CPU's. It's possible
>>> >>> hyperthreading has been disabled on some of these nodes though. The 
>>> >>> 3-0-n nodes are all IBM x3550
>>> >>> nodes while the 3-6-n nodes are all blade nodes.
>>> >>>
>>> >>> I have run this exact same test code successfully in the past on 
>>> >>> another cluster (~200 nodes of Sunfire X2100
>>> >>> 2x dual-core Opterons) w/no issues on upwards of 390 slots. I even 
>>> >>> tested it recently on OpenMPI 1.8.5
>>> >>> on another smaller R&D cluster consisting of 10 Sunfire X2100 nodes 
>>> >>> (w/2 dual core Opterons apiece).
>>> >>> On this particular cluster I've had success running this code on < 132 
>>> >>> slots.
>>> >>>
>>> >>> Anyway, here's the results of the following mpirun:
>>> >>>
>>> >>> mpirun -np 132 -display-devel-map --prefix /hpc/apps/mpi/openmpi/1.8.6/ 
>>> >>> --hostfile hostfile-noslots --mca btl_tcp_if_include eth0 
>>> >>> --hetero-nodes --bind-to core 
>>> >>> /hpc/home/lanew/mpi/openmpi/ProcessColors3 >> out.txt 2>&1
>>> >>>
>>> >>> --------------------------------------------------------------------------
>>> >>> WARNING: a request was made to bind a process. While the system
>>> >>> supports binding the process itself, at least one node does NOT
>>> >>> support binding memory to the process location.
>>> >>>
>>> >>>  Node:  csclprd3-6-1
>>> >>>
>>> >>> This usually is due to not having the required NUMA support installed
>>> >>> on the node. In some Linux distributions, the required support is
>>> >>> contained in the libnumactl and libnumactl-devel packages.
>>> >>> This is a warning only; your job will continue, though performance may 
>>> >>> be degraded.
>>> >>> --------------------------------------------------------------------------
>>> >>> Data for JOB [51718,1] offset 0
>>> >>>
>>> >>> Mapper requested: NULL  Last mapper: round_robin  Mapping policy: 
>>> >>> BYSOCKET  Ranking policy: SLOT
>>> >>> Binding policy: CORE  Cpu set: NULL  PPR: NULL  Cpus-per-rank: 1
>>> >>>     Num new daemons: 0    New daemon starting vpid INVALID
>>> >>>     Num nodes: 15
>>> >>>
>>> >>> Data for node: csclprd3-6-1         Launch id: -1    State: 0
>>> >>>     Daemon: [[51718,0],1]    Daemon launched: True
>>> >>>     Num slots: 4    Slots in use: 4    Oversubscribed: FALSE
>>> >>>     Num slots allocated: 4    Max slots: 0
>>> >>>     Username on node: NULL
>>> >>>     Num procs: 4    Next node_rank: 4
>>> >>>     Data for proc: [[51718,1],0]
>>> >>>         Pid: 0    Local rank: 0    Node rank: 0    App rank: 0
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [B/B][./.]
>>> >>>         Binding: [B/.][./.]
>>> >>>     Data for proc: [[51718,1],1]
>>> >>>         Pid: 0    Local rank: 1    Node rank: 1    App rank: 1
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [./.][B/B]
>>> >>>         Binding: [./.][B/.]
>>> >>>     Data for proc: [[51718,1],2]
>>> >>>         Pid: 0    Local rank: 2    Node rank: 2    App rank: 2
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [B/B][./.]
>>> >>>         Binding: [./B][./.]
>>> >>>     Data for proc: [[51718,1],3]
>>> >>>         Pid: 0    Local rank: 3    Node rank: 3    App rank: 3
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [./.][B/B]
>>> >>>         Binding: [./.][./B]
>>> >>>
>>> >>> Data for node: csclprd3-6-5         Launch id: -1    State: 0
>>> >>>     Daemon: [[51718,0],2]    Daemon launched: True
>>> >>>     Num slots: 4    Slots in use: 4    Oversubscribed: FALSE
>>> >>>     Num slots allocated: 4    Max slots: 0
>>> >>>     Username on node: NULL
>>> >>>     Num procs: 4    Next node_rank: 4
>>> >>>     Data for proc: [[51718,1],4]
>>> >>>         Pid: 0    Local rank: 0    Node rank: 0    App rank: 4
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [B/B][./.]
>>> >>>         Binding: [B/.][./.]
>>> >>>     Data for proc: [[51718,1],5]
>>> >>>         Pid: 0    Local rank: 1    Node rank: 1    App rank: 5
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [./.][B/B]
>>> >>>         Binding: [./.][B/.]
>>> >>>     Data for proc: [[51718,1],6]
>>> >>>         Pid: 0    Local rank: 2    Node rank: 2    App rank: 6
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [B/B][./.]
>>> >>>         Binding: [./B][./.]
>>> >>>     Data for proc: [[51718,1],7]
>>> >>>         Pid: 0    Local rank: 3    Node rank: 3    App rank: 7
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [./.][B/B]
>>> >>>         Binding: [./.][./B]
>>> >>>
>>> >>> Data for node: csclprd3-0-0         Launch id: -1    State: 0
>>> >>>     Daemon: [[51718,0],3]    Daemon launched: True
>>> >>>     Num slots: 12    Slots in use: 12    Oversubscribed: FALSE
>>> >>>     Num slots allocated: 12    Max slots: 0
>>> >>>     Username on node: NULL
>>> >>>     Num procs: 12    Next node_rank: 12
>>> >>>     Data for proc: [[51718,1],8]
>>> >>>         Pid: 0    Local rank: 0    Node rank: 0    App rank: 8
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [B/B/B/B/B/B][./././././.]
>>> >>>         Binding: [B/././././.][./././././.]
>>> >>>     Data for proc: [[51718,1],9]
>>> >>>         Pid: 0    Local rank: 1    Node rank: 1    App rank: 9
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [./././././.][B/B/B/B/B/B]
>>> >>>         Binding: [./././././.][B/././././.]
>>> >>>     Data for proc: [[51718,1],10]
>>> >>>         Pid: 0    Local rank: 2    Node rank: 2    App rank: 10
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [B/B/B/B/B/B][./././././.]
>>> >>>         Binding: [./B/./././.][./././././.]
>>> >>>     Data for proc: [[51718,1],11]
>>> >>>         Pid: 0    Local rank: 3    Node rank: 3    App rank: 11
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [./././././.][B/B/B/B/B/B]
>>> >>>         Binding: [./././././.][./B/./././.]
>>> >>>     Data for proc: [[51718,1],12]
>>> >>>         Pid: 0    Local rank: 4    Node rank: 4    App rank: 12
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [B/B/B/B/B/B][./././././.]
>>> >>>         Binding: [././B/././.][./././././.]
>>> >>>     Data for proc: [[51718,1],13]
>>> >>>         Pid: 0    Local rank: 5    Node rank: 5    App rank: 13
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [./././././.][B/B/B/B/B/B]
>>> >>>         Binding: [./././././.][././B/././.]
>>> >>>     Data for proc: [[51718,1],14]
>>> >>>         Pid: 0    Local rank: 6    Node rank: 6    App rank: 14
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [B/B/B/B/B/B][./././././.]
>>> >>>         Binding: [./././B/./.][./././././.]
>>> >>>     Data for proc: [[51718,1],15]
>>> >>>         Pid: 0    Local rank: 7    Node rank: 7    App rank: 15
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [./././././.][B/B/B/B/B/B]
>>> >>>         Binding: [./././././.][./././B/./.]
>>> >>>     Data for proc: [[51718,1],16]
>>> >>>         Pid: 0    Local rank: 8    Node rank: 8    App rank: 16
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [B/B/B/B/B/B][./././././.]
>>> >>>         Binding: [././././B/.][./././././.]
>>> >>>     Data for proc: [[51718,1],17]
>>> >>>         Pid: 0    Local rank: 9    Node rank: 9    App rank: 17
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [./././././.][B/B/B/B/B/B]
>>> >>>         Binding: [./././././.][././././B/.]
>>> >>>     Data for proc: [[51718,1],18]
>>> >>>         Pid: 0    Local rank: 10    Node rank: 10    App rank: 18
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [B/B/B/B/B/B][./././././.]
>>> >>>         Binding: [./././././B][./././././.]
>>> >>>     Data for proc: [[51718,1],19]
>>> >>>         Pid: 0    Local rank: 11    Node rank: 11    App rank: 19
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [./././././.][B/B/B/B/B/B]
>>> >>>         Binding: [./././././.][./././././B]
>>> >>>
>>> >>> Data for node: csclprd3-0-1         Launch id: -1    State: 0
>>> >>>     Daemon: [[51718,0],4]    Daemon launched: True
>>> >>>     Num slots: 6    Slots in use: 6    Oversubscribed: FALSE
>>> >>>     Num slots allocated: 6    Max slots: 0
>>> >>>     Username on node: NULL
>>> >>>     Num procs: 6    Next node_rank: 6
>>> >>>     Data for proc: [[51718,1],20]
>>> >>>         Pid: 0    Local rank: 0    Node rank: 0    App rank: 20
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [B/././././.]
>>> >>>     Data for proc: [[51718,1],21]
>>> >>>         Pid: 0    Local rank: 1    Node rank: 1    App rank: 21
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [./B/./././.]
>>> >>>     Data for proc: [[51718,1],22]
>>> >>>         Pid: 0    Local rank: 2    Node rank: 2    App rank: 22
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [././B/././.]
>>> >>>     Data for proc: [[51718,1],23]
>>> >>>         Pid: 0    Local rank: 3    Node rank: 3    App rank: 23
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [./././B/./.]
>>> >>>     Data for proc: [[51718,1],24]
>>> >>>         Pid: 0    Local rank: 4    Node rank: 4    App rank: 24
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [././././B/.]
>>> >>>     Data for proc: [[51718,1],25]
>>> >>>         Pid: 0    Local rank: 5    Node rank: 5    App rank: 25
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [./././././B]
>>> >>>
>>> >>> Data for node: csclprd3-0-2         Launch id: -1    State: 0
>>> >>>     Daemon: [[51718,0],5]    Daemon launched: True
>>> >>>     Num slots: 6    Slots in use: 6    Oversubscribed: FALSE
>>> >>>     Num slots allocated: 6    Max slots: 0
>>> >>>     Username on node: NULL
>>> >>>     Num procs: 6    Next node_rank: 6
>>> >>>     Data for proc: [[51718,1],26]
>>> >>>         Pid: 0    Local rank: 0    Node rank: 0    App rank: 26
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [B/././././.]
>>> >>>     Data for proc: [[51718,1],27]
>>> >>>         Pid: 0    Local rank: 1    Node rank: 1    App rank: 27
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [./B/./././.]
>>> >>>     Data for proc: [[51718,1],28]
>>> >>>         Pid: 0    Local rank: 2    Node rank: 2    App rank: 28
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [././B/././.]
>>> >>>     Data for proc: [[51718,1],29]
>>> >>>         Pid: 0    Local rank: 3    Node rank: 3    App rank: 29
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [./././B/./.]
>>> >>>     Data for proc: [[51718,1],30]
>>> >>>         Pid: 0    Local rank: 4    Node rank: 4    App rank: 30
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [././././B/.]
>>> >>>     Data for proc: [[51718,1],31]
>>> >>>         Pid: 0    Local rank: 5    Node rank: 5    App rank: 31
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [./././././B]
>>> >>>
>>> >>> Data for node: csclprd3-0-3         Launch id: -1    State: 0
>>> >>>     Daemon: [[51718,0],6]    Daemon launched: True
>>> >>>     Num slots: 6    Slots in use: 6    Oversubscribed: FALSE
>>> >>>     Num slots allocated: 6    Max slots: 0
>>> >>>     Username on node: NULL
>>> >>>     Num procs: 6    Next node_rank: 6
>>> >>>     Data for proc: [[51718,1],32]
>>> >>>         Pid: 0    Local rank: 0    Node rank: 0    App rank: 32
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [B/././././.]
>>> >>>     Data for proc: [[51718,1],33]
>>> >>>         Pid: 0    Local rank: 1    Node rank: 1    App rank: 33
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [./B/./././.]
>>> >>>     Data for proc: [[51718,1],34]
>>> >>>         Pid: 0    Local rank: 2    Node rank: 2    App rank: 34
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [././B/././.]
>>> >>>     Data for proc: [[51718,1],35]
>>> >>>         Pid: 0    Local rank: 3    Node rank: 3    App rank: 35
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [./././B/./.]
>>> >>>     Data for proc: [[51718,1],36]
>>> >>>         Pid: 0    Local rank: 4    Node rank: 4    App rank: 36
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [././././B/.]
>>> >>>     Data for proc: [[51718,1],37]
>>> >>>         Pid: 0    Local rank: 5    Node rank: 5    App rank: 37
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [./././././B]
>>> >>>
>>> >>> Data for node: csclprd3-0-4         Launch id: -1    State: 0
>>> >>>     Daemon: [[51718,0],7]    Daemon launched: True
>>> >>>     Num slots: 6    Slots in use: 6    Oversubscribed: FALSE
>>> >>>     Num slots allocated: 6    Max slots: 0
>>> >>>     Username on node: NULL
>>> >>>     Num procs: 6    Next node_rank: 6
>>> >>>     Data for proc: [[51718,1],38]
>>> >>>         Pid: 0    Local rank: 0    Node rank: 0    App rank: 38
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [B/././././.]
>>> >>>     Data for proc: [[51718,1],39]
>>> >>>         Pid: 0    Local rank: 1    Node rank: 1    App rank: 39
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [./B/./././.]
>>> >>>     Data for proc: [[51718,1],40]
>>> >>>         Pid: 0    Local rank: 2    Node rank: 2    App rank: 40
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [././B/././.]
>>> >>>     Data for proc: [[51718,1],41]
>>> >>>         Pid: 0    Local rank: 3    Node rank: 3    App rank: 41
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [./././B/./.]
>>> >>>     Data for proc: [[51718,1],42]
>>> >>>         Pid: 0    Local rank: 4    Node rank: 4    App rank: 42
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [././././B/.]
>>> >>>     Data for proc: [[51718,1],43]
>>> >>>         Pid: 0    Local rank: 5    Node rank: 5    App rank: 43
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [./././././B]
>>> >>>
>>> >>> Data for node: csclprd3-0-5         Launch id: -1    State: 0
>>> >>>     Daemon: [[51718,0],8]    Daemon launched: True
>>> >>>     Num slots: 6    Slots in use: 6    Oversubscribed: FALSE
>>> >>>     Num slots allocated: 6    Max slots: 0
>>> >>>     Username on node: NULL
>>> >>>     Num procs: 6    Next node_rank: 6
>>> >>>     Data for proc: [[51718,1],44]
>>> >>>         Pid: 0    Local rank: 0    Node rank: 0    App rank: 44
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [B/././././.]
>>> >>>     Data for proc: [[51718,1],45]
>>> >>>         Pid: 0    Local rank: 1    Node rank: 1    App rank: 45
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [./B/./././.]
>>> >>>     Data for proc: [[51718,1],46]
>>> >>>         Pid: 0    Local rank: 2    Node rank: 2    App rank: 46
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [././B/././.]
>>> >>>     Data for proc: [[51718,1],47]
>>> >>>         Pid: 0    Local rank: 3    Node rank: 3    App rank: 47
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [./././B/./.]
>>> >>>     Data for proc: [[51718,1],48]
>>> >>>         Pid: 0    Local rank: 4    Node rank: 4    App rank: 48
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [././././B/.]
>>> >>>     Data for proc: [[51718,1],49]
>>> >>>         Pid: 0    Local rank: 5    Node rank: 5    App rank: 49
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [./././././B]
>>> >>>
>>> >>> Data for node: csclprd3-0-6         Launch id: -1    State: 0
>>> >>>     Daemon: [[51718,0],9]    Daemon launched: True
>>> >>>     Num slots: 6    Slots in use: 6    Oversubscribed: FALSE
>>> >>>     Num slots allocated: 6    Max slots: 0
>>> >>>     Username on node: NULL
>>> >>>     Num procs: 6    Next node_rank: 6
>>> >>>     Data for proc: [[51718,1],50]
>>> >>>         Pid: 0    Local rank: 0    Node rank: 0    App rank: 50
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [B/././././.]
>>> >>>     Data for proc: [[51718,1],51]
>>> >>>         Pid: 0    Local rank: 1    Node rank: 1    App rank: 51
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [./B/./././.]
>>> >>>     Data for proc: [[51718,1],52]
>>> >>>         Pid: 0    Local rank: 2    Node rank: 2    App rank: 52
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [././B/././.]
>>> >>>     Data for proc: [[51718,1],53]
>>> >>>         Pid: 0    Local rank: 3    Node rank: 3    App rank: 53
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [./././B/./.]
>>> >>>     Data for proc: [[51718,1],54]
>>> >>>         Pid: 0    Local rank: 4    Node rank: 4    App rank: 54
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [././././B/.]
>>> >>>     Data for proc: [[51718,1],55]
>>> >>>         Pid: 0    Local rank: 5    Node rank: 5    App rank: 55
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [./././././B]
>>> >>>
>>> >>> Data for node: csclprd3-0-7         Launch id: -1    State: 0
>>> >>>     Daemon: [[51718,0],10]    Daemon launched: True
>>> >>>     Num slots: 16    Slots in use: 16    Oversubscribed: FALSE
>>> >>>     Num slots allocated: 16    Max slots: 0
>>> >>>     Username on node: NULL
>>> >>>     Num procs: 16    Next node_rank: 16
>>> >>>     Data for proc: [[51718,1],56]
>>> >>>         Pid: 0    Local rank: 0    Node rank: 0    App rank: 56
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [BB/../../../../../../..][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],57]
>>> >>>         Pid: 0    Local rank: 1    Node rank: 1    App rank: 57
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][BB/../../../../../../..]
>>> >>>     Data for proc: [[51718,1],58]
>>> >>>         Pid: 0    Local rank: 2    Node rank: 2    App rank: 58
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [../BB/../../../../../..][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],59]
>>> >>>         Pid: 0    Local rank: 3    Node rank: 3    App rank: 59
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][../BB/../../../../../..]
>>> >>>     Data for proc: [[51718,1],60]
>>> >>>         Pid: 0    Local rank: 4    Node rank: 4    App rank: 60
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [../../BB/../../../../..][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],61]
>>> >>>         Pid: 0    Local rank: 5    Node rank: 5    App rank: 61
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][../../BB/../../../../..]
>>> >>>     Data for proc: [[51718,1],62]
>>> >>>         Pid: 0    Local rank: 6    Node rank: 6    App rank: 62
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [../../../BB/../../../..][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],63]
>>> >>>         Pid: 0    Local rank: 7    Node rank: 7    App rank: 63
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][../../../BB/../../../..]
>>> >>>     Data for proc: [[51718,1],64]
>>> >>>         Pid: 0    Local rank: 8    Node rank: 8    App rank: 64
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [../../../../BB/../../..][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],65]
>>> >>>         Pid: 0    Local rank: 9    Node rank: 9    App rank: 65
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][../../../../BB/../../..]
>>> >>>     Data for proc: [[51718,1],66]
>>> >>>         Pid: 0    Local rank: 10    Node rank: 10    App rank: 66
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [../../../../../BB/../..][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],67]
>>> >>>         Pid: 0    Local rank: 11    Node rank: 11    App rank: 67
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][../../../../../BB/../..]
>>> >>>     Data for proc: [[51718,1],68]
>>> >>>         Pid: 0    Local rank: 12    Node rank: 12    App rank: 68
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [../../../../../../BB/..][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],69]
>>> >>>         Pid: 0    Local rank: 13    Node rank: 13    App rank: 69
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][../../../../../../BB/..]
>>> >>>     Data for proc: [[51718,1],70]
>>> >>>         Pid: 0    Local rank: 14    Node rank: 14    App rank: 70
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [../../../../../../../BB][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],71]
>>> >>>         Pid: 0    Local rank: 15    Node rank: 15    App rank: 71
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][../../../../../../../BB]
>>> >>>
>>> >>> Data for node: csclprd3-0-8         Launch id: -1    State: 0
>>> >>>     Daemon: [[51718,0],11]    Daemon launched: True
>>> >>>     Num slots: 16    Slots in use: 16    Oversubscribed: FALSE
>>> >>>     Num slots allocated: 16    Max slots: 0
>>> >>>     Username on node: NULL
>>> >>>     Num procs: 16    Next node_rank: 16
>>> >>>     Data for proc: [[51718,1],72]
>>> >>>         Pid: 0    Local rank: 0    Node rank: 0    App rank: 72
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [BB/../../../../../../..][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],73]
>>> >>>         Pid: 0    Local rank: 1    Node rank: 1    App rank: 73
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][BB/../../../../../../..]
>>> >>>     Data for proc: [[51718,1],74]
>>> >>>         Pid: 0    Local rank: 2    Node rank: 2    App rank: 74
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [../BB/../../../../../..][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],75]
>>> >>>         Pid: 0    Local rank: 3    Node rank: 3    App rank: 75
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][../BB/../../../../../..]
>>> >>>     Data for proc: [[51718,1],76]
>>> >>>         Pid: 0    Local rank: 4    Node rank: 4    App rank: 76
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [../../BB/../../../../..][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],77]
>>> >>>         Pid: 0    Local rank: 5    Node rank: 5    App rank: 77
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][../../BB/../../../../..]
>>> >>>     Data for proc: [[51718,1],78]
>>> >>>         Pid: 0    Local rank: 6    Node rank: 6    App rank: 78
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [../../../BB/../../../..][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],79]
>>> >>>         Pid: 0    Local rank: 7    Node rank: 7    App rank: 79
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][../../../BB/../../../..]
>>> >>>     Data for proc: [[51718,1],80]
>>> >>>         Pid: 0    Local rank: 8    Node rank: 8    App rank: 80
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [../../../../BB/../../..][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],81]
>>> >>>         Pid: 0    Local rank: 9    Node rank: 9    App rank: 81
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][../../../../BB/../../..]
>>> >>>     Data for proc: [[51718,1],82]
>>> >>>         Pid: 0    Local rank: 10    Node rank: 10    App rank: 82
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [../../../../../BB/../..][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],83]
>>> >>>         Pid: 0    Local rank: 11    Node rank: 11    App rank: 83
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][../../../../../BB/../..]
>>> >>>     Data for proc: [[51718,1],84]
>>> >>>         Pid: 0    Local rank: 12    Node rank: 12    App rank: 84
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [../../../../../../BB/..][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],85]
>>> >>>         Pid: 0    Local rank: 13    Node rank: 13    App rank: 85
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][../../../../../../BB/..]
>>> >>>     Data for proc: [[51718,1],86]
>>> >>>         Pid: 0    Local rank: 14    Node rank: 14    App rank: 86
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [../../../../../../../BB][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],87]
>>> >>>         Pid: 0    Local rank: 15    Node rank: 15    App rank: 87
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][../../../../../../../BB]
>>> >>>
>>> >>> Data for node: csclprd3-0-10         Launch id: -1    State: 0
>>> >>>     Daemon: [[51718,0],12]    Daemon launched: True
>>> >>>     Num slots: 16    Slots in use: 16    Oversubscribed: FALSE
>>> >>>     Num slots allocated: 16    Max slots: 0
>>> >>>     Username on node: NULL
>>> >>>     Num procs: 16    Next node_rank: 16
>>> >>>     Data for proc: [[51718,1],88]
>>> >>>         Pid: 0    Local rank: 0    Node rank: 0    App rank: 88
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [BB/../../../../../../..][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],89]
>>> >>>         Pid: 0    Local rank: 1    Node rank: 1    App rank: 89
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][BB/../../../../../../..]
>>> >>>     Data for proc: [[51718,1],90]
>>> >>>         Pid: 0    Local rank: 2    Node rank: 2    App rank: 90
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [../BB/../../../../../..][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],91]
>>> >>>         Pid: 0    Local rank: 3    Node rank: 3    App rank: 91
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][../BB/../../../../../..]
>>> >>>     Data for proc: [[51718,1],92]
>>> >>>         Pid: 0    Local rank: 4    Node rank: 4    App rank: 92
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [../../BB/../../../../..][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],93]
>>> >>>         Pid: 0    Local rank: 5    Node rank: 5    App rank: 93
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][../../BB/../../../../..]
>>> >>>     Data for proc: [[51718,1],94]
>>> >>>         Pid: 0    Local rank: 6    Node rank: 6    App rank: 94
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [../../../BB/../../../..][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],95]
>>> >>>         Pid: 0    Local rank: 7    Node rank: 7    App rank: 95
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][../../../BB/../../../..]
>>> >>>     Data for proc: [[51718,1],96]
>>> >>>         Pid: 0    Local rank: 8    Node rank: 8    App rank: 96
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [../../../../BB/../../..][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],97]
>>> >>>         Pid: 0    Local rank: 9    Node rank: 9    App rank: 97
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][../../../../BB/../../..]
>>> >>>     Data for proc: [[51718,1],98]
>>> >>>         Pid: 0    Local rank: 10    Node rank: 10    App rank: 98
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [../../../../../BB/../..][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],99]
>>> >>>         Pid: 0    Local rank: 11    Node rank: 11    App rank: 99
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][../../../../../BB/../..]
>>> >>>     Data for proc: [[51718,1],100]
>>> >>>         Pid: 0    Local rank: 12    Node rank: 12    App rank: 100
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [../../../../../../BB/..][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],101]
>>> >>>         Pid: 0    Local rank: 13    Node rank: 13    App rank: 101
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][../../../../../../BB/..]
>>> >>>     Data for proc: [[51718,1],102]
>>> >>>         Pid: 0    Local rank: 14    Node rank: 14    App rank: 102
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [../../../../../../../BB][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],103]
>>> >>>         Pid: 0    Local rank: 15    Node rank: 15    App rank: 103
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][../../../../../../../BB]
>>> >>>
>>> >>> Data for node: csclprd3-0-11         Launch id: -1    State: 0
>>> >>>     Daemon: [[51718,0],13]    Daemon launched: True
>>> >>>     Num slots: 16    Slots in use: 16    Oversubscribed: FALSE
>>> >>>     Num slots allocated: 16    Max slots: 0
>>> >>>     Username on node: NULL
>>> >>>     Num procs: 16    Next node_rank: 16
>>> >>>     Data for proc: [[51718,1],104]
>>> >>>         Pid: 0    Local rank: 0    Node rank: 0    App rank: 104
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [BB/../../../../../../..][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],105]
>>> >>>         Pid: 0    Local rank: 1    Node rank: 1    App rank: 105
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][BB/../../../../../../..]
>>> >>>     Data for proc: [[51718,1],106]
>>> >>>         Pid: 0    Local rank: 2    Node rank: 2    App rank: 106
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [../BB/../../../../../..][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],107]
>>> >>>         Pid: 0    Local rank: 3    Node rank: 3    App rank: 107
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][../BB/../../../../../..]
>>> >>>     Data for proc: [[51718,1],108]
>>> >>>         Pid: 0    Local rank: 4    Node rank: 4    App rank: 108
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [../../BB/../../../../..][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],109]
>>> >>>         Pid: 0    Local rank: 5    Node rank: 5    App rank: 109
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][../../BB/../../../../..]
>>> >>>     Data for proc: [[51718,1],110]
>>> >>>         Pid: 0    Local rank: 6    Node rank: 6    App rank: 110
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [../../../BB/../../../..][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],111]
>>> >>>         Pid: 0    Local rank: 7    Node rank: 7    App rank: 111
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][../../../BB/../../../..]
>>> >>>     Data for proc: [[51718,1],112]
>>> >>>         Pid: 0    Local rank: 8    Node rank: 8    App rank: 112
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [../../../../BB/../../..][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],113]
>>> >>>         Pid: 0    Local rank: 9    Node rank: 9    App rank: 113
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][../../../../BB/../../..]
>>> >>>     Data for proc: [[51718,1],114]
>>> >>>         Pid: 0    Local rank: 10    Node rank: 10    App rank: 114
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [../../../../../BB/../..][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],115]
>>> >>>         Pid: 0    Local rank: 11    Node rank: 11    App rank: 115
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][../../../../../BB/../..]
>>> >>>     Data for proc: [[51718,1],116]
>>> >>>         Pid: 0    Local rank: 12    Node rank: 12    App rank: 116
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [../../../../../../BB/..][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],117]
>>> >>>         Pid: 0    Local rank: 13    Node rank: 13    App rank: 117
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][../../../../../../BB/..]
>>> >>>     Data for proc: [[51718,1],118]
>>> >>>         Pid: 0    Local rank: 14    Node rank: 14    App rank: 118
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>>         Binding: [../../../../../../../BB][../../../../../../../..]
>>> >>>     Data for proc: [[51718,1],119]
>>> >>>         Pid: 0    Local rank: 15    Node rank: 15    App rank: 119
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../../../..][../../../../../../../BB]
>>> >>>
>>> >>> Data for node: csclprd3-0-12         Launch id: -1    State: 0
>>> >>>     Daemon: [[51718,0],14]    Daemon launched: True
>>> >>>     Num slots: 6    Slots in use: 6    Oversubscribed: FALSE
>>> >>>     Num slots allocated: 6    Max slots: 0
>>> >>>     Username on node: NULL
>>> >>>     Num procs: 6    Next node_rank: 6
>>> >>>     Data for proc: [[51718,1],120]
>>> >>>         Pid: 0    Local rank: 0    Node rank: 0    App rank: 120
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [BB/../../../../..]
>>> >>>     Data for proc: [[51718,1],121]
>>> >>>         Pid: 0    Local rank: 1    Node rank: 1    App rank: 121
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [../BB/../../../..]
>>> >>>     Data for proc: [[51718,1],122]
>>> >>>         Pid: 0    Local rank: 2    Node rank: 2    App rank: 122
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [../../BB/../../..]
>>> >>>     Data for proc: [[51718,1],123]
>>> >>>         Pid: 0    Local rank: 3    Node rank: 3    App rank: 123
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [../../../BB/../..]
>>> >>>     Data for proc: [[51718,1],124]
>>> >>>         Pid: 0    Local rank: 4    Node rank: 4    App rank: 124
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [../../../../BB/..]
>>> >>>     Data for proc: [[51718,1],125]
>>> >>>         Pid: 0    Local rank: 5    Node rank: 5    App rank: 125
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: UNKNOWN
>>> >>>         Binding: [../../../../../BB]
>>> >>>
>>> >>> Data for node: csclprd3-0-13         Launch id: -1    State: 0
>>> >>>     Daemon: [[51718,0],15]    Daemon launched: True
>>> >>>     Num slots: 12    Slots in use: 6    Oversubscribed: FALSE
>>> >>>     Num slots allocated: 12    Max slots: 0
>>> >>>     Username on node: NULL
>>> >>>     Num procs: 6    Next node_rank: 6
>>> >>>     Data for proc: [[51718,1],126]
>>> >>>         Pid: 0    Local rank: 0    Node rank: 0    App rank: 126
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB][../../../../../..]
>>> >>>         Binding: [BB/../../../../..][../../../../../..]
>>> >>>     Data for proc: [[51718,1],127]
>>> >>>         Pid: 0    Local rank: 1    Node rank: 1    App rank: 127
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../..][BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../..][BB/../../../../..]
>>> >>>     Data for proc: [[51718,1],128]
>>> >>>         Pid: 0    Local rank: 2    Node rank: 2    App rank: 128
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB][../../../../../..]
>>> >>>         Binding: [../BB/../../../..][../../../../../..]
>>> >>>     Data for proc: [[51718,1],129]
>>> >>>         Pid: 0    Local rank: 3    Node rank: 3    App rank: 129
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../..][BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../..][../BB/../../../..]
>>> >>>     Data for proc: [[51718,1],130]
>>> >>>         Pid: 0    Local rank: 4    Node rank: 4    App rank: 130
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [BB/BB/BB/BB/BB/BB][../../../../../..]
>>> >>>         Binding: [../../BB/../../..][../../../../../..]
>>> >>>     Data for proc: [[51718,1],131]
>>> >>>         Pid: 0    Local rank: 5    Node rank: 5    App rank: 131
>>> >>>         State: INITIALIZED    App_context: 0
>>> >>>         Locale: [../../../../../..][BB/BB/BB/BB/BB/BB]
>>> >>>         Binding: [../../../../../..][../../BB/../../..]
>>> >>> [csclprd3-0-13:31619] *** Process received signal ***
>>> >>> [csclprd3-0-13:31619] Signal: Bus error (7)
>>> >>> [csclprd3-0-13:31619] Signal code: Non-existant physical address (2)
>>> >>> [csclprd3-0-13:31619] Failing at address: 0x7f1374267a00
>>> >>> [csclprd3-0-13:31620] *** Process received signal ***
>>> >>> [csclprd3-0-13:31620] Signal: Bus error (7)
>>> >>> [csclprd3-0-13:31620] Signal code: Non-existant physical address (2)
>>> >>> [csclprd3-0-13:31620] Failing at address: 0x7fcc702a7980
>>> >>> [csclprd3-0-13:31615] *** Process received signal ***
>>> >>> [csclprd3-0-13:31615] Signal: Bus error (7)
>>> >>> [csclprd3-0-13:31615] Signal code: Non-existant physical address (2)
>>> >>> [csclprd3-0-13:31615] Failing at address: 0x7f8128367880
>>> >>> [csclprd3-0-13:31616] *** Process received signal ***
>>> >>> [csclprd3-0-13:31616] Signal: Bus error (7)
>>> >>> [csclprd3-0-13:31616] Signal code: Non-existant physical address (2)
>>> >>> [csclprd3-0-13:31616] Failing at address: 0x7fe674227a00
>>> >>> [csclprd3-0-13:31617] *** Process received signal ***
>>> >>> [csclprd3-0-13:31617] Signal: Bus error (7)
>>> >>> [csclprd3-0-13:31617] Signal code: Non-existant physical address (2)
>>> >>> [csclprd3-0-13:31617] Failing at address: 0x7f061c32db80
>>> >>> [csclprd3-0-13:31618] *** Process received signal ***
>>> >>> [csclprd3-0-13:31618] Signal: Bus error (7)
>>> >>> [csclprd3-0-13:31618] Signal code: Non-existant physical address (2)
>>> >>> [csclprd3-0-13:31618] Failing at address: 0x7fb8402eaa80
>>> >>> [csclprd3-0-13:31618] [ 0] 
>>> >>> /lib64/libpthread.so.0(+0xf500)[0x7fb851851500]
>>> >>> [csclprd3-0-13:31618] [ 1] [csclprd3-0-13:31616] [ 0] 
>>> >>> /lib64/libpthread.so.0(+0xf500)[0x7fe6843a4500]
>>> >>> [csclprd3-0-13:31616] [ 1] [csclprd3-0-13:31620] [ 0] 
>>> >>> /lib64/libpthread.so.0(+0xf500)[0x7fcc80c54500]
>>> >>> [csclprd3-0-13:31620] [ 1] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(+0x167f61)[0x7fcc80fc9f61]
>>> >>> [csclprd3-0-13:31620] [ 2] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(+0x168047)[0x7fcc80fca047]
>>> >>> [csclprd3-0-13:31620] [ 3] [csclprd3-0-13:31615] [ 0] 
>>> >>> /lib64/libpthread.so.0(+0xf500)[0x7f81385ca500]
>>> >>> [csclprd3-0-13:31615] [ 1] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(+0x167f61)[0x7f813893ff61]
>>> >>> [csclprd3-0-13:31615] [ 2] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(+0x168047)[0x7f8138940047]
>>> >>> [csclprd3-0-13:31615] [ 3] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(+0x167f61)[0x7fb851bc6f61]
>>> >>> [csclprd3-0-13:31618] [ 2] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(+0x168047)[0x7fb851bc7047]
>>> >>> [csclprd3-0-13:31618] [ 3] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(+0x55670)[0x7fb851ab4670]
>>> >>> [csclprd3-0-13:31618] [ 4] [csclprd3-0-13:31617] [ 0] 
>>> >>> /lib64/libpthread.so.0(+0xf500)[0x7f062cfe5500]
>>> >>> [csclprd3-0-13:31617] [ 1] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(+0x167f61)[0x7f062d35af61]
>>> >>> [csclprd3-0-13:31617] [ 2] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(+0x168047)[0x7f062d35b047]
>>> >>> [csclprd3-0-13:31617] [ 3] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(+0x55670)[0x7f062d248670]
>>> >>> [csclprd3-0-13:31617] [ 4] [csclprd3-0-13:31619] [ 0] 
>>> >>> /lib64/libpthread.so.0(+0xf500)[0x7f1384fde500]
>>> >>> [csclprd3-0-13:31619] [ 1] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(+0x167f61)[0x7f1385353f61]
>>> >>> [csclprd3-0-13:31619] [ 2] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(+0x167f61)[0x7fe684719f61]
>>> >>> [csclprd3-0-13:31616] [ 2] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(+0x168047)[0x7fe68471a047]
>>> >>> [csclprd3-0-13:31616] [ 3] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(+0x55670)[0x7fe684607670]
>>> >>> [csclprd3-0-13:31616] [ 4] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(+0x168047)[0x7f1385354047]
>>> >>> [csclprd3-0-13:31619] [ 3] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(+0x55670)[0x7f1385241670]
>>> >>> [csclprd3-0-13:31619] [ 4] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(ompi_free_list_grow+0x3b9)[0x7f13852425ab]
>>> >>> [csclprd3-0-13:31619] [ 5] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(ompi_free_list_resize_mt+0xfb)[0x7f1385242751]
>>> >>> [csclprd3-0-13:31619] [ 6] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(mca_btl_sm_add_procs+0x671)[0x7f13853501c9]
>>> >>> [csclprd3-0-13:31619] [ 7] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(+0x14a628)[0x7f1385336628]
>>> >>> [csclprd3-0-13:31619] [ 8] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(+0x55670)[0x7fcc80eb7670]
>>> >>> [csclprd3-0-13:31620] [ 4] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(ompi_free_list_grow+0x3b9)[0x7fcc80eb85ab]
>>> >>> [csclprd3-0-13:31620] [ 5] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(ompi_free_list_resize_mt+0xfb)[0x7fcc80eb8751]
>>> >>> [csclprd3-0-13:31620] [ 6] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(mca_btl_sm_add_procs+0x671)[0x7fcc80fc61c9]
>>> >>> [csclprd3-0-13:31620] [ 7] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(+0x14a628)[0x7fcc80fac628]
>>> >>> [csclprd3-0-13:31620] [ 8] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(mca_pml_ob1_add_procs+0xff)[0x7fcc8111fd61]
>>> >>> [csclprd3-0-13:31620] [ 9] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(+0x55670)[0x7f813882d670]
>>> >>> [csclprd3-0-13:31615] [ 4] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(ompi_free_list_grow+0x3b9)[0x7f813882e5ab]
>>> >>> [csclprd3-0-13:31615] [ 5] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(ompi_free_list_resize_mt+0xfb)[0x7f813882e751]
>>> >>> [csclprd3-0-13:31615] [ 6] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(mca_btl_sm_add_procs+0x671)[0x7f813893c1c9]
>>> >>> [csclprd3-0-13:31615] [ 7] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(+0x14a628)[0x7f8138922628]
>>> >>> [csclprd3-0-13:31615] [ 8] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(mca_pml_ob1_add_procs+0xff)[0x7f8138a95d61]
>>> >>> [csclprd3-0-13:31615] [ 9] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(ompi_mpi_init+0xbda)[0x7f813885d747]
>>> >>> [csclprd3-0-13:31615] [10] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(ompi_free_list_grow+0x3b9)[0x7fb851ab55ab]
>>> >>> [csclprd3-0-13:31618] [ 5] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(ompi_free_list_resize_mt+0xfb)[0x7fb851ab5751]
>>> >>> [csclprd3-0-13:31618] [ 6] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(mca_btl_sm_add_procs+0x671)[0x7fb851bc31c9]
>>> >>> [csclprd3-0-13:31618] [ 7] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(+0x14a628)[0x7fb851ba9628]
>>> >>> [csclprd3-0-13:31618] [ 8] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(mca_pml_ob1_add_procs+0xff)[0x7fb851d1cd61]
>>> >>> [csclprd3-0-13:31618] [ 9] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(ompi_mpi_init+0xbda)[0x7fb851ae4747]
>>> >>> [csclprd3-0-13:31618] [10] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(ompi_free_list_grow+0x3b9)[0x7f062d2495ab]
>>> >>> [csclprd3-0-13:31617] [ 5] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(ompi_free_list_resize_mt+0xfb)[0x7f062d249751]
>>> >>> [csclprd3-0-13:31617] [ 6] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(mca_btl_sm_add_procs+0x671)[0x7f062d3571c9]
>>> >>> [csclprd3-0-13:31617] [ 7] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(+0x14a628)[0x7f062d33d628]
>>> >>> [csclprd3-0-13:31617] [ 8] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(mca_pml_ob1_add_procs+0xff)[0x7f062d4b0d61]
>>> >>> [csclprd3-0-13:31617] [ 9] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(ompi_mpi_init+0xbda)[0x7f062d278747]
>>> >>> [csclprd3-0-13:31617] [10] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(ompi_free_list_grow+0x3b9)[0x7fe6846085ab]
>>> >>> [csclprd3-0-13:31616] [ 5] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(ompi_free_list_resize_mt+0xfb)[0x7fe684608751]
>>> >>> [csclprd3-0-13:31616] [ 6] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(mca_btl_sm_add_procs+0x671)[0x7fe6847161c9]
>>> >>> [csclprd3-0-13:31616] [ 7] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(+0x14a628)[0x7fe6846fc628]
>>> >>> [csclprd3-0-13:31616] [ 8] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(mca_pml_ob1_add_procs+0xff)[0x7fe68486fd61]
>>> >>> [csclprd3-0-13:31616] [ 9] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(ompi_mpi_init+0xbda)[0x7fe684637747]
>>> >>> [csclprd3-0-13:31616] [10] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(MPI_Init+0x185)[0x7fe68467750b]
>>> >>> [csclprd3-0-13:31616] [11] 
>>> >>> /hpc/home/lanew/mpi/openmpi/ProcessColors3[0x400ad0]
>>> >>> [csclprd3-0-13:31616] [12] 
>>> >>> /lib64/libc.so.6(__libc_start_main+0xfd)[0x7fe684021cdd]
>>> >>> [csclprd3-0-13:31616] [13] 
>>> >>> /hpc/home/lanew/mpi/openmpi/ProcessColors3[0x400999]
>>> >>> [csclprd3-0-13:31616] *** End of error message ***
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(MPI_Init+0x185)[0x7f062d2b850b]
>>> >>> [csclprd3-0-13:31617] [11] 
>>> >>> /hpc/home/lanew/mpi/openmpi/ProcessColors3[0x400ad0]
>>> >>> [csclprd3-0-13:31617] [12] 
>>> >>> /lib64/libc.so.6(__libc_start_main+0xfd)[0x7f062cc62cdd]
>>> >>> [csclprd3-0-13:31617] [13] 
>>> >>> /hpc/home/lanew/mpi/openmpi/ProcessColors3[0x400999]
>>> >>> [csclprd3-0-13:31617] *** End of error message ***
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(mca_pml_ob1_add_procs+0xff)[0x7f13854a9d61]
>>> >>> [csclprd3-0-13:31619] [ 9] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(ompi_mpi_init+0xbda)[0x7f1385271747]
>>> >>> [csclprd3-0-13:31619] [10] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(MPI_Init+0x185)[0x7f13852b150b]
>>> >>> [csclprd3-0-13:31619] [11] 
>>> >>> /hpc/home/lanew/mpi/openmpi/ProcessColors3[0x400ad0]
>>> >>> [csclprd3-0-13:31619] [12] 
>>> >>> /lib64/libc.so.6(__libc_start_main+0xfd)[0x7f1384c5bcdd]
>>> >>> [csclprd3-0-13:31619] [13] 
>>> >>> /hpc/home/lanew/mpi/openmpi/ProcessColors3[0x400999]
>>> >>> [csclprd3-0-13:31619] *** End of error message ***
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(ompi_mpi_init+0xbda)[0x7fcc80ee7747]
>>> >>> [csclprd3-0-13:31620] [10] 
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(MPI_Init+0x185)[0x7fcc80f2750b]
>>> >>> [csclprd3-0-13:31620] [11] 
>>> >>> /hpc/home/lanew/mpi/openmpi/ProcessColors3[0x400ad0]
>>> >>> [csclprd3-0-13:31620] [12] 
>>> >>> /lib64/libc.so.6(__libc_start_main+0xfd)[0x7fcc808d1cdd]
>>> >>> [csclprd3-0-13:31620] [13] 
>>> >>> /hpc/home/lanew/mpi/openmpi/ProcessColors3[0x400999]
>>> >>> [csclprd3-0-13:31620] *** End of error message ***
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(MPI_Init+0x185)[0x7f813889d50b]
>>> >>> [csclprd3-0-13:31615] [11] 
>>> >>> /hpc/home/lanew/mpi/openmpi/ProcessColors3[0x400ad0]
>>> >>> [csclprd3-0-13:31615] [12] 
>>> >>> /lib64/libc.so.6(__libc_start_main+0xfd)[0x7f8138247cdd]
>>> >>> [csclprd3-0-13:31615] [13] 
>>> >>> /hpc/home/lanew/mpi/openmpi/ProcessColors3[0x400999]
>>> >>> [csclprd3-0-13:31615] *** End of error message ***
>>> >>> /hpc/apps/mpi/openmpi/1.8.6/lib/libmpi.so.1(MPI_Init+0x185)[0x7fb851b2450b]
>>> >>> [csclprd3-0-13:31618] [11] 
>>> >>> /hpc/home/lanew/mpi/openmpi/ProcessColors3[0x400ad0]
>>> >>> [csclprd3-0-13:31618] [12] 
>>> >>> /lib64/libc.so.6(__libc_start_main+0xfd)[0x7fb8514cecdd]
>>> >>> [csclprd3-0-13:31618] [13] 
>>> >>> /hpc/home/lanew/mpi/openmpi/ProcessColors3[0x400999]
>>> >>> [csclprd3-0-13:31618] *** End of error message ***
>>> >>> --------------------------------------------------------------------------
>>> >>> mpirun noticed that process rank 126 with PID 0 on node csclprd3-0-13 
>>> >>> exited on signal 7 (Bus error).
>>> >>> --------------------------------------------------------------------------
>>> >>>
>>> >>> From: users [users-boun...@open-mpi.org 
>>> >>> <mailto:users-boun...@open-mpi.org>] on behalf of Ralph Castain 
>>> >>> [r...@open-mpi.org <mailto:r...@open-mpi.org>]
>>> >>> Sent: Tuesday, June 23, 2015 6:20 PM
>>> >>> To: Open MPI Users
>>> >>> Subject: Re: [OMPI users] OpenMPI 1.8.6, CentOS 6.3, too many slots = 
>>> >>> crash
>>> >>>
>>> >>> Wow - that is one sick puppy! I see that some nodes are reporting 
>>> >>> not-bound for their procs, and the rest are binding to socket (as they 
>>> >>> should). Some of your nodes clearly do not have hyper threads enabled 
>>> >>> (or only have single-thread cores on them), and have 2 cores/socket. 
>>> >>> Other nodes have 8 cores/socket with hyper threads enabled, while still 
>>> >>> others have 6 cores/socket and HT enabled.
>>> >>>
>>> >>> I don't see anyone binding to a single HT if multiple HTs/core are 
>>> >>> available. I think you are being fooled by those nodes that either 
>>> >>> don't have HT enabled, or have only 1 HT/core.
>>> >>>
>>> >>> In both cases, it is node 13 that is the node that fails. I also note 
>>> >>> that you said everything works okay with < 132 ranks, and node 13 hosts 
>>> >>> ranks 127-131. So node 13 would host ranks even if you reduced the 
>>> >>> number in the job to 131. This would imply that it probably isn't 
>>> >>> something wrong with the node itself.
>>> >>>
>>> >>> Is there any way you could run a job of this size on a homogeneous 
>>> >>> cluster? The procs all show bindings that look right, but I'm wondering 
>>> >>> if the heterogeneity is the source of the trouble here. We may be 
>>> >>> communicating the binding pattern incorrectly and giving bad info to 
>>> >>> the backend daemon.
>>> >>>
>>> >>> Also, rather than --report-bindings, use "--display-devel-map" on the 
>>> >>> command line and let's see what the mapper thinks it did. If there is a 
>>> >>> problem with placement, that is where it would exist.
>>> >>>
>>> >>>
>>> >>> On Tue, Jun 23, 2015 at 5:12 PM, Lane, William <william.l...@cshs.org 
>>> >>> <mailto:william.l...@cshs.org>> wrote:
>>> >>> Ralph,
>>> >>>
>>> >>> There is something funny going on, the trace from the
>>> >>> runs w/the debug build aren't showing any differences from
>>> >>> what I got earlier. However, I did do a run w/the --bind-to core
>>> >>> switch and was surprised to see that hyperthreading cores were
>>> >>> sometimes being used.
>>> >>>
>>> >>> Here's the traces that I have:
>>> >>>
>>> >>> mpirun -np 132 -report-bindings --prefix /hpc/apps/mpi/openmpi/1.8.6/ 
>>> >>> --hostfile hostfile-noslots --mca btl_tcp_if_include eth0 
>>> >>> --hetero-nodes /hpc/home/lanew/mpi/openmpi/ProcessColors3
>>> >>> [csclprd3-0-5:16802] MCW rank 44 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-5:16802] MCW rank 45 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-5:16802] MCW rank 46 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-6-5:12480] MCW rank 4 bound to socket 0[core 0[hwt 0]], 
>>> >>> socket 0[core 1[hwt 0]]: [B/B][./.]
>>> >>> [csclprd3-6-5:12480] MCW rank 5 bound to socket 1[core 2[hwt 0]], 
>>> >>> socket 1[core 3[hwt 0]]: [./.][B/B]
>>> >>> [csclprd3-6-5:12480] MCW rank 6 bound to socket 0[core 0[hwt 0]], 
>>> >>> socket 0[core 1[hwt 0]]: [B/B][./.]
>>> >>> [csclprd3-6-5:12480] MCW rank 7 bound to socket 1[core 2[hwt 0]], 
>>> >>> socket 1[core 3[hwt 0]]: [./.][B/B]
>>> >>> [csclprd3-0-5:16802] MCW rank 47 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-5:16802] MCW rank 48 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-5:16802] MCW rank 49 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-1:14318] MCW rank 22 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-1:14318] MCW rank 23 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-1:14318] MCW rank 24 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-6-1:24682] MCW rank 3 bound to socket 1[core 2[hwt 0]], 
>>> >>> socket 1[core 3[hwt 0]]: [./.][B/B]
>>> >>> [csclprd3-6-1:24682] MCW rank 0 bound to socket 0[core 0[hwt 0]], 
>>> >>> socket 0[core 1[hwt 0]]: [B/B][./.]
>>> >>> [csclprd3-0-1:14318] MCW rank 25 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-1:14318] MCW rank 20 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-3:13827] MCW rank 34 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-1:14318] MCW rank 21 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-3:13827] MCW rank 35 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-6-1:24682] MCW rank 1 bound to socket 1[core 2[hwt 0]], 
>>> >>> socket 1[core 3[hwt 0]]: [./.][B/B]
>>> >>> [csclprd3-0-3:13827] MCW rank 36 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-6-1:24682] MCW rank 2 bound to socket 0[core 0[hwt 0]], 
>>> >>> socket 0[core 1[hwt 0]]: [B/B][./.]
>>> >>> [csclprd3-0-6:30371] MCW rank 51 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-6:30371] MCW rank 52 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-6:30371] MCW rank 53 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-2:05825] MCW rank 30 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-6:30371] MCW rank 54 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-3:13827] MCW rank 37 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-2:05825] MCW rank 31 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-3:13827] MCW rank 32 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-6:30371] MCW rank 55 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-3:13827] MCW rank 33 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-6:30371] MCW rank 50 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-2:05825] MCW rank 26 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-2:05825] MCW rank 27 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-2:05825] MCW rank 28 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-2:05825] MCW rank 29 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-12:12383] MCW rank 121 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-12:12383] MCW rank 122 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-12:12383] MCW rank 123 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-12:12383] MCW rank 124 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-12:12383] MCW rank 125 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-12:12383] MCW rank 120 is not bound (or bound to all 
>>> >>> available processors)
>>> >>> [csclprd3-0-0:31079] MCW rank 13 bound to socket 1[core 6[hwt 0]], 
>>> >>> socket 1[core 7[hwt 0]], socket 1[core 8[hwt 0]], socket 1[core 9[hwt 
>>> >>> 0]], socket 1[core 10[hwt 0]], socket 1[core 11[hwt 0]]: 
>>> >>> [./././././.][B/B/B/B/B/B]
>>> >>> [csclprd3-0-0:31079] MCW rank 14 bound to socket 0[core 0[hwt 0]], 
>>> >>> socket 0[core 1[hwt 0]], socket 0[core 2[hwt 0]], socket 0[core 3[hwt 
>>> >>> 0]], socket 0[core 4[hwt 0]], socket 0[core 5[hwt 0]]: 
>>> >>> [B/B/B/B/B/B][./././././.]
>>> >>> [csclprd3-0-0:31079] MCW rank 15 bound to socket 1[core 6[hwt 0]], 
>>> >>> socket 1[core 7[hwt 0]], socket 1[core 8[hwt 0]], socket 1[core 9[hwt 
>>> >>> 0]], socket 1[core 10[hwt 0]], socket 1[core 11[hwt 0]]: 
>>> >>> [./././././.][B/B/B/B/B/B]
>>> >>> [csclprd3-0-0:31079] MCW rank 16 bound to socket 0[core 0[hwt 0]], 
>>> >>> socket 0[core 1[hwt 0]], socket 0[core 2[hwt 0]], socket 0[core 3[hwt 
>>> >>> 0]], socket 0[core 4[hwt 0]], socket 0[core 5[hwt 0]]: 
>>> >>> [B/B/B/B/B/B][./././././.]
>>> >>> [csclprd3-0-7:20515] MCW rank 68 bound to socket 0[core 0[hwt 0-1]], 
>>> >>> socket 0[core 1[hwt 0-1]], socket 0[core 2[hwt 0-1]], socket 0[core 
>>> >>> 3[hwt 0-1]], socket 0[core 4[hwt 0-1]], socket 0[core 5[hwt 0-1]], 
>>> >>> socket 0[core 6[hwt 0-1]], socket 0[core 7[hwt 0-1]]: 
>>> >>> [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>> [csclprd3-0-10:19096] MCW rank 100 bound to socket 0[core 0[hwt 0-1]], 
>>> >>> socket 0[core 1[hwt 0-1]], socket 0[core 2[hwt 0-1]], socket 0[core 
>>> >>> 3[hwt 0-1]], socket 0[core 4[hwt 0-1]], socket 0[core 5[hwt 0-1]], 
>>> >>> socket 0[core 6[hwt 0-1]], socket 0[core 7[hwt 0-1]]: 
>>> >>> [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>> [csclprd3-0-7:20515] MCW rank 69 bound to socket 1[core 8[hwt 0-1]], 
>>> >>> socket 1[core 9[hwt 0-1]], socket 1[core 10[hwt 0-1]], socket 1[core 
>>> >>> 11[hwt 0-1]], socket 1[core 12[hwt 0-1]], socket 1[core 13[hwt 0-1]], 
>>> >>> socket 1[core 14[hwt 0-1]], socket 1[core 15[hwt 0-1]]: 
>>> >>> [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>> [csclprd3-0-10:19096] MCW rank 101 bound to socket 1[core 8[hwt 0-1]], 
>>> >>> socket 1[core 9[hwt 0-1]], socket 1[core 10[hwt 0-1]], socket 1[core 
>>> >>> 11[hwt 0-1]], socket 1[core 12[hwt 0-1]], socket 1[core 13[hwt 0-1]], 
>>> >>> socket 1[core 14[hwt 0-1]], socket 1[core 15[hwt 0-1]]: 
>>> >>> [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>> [csclprd3-0-0:31079] MCW rank 17 bound to socket 1[core 6[hwt 0]], 
>>> >>> socket 1[core 7[hwt 0]], socket 1[core 8[hwt 0]], socket 1[core 9[hwt 
>>> >>> 0]], socket 1[core 10[hwt 0]], socket 1[core 11[hwt 0]]: 
>>> >>> [./././././.][B/B/B/B/B/B]
>>> >>> [csclprd3-0-7:20515] MCW rank 70 bound to socket 0[core 0[hwt 0-1]], 
>>> >>> socket 0[core 1[hwt 0-1]], socket 0[core 2[hwt 0-1]], socket 0[core 
>>> >>> 3[hwt 0-1]], socket 0[core 4[hwt 0-1]], socket 0[core 5[hwt 0-1]], 
>>> >>> socket 0[core 6[hwt 0-1]], socket 0[core 7[hwt 0-1]]: 
>>> >>> [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>> [csclprd3-0-10:19096] MCW rank 102 bound to socket 0[core 0[hwt 0-1]], 
>>> >>> socket 0[core 1[hwt 0-1]], socket 0[core 2[hwt 0-1]], socket 0[core 
>>> >>> 3[hwt 0-1]], socket 0[core 4[hwt 0-1]], socket 0[core 5[hwt 0-1]], 
>>> >>> socket 0[core 6[hwt 0-1]], socket 0[core 7[hwt 0-1]]: 
>>> >>> [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>> [csclprd3-0-11:31636] MCW rank 116 bound to socket 0[core 0[hwt 0-1]], 
>>> >>> socket 0[core 1[hwt 0-1]], socket 0[core 2[hwt 0-1]], socket 0[core 
>>> >>> 3[hwt 0-1]], socket 0[core 4[hwt 0-1]], socket 0[core 5[hwt 0-1]], 
>>> >>> socket 0[core 6[hwt 0-1]], socket 0[core 7[hwt 0-1]]: 
>>> >>> [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>> [csclprd3-0-11:31636] MCW rank 117 bound to socket 1[core 8[hwt 0-1]], 
>>> >>> socket 1[core 9[hwt 0-1]], socket 1[core 10[hwt 0-1]], socket 1[core 
>>> >>> 11[hwt 0-1]], socket 1[core 12[hwt 0-1]], socket 1[core 13[hwt 0-1]], 
>>> >>> socket 1[core 14[hwt 0-1]], socket 1[core 15[hwt 0-1]]: 
>>> >>> [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>> [csclprd3-0-0:31079] MCW rank 18 bound to socket 0[core 0[hwt 0]], 
>>> >>> socket 0[core 1[hwt 0]], socket 0[core 2[hwt 0]], socket 0[core 3[hwt 
>>> >>> 0]], socket 0[core 4[hwt 0]], socket 0[core 5[hwt 0]]: 
>>> >>> [B/B/B/B/B/B][./././././.]
>>> >>> [csclprd3-0-11:31636] MCW rank 118 bound to socket 0[core 0[hwt 0-1]], 
>>> >>> socket 0[core 1[hwt 0-1]], socket 0[core 2[hwt 0-1]], socket 0[core 
>>> >>> 3[hwt 0-1]], socket 0[core 4[hwt 0-1]], socket 0[core 5[hwt 0-1]], 
>>> >>> socket 0[core 6[hwt 0-1]], socket 0[core 7[hwt 0-1]]: 
>>> >>> [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>> [csclprd3-0-0:31079] MCW rank 19 bound to socket 1[core 6[hwt 0]], 
>>> >>> socket 1[core 7[hwt 0]], socket 1[core 8[hwt 0]], socket 1[core 9[hwt 
>>> >>> 0]], socket 1[core 10[hwt 0]], socket 1[core 11[hwt 0]]: 
>>> >>> [./././././.][B/B/B/B/B/B]
>>> >>> [csclprd3-0-7:20515] MCW rank 71 bound to socket 1[core 8[hwt 0-1]], 
>>> >>> socket 1[core 9[hwt 0-1]], socket 1[core 10[hwt 0-1]], socket 1[core 
>>> >>> 11[hwt 0-1]], socket 1[core 12[hwt 0-1]], socket 1[core 13[hwt 0-1]], 
>>> >>> socket 1[core 14[hwt 0-1]], socket 1[core 15[hwt 0-1]]: 
>>> >>> [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>> [csclprd3-0-10:19096] MCW rank 103 bound to socket 1[core 8[hwt 0-1]], 
>>> >>> socket 1[core 9[hwt 0-1]], socket 1[core 10[hwt 0-1]], socket 1[core 
>>> >>> 11[hwt 0-1]], socket 1[core 12[hwt 0-1]], socket 1[core 13[hwt 0-1]], 
>>> >>> socket 1[core 14[hwt 0-1]], socket 1[core 15[hwt 0-1]]: 
>>> >>> [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>> [csclprd3-0-0:31079] MCW rank 8 bound to socket 0[core 0[hwt 0]], 
>>> >>> socket 0[core 1[hwt 0]], socket 0[core 2[hwt 0]], socket 0[core 3[hwt 
>>> >>> 0]], socket 0[core 4[hwt 0]], socket 0[core 5[hwt 0]]: 
>>> >>> [B/B/B/B/B/B][./././././.]
>>> >>> [csclprd3-0-0:31079] MCW rank 9 bound to socket 1[core 6[hwt 0]], 
>>> >>> socket 1[core 7[hwt 0]], socket 1[core 8[hwt 0]], socket 1[core 9[hwt 
>>> >>> 0]], socket 1[core 10[hwt 0]], socket 1[core 11[hwt 0]]: 
>>> >>> [./././././.][B/B/B/B/B/B]
>>> >>> [csclprd3-0-10:19096] MCW rank 88 bound to socket 0[core 0[hwt 0-1]], 
>>> >>> socket 0[core 1[hwt 0-1]], socket 0[core 2[hwt 0-1]], socket 0[core 
>>> >>> 3[hwt 0-1]], socket 0[core 4[hwt 0-1]], socket 0[core 5[hwt 0-1]], 
>>> >>> socket 0[core 6[hwt 0-1]], socket 0[core 7[hwt 0-1]]: 
>>> >>> [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>> [csclprd3-0-11:31636] MCW rank 119 bound to socket 1[core 8[hwt 0-1]], 
>>> >>> socket 1[core 9[hwt 0-1]], socket 1[core 10[hwt 0-1]], socket 1[core 
>>> >>> 11[hwt 0-1]], socket 1[core 12[hwt 0-1]], socket 1[core 13[hwt 0-1]], 
>>> >>> socket 1[core 14[hwt 0-1]], socket 1[core 15[hwt 0-1]]: 
>>> >>> [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>> [csclprd3-0-7:20515] MCW rank 56 bound to socket 0[core 0[hwt 0-1]], 
>>> >>> socket 0[core 1[hwt 0-1]], socket 0[core 2[hwt 0-1]], socket 0[core 
>>> >>> 3[hwt 0-1]], socket 0[core 4[hwt 0-1]], socket 0[core 5[hwt 0-1]], 
>>> >>> socket 0[core 6[hwt 0-1]], socket 0[core 7[hwt 0-1]]: 
>>> >>> [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>> [csclprd3-0-0:31079] MCW rank 10 bound to socket 0[core 0[hwt 0]], 
>>> >>> socket 0[core 1[hwt 0]], socket 0[core 2[hwt 0]], socket 0[core 3[hwt 
>>> >>> 0]], socket 0[core 4[hwt 0]], socket 0[core 5[hwt 0]]: 
>>> >>> [B/B/B/B/B/B][./././././.]
>>> >>> [csclprd3-0-7:20515] MCW rank 57 bound to socket 1[core 8[hwt 0-1]], 
>>> >>> socket 1[core 9[hwt 0-1]], socket 1[core 10[hwt 0-1]], socket 1[core 
>>> >>> 11[hwt 0-1]], socket 1[core 12[hwt 0-1]], socket 1[core 13[hwt 0-1]], 
>>> >>> socket 1[core 14[hwt 0-1]], socket 1[core 15[hwt 0-1]]: 
>>> >>> [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>> [csclprd3-0-10:19096] MCW rank 89 bound to socket 1[core 8[hwt 0-1]], 
>>> >>> socket 1[core 9[hwt 0-1]], socket 1[core 10[hwt 0-1]], socket 1[core 
>>> >>> 11[hwt 0-1]], socket 1[core 12[hwt 0-1]], socket 1[core 13[hwt 0-1]], 
>>> >>> socket 1[core 14[hwt 0-1]], socket 1[core 15[hwt 0-1]]: 
>>> >>> [../../../../../../../..][BB/BB/BB/BB/BB/BB/BB/BB]
>>> >>> [csclprd3-0-11:31636] MCW rank 104 bound to socket 0[core 0[hwt 0-1]], 
>>> >>> socket 0[core 1[hwt 0-1]], socket 0[core 2[hwt 0-1]], socket 0[core 
>>> >>> 3[hwt 0-1]], socket 0[core 4[hwt 0-1]], socket 0[core 5[hwt 0-1]], 
>>> >>> socket 0[core 6[hwt 0-1]], socket 0[core 7[hwt 0-1]]: 
>>> >>> [BB/BB/BB/BB/BB/BB/BB/BB][../../../../../../../..]
>>> >>> [csclprd3-0-0:31079] MCW rank 11 bound to socket 1[core 6[hwt 0]], 
>>> >>> socket 1[core 7[hwt 0]], socket 1[core 8[hwt 0]], socket 1[core 9[hwt 
>>> >>> 0]], socket 1[core 10[hwt 0]], socket 1[core 11[hwt 0]]: 
>>> >>> [./././././.][B/B/B/B/B/B]
>>> >>> [csclprd3-0-0:31079] MCW rank 12 bound to socket 0[core 0[hwt 0]], 
>>> >>> socket 0[core 1[hwt 0]], socket 0[core 2[hwt 0]], socket 0[core 3[hwt 
>>> >>> 0]], socket 0[core 4[hwt 0]], socket 0[core 5[hwt 0]]: 
>>> >>> [B/B/B/B/B/B][./././././.]
>>> >>> [csclprd3-0-4:30348] MCW rank 42 is not bound (or bound to all
>>> >>>
>>> >>> _______________________________________________
>>> >>> users mailing list
>>> >>> us...@open-mpi.org <mailto:us...@open-mpi.org>
>>> >>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users 
>>> >>> <http://www.open-mpi.org/mailman/listinfo.cgi/users>
>>> >>> Link to this post: 
>>> >>> http://www.open-mpi.org/community/lists/users/2015/06/27185.php 
>>> >>> <http://www.open-mpi.org/community/lists/users/2015/06/27185.php>
>>> >>>
>>> >>> IMPORTANT WARNING: This message is intended for the use of the person 
>>> >>> or entity to which it is addressed and may contain information that is 
>>> >>> privileged and confidential, the disclosure of which is governed by 
>>> >>> applicable law. If the reader of this message is not the intended 
>>> >>> recipient, or the employee or agent responsible for delivering it to 
>>> >>> the intended recipient, you are hereby notified that any dissemination, 
>>> >>> distribution or copying of this information is strictly prohibited. 
>>> >>> Thank you for your cooperation.
>>> >>> _______________________________________________
>>> >>> users mailing list
>>> >>> us...@open-mpi.org <mailto:us...@open-mpi.org>
>>> >>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users 
>>> >>> <http://www.open-mpi.org/mailman/listinfo.cgi/users>
>>> >>> Link to this post: 
>>> >>> http://www.open-mpi.org/community/lists/users/2015/06/27204.php 
>>> >>> <http://www.open-mpi.org/community/lists/users/2015/06/27204.php>
>>> >>
>>> >> IMPORTANT WARNING: This message is intended for the use of the person or 
>>> >> entity to which it is addressed and may contain information that is 
>>> >> privileged and confidential, the disclosure of which is governed by 
>>> >> applicable law. If the reader of this message is not the intended 
>>> >> recipient, or the employee or agent responsible for delivering it to the 
>>> >> intended recipient, you are hereby notified that any dissemination, 
>>> >> distribution or copying of this information is strictly prohibited. 
>>> >> Thank you for your cooperation.
>>> >> _______________________________________________
>>> >> users mailing list
>>> >> us...@open-mpi.org <mailto:us...@open-mpi.org>
>>> >> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users 
>>> >> <http://www.open-mpi.org/mailman/listinfo.cgi/users>
>>> >> Link to this post: 
>>> >> http://www.open-mpi.org/community/lists/users/2015/06/27220.php 
>>> >> <http://www.open-mpi.org/community/lists/users/2015/06/27220.php>
>>> >
>>> >
>>> > --
>>> > Jeff Squyres
>>> > jsquy...@cisco.com <mailto:jsquy...@cisco.com>
>>> > For corporate legal information go to: 
>>> > http://www.cisco.com/web/about/doing_business/legal/cri/ 
>>> > <http://www.cisco.com/web/about/doing_business/legal/cri/>
>>> >
>>> > _______________________________________________
>>> > users mailing list
>>> > us...@open-mpi.org <mailto:us...@open-mpi.org>
>>> > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users 
>>> > <http://www.open-mpi.org/mailman/listinfo.cgi/users>
>>> > Link to this post: 
>>> > http://www.open-mpi.org/community/lists/users/2015/06/27222.php 
>>> > <http://www.open-mpi.org/community/lists/users/2015/06/27222.php>
>>> 
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org <mailto:us...@open-mpi.org>
>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users 
>>> <http://www.open-mpi.org/mailman/listinfo.cgi/users>
>>> Link to this post: 
>>> http://www.open-mpi.org/community/lists/users/2015/07/27261.php 
>>> <http://www.open-mpi.org/community/lists/users/2015/07/27261.php>
>>> IMPORTANT WARNING: This message is intended for the use of the person or 
>>> entity to which it is addressed and may contain information that is 
>>> privileged and confidential, the disclosure of which is governed by 
>>> applicable law. If the reader of this message is not the intended 
>>> recipient, or the employee or agent responsible for delivering it to the 
>>> intended recipient, you are hereby notified that any dissemination, 
>>> distribution or copying of this information is strictly prohibited. Thank 
>>> you for your cooperation.
>>> 
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org <mailto:us...@open-mpi.org>
>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users 
>>> <http://www.open-mpi.org/mailman/listinfo.cgi/users>
>>> Link to this post: 
>>> http://www.open-mpi.org/community/lists/users/2015/07/27263.php 
>>> <http://www.open-mpi.org/community/lists/users/2015/07/27263.php>
>>> 
>>> IMPORTANT WARNING: This message is intended for the use of the person or 
>>> entity to which it is addressed and may contain information that is 
>>> privileged and confidential, the disclosure of which is governed by 
>>> applicable law. If the reader of this message is not the intended 
>>> recipient, or the employee or agent responsible for delivering it to the 
>>> intended recipient, you are hereby notified that any dissemination, 
>>> distribution or copying of this information is strictly prohibited. Thank 
>>> you for your cooperation. _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org <mailto:us...@open-mpi.org>
>>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users 
>>> <http://www.open-mpi.org/mailman/listinfo.cgi/users>
>>> Link to this post: 
>>> http://www.open-mpi.org/community/lists/users/2015/07/27279.php 
>>> <http://www.open-mpi.org/community/lists/users/2015/07/27279.php>
>> IMPORTANT WARNING: This message is intended for the use of the person or 
>> entity to which it is addressed and may contain information that is 
>> privileged and confidential, the disclosure of which is governed by 
>> applicable law. If the reader of this message is not the intended recipient, 
>> or the employee or agent responsible for delivering it to the intended 
>> recipient, you are hereby notified that any dissemination, distribution or 
>> copying of this information is strictly prohibited. Thank you for your 
>> cooperation. _______________________________________________
>> users mailing list
>> us...@open-mpi.org <mailto:us...@open-mpi.org>
>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users 
>> <http://www.open-mpi.org/mailman/listinfo.cgi/users>
>> Link to this post: 
>> http://www.open-mpi.org/community/lists/users/2015/07/27281.php 
>> <http://www.open-mpi.org/community/lists/users/2015/07/27281.php>
> IMPORTANT WARNING: This message is intended for the use of the person or 
> entity to which it is addressed and may contain information that is 
> privileged and confidential, the disclosure of which is governed by 
> applicable law. If the reader of this message is not the intended recipient, 
> or the employee or agent responsible for delivering it to the intended 
> recipient, you are hereby notified that any dissemination, distribution or 
> copying of this information is strictly prohibited. Thank you for your 
> cooperation.
> IMPORTANT WARNING: This message is intended for the use of the person or 
> entity to which it is addressed and may contain information that is 
> privileged and confidential, the disclosure of which is governed by 
> applicable law. If the reader of this message is not the intended recipient, 
> or the employee or agent responsible for delivering it to the intended 
> recipient, you are hereby notified that any dissemination, distribution or 
> copying of this information is strictly prohibited. Thank you for your 
> cooperation.<out.txt>_______________________________________________
> users mailing list
> us...@open-mpi.org <mailto:us...@open-mpi.org>
> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users 
> <http://www.open-mpi.org/mailman/listinfo.cgi/users>
> Link to this post: 
> http://www.open-mpi.org/community/lists/users/2015/07/27325.php 
> <http://www.open-mpi.org/community/lists/users/2015/07/27325.php>

Reply via email to