I'll make it work so that NUMA can be either above or below socket

On Dec 20, 2013, at 2:57 AM, tmish...@jcity.maeda.co.jp wrote:

> 
> 
> Hi Brice,
> 
> Thank you for your comment. I understand what you mean.
> 
> My opinion was made just considering easy way to adjust the code for
> inversion of hierarchy in object tree.
> 
> Tetsuya Mishima
> 
> 
>> I don't think there's any such difference.
>> Also, all these NUMA architectures are reported the same by hwloc, and
>> therefore used the same in Open MPI.
>> 
>> And yes, L3 and NUMA are topologically-identical on AMD Magny-Cours (and
>> most recent AMD and Intel platforms).
>> 
>> Brice
>> 
>> 
>> 
>> Le 20/12/2013 11:33, tmish...@jcity.maeda.co.jp a écrit :
>>> 
>>> Hi Ralph,
>>> 
>>> The numa-node in AMD Mangy-Cours/Interlagos is so called cc(cache
>>> coherent)NUMA,
>>> which seems to be a little bit different from the traditional numa
> defined
>>> in openmpi.
>>> 
>>> I notice that ccNUMA object is almost same as L3cache object.
>>> So "-bind-to l3cache" or "-map-by l3cache" is valid for what I want to
> do.
>>> Therefore, "do not touch it" is one of the solution, I think ...
>>> 
>>> Anyway, mixing up these two types of numa is the problem.
>>> 
>>> Regards,
>>> Tetsuya Mishima
>>> 
>>>> I can wait it'll be fixed in 1.7.5 or later, because putting "-bind-to
>>>> numa"
>>>> and "-map-by numa" at the same time works as a workaround.
>>>> 
>>>> Thanks,
>>>> Tetsuya Mishima
>>>> 
>>>>> Yeah, it will impact everything that uses hwloc topology maps, I
> fear.
>>>>> 
>>>>> One side note: you'll need to add --hetero-nodes to your cmd line. If
>>> we
>>>> don't see that, we assume that all the node topologies are identical -
>>>> which clearly isn't true here.
>>>>> I'll try to resolve the hier inversion over the holiday - won't be
> for
>>>> 1.7.4, but hopefully for 1.7.5
>>>>> Thanks
>>>>> Ralph
>>>>> 
>>>>> On Dec 18, 2013, at 9:44 PM, tmish...@jcity.maeda.co.jp wrote:
>>>>> 
>>>>>> 
>>>>>> I think it's normal for AMD opteron having 8/16 cores such as
>>>>>> magny cours or interlagos. Because it usually has 2 numa nodes
>>>>>> in a cpu(socket), numa-node can not include a socket. This type
>>>>>> of hierarchy would be natural.
>>>>>> 
>>>>>> (node03 is Dell PowerEdge R815 and maybe quite common, I guess)
>>>>>> 
>>>>>> By the way, I think this inversion should affect rmaps_lama mapping.
>>>>>> 
>>>>>> Tetsuya Mishima
>>>>>> 
>>>>>>> Ick - yeah, that would be a problem. I haven't seen that type of
>>>>>> hierarchical inversion before - is node03 a different type of chip?
>>>>>>> Might take awhile for me to adjust the code to handle hier
>>>>>> inversion... :-(
>>>>>>> On Dec 18, 2013, at 9:05 PM, tmish...@jcity.maeda.co.jp wrote:
>>>>>>> 
>>>>>>>> 
>>>>>>>> Hi Ralph,
>>>>>>>> 
>>>>>>>> I found the reason. I attached the main part of output with 32
>>>>>>>> core node(node03) and 8 core node(node05) at the bottom.
>>>>>>>> 
>>>>>>>> From this information, socket of node03 includes numa-node.
>>>>>>>> On the other hand, numa-node of node05 includes socket.
>>>>>>>> The direction of object tree is opposite.
>>>>>>>> 
>>>>>>>> Since "-map-by socket" may be assumed as default,
>>>>>>>> for node05, "-bind-to numa and -map-by socket" means
>>>>>>>> upward search. For node03, this should be downward.
>>>>>>>> 
>>>>>>>> I guess that openmpi-1.7.4rc1 will always assume numa-node
>>>>>>>> includes socket. Is it right? Then, upward search is assumed
>>>>>>>> in orte_rmaps_base_compute_bindings even for node03 when I
>>>>>>>> put "-bind-to numa and -map-by socket" option.
>>>>>>>> 
>>>>>>>> [node03.cluster:15508] [[38286,0],0] rmaps:base:compute_usage
>>>>>>>> [node03.cluster:15508] mca:rmaps: compute bindings for job
>>> [38286,1]
>>>>>> with
>>>>>>>> policy NUMA
>>>>>>>> [node03.cluster:15508] mca:rmaps: bind upwards for job [38286,1]
>>> with
>>>>>>>> bindings NUMA
>>>>>>>> [node03.cluster:15508] [[38286,0],0] bind:upward target NUMANode
>>> type
>>>>>>>> Machine
>>>>>>>> 
>>>>>>>> That's the reason of this trouble. Therefore, adding "-map-by
> core"
>>>>>> works.
>>>>>>>> (mapping pattern seems to be strange ...)
>>>>>>>> 
>>>>>>>> [mishima@node03 demos]$ mpirun -np 8 -bind-to numa -map-by core
>>>>>>>> -report-bindings myprog
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>> Cache
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>> Cache
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>> Cache
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>>>> NUMANode
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>> Cache
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>> Cache
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>> Cache
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>>>> NUMANode
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>> Cache
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>> Cache
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>> Cache
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>>>> NUMANode
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>> Cache
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>> Cache
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>> Cache
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>>>> NUMANode
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>> Cache
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>> Cache
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>> Cache
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>>>> NUMANode
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>> Cache
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>> Cache
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>> Cache
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>>>> NUMANode
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>> Cache
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>> Cache
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>> Cache
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>>>> NUMANode
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>> Cache
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>> Cache
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>> Cache
>>>>>>>> [node03.cluster:15885] [[38679,0],0] bind:upward target NUMANode
>>> type
>>>>>>>> NUMANode
>>>>>>>> [node03.cluster:15885] MCW rank 2 bound to socket 0[core 0[hwt
> 0]],
>>>>>> socket
>>>>>>>> 0[core 1[hwt 0]], socket 0[core 2[hwt 0]], so
>>>>>>>> cket 0[core 3[hwt 0]]:
>>>>>>>> 
>>> [B/B/B/B/./././.][./././././././.][./././././././.][./././././././.]
>>>>>>>> [node03.cluster:15885] MCW rank 3 bound to socket 0[core 0[hwt
> 0]],
>>>>>> socket
>>>>>>>> 0[core 1[hwt 0]], socket 0[core 2[hwt 0]], so
>>>>>>>> cket 0[core 3[hwt 0]]:
>>>>>>>> 
>>> [B/B/B/B/./././.][./././././././.][./././././././.][./././././././.]
>>>>>>>> [node03.cluster:15885] MCW rank 4 bound to socket 0[core 4[hwt
> 0]],
>>>>>> socket
>>>>>>>> 0[core 5[hwt 0]], socket 0[core 6[hwt 0]], so
>>>>>>>> cket 0[core 7[hwt 0]]:
>>>>>>>> 
>>> [././././B/B/B/B][./././././././.][./././././././.][./././././././.]
>>>>>>>> [node03.cluster:15885] MCW rank 5 bound to socket 0[core 4[hwt
> 0]],
>>>>>> socket
>>>>>>>> 0[core 5[hwt 0]], socket 0[core 6[hwt 0]], so
>>>>>>>> cket 0[core 7[hwt 0]]:
>>>>>>>> 
>>> [././././B/B/B/B][./././././././.][./././././././.][./././././././.]
>>>>>>>> [node03.cluster:15885] MCW rank 6 bound to socket 0[core 4[hwt
> 0]],
>>>>>> socket
>>>>>>>> 0[core 5[hwt 0]], socket 0[core 6[hwt 0]], so
>>>>>>>> cket 0[core 7[hwt 0]]:
>>>>>>>> 
>>> [././././B/B/B/B][./././././././.][./././././././.][./././././././.]
>>>>>>>> [node03.cluster:15885] MCW rank 7 bound to socket 0[core 4[hwt
> 0]],
>>>>>> socket
>>>>>>>> 0[core 5[hwt 0]], socket 0[core 6[hwt 0]], so
>>>>>>>> cket 0[core 7[hwt 0]]:
>>>>>>>> 
>>> [././././B/B/B/B][./././././././.][./././././././.][./././././././.]
>>>>>>>> [node03.cluster:15885] MCW rank 0 bound to socket 0[core 0[hwt
> 0]],
>>>>>> socket
>>>>>>>> 0[core 1[hwt 0]], socket 0[core 2[hwt 0]], so
>>>>>>>> cket 0[core 3[hwt 0]]:
>>>>>>>> 
>>> [B/B/B/B/./././.][./././././././.][./././././././.][./././././././.]
>>>>>>>> [node03.cluster:15885] MCW rank 1 bound to socket 0[core 0[hwt
> 0]],
>>>>>> socket
>>>>>>>> 0[core 1[hwt 0]], socket 0[core 2[hwt 0]], so
>>>>>>>> cket 0[core 3[hwt 0]]:
>>>>>>>> 
>>> [B/B/B/B/./././.][./././././././.][./././././././.][./././././././.]
>>>>>>>> Hello world from process 6 of 8
>>>>>>>> Hello world from process 5 of 8
>>>>>>>> Hello world from process 0 of 8
>>>>>>>> Hello world from process 7 of 8
>>>>>>>> Hello world from process 3 of 8
>>>>>>>> Hello world from process 4 of 8
>>>>>>>> Hello world from process 2 of 8
>>>>>>>> Hello world from process 1 of 8
>>>>>>>> 
>>>>>>>> Regards,
>>>>>>>> Tetsuya Mishima
>>>>>>>> 
>>>>>>>> [node03.cluster:15508] Type: Machine Number of child objects: 4
>>>>>>>>      Name=NULL
>>>>>>>>      total=132358820KB
>>>>>>>>      Backend=Linux
>>>>>>>>      OSName=Linux
>>>>>>>>      OSRelease=2.6.18-308.16.1.el5
>>>>>>>>      OSVersion="#1 SMP Tue Oct 2 22:01:43 EDT 2012"
>>>>>>>>      Architecture=x86_64
>>>>>>>>      Cpuset:  0xffffffff
>>>>>>>>      Online:  0xffffffff
>>>>>>>>      Allowed: 0xffffffff
>>>>>>>>      Bind CPU proc:   TRUE
>>>>>>>>      Bind CPU thread: TRUE
>>>>>>>>      Bind MEM proc:   FALSE
>>>>>>>>      Bind MEM thread: TRUE
>>>>>>>>      Type: Socket Number of child objects: 2
>>>>>>>>              Name=NULL
>>>>>>>>              total=33071780KB
>>>>>>>>              CPUModel="AMD Opteron(tm) Processor 6136"
>>>>>>>>              Cpuset:  0x000000ff
>>>>>>>>              Online:  0x000000ff
>>>>>>>>              Allowed: 0x000000ff
>>>>>>>>              Type: NUMANode Number of child objects: 1
>>>>>>>> 
>>>>>>>> 
>>>>>>>> [node05.cluster:21750] Type: Machine Number of child objects: 2
>>>>>>>>      Name=NULL
>>>>>>>>      total=33080072KB
>>>>>>>>      Backend=Linux
>>>>>>>>      OSName=Linux
>>>>>>>>      OSRelease=2.6.18-308.16.1.el5
>>>>>>>>      OSVersion="#1 SMP Tue Oct 2 22:01:43 EDT 2012"
>>>>>>>>      Architecture=x86_64
>>>>>>>>      Cpuset:  0x000000ff
>>>>>>>>      Online:  0x000000ff
>>>>>>>>      Allowed: 0x000000ff
>>>>>>>>      Bind CPU proc:   TRUE
>>>>>>>>      Bind CPU thread: TRUE
>>>>>>>>      Bind MEM proc:   FALSE
>>>>>>>>      Bind MEM thread: TRUE
>>>>>>>>      Type: NUMANode Number of child objects: 1
>>>>>>>>              Name=NULL
>>>>>>>>              local=16532232KB
>>>>>>>>              total=16532232KB
>>>>>>>>              Cpuset:  0x0000000f
>>>>>>>>              Online:  0x0000000f
>>>>>>>>              Allowed: 0x0000000f
>>>>>>>>              Type: Socket Number of child objects: 1
>>>>>>>> 
>>>>>>>> 
>>>>>>>>> Hmm...try adding "-mca rmaps_base_verbose 10 -mca
> ess_base_verbose
>>>> 5"
>>>>>> to
>>>>>>>> your cmd line and let's see what it thinks it found.
>>>>>>>>> 
>>>>>>>>> On Dec 18, 2013, at 6:55 PM, tmish...@jcity.maeda.co.jp wrote:
>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>>> Hi, I report one more problem with openmpi-1.7.4rc1,
>>>>>>>>>> which is more serious.
>>>>>>>>>> 
>>>>>>>>>> For our 32 core nodes(AMD magny cours based) which has
>>>>>>>>>> 8 numa-nodes, "-bind-to numa" does not work. Without
>>>>>>>>>> this option, it works. For your infomation, at the
>>>>>>>>>> bottom of this mail, I added the lstopo information
>>>>>>>>>> of the node.
>>>>>>>>>> 
>>>>>>>>>> Regards,
>>>>>>>>>> Tetsuya Mishima
>>>>>>>>>> 
>>>>>>>>>> [mishima@manage ~]$ qsub -I -l nodes=1:ppn=32
>>>>>>>>>> qsub: waiting for job 8352.manage.cluster to start
>>>>>>>>>> qsub: job 8352.manage.cluster ready
>>>>>>>>>> 
>>>>>>>>>> [mishima@node03 demos]$ mpirun -np 8 -report-bindings -bind-to
>>> numa
>>>>>>>> myprog
>>>>>>>>>> [node03.cluster:15316] [[37582,0],0] bind:upward target NUMANode
>>>> type
>>>>>>>>>> Machine
>>>>>>>>>> 
>>> 
> --------------------------------------------------------------------------
>>>>>>>>>> A request was made to bind to NUMA, but an appropriate target
>>> could
>>>>>> not
>>>>>>>>>> be found on node node03.
>>>>>>>>>> 
>>> 
> --------------------------------------------------------------------------
>>>>>>>>>> [mishima@node03 ~]$ cd ~/Desktop/openmpi-1.7/demos/
>>>>>>>>>> [mishima@node03 demos]$ mpirun -np 8 -report-bindings myprog
>>>>>>>>>> [node03.cluster:15282] MCW rank 2 bound to socket 1[core 8[hwt
>>> 0]]:
>>>>>>>>>> [./././././././.][B/././././././.][./././././././.][
>>>>>>>>>> ./././././././.]
>>>>>>>>>> [node03.cluster:15282] MCW rank 3 bound to socket 1[core 9[hwt
>>> 0]]:
>>>>>>>>>> [./././././././.][./B/./././././.][./././././././.][
>>>>>>>>>> ./././././././.]
>>>>>>>>>> [node03.cluster:15282] MCW rank 4 bound to socket 2[core 16[hwt
>>>> 0]]:
>>>>>>>>>> [./././././././.][./././././././.][B/././././././.]
>>>>>>>>>> [./././././././.]
>>>>>>>>>> [node03.cluster:15282] MCW rank 5 bound to socket 2[core 17[hwt
>>>> 0]]:
>>>>>>>>>> [./././././././.][./././././././.][./B/./././././.]
>>>>>>>>>> [./././././././.]
>>>>>>>>>> [node03.cluster:15282] MCW rank 6 bound to socket 3[core 24[hwt
>>>> 0]]:
>>>>>>>>>> [./././././././.][./././././././.][./././././././.]
>>>>>>>>>> [B/././././././.]
>>>>>>>>>> [node03.cluster:15282] MCW rank 7 bound to socket 3[core 25[hwt
>>>> 0]]:
>>>>>>>>>> [./././././././.][./././././././.][./././././././.]
>>>>>>>>>> [./B/./././././.]
>>>>>>>>>> [node03.cluster:15282] MCW rank 0 bound to socket 0[core 0[hwt
>>> 0]]:
>>>>>>>>>> [B/././././././.][./././././././.][./././././././.][
>>>>>>>>>> ./././././././.]
>>>>>>>>>> [node03.cluster:15282] MCW rank 1 bound to socket 0[core 1[hwt
>>> 0]]:
>>>>>>>>>> [./B/./././././.][./././././././.][./././././././.][
>>>>>>>>>> ./././././././.]
>>>>>>>>>> Hello world from process 2 of 8
>>>>>>>>>> Hello world from process 5 of 8
>>>>>>>>>> Hello world from process 4 of 8
>>>>>>>>>> Hello world from process 3 of 8
>>>>>>>>>> Hello world from process 1 of 8
>>>>>>>>>> Hello world from process 7 of 8
>>>>>>>>>> Hello world from process 6 of 8
>>>>>>>>>> Hello world from process 0 of 8
>>>>>>>>>> [mishima@node03 demos]$ ~/opt/hwloc/bin/lstopo-no-graphics
>>>>>>>>>> Machine (126GB)
>>>>>>>>>> Socket L#0 (32GB)
>>>>>>>>>> NUMANode L#0 (P#0 16GB) + L3 L#0 (5118KB)
>>>>>>>>>>   L2 L#0 (512KB) + L1d L#0 (64KB) + L1i L#0 (64KB) + Core L#0 +
>>> PU
>>>>>>>> L#0
>>>>>>>>>> (P#0)
>>>>>>>>>>   L2 L#1 (512KB) + L1d L#1 (64KB) + L1i L#1 (64KB) + Core L#1 +
>>> PU
>>>>>>>> L#1
>>>>>>>>>> (P#1)
>>>>>>>>>>   L2 L#2 (512KB) + L1d L#2 (64KB) + L1i L#2 (64KB) + Core L#2 +
>>> PU
>>>>>>>> L#2
>>>>>>>>>> (P#2)
>>>>>>>>>>   L2 L#3 (512KB) + L1d L#3 (64KB) + L1i L#3 (64KB) + Core L#3 +
>>> PU
>>>>>>>> L#3
>>>>>>>>>> (P#3)
>>>>>>>>>> NUMANode L#1 (P#1 16GB) + L3 L#1 (5118KB)
>>>>>>>>>>   L2 L#4 (512KB) + L1d L#4 (64KB) + L1i L#4 (64KB) + Core L#4 +
>>> PU
>>>>>>>> L#4
>>>>>>>>>> (P#4)
>>>>>>>>>>   L2 L#5 (512KB) + L1d L#5 (64KB) + L1i L#5 (64KB) + Core L#5 +
>>> PU
>>>>>>>> L#5
>>>>>>>>>> (P#5)
>>>>>>>>>>   L2 L#6 (512KB) + L1d L#6 (64KB) + L1i L#6 (64KB) + Core L#6 +
>>> PU
>>>>>>>> L#6
>>>>>>>>>> (P#6)
>>>>>>>>>>   L2 L#7 (512KB) + L1d L#7 (64KB) + L1i L#7 (64KB) + Core L#7 +
>>> PU>>>>>> L#7
>>>>>>>>>> (P#7)
>>>>>>>>>> Socket L#1 (32GB)
>>>>>>>>>> NUMANode L#2 (P#6 16GB) + L3 L#2 (5118KB)
>>>>>>>>>>   L2 L#8 (512KB) + L1d L#8 (64KB) + L1i L#8 (64KB) + Core L#8 +
>>> PU
>>>>>>>> L#8
>>>>>>>>>> (P#8)
>>>>>>>>>>   L2 L#9 (512KB) + L1d L#9 (64KB) + L1i L#9 (64KB) + Core L#9 +
>>> PU
>>>>>>>> L#9
>>>>>>>>>> (P#9)
>>>>>>>>>>   L2 L#10 (512KB) + L1d L#10 (64KB) + L1i L#10 (64KB) + Core
>>> L#10
>>>> +
>>>>>>>> PU
>>>>>>>>>> L#10 (P#10)
>>>>>>>>>>   L2 L#11 (512KB) + L1d L#11 (64KB) + L1i L#11 (64KB) + Core
>>> L#11
>>>> +
>>>>>>>> PU
>>>>>>>>>> L#11 (P#11)
>>>>>>>>>> NUMANode L#3 (P#7 16GB) + L3 L#3 (5118KB)
>>>>>>>>>>   L2 L#12 (512KB) + L1d L#12 (64KB) + L1i L#12 (64KB) + Core
>>> L#12
>>>> +
>>>>>>>> PU
>>>>>>>>>> L#12 (P#12)
>>>>>>>>>>   L2 L#13 (512KB) + L1d L#13 (64KB) + L1i L#13 (64KB) + Core
>>> L#13
>>>> +
>>>>>>>> PU
>>>>>>>>>> L#13 (P#13)
>>>>>>>>>>   L2 L#14 (512KB) + L1d L#14 (64KB) + L1i L#14 (64KB) + Core
>>> L#14
>>>> +
>>>>>>>> PU
>>>>>>>>>> L#14 (P#14)
>>>>>>>>>>   L2 L#15 (512KB) + L1d L#15 (64KB) + L1i L#15 (64KB) + Core
>>> L#15
>>>> +
>>>>>>>> PU
>>>>>>>>>> L#15 (P#15)
>>>>>>>>>> Socket L#2 (32GB)
>>>>>>>>>> NUMANode L#4 (P#4 16GB) + L3 L#4 (5118KB)
>>>>>>>>>>   L2 L#16 (512KB) + L1d L#16 (64KB) + L1i L#16 (64KB) + Core
>>> L#16
>>>> +
>>>>>>>> PU
>>>>>>>>>> L#16 (P#16)
>>>>>>>>>>   L2 L#17 (512KB) + L1d L#17 (64KB) + L1i L#17 (64KB) + Core
>>> L#17
>>>> +
>>>>>>>> PU
>>>>>>>>>> L#17 (P#17)> >>>>>    L2 L#18 (512KB) + L1d L#18 (64KB) + L1i
>>> L#18 (64KB) + Core L#18
>>>> +
>>>>>>>> PU
>>>>>>>>>> L#18 (P#18)
>>>>>>>>>>   L2 L#19 (512KB) + L1d L#19 (64KB) + L1i L#19 (64KB) + Core
>>> L#19
>>>> +
>>>>>>>> PU
>>>>>>>>>> L#19 (P#19)
>>>>>>>>>> NUMANode L#5 (P#5 16GB) + L3 L#5 (5118KB)
>>>>>>>>>>   L2 L#20 (512KB) + L1d L#20 (64KB) + L1i L#20 (64KB) + Core
>>> L#20
>>>> +
>>>>>>>> PU
>>>>>>>>>> L#20 (P#20)
>>>>>>>>>>   L2 L#21 (512KB) + L1d L#21 (64KB) + L1i L#21 (64KB) + Core
>>> L#21
>>>> +
>>>>>>>> PU
>>>>>>>>>> L#21 (P#21)
>>>>>>>>>>   L2 L#22 (512KB) + L1d L#22 (64KB) + L1i L#22 (64KB) + Core
>>> L#22
>>>> +
>>>>>>>> PU
>>>>>>>>>> L#22 (P#22)
>>>>>>>>>>   L2 L#23 (512KB) + L1d L#23 (64KB) + L1i L#23 (64KB) + Core
>>> L#23
>>>> +
>>>>>>>> PU
>>>>>>>>>> L#23 (P#23)
>>>>>>>>>> Socket L#3 (32GB)
>>>>>>>>>> NUMANode L#6 (P#2 16GB) + L3 L#6 (5118KB)
>>>>>>>>>>   L2 L#24 (512KB) + L1d L#24 (64KB) + L1i L#24 (64KB) + Core
>>> L#24
>>>> +
>>>>>>>> PU
>>>>>>>>>> L#24 (P#24)>>>>>    L2 L#25 (512KB) + L1d L#25 (64KB) + L1i L#25
>>>> (64KB) + Core L#25 +
>>>>>>>> PU
>>>>>>>>>> L#25 (P#25)
>>>>>>>>>>   L2 L#26 (512KB) + L1d L#26 (64KB) + L1i L#26 (64KB) + Core
>>> L#26
>>>> +
>>>>>>>> PU
>>>>>>>>>> L#26 (P#26)
>>>>>>>>>>   L2 L#27 (512KB) + L1d L#27 (64KB) + L1i L#27 (64KB) + Core
>>> L#27
>>>> +
>>>>>>>> PU
>>>>>>>>>> L#27 (P#27)
>>>>>>>>>> NUMANode L#7 (P#3 16GB) + L3 L#7 (5118KB)
>>>>>>>>>>   L2 L#28 (512KB) + L1d L#28 (64KB) + L1i L#28 (64KB) + Core
>>> L#28
>>>> +
>>>>>>>> PU
>>>>>>>>>> L#28 (P#28)
>>>>>>>>>>   L2 L#29 (512KB) + L1d L#29 (64KB) + L1i L#29 (64KB) + Core
>>> L#29
>>>> +
>>>>>>>> PU
>>>>>>>>>> L#29 (P#29)
>>>>>>>>>>   L2 L#30 (512KB) + L1d L#30 (64KB) + L1i L#30 (64KB) + Core
>>> L#30
>>>> +
>>>>>>>> PU
>>>>>>>>>> L#30 (P#30)
>>>>>>>>>>   L2 L#31 (512KB) + L1d L#31 (64KB) + L1i L#31 (64KB) + Core
>>> L#31
>>>> +
>>>>>>>> PU
>>>>>>>>>> L#31 (P#31)
>>>>>>>>>> HostBridge L#0
>>>>>>>>>> PCIBridge
>>>>>>>>>>   PCI 14e4:1639
>>>>>>>>>>     Net L#0 "eth0"
>>>>>>>>>>   PCI 14e4:1639
>>>>>>>>>>     Net L#1 "eth1"
>>>>>>>>>> PCIBridge
>>>>>>>>>>   PCI 14e4:1639
>>>>>>>>>>     Net L#2 "eth2"
>>>>>>>>>>   PCI 14e4:1639
>>>>>>>>>>     Net L#3 "eth3"
>>>>>>>>>> PCIBridge
>>>>>>>>>>   PCIBridge
>>>>>>>>>>     PCIBridge
>>>>>>>>>>       PCI 1000:0072
>>>>>>>>>>         Block L#4 "sdb"
>>>>>>>>>>         Block L#5 "sda"
>>>>>>>>>> PCI 1002:4390
>>>>>>>>>>   Block L#6 "sr0"
>>>>>>>>>> PCIBridge
>>>>>>>>>>   PCI 102b:0532
>>>>>>>>>> HostBridge L#7
>>>>>>>>>> PCIBridge
>>>>>>>>>>   PCI 15b3:6274
>>>>>>>>>>     Net L#7 "ib0"
>>>>>>>>>>     OpenFabrics L#8 "mthca0"
>>>>>>>>>> 
>>>>>>>>>> _______________________________________________
>>>>>>>>>> users mailing list
>>>>>>>>>> us...@open-mpi.org
>>>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>>> _______________________________________________
>>>>>>>>> users mailing list
>>>>>>>>> us...@open-mpi.org>>
>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>>> _______________________________________________
>>>>>>>> users mailing list
>>>>>>>> us...@open-mpi.org
>>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>>> _______________________________________________
>>>>>>> users mailing list
>>>>>>> us...@open-mpi.org
>>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>>> _______________________________________________
>>>>>> users mailing list
>>>>>> us...@open-mpi.org
>>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>>> _______________________________________________
>>>>> users mailing list
>>>>> us...@open-mpi.org
>>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>>> _______________________________________________
>>>> users mailing list
>>>> us...@open-mpi.org
>>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>>> _______________________________________________
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
>> _______________________________________________
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users

Reply via email to