Brian,

i have no doubt this was enough to get rid of the warning messages.

out of curiosity, are you now able to experience performances close to
native runs ?
if i understand correctly, the linux kernel allocates memory on the
closest NUMA domain (e.g. socket if i oversimplify), and since
MPI tasks are bound by orted/mpirun before they are execv'ed, i have
some hard time understanding how not binding MPI tasks to
memory can have a significant impact on performances as long as they
are bound on cores.

Cheers,

Gilles


On Sat, Dec 23, 2017 at 7:27 AM, Brian Dobbins <bdobb...@gmail.com> wrote:
>
> Hi Ralph,
>
>   Well, this gets chalked up to user error - the default AMI images come
> without the NUMA-dev libraries, so OpenMPI didn't get built with it (and in
> my haste, I hadn't checked).  Oops.  Things seem to be working correctly
> now.
>
>   Thanks again for your help,
>   - Brian
>
>
> On Fri, Dec 22, 2017 at 2:14 PM, r...@open-mpi.org <r...@open-mpi.org> wrote:
>>
>> I honestly don’t know - will have to defer to Brian, who is likely out for
>> at least the extended weekend. I’ll point this one to him when he returns.
>>
>>
>> On Dec 22, 2017, at 1:08 PM, Brian Dobbins <bdobb...@gmail.com> wrote:
>>
>>
>>   Hi Ralph,
>>
>>   OK, that certainly makes sense - so the next question is, what prevents
>> binding memory to be local to particular cores?  Is this possible in a
>> virtualized environment like AWS HVM instances?
>>
>>   And does this apply only to dynamic allocations within an instance, or
>> static as well?  I'm pretty unfamiliar with how the hypervisor (KVM-based, I
>> believe) maps out 'real' hardware, including memory, to particular
>> instances.  We've seen some parts of the code (bandwidth heavy) run ~10x
>> faster on bare-metal hardware, though, presumably from memory locality, so
>> it certainly has a big impact.
>>
>>   Thanks again, and merry Christmas!
>>   - Brian
>>
>>
>> On Fri, Dec 22, 2017 at 1:53 PM, r...@open-mpi.org <r...@open-mpi.org>
>> wrote:
>>>
>>> Actually, that message is telling you that binding to core is available,
>>> but that we cannot bind memory to be local to that core. You can verify the
>>> binding pattern by adding --report-bindings to your cmd line.
>>>
>>>
>>> On Dec 22, 2017, at 11:58 AM, Brian Dobbins <bdobb...@gmail.com> wrote:
>>>
>>>
>>> Hi all,
>>>
>>>   We're testing a model on AWS using C4/C5 nodes and some of our timers,
>>> in a part of the code with no communication, show really poor performance
>>> compared to native runs.  We think this is because we're not binding to a
>>> core properly and thus not caching, and a quick 'mpirun --bind-to core
>>> hostname' does suggest issues with this on AWS:
>>>
>>> [bdobbins@head run]$ mpirun --bind-to core hostname
>>>
>>> --------------------------------------------------------------------------
>>> WARNING: a request was made to bind a process. While the system
>>> supports binding the process itself, at least one node does NOT
>>> support binding memory to the process location.
>>>
>>>   Node:  head
>>>
>>> Open MPI uses the "hwloc" library to perform process and memory
>>> binding. This error message means that hwloc has indicated that
>>> processor binding support is not available on this machine.
>>>
>>>   (It also happens on compute nodes, and with real executables.)
>>>
>>>   Does anyone know how to enforce binding to cores on AWS instances?  Any
>>> insight would be great.
>>>
>>>   Thanks,
>>>   - Brian
>>> _______________________________________________
>>> users mailing list
>>> users@lists.open-mpi.org
>>> https://lists.open-mpi.org/mailman/listinfo/users
>>>
>>>
>>>
>>> _______________________________________________
>>> users mailing list
>>> users@lists.open-mpi.org
>>> https://lists.open-mpi.org/mailman/listinfo/users
>>
>>
>> _______________________________________________
>> users mailing list
>> users@lists.open-mpi.org
>> https://lists.open-mpi.org/mailman/listinfo/users
>>
>>
>>
>> _______________________________________________
>> users mailing list
>> users@lists.open-mpi.org
>> https://lists.open-mpi.org/mailman/listinfo/users
>
>
>
> _______________________________________________
> users mailing list
> users@lists.open-mpi.org
> https://lists.open-mpi.org/mailman/listinfo/users
_______________________________________________
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users

Reply via email to