I honestly don’t know - will have to defer to Brian, who is likely out for at least the extended weekend. I’ll point this one to him when he returns.
> On Dec 22, 2017, at 1:08 PM, Brian Dobbins <bdobb...@gmail.com> wrote: > > > Hi Ralph, > > OK, that certainly makes sense - so the next question is, what prevents > binding memory to be local to particular cores? Is this possible in a > virtualized environment like AWS HVM instances? > > And does this apply only to dynamic allocations within an instance, or > static as well? I'm pretty unfamiliar with how the hypervisor (KVM-based, I > believe) maps out 'real' hardware, including memory, to particular instances. > We've seen some parts of the code (bandwidth heavy) run ~10x faster on > bare-metal hardware, though, presumably from memory locality, so it certainly > has a big impact. > > Thanks again, and merry Christmas! > - Brian > > > On Fri, Dec 22, 2017 at 1:53 PM, r...@open-mpi.org <mailto:r...@open-mpi.org> > <r...@open-mpi.org <mailto:r...@open-mpi.org>> wrote: > Actually, that message is telling you that binding to core is available, but > that we cannot bind memory to be local to that core. You can verify the > binding pattern by adding --report-bindings to your cmd line. > > >> On Dec 22, 2017, at 11:58 AM, Brian Dobbins <bdobb...@gmail.com >> <mailto:bdobb...@gmail.com>> wrote: >> >> >> Hi all, >> >> We're testing a model on AWS using C4/C5 nodes and some of our timers, in >> a part of the code with no communication, show really poor performance >> compared to native runs. We think this is because we're not binding to a >> core properly and thus not caching, and a quick 'mpirun --bind-to core >> hostname' does suggest issues with this on AWS: >> >> [bdobbins@head run]$ mpirun --bind-to core hostname >> -------------------------------------------------------------------------- >> WARNING: a request was made to bind a process. While the system >> supports binding the process itself, at least one node does NOT >> support binding memory to the process location. >> >> Node: head >> >> Open MPI uses the "hwloc" library to perform process and memory >> binding. This error message means that hwloc has indicated that >> processor binding support is not available on this machine. >> >> (It also happens on compute nodes, and with real executables.) >> >> Does anyone know how to enforce binding to cores on AWS instances? Any >> insight would be great. >> >> Thanks, >> - Brian >> _______________________________________________ >> users mailing list >> users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> >> https://lists.open-mpi.org/mailman/listinfo/users >> <https://lists.open-mpi.org/mailman/listinfo/users> > > _______________________________________________ > users mailing list > users@lists.open-mpi.org <mailto:users@lists.open-mpi.org> > https://lists.open-mpi.org/mailman/listinfo/users > <https://lists.open-mpi.org/mailman/listinfo/users> > > _______________________________________________ > users mailing list > users@lists.open-mpi.org > https://lists.open-mpi.org/mailman/listinfo/users
_______________________________________________ users mailing list users@lists.open-mpi.org https://lists.open-mpi.org/mailman/listinfo/users