On 10/20/2017 12:24 PM, Dave Love wrote:
> Paul Kapinos <kapi...@itc.rwth-aachen.de> writes:
> 
>> Hi all,
>> sorry for the long long latency - this message was buried in my mailbox for
>> months....
>>
>>
>>
>> On 03/16/2017 10:35 AM, Alfio Lazzaro wrote:
>>> Hello Dave and others,
>>> we jump in the discussion as CP2K developers.
>>> We would like to ask you which version of CP2K you are using in your tests 
>> version 4.1 (release)
>>
>>> and
>>> if you can share with us your input file and output log.
>>
>> The input file is property of Mathias Schumacher (CC:) and we need a 
>> permission
>> of him to provide it.
> 
> I lost track of this, but the problem went away using libfabric instead
> of openib, so I left it at that, though libfabric hurt IMB pingpong
> latency compared with openib.
> 
> I seem to remember there's a workaround in the cp2k development source,
> but that obviously doesn't solve the general problem.
> 

The issue has two facing pages:
- CP2K used a lot of MPI_Alloc_mem() / MPI_Free_mem() calls. This was addressed
by CP2K developers, (private mail by Alfio Lazzaro):
> in the new CP2K release (next week will have version 5.1), I have reduced the
> amount of MPI allocations. I have also added a flag to avoid any MPI
> allocations, that you can add in the CP2K input file:
> &GLOBAL
>  &DBCSR
>     use_mpi_allocator F
>  &END
> &END GLOBAL
We will test the new release after it is available. (Note, the user still has to
think on disabling use_mpi_allocator, if in doubt.)



- in Open MPI compiled for both(1) IB and OPA, on a node with OPA, using
*default* configuration (failback to 'openib' *not prohibited*), MPI_Free_mem()
calls suddenly lasts 10000 or so times longer, starting to dominate the
application run time.
Known workaround: prohibit the failback to 'openib' BTL by '-mca btl
^tcp,openib' - that's what we implemented.

It's up to Open MPI developers if they would like to follow-up this 'small'
performance issue.

Best,
Paul Kapinos

P.S. It was a hard work even to locate this issue, as only 2 tools (of 7 or 8
tried) were able to point to the evil call...

(1) yes we use the same Open MPI installation on islands with InfiniBand, with
OmniPath, and even on ethernet-only.


-- 
Dipl.-Inform. Paul Kapinos   -   High Performance Computing,
RWTH Aachen University, IT Center
Seffenter Weg 23,  D 52074  Aachen (Germany)
Tel: +49 241/80-24915

Attachment: smime.p7s
Description: S/MIME Cryptographic Signature

_______________________________________________
users mailing list
users@lists.open-mpi.org
https://lists.open-mpi.org/mailman/listinfo/users

Reply via email to