Dear PETSc team,

A few years ago we were having some issue with MPI communications with large 
numbers of processes and subcomms, see this thread here:

https://urldefense.us/v3/__https://lists.mcs.anl.gov/mailman/htdig/petsc-users/2020-April/040976.html__;!!G_uCfscf7eWS!fyPrzMKC4KZmxGO-HI0xUlOCbgwXod4O8q2h_6MjHqPLPj9ppLkgFkJUig-KqXgu6AX7pMhYtEpWOP_cCesCWcCk_Q$
 

We are once again encountering strange issues when running our code on a new 
cluster and after a month of various tests we have not found a solution, but we 
think it has something to do with network traffic and high MPI communications, 
similar perhaps to the thread from 3 years ago.

Is it still possible to change the communication pattern with the option 
-build_twosided_allreduce (and is that the right syntax?).

Are there other runtime options that we can try to change the MPI communication 
type for all underlying communications?


Thank you,

Randy M.


Reply via email to