1) is correct. coll/sm is disqualified if the communicator is an inter communicator or the communicator spans on several nodes.

you can have a look at the source code, and you will not that bcast does not use send/recv. instead, it uses a shared memory, so hopefully, it is faster than other modules


Cheers,


Gilles

On 6/30/2016 3:04 PM, Saliya Ekanayake wrote:
Hi,

Looking at the *ompi/mca/coll/sm/coll_sm_module.c* it seems this module will be used only if the calling communicator solely groups processes within a node. I've got two questions here.

1. So is my understanding correct that for something like MPI_COMM_WORLD where world is multiple processes within a node across many nodes, this module will not be used?

2. If 1, is correct then are there any shared memory optimizations that happen when a collective like bcast or allreduce is called, so that communicating within a node is done efficiently through memory?

Thank you,
Saliya


--
Saliya Ekanayake
Ph.D. Candidate | Research Assistant
School of Informatics and Computing | Digital Science Center
Indiana University, Bloomington



_______________________________________________
users mailing list
us...@open-mpi.org
Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/users
Link to this post: 
http://www.open-mpi.org/community/lists/users/2016/06/29564.php

Reply via email to