Were using MLNX_OFED 4.7.3. It supplies UCX 1.7.0.
We have OpenMPI 4.02 compiled against the Mellanox OFED 4.7.3 provided versions of UCX, KNEM and
HCOLL, along with HWLOC 2.1.0 from the OpenMPI site.
I mirrored the build to be what Mellanox used to configure OpenMPI in HPC-X 2.5.
I have users using GCC, PGI, Intel and AOCC compilers with this config. PGI was the only one that
was a challenge to build due to conflicts with HCOLL.
-Ray Muno
On 2/7/20 10:04 AM, Michael Di Domenico via users wrote:
i haven't compiled openmpi in a while, but i'm in the process of
upgrading our cluster.
the last time i did this there were specific versions of mpi/pmix/ucx
that were all tested and supposed to work together. my understanding
of this was because pmi/ucx was under rapid development and the api's
were changing
is that still an issue or can i take the latest stable branches from
git for each and have a relatively good shot at it all working
together?
the one semi-immovable i have right now is ucx which is at 1.7.0 as
installed by mellanox ofed. if the above is true, is there a matrix
of versions i should be using for all the others? nothing jumped out
at me on the openmpi website
--
Ray Muno
IT Manager
e-mail: m...@aem.umn.edu
University of Minnesota
Aerospace Engineering and Mechanics Mechanical Engineering