Jorge, I am not that familiar with UCX, but I hope that will help:
The changes I mentioned were introduced by https://github.com/open-mpi/ompi/pull/8549 I suspect mpirun --mca pml_ucx_tls any --mca pml_ucx_devices --mca pml ucx ... will do what you expect Cheers, Gilles On Mon, Sep 13, 2021 at 9:05 PM Jorge D'Elia via users < users@lists.open-mpi.org> wrote: > Dear Gilles, > > Despite my last answer (see below), I am noticing that > some tests with a coarray fortran code on a laptop show a > performance drop of the order of 20% using the 4.1.1 version > (with --mca pml ucx disabled), versus the 4.1.0 one > (with --mca pml ucx enabled). > > I would like to experiment with pml/ucx framework using the 4.1.0 > version on that laptop. Then, please, how do I manually re-enable > those providers? (e.g. perhaps, is it during the construction > stage?) or where can I find out how to do it? Thanks in advance. > > Regards. > Jorge. > > ----- Mensaje original ----- > > De: "Open MPI Users" <users@lists.open-mpi.org> > > Para: "Open MPI Users" <users@lists.open-mpi.org> > > CC: "Jorge D'Elia" > > Enviado: Sábado, 29 de Mayo 2021 7:18:23 > > Asunto: Re: [OMPI users] (Fedora 34, x86_64-pc-linux-gnu, > openmpi-4.1.1.tar.gz): PML ucx cannot be selected > > > > Dear Gilles, > > > > Ahhh ... now the new behavior is better understood. > > The intention of using pml/ucx was simply for preliminary > > testing, and does not merit re-enabling these providers in > > this notebook. > > > > Thank you very much for the clarification. > > > > Regards, > > Jorge. > > > > ----- Mensaje original ----- > >> De: "Gilles Gouaillardet" > >> Para: "Jorge D'Elia" , "Open MPI Users" <users@lists.open-mpi.org> > >> Enviado: Viernes, 28 de Mayo 2021 23:35:37 > >> Asunto: Re: [OMPI users] (Fedora 34, x86_64-pc-linux-gnu, > openmpi-4.1.1.tar.gz): > >> PML ucx cannot be selected > >> > >> Jorge, > >> > >> pml/ucx used to be selected when no fast interconnect were detected > >> (since ucx provides driver for both TCP and shared memory). > >> These providers are now disabled by default, so unless your machine > >> has a supported fast interconnect (such as Infiniband), > >> pml/ucx cannot be used out of the box anymore. > >> > >> if you really want to use pml/ucx on your notebook, you need to > >> manually re-enable these providers. > >> > >> That being said, your best choice here is really not to force any pml, > >> and let Open MPI use pml/ob1 > >> (that has support for both TCP and shared memory) > >> > >> Cheers, > >> > >> Gilles > >> > >> On Sat, May 29, 2021 at 11:19 AM Jorge D'Elia via users > >> <users@lists.open-mpi.org> wrote: > >>> > >>> Hi, > >>> > >>> We routinely build OpenMPI on x86_64-pc-linux-gnu machines from > >>> the sources using gcc and usually everything works fine. > >>> > >>> In one case we recently installed Fedora 34 from scratch on an > >>> ASUS G53SX notebook (Intel Core i7-2630QM CPU 2.00GHz ×4 cores, > >>> without any IB device). Next we build OpenMPI using the file > >>> openmpi-4.1.1.tar.gz and the GCC 12.0.0 20210524 (experimental) > >>> compiler. > >>> > >>> However, when trying to experiment OpenMPI using UCX > >>> with a simple test, we get the runtime errors: > >>> > >>> No components were able to be opened in the btl framework. > >>> PML ucx cannot be selected > >>> > >>> while the test worked fine until Fedora 33 on the same > >>> machine using the same OpenMPI configuration. > >>> > >>> We attach below some info about a simple test run. > >>> > >>> Please, any clues where to check or maybe something is missing? > >>> Thanks in advance. > >>> > >>> Regards > >>> Jorge. > >>> > >>> -- > >>> $ cat /proc/version > >>> Linux version 5.12.7-300.fc34.x86_64 > >>> (mockbu...@bkernel01.iad2.fedoraproject.org) (gcc (GCC) 11.1.1 > 20210428 (Red > >>> Hat 11.1.1-1), GNU ld version 2.35.1-41.fc34) #1 SMP Wed May 26 > 12:58:58 UTC > >>> 2021 > >>> > >>> $ mpifort --version > >>> GNU Fortran (GCC) 12.0.0 20210524 (experimental) > >>> Copyright (C) 2021 Free Software Foundation, Inc. > >>> > >>> $ which mpifort > >>> /usr/beta/openmpi/bin/mpifort > >>> > >>> $ mpifort -o hello_usempi_f08.exe hello_usempi_f08.f90 > >>> > >>> $ mpirun --mca orte_base_help_aggregate 0 --mca btl self,vader,tcp > --map-by node > >>> --report-bindings --machinefile ~/machi-openmpi.dat --np 2 > >>> hello_usempi_f08.exe > >>> [verne:200650] MCW rank 0 bound to socket 0[core 0[hwt 0]]: [B/././.] > >>> [verne:200650] MCW rank 1 bound to socket 0[core 1[hwt 0]]: [./B/./.] > >>> Hello, world, I am 0 of 2: Open MPI v4.1.1, package: Open MPI > bigpack@verne > >>> Distribution, ident: 4.1.1, repo rev: v4.1.1, Apr 24, 2021 > >>> Hello, world, I am 1 of 2: Open MPI v4.1.1, package: Open MPI > bigpack@verne > >>> Distribution, ident: 4.1.1, repo rev: v4.1.1, Apr 24, 2021 > >>> > >>> $ mpirun --mca orte_base_help_aggregate 0 --mca pml ucx --mca btl > >>> ^self,vader,tcp --map-by node --report-bindings --machinefile > >>> ~/machi-openmpi.dat --np 2 hello_usempi_f08.exe > >>> [verne:200772] MCW rank 0 bound to socket 0[core 0[hwt 0]]: [B/././.] > >>> [verne:200772] MCW rank 1 bound to socket 0[core 1[hwt 0]]: [./B/./.] > >>> > -------------------------------------------------------------------------- > >>> No components were able to be opened in the btl framework. > >>> > >>> This typically means that either no components of this type were > >>> installed, or none of the installed components can be loaded. > >>> Sometimes this means that shared libraries required by these > >>> components are unable to be found/loaded. > >>> > >>> Host: verne > >>> Framework: btl > >>> > -------------------------------------------------------------------------- > >>> > -------------------------------------------------------------------------- > >>> No components were able to be opened in the btl framework. > >>> > >>> This typically means that either no components of this type were > >>> installed, or none of the installed components can be loaded. > >>> Sometimes this means that shared libraries required by these > >>> components are unable to be found/loaded. > >>> > >>> Host: verne > >>> Framework: btl > >>> > -------------------------------------------------------------------------- > >>> > -------------------------------------------------------------------------- > >>> No components were able to be opened in the pml framework. > >>> > >>> This typically means that either no components of this type were > >>> installed, or none of the installed components can be loaded. > >>> Sometimes this means that shared libraries required by these > >>> components are unable to be found/loaded. > >>> > >>> Host: verne > >>> Framework: pml > >>> > -------------------------------------------------------------------------- > >>> [verne:200777] PML ucx cannot be selected > >>> > -------------------------------------------------------------------------- > >>> No components were able to be opened in the pml framework. > >>> > >>> This typically means that either no components of this type were > >>> installed, or none of the installed components can be loaded. > >>> Sometimes this means that shared libraries required by these > >>> components are unable to be found/loaded. > >>> > >>> Host: verne > >>> Framework: pml > >>> > -------------------------------------------------------------------------- > >>> [verne:200772] PMIX ERROR: UNREACHABLE in file > >>> > ../../../../../../../opal/mca/pmix/pmix3x/pmix/src/server/pmix_server.c at > line > >>> 2198 > >>> > >>> > >>> $ ompi_info | grep ucx > >>> Configure command line: '--enable-ipv6' '--enable-sparse-groups' > >>> '--enable-mpi-ext' '--enable-mpi-cxx' '--enable-oshmem' > >>> '--with-libevent=internal' '--with-ucx' '--with-pmix=internal' > >>> '--without-libfabric' '--prefix=/usr/beta/openmpi' > >>> MCA osc: ucx (MCA v2.1.0, API v3.0.0, Component > v4.1.1) > >>> MCA pml: ucx (MCA v2.1.0, API v2.0.0, Component > >>> v4.1.1) > >>> > >>> $ ompi_info --param all all --level 9 | grep ucx > >>> MCA osc: ucx (MCA v2.1.0, API v3.0.0, Component > v4.1.1) > >>> MCA pml: ucx (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA osc ucx: > --------------------------------------------------- > >>> MCA osc ucx: parameter "osc_ucx_priority" (current value: > "60", data source: > >>> default, level: 3 user/all, type: unsigned_int) > >>> Priority of the osc/ucx component (default: > 60) > >>> MCA osc ucx: parameter "osc_ucx_verbose" (current value: > "0", data source: > >>> default, level: 3 user/all, type: int, synonym of: > opal_common_ucx_verbose) > >>> MCA osc ucx: parameter "osc_ucx_progress_iterations" > (current value: "100", data > >>> source: default, level: 3 user/all, type: int, synonym of: > >>> opal_common_ucx_progress_iterations) > >>> MCA osc ucx: parameter "osc_ucx_opal_mem_hooks" (current > value: "false", data > >>> source: default, level: 3 user/all, type: bool, synonym > of: > >>> opal_common_ucx_opal_mem_hooks) > >>> MCA osc ucx: parameter "osc_ucx_tls" (current value: > >>> "rc_verbs,ud_verbs,rc_mlx5,dc_mlx5,cuda_ipc,rocm_ipc", > data source: default, > >>> level: 3 user/all, type: string, synonym of: > opal_common_ucx_tls) > >>> MCA osc ucx: parameter "osc_ucx_devices" (current value: > "mlx*", data source: > >>> default, level: 3 user/all, type: string, synonym of: > opal_common_ucx_devices) > >>> MCA pml ucx: > --------------------------------------------------- > >>> MCA pml ucx: parameter "pml_ucx_priority" (current value: > "51", data source: > >>> default, level: 3 user/all, type: int) > >>> MCA pml ucx: parameter "pml_ucx_num_disconnect" (current > value: "1", data > >>> source: default, level: 3 user/all, type: int) > >>> MCA pml ucx: parameter "pml_ucx_verbose" (current value: > "0", data source: > >>> default, level: 3 user/all, type: int, synonym of: > opal_common_ucx_verbose) > >>> MCA pml ucx: parameter "pml_ucx_progress_iterations" > (current value: "100", data > >>> source: default, level: 3 user/all, type: int, synonym of: > >>> opal_common_ucx_progress_iterations) > >>> MCA pml ucx: parameter "pml_ucx_opal_mem_hooks" (current > value: "false", data > >>> source: default, level: 3 user/all, type: bool, synonym > of: > >>> opal_common_ucx_opal_mem_hooks) > >>> MCA pml ucx: parameter "pml_ucx_tls" (current value: > >>> "rc_verbs,ud_verbs,rc_mlx5,dc_mlx5,cuda_ipc,rocm_ipc", > data source: default, > >>> level: 3 user/all, type: string, synonym of: > opal_common_ucx_tls) > >>> MCA pml ucx: parameter "pml_ucx_devices" (current value: > "mlx*", data source: > >>> default, level: 3 user/all, type: string, synonym of: > opal_common_ucx_devices) > >>> > >>> $ ompi_info > >>> Package: Open MPI bigpack@verne Distribution > >>> Open MPI: 4.1.1 > >>> Open MPI repo revision: v4.1.1 > >>> Open MPI release date: Apr 24, 2021 > >>> Open RTE: 4.1.1 > >>> Open RTE repo revision: v4.1.1 > >>> Open RTE release date: Apr 24, 2021 > >>> OPAL: 4.1.1 > >>> OPAL repo revision: v4.1.1 > >>> OPAL release date: Apr 24, 2021 > >>> MPI API: 3.1.0 > >>> Ident string: 4.1.1 > >>> Prefix: /usr/beta/openmpi > >>> Configured architecture: x86_64-pc-linux-gnu > >>> Configure host: verne > >>> Configured by: bigpack > >>> Configured on: Tue May 25 17:16:38 UTC 2021 > >>> Configure host: verne > >>> Configure command line: '--enable-ipv6' '--enable-sparse-groups' > >>> '--enable-mpi-ext' '--enable-mpi-cxx' > >>> '--enable-oshmem' '--with-libevent=internal' > >>> '--with-ucx' '--with-pmix=internal' > >>> '--without-libfabric' > '--prefix=/usr/beta/openmpi' > >>> Built by: bigpack > >>> Built on: Tue 25 May 17:57:46 UTC 2021 > >>> Built host: verne > >>> C bindings: yes > >>> C++ bindings: yes > >>> Fort mpif.h: yes (all) > >>> Fort use mpi: yes (full: ignore TKR) > >>> Fort use mpi size: deprecated-ompi-info-value > >>> Fort use mpi_f08: yes > >>> Fort mpi_f08 compliance: The mpi_f08 module is available, but due to > >>> limitations in the gfortran compiler and/or > Open > >>> MPI, does not support the following: array > >>> subsections, direct passthru (where > possible) to > >>> underlying Open MPI's C functionality > >>> Fort mpi_f08 subarrays: no > >>> Java bindings: no > >>> Wrapper compiler rpath: runpath > >>> C compiler: gcc > >>> C compiler absolute: /usr/beta/gcc-trunk/bin/gcc > >>> C compiler family name: GNU > >>> C compiler version: 12.0.0 > >>> C++ compiler: g++ > >>> C++ compiler absolute: /usr/beta/gcc-trunk/bin/g++ > >>> Fort compiler: gfortran > >>> Fort compiler abs: /usr/beta/gcc-trunk/bin/gfortran > >>> Fort ignore TKR: yes (!GCC$ ATTRIBUTES NO_ARG_CHECK ::) > >>> Fort 08 assumed shape: yes > >>> Fort optional args: yes > >>> Fort INTERFACE: yes > >>> Fort ISO_FORTRAN_ENV: yes > >>> Fort STORAGE_SIZE: yes > >>> Fort BIND(C) (all): yes > >>> Fort ISO_C_BINDING: yes > >>> Fort SUBROUTINE BIND(C): yes > >>> Fort TYPE,BIND(C): yes > >>> Fort T,BIND(C,name="a"): yes > >>> Fort PRIVATE: yes > >>> Fort PROTECTED: yes > >>> Fort ABSTRACT: yes > >>> Fort ASYNCHRONOUS: yes > >>> Fort PROCEDURE: yes > >>> Fort USE...ONLY: yes > >>> Fort C_FUNLOC: yes > >>> Fort f08 using wrappers: yes > >>> Fort MPI_SIZEOF: yes > >>> C profiling: yes > >>> C++ profiling: yes > >>> Fort mpif.h profiling: yes > >>> Fort use mpi profiling: yes > >>> Fort use mpi_f08 prof: yes > >>> C++ exceptions: no > >>> Thread support: posix (MPI_THREAD_MULTIPLE: yes, OPAL > support: yes, > >>> OMPI progress: no, ORTE progress: yes, Event > lib: > >>> yes) > >>> Sparse Groups: yes > >>> Internal debug support: no > >>> MPI interface warnings: yes > >>> MPI parameter check: runtime > >>> Memory profiling support: no > >>> Memory debugging support: no > >>> dl support: yes > >>> Heterogeneous support: no > >>> mpirun default --prefix: no > >>> MPI_WTIME support: native > >>> Symbol vis. support: yes > >>> Host topology support: yes > >>> IPv6 support: yes > >>> MPI1 compatibility: no > >>> MPI extensions: affinity, cuda, pcollreq > >>> FT Checkpoint support: no (checkpoint thread: no) > >>> C/R Enabled Debugging: no > >>> MPI_MAX_PROCESSOR_NAME: 256 > >>> MPI_MAX_ERROR_STRING: 256 > >>> MPI_MAX_OBJECT_NAME: 64 > >>> MPI_MAX_INFO_KEY: 36 > >>> MPI_MAX_INFO_VAL: 256 > >>> MPI_MAX_PORT_NAME: 1024 > >>> MPI_MAX_DATAREP_STRING: 128 > >>> MCA allocator: basic (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA allocator: bucket (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA backtrace: execinfo (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA btl: self (MCA v2.1.0, API v3.1.0, Component > v4.1.1) > >>> MCA btl: openib (MCA v2.1.0, API v3.1.0, Component > v4.1.1) > >>> MCA btl: tcp (MCA v2.1.0, API v3.1.0, Component > v4.1.1) > >>> MCA btl: vader (MCA v2.1.0, API v3.1.0, Component > v4.1.1) > >>> MCA compress: bzip (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA compress: gzip (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA crs: none (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA dl: dlopen (MCA v2.1.0, API v1.0.0, Component > v4.1.1) > >>> MCA event: libevent2022 (MCA v2.1.0, API v2.0.0, > Component > >>> v4.1.1) > >>> MCA hwloc: external (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA if: linux_ipv6 (MCA v2.1.0, API v2.0.0, Component > >>> v4.1.1) > >>> MCA if: posix_ipv4 (MCA v2.1.0, API v2.0.0, Component > >>> v4.1.1) > >>> MCA installdirs: env (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA installdirs: config (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA memory: patcher (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA mpool: hugepage (MCA v2.1.0, API v3.0.0, Component > v4.1.1) > >>> MCA patcher: overwrite (MCA v2.1.0, API v1.0.0, Component > >>> v4.1.1) > >>> MCA pmix: isolated (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA pmix: flux (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA pmix: pmix3x (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA pstat: linux (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA rcache: grdma (MCA v2.1.0, API v3.3.0, Component > v4.1.1) > >>> MCA reachable: weighted (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA shmem: mmap (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA shmem: posix (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA shmem: sysv (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA timer: linux (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA errmgr: default_app (MCA v2.1.0, API v3.0.0, > Component > >>> v4.1.1) > >>> MCA errmgr: default_hnp (MCA v2.1.0, API v3.0.0, > Component > >>> v4.1.1) > >>> MCA errmgr: default_orted (MCA v2.1.0, API v3.0.0, > Component > >>> v4.1.1) > >>> MCA errmgr: default_tool (MCA v2.1.0, API v3.0.0, > Component > >>> v4.1.1) > >>> MCA ess: env (MCA v2.1.0, API v3.0.0, Component > v4.1.1) > >>> MCA ess: hnp (MCA v2.1.0, API v3.0.0, Component > v4.1.1) > >>> MCA ess: pmi (MCA v2.1.0, API v3.0.0, Component > v4.1.1) > >>> MCA ess: singleton (MCA v2.1.0, API v3.0.0, Component > >>> v4.1.1) > >>> MCA ess: tool (MCA v2.1.0, API v3.0.0, Component > v4.1.1) > >>> MCA ess: slurm (MCA v2.1.0, API v3.0.0, Component > v4.1.1) > >>> MCA filem: raw (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA grpcomm: direct (MCA v2.1.0, API v3.0.0, Component > v4.1.1) > >>> MCA iof: hnp (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA iof: orted (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA iof: tool (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA odls: default (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA odls: pspawn (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA oob: tcp (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA plm: isolated (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA plm: rsh (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA plm: slurm (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA ras: simulator (MCA v2.1.0, API v2.0.0, Component > >>> v4.1.1) > >>> MCA ras: slurm (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA regx: fwd (MCA v2.1.0, API v1.0.0, Component > v4.1.1) > >>> MCA regx: naive (MCA v2.1.0, API v1.0.0, Component > v4.1.1) > >>> MCA regx: reverse (MCA v2.1.0, API v1.0.0, Component > v4.1.1) > >>> MCA rmaps: mindist (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA rmaps: ppr (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA rmaps: rank_file (MCA v2.1.0, API v2.0.0, Component > >>> v4.1.1) > >>> MCA rmaps: resilient (MCA v2.1.0, API v2.0.0, Component > >>> v4.1.1) > >>> MCA rmaps: round_robin (MCA v2.1.0, API v2.0.0, > Component > >>> v4.1.1) > >>> MCA rmaps: seq (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA rml: oob (MCA v2.1.0, API v3.0.0, Component > v4.1.1) > >>> MCA routed: binomial (MCA v2.1.0, API v3.0.0, Component > v4.1.1) > >>> MCA routed: direct (MCA v2.1.0, API v3.0.0, Component > v4.1.1) > >>> MCA routed: radix (MCA v2.1.0, API v3.0.0, Component > v4.1.1) > >>> MCA rtc: hwloc (MCA v2.1.0, API v1.0.0, Component > v4.1.1) > >>> MCA schizo: flux (MCA v2.1.0, API v1.0.0, Component > v4.1.1) > >>> MCA schizo: ompi (MCA v2.1.0, API v1.0.0, Component > v4.1.1) > >>> MCA schizo: orte (MCA v2.1.0, API v1.0.0, Component > v4.1.1) > >>> MCA schizo: jsm (MCA v2.1.0, API v1.0.0, Component > v4.1.1) > >>> MCA schizo: slurm (MCA v2.1.0, API v1.0.0, Component > v4.1.1) > >>> MCA state: app (MCA v2.1.0, API v1.0.0, Component > v4.1.1) > >>> MCA state: hnp (MCA v2.1.0, API v1.0.0, Component > v4.1.1) > >>> MCA state: novm (MCA v2.1.0, API v1.0.0, Component > v4.1.1) > >>> MCA state: orted (MCA v2.1.0, API v1.0.0, Component > v4.1.1) > >>> MCA state: tool (MCA v2.1.0, API v1.0.0, Component > v4.1.1) > >>> MCA bml: r2 (MCA v2.1.0, API v2.0.0, Component v4.1.1) > >>> MCA coll: adapt (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA coll: basic (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA coll: han (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA coll: inter (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA coll: libnbc (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA coll: self (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA coll: sm (MCA v2.1.0, API v2.0.0, Component v4.1.1) > >>> MCA coll: sync (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA coll: tuned (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA coll: monitoring (MCA v2.1.0, API v2.0.0, Component > >>> v4.1.1) > >>> MCA fbtl: posix (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA fcoll: dynamic (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA fcoll: dynamic_gen2 (MCA v2.1.0, API v2.0.0, > Component > >>> v4.1.1) > >>> MCA fcoll: individual (MCA v2.1.0, API v2.0.0, Component > >>> v4.1.1) > >>> MCA fcoll: two_phase (MCA v2.1.0, API v2.0.0, Component > >>> v4.1.1) > >>> MCA fcoll: vulcan (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA fs: ufs (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA io: ompio (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA io: romio321 (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA op: avx (MCA v2.1.0, API v1.0.0, Component > v4.1.1) > >>> MCA osc: sm (MCA v2.1.0, API v3.0.0, Component v4.1.1) > >>> MCA osc: monitoring (MCA v2.1.0, API v3.0.0, Component > >>> v4.1.1) > >>> MCA osc: pt2pt (MCA v2.1.0, API v3.0.0, Component > v4.1.1) > >>> MCA osc: rdma (MCA v2.1.0, API v3.0.0, Component > v4.1.1) > >>> MCA osc: ucx (MCA v2.1.0, API v3.0.0, Component > v4.1.1) > >>> MCA pml: v (MCA v2.1.0, API v2.0.0, Component v4.1.1) > >>> MCA pml: cm (MCA v2.1.0, API v2.0.0, Component v4.1.1) > >>> MCA pml: monitoring (MCA v2.1.0, API v2.0.0, Component > >>> v4.1.1) > >>> MCA pml: ob1 (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA pml: ucx (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA rte: orte (MCA v2.1.0, API v2.0.0, Component > v4.1.1) > >>> MCA sharedfp: individual (MCA v2.1.0, API v2.0.0, Component > >>> v4.1.1) > >>> MCA sharedfp: lockedfile (MCA v2.1.0, API v2.0.0, Component > >>> v4.1.1) > >>> MCA sharedfp: sm (MCA v2.1.0, API v2.0.0, Component v4.1.1) > >>> MCA topo: basic (MCA v2.1.0, API v2.2.0, Component > v4.1.1) > >>> MCA topo: treematch (MCA v2.1.0, API v2.2.0, Component > >>> v4.1.1) > >>> MCA vprotocol: pessimist (MCA v2.1.0, API v2.0.0, Component > >>> v4.1.1) > >>> > > > > #end >