Hi Gilles, in attach the outpuf of:

mpirun --mca btl_base_verbose 100 -np 2 ...

2017-05-19 9:43 GMT+02:00 Gilles Gouaillardet <gil...@rist.or.jp>:

> Gabriele,
>
>
> can you
>
> mpirun --mca btl_base_verbose 100 -np 2 ...
>
>
> so we can figure out why nor sm nor vader is used ?
>
>
> Cheers,
>
>
> Gilles
>
>
>
> On 5/19/2017 4:23 PM, Gabriele Fatigati wrote:
>
>> Oh no, by using two procs:
>>
>>
>> findActiveDevices Error
>> We found no active IB device ports
>> findActiveDevices Error
>> We found no active IB device ports
>> ------------------------------------------------------------
>> --------------
>> At least one pair of MPI processes are unable to reach each other for
>> MPI communications.  This means that no Open MPI device has indicated
>> that it can be used to communicate between these processes.  This is
>> an error; Open MPI requires that all MPI processes be able to reach
>> each other.  This error can sometimes be the result of forgetting to
>> specify the "self" BTL.
>>
>>   Process 1 ([[12380,1],0]) is on host: openpower
>>   Process 2 ([[12380,1],1]) is on host: openpower
>>   BTLs attempted: self
>>
>> Your MPI job is now going to abort; sorry.
>> ------------------------------------------------------------
>> --------------
>> *** An error occurred in MPI_Init
>> *** on a NULL communicator
>> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
>> ***    and potentially your MPI job)
>> *** An error occurred in MPI_Init
>> *** on a NULL communicator
>> *** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
>> ***    and potentially your MPI job)
>> ------------------------------------------------------------
>> --------------
>> MPI_INIT has failed because at least one MPI process is unreachable
>> from another.  This *usually* means that an underlying communication
>> plugin -- such as a BTL or an MTL -- has either not loaded or not
>> allowed itself to be used.  Your MPI job will now abort.
>>
>> You may wish to try to narrow down the problem;
>>  * Check the output of ompi_info to see which BTL/MTL plugins are
>>    available.
>>  * Run your application with MPI_THREAD_SINGLE.
>>  * Set the MCA parameter btl_base_verbose to 100 (or mtl_base_verbose,
>>    if using MTL-based communications) to see exactly which
>>    communication plugins were considered and/or discarded.
>> ------------------------------------------------------------
>> --------------
>> [openpower:88867] 1 more process has sent help message
>> help-mca-bml-r2.txt / unreachable proc
>> [openpower:88867] Set MCA parameter "orte_base_help_aggregate" to 0 to
>> see all help / error messages
>> [openpower:88867] 1 more process has sent help message
>> help-mpi-runtime.txt / mpi_init:startup:pml-add-procs-fail
>>
>>
>>
>>
>>
>> 2017-05-19 9:22 GMT+02:00 Gabriele Fatigati <g.fatig...@cineca.it
>> <mailto:g.fatig...@cineca.it>>:
>>
>>     Hi GIlles,
>>
>>     using your command with one MPI procs I get:
>>
>>     findActiveDevices Error
>>     We found no active IB device ports
>>     Hello world from rank 0  out of 1 processors
>>
>>     So it seems to work apart the error message.
>>
>>
>>     2017-05-19 9:10 GMT+02:00 Gilles Gouaillardet <gil...@rist.or.jp
>>     <mailto:gil...@rist.or.jp>>:
>>
>>         Gabriele,
>>
>>
>>         so it seems pml/pami assumes there is an infiniband card
>>         available (!)
>>
>>         i guess IBM folks will comment on that shortly.
>>
>>
>>         meanwhile, you do not need pami since you are running on a
>>         single node
>>
>>         mpirun --mca pml ^pami ...
>>
>>         should do the trick
>>
>>         (if it does not work, can run and post the logs)
>>
>>         mpirun --mca pml ^pami --mca pml_base_verbose 100 ...
>>
>>
>>         Cheers,
>>
>>
>>         Gilles
>>
>>
>>         On 5/19/2017 4:01 PM, Gabriele Fatigati wrote:
>>
>>             Hi John,
>>             Infiniband is not used, there is a single node on this
>>             machine.
>>
>>             2017-05-19 8:50 GMT+02:00 John Hearns via users
>>             <users@lists.open-mpi.org
>>             <mailto:users@lists.open-mpi.org>
>>             <mailto:users@lists.open-mpi.org
>>             <mailto:users@lists.open-mpi.org>>>:
>>
>>                 Gabriele,   pleae run  'ibv_devinfo'
>>                 It looks to me like you may have the physical
>>             interface cards in
>>                 these systems, but you do not have the correct drivers or
>>                 libraries loaded.
>>
>>                 I have had similar messages when using Infiniband on
>>             x86 systems -
>>                 which did not have libibverbs installed.
>>
>>
>>                 On 19 May 2017 at 08:41, Gabriele Fatigati
>>             <g.fatig...@cineca.it <mailto:g.fatig...@cineca.it>
>>                 <mailto:g.fatig...@cineca.it
>>             <mailto:g.fatig...@cineca.it>>> wrote:
>>
>>                     Hi Gilles, using your command:
>>
>>                     [openpower:88536] mca: base: components_register:
>>             registering
>>                     framework pml components
>>                     [openpower:88536] mca: base: components_register:
>>             found loaded
>>                     component pami
>>                     [openpower:88536] mca: base: components_register:
>>             component
>>                     pami register function successful
>>                     [openpower:88536] mca: base: components_open:
>>             opening pml
>>                     components
>>                     [openpower:88536] mca: base: components_open:
>>             found loaded
>>                     component pami
>>                     [openpower:88536] mca: base: components_open:
>>             component pami
>>                     open function successful
>>                     [openpower:88536] select: initializing pml
>>             component pami
>>                     findActiveDevices Error
>>                     We found no active IB device ports
>>                     [openpower:88536] select: init returned failure
>>             for component pami
>>                     [openpower:88536] PML pami cannot be selected
>>                                -----------------------------
>> ---------------------------------------------
>>                     No components were able to be opened in the pml
>>             framework.
>>
>>                     This typically means that either no components of
>>             this type were
>>                     installed, or none of the installed componnets can
>>             be loaded.
>>                     Sometimes this means that shared libraries
>>             required by these
>>                     components are unable to be found/loaded.
>>
>>                       Host:      openpower
>>                       Framework: pml
>>                                -----------------------------
>> ---------------------------------------------
>>
>>
>>                     2017-05-19 7:03 GMT+02:00 Gilles Gouaillardet
>>                     <gil...@rist.or.jp <mailto:gil...@rist.or.jp>
>>             <mailto:gil...@rist.or.jp <mailto:gil...@rist.or.jp>>>:
>>
>>
>>                         Gabriele,
>>
>>
>>                         pml/pami is here, at least according to ompi_info
>>
>>
>>                         can you update your mpirun command like this
>>
>>                         mpirun --mca pml_base_verbose 100 ..
>>
>>
>>                         and post the output ?
>>
>>
>>                         Cheers,
>>
>>                         Gilles
>>
>>                         On 5/18/2017 10:41 PM, Gabriele Fatigati wrote:
>>
>>                             Hi Gilles, attached the requested info
>>
>>                             2017-05-18 15:04 GMT+02:00 Gilles Gouaillardet
>>                             <gilles.gouaillar...@gmail.com
>>             <mailto:gilles.gouaillar...@gmail.com>
>>                             <mailto:gilles.gouaillar...@gmail.com
>>             <mailto:gilles.gouaillar...@gmail.com>>
>>                             <mailto:gilles.gouaillar...@gmail.com
>>             <mailto:gilles.gouaillar...@gmail.com>
>>                             <mailto:gilles.gouaillar...@gmail.com
>>             <mailto:gilles.gouaillar...@gmail.com>>>>:
>>
>>                                 Gabriele,
>>
>>                                 can you
>>                                 ompi_info --all | grep pml
>>
>>                                 also, make sure there is nothing in your
>>                             environment pointing to
>>                                 an other Open MPI install
>>                                 for example
>>                                 ldd a.out
>>                                 should only point to IBM libraries
>>
>>                                 Cheers,
>>
>>                                 Gilles
>>
>>
>>                                 On Thursday, May 18, 2017, Gabriele
>>             Fatigati
>>                             <g.fatig...@cineca.it
>>             <mailto:g.fatig...@cineca.it> <mailto:g.fatig...@cineca.it
>>             <mailto:g.fatig...@cineca.it>>
>>                                 <mailto:g.fatig...@cineca.it
>>             <mailto:g.fatig...@cineca.it>
>>
>>                             <mailto:g.fatig...@cineca.it
>>             <mailto:g.fatig...@cineca.it>>>> wrote:
>>
>>                                     Dear OpenMPI users and developers,
>>             I'm using
>>                             IBM Spectrum MPI
>>                                     10.1.0 based on OpenMPI, so I hope
>>             there are
>>                             some MPI expert
>>                                     can help me to solve the problem.
>>
>>                                     When I run a simple Hello World
>>             MPI program, I
>>                             get the follow
>>                                     error message:
>>
>>
>>                                     A requested component was not
>>             found, or was
>>                             unable to be
>>                                     opened.  This
>>                                     means that this component is
>>             either not
>>                             installed or is unable
>>                                     to be
>>                                     used on your system (e.g.,
>>             sometimes this
>>                             means that shared
>>                                     libraries
>>                                     that the component requires are
>>             unable to be
>>                             found/loaded).         Note that
>>                                     Open MPI stopped checking at the first
>>                             component that it did
>>                                     not find.
>>
>>                                     Host:      openpower
>>                                     Framework: pml
>>                                     Component: pami
>>             ------------------------------------------------------------
>> --------------
>>             ------------------------------------------------------------
>> --------------
>>                                     It looks like MPI_INIT failed for
>>             some reason;
>>                             your parallel
>>                                     process is
>>                                     likely to abort. There are many
>>             reasons that a
>>                             parallel
>>                                     process can
>>                                     fail during MPI_INIT; some of
>>             which are due to
>>                             configuration
>>                                     or environment
>>                                     problems.  This failure appears to
>>             be an
>>                             internal failure;
>>                                     here's some
>>                                     additional information (which may
>>             only be
>>                             relevant to an Open MPI
>>                                     developer):
>>
>>                                     mca_pml_base_open() failed
>>                                       --> Returned "Not found" (-13)
>>             instead of
>>                             "Success" (0)
>>             ------------------------------------------------------------
>> --------------
>>                                     *** An error occurred in MPI_Init
>>                                     *** on a NULL communicator
>>                                     *** MPI_ERRORS_ARE_FATAL
>>             (processes in this
>>                             communicator will
>>                                     now abort,
>>                                     ***    and potentially your MPI job)
>>
>>                                     My sysadmin used official IBM Spectrum
>>                             packages to install
>>                                     MPI, so It's quite strange that
>>             there are some
>>                             components
>>                                     missing (pami). Any help? Thanks
>>
>>
>>                                     --         Ing. Gabriele Fatigati
>>
>>                                     HPC specialist
>>
>>                                     SuperComputing Applications and
>>             Innovation
>>                             Department
>>
>>                                     Via Magnanelli 6/3, Casalecchio di
>>             Reno (BO) Italy
>>
>>             www.cineca.it <http://www.cineca.it> <http://www.cineca.it>
>>                             <http://www.cineca.it>             Tel: +39
>>             051 6171722 <tel:051%206171722> <tel:051%206171722>
>>
>>             <tel:051%20617%201722>
>>
>>                                     g.fatigati [AT] cineca.it
>>             <http://cineca.it> <http://cineca.it>
>>                             <http://cineca.it>
>>
>>
>>             _______________________________________________
>>                                 users mailing list
>>             users@lists.open-mpi.org <mailto:users@lists.open-mpi.org>
>>                             <mailto:users@lists.open-mpi.org
>>             <mailto:users@lists.open-mpi.org>>
>>                             <mailto:users@lists.open-mpi.org
>>             <mailto:users@lists.open-mpi.org>
>>                             <mailto:users@lists.open-mpi.org
>>             <mailto:users@lists.open-mpi.org>>>
>>             https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>
>>                                        <https://rfd.newmexicoconsort
>> ium.org/mailman/listinfo/users
>>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>>
>>                                                           <
>> https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>
>>                                        <https://rfd.newmexicoconsort
>> ium.org/mailman/listinfo/users
>>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>
>> >>
>>
>>
>>
>>
>>                             --                 Ing. Gabriele Fatigati
>>
>>                             HPC specialist
>>
>>                             SuperComputing Applications and Innovation
>>             Department
>>
>>                             Via Magnanelli 6/3, Casalecchio di Reno
>>             (BO) Italy
>>
>>             www.cineca.it <http://www.cineca.it> <http://www.cineca.it>
>>                             <http://www.cineca.it> Tel: +39 051
>>             6171722 <tel:%2B39%20051%206171722>
>>             <tel:%2B39%20051%206171722>
>>
>>                             g.fatigati [AT] cineca.it
>>             <http://cineca.it> <http://cineca.it>
>>                             <http://cineca.it>
>>
>>
>>                                        _____________________________
>> __________________
>>                             users mailing list
>>             users@lists.open-mpi.org <mailto:users@lists.open-mpi.org>
>>             <mailto:users@lists.open-mpi.org
>>             <mailto:users@lists.open-mpi.org>>
>>             https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>
>>                                        <https://rfd.newmexicoconsort
>> ium.org/mailman/listinfo/users
>>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>>
>>
>>
>>
>>
>>
>>                     --         Ing. Gabriele Fatigati
>>
>>                     HPC specialist
>>
>>                     SuperComputing Applications and Innovation Department
>>
>>                     Via Magnanelli 6/3, Casalecchio di Reno (BO) Italy
>>
>>             www.cineca.it <http://www.cineca.it>
>>             <http://www.cineca.it>                  Tel:
>>             +39 051 6171722 <tel:%2B39%20051%206171722>
>>             <tel:+39%20051%20617%201722>
>>
>>
>>                     g.fatigati [AT] cineca.it <http://cineca.it>
>>             <http://cineca.it>
>>
>>                     _______________________________________________
>>                     users mailing list
>>             users@lists.open-mpi.org <mailto:users@lists.open-mpi.org>
>>             <mailto:users@lists.open-mpi.org
>>             <mailto:users@lists.open-mpi.org>>
>>             https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>
>>                                <https://rfd.newmexicoconsort
>> ium.org/mailman/listinfo/users
>>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>>
>>
>>
>>
>>                 _______________________________________________
>>                 users mailing list
>>             users@lists.open-mpi.org <mailto:users@lists.open-mpi.org>
>>             <mailto:users@lists.open-mpi.org
>>             <mailto:users@lists.open-mpi.org>>
>>             https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>
>>                            <https://rfd.newmexicoconsort
>> ium.org/mailman/listinfo/users
>>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>>
>>
>>
>>
>>
>>             --             Ing. Gabriele Fatigati
>>
>>             HPC specialist
>>
>>             SuperComputing Applications and Innovation Department
>>
>>             Via Magnanelli 6/3, Casalecchio di Reno (BO) Italy
>>
>>             www.cineca.it <http://www.cineca.it>
>>             <http://www.cineca.it> Tel: +39 051 6171722
>>             <tel:%2B39%20051%206171722>
>>
>>             g.fatigati [AT] cineca.it <http://cineca.it>
>>             <http://cineca.it>
>>
>>
>>             _______________________________________________
>>             users mailing list
>>             users@lists.open-mpi.org <mailto:users@lists.open-mpi.org>
>>             https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>             <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>
>>
>>
>>         _______________________________________________
>>         users mailing list
>>         users@lists.open-mpi.org <mailto:users@lists.open-mpi.org>
>>         https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>         <https://rfd.newmexicoconsortium.org/mailman/listinfo/users>
>>
>>
>>
>>
>>     --     Ing. Gabriele Fatigati
>>
>>     HPC specialist
>>
>>     SuperComputing Applications and Innovation Department
>>
>>     Via Magnanelli 6/3, Casalecchio di Reno (BO) Italy
>>
>>     www.cineca.it <http://www.cineca.it>       Tel: +39 051 6171722
>>     <tel:051%20617%201722>
>>
>>     g.fatigati [AT] cineca.it <http://cineca.it>
>>
>>
>>
>>
>> --
>> Ing. Gabriele Fatigati
>>
>> HPC specialist
>>
>> SuperComputing Applications and Innovation Department
>>
>> Via Magnanelli 6/3, Casalecchio di Reno (BO) Italy
>>
>> www.cineca.it <http://www.cineca.it> Tel:   +39 051 6171722
>>
>> g.fatigati [AT] cineca.it <http://cineca.it>
>>
>>
>> _______________________________________________
>> users mailing list
>> users@lists.open-mpi.org
>> https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>>
>
> _______________________________________________
> users mailing list
> users@lists.open-mpi.org
> https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>



-- 
Ing. Gabriele Fatigati

HPC specialist

SuperComputing Applications and Innovation Department

Via Magnanelli 6/3, Casalecchio di Reno (BO) Italy

www.cineca.it                    Tel:   +39 051 6171722

g.fatigati [AT] cineca.it

Attachment: output_mpirun
Description: Binary data

_______________________________________________
users mailing list
users@lists.open-mpi.org
https://rfd.newmexicoconsortium.org/mailman/listinfo/users

Reply via email to