Odd - it would appear that none of the pmix components built? Can you send along the output from ompi_info? Or just send a listing of the files in the <prefix>/lib/openmpi directory?
> On Apr 21, 2016, at 1:27 AM, Siegmar Gross > <siegmar.gr...@informatik.hs-fulda.de> wrote: > > Hi Ralph, > > Am 21.04.2016 um 00:18 schrieb Ralph Castain: >> Could you please rerun these test and add “-mca pmix_base_verbose 10 >> -mca pmix_server_verbose 5” to your cmd line? I need to see why the >> pmix components failed. > > > tyr spawn 111 mpiexec -np 1 --host tyr,sunpc1,linpc1,ruester -mca > pmix_base_verbose 10 -mca pmix_server_verbose 5 spawn_multiple_master > [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>:26652] mca: > base: components_register: registering framework pmix components > [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>:26652] mca: > base: components_open: opening pmix components > [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>:26652] > mca:base:select: Auto-selecting pmix components > [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>:26652] > mca:base:select:( pmix) No component selected! > [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>:26652] > [[52794,0],0] ORTE_ERROR_LOG: Not found in file > ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c > at line 638 > -------------------------------------------------------------------------- > It looks like orte_init failed for some reason; your parallel process is > likely to abort. There are many reasons that a parallel process can > fail during orte_init; some of which are due to configuration or > environment problems. This failure appears to be an internal failure; > here's some additional information (which may only be relevant to an > Open MPI developer): > > opal_pmix_base_select failed > --> Returned value Not found (-13) instead of ORTE_SUCCESS > -------------------------------------------------------------------------- > tyr spawn 112 > > > > > tyr hello_1 116 mpiexec -np 1 --host tyr,sunpc1,linpc1,ruester -mca > pmix_base_verbose 10 -mca pmix_server_verbose 5 hello_1_mpi > [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>:27261] mca: > base: components_register: registering framework pmix components > [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>:27261] mca: > base: components_open: opening pmix components > [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>:27261] > mca:base:select: Auto-selecting pmix components > [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>:27261] > mca:base:select:( pmix) No component selected! > [tyr.informatik.hs-fulda.de <http://tyr.informatik.hs-fulda.de/>:27261] > [[52315,0],0] ORTE_ERROR_LOG: Not found in file > ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c > at line 638 > -------------------------------------------------------------------------- > It looks like orte_init failed for some reason; your parallel process is > likely to abort. There are many reasons that a parallel process can > fail during orte_init; some of which are due to configuration or > environment problems. This failure appears to be an internal failure; > here's some additional information (which may only be relevant to an > Open MPI developer): > > opal_pmix_base_select failed > --> Returned value Not found (-13) instead of ORTE_SUCCESS > -------------------------------------------------------------------------- > tyr hello_1 117 > > > > Thank you very much for your help. > > > Kind regards > > Siegmar > > > >> >> Thanks >> Ralph >> >>> On Apr 20, 2016, at 10:12 AM, Siegmar Gross >>> <siegmar.gr...@informatik.hs-fulda.de> wrote: >>> >>> Hi, >>> >>> I have built openmpi-v2.x-dev-1280-gc110ae8 on my machines >>> (Solaris 10 Sparc, Solaris 10 x86_64, and openSUSE Linux >>> 12.1 x86_64) with gcc-5.1.0 and Sun C 5.13. Unfortunately I get >>> runtime errors for some programs. >>> >>> >>> Sun C 5.13: >>> =========== >>> >>> For all my test programs I get the same error on Solaris Sparc and >>> Solaris x86_64, while the programs work fine on Linux. >>> >>> tyr hello_1 115 mpiexec -np 2 hello_1_mpi >>> [tyr.informatik.hs-fulda.de:22373] [[61763,0],0] ORTE_ERROR_LOG: Not found >>> in file >>> ../../../../../openmpi-v2.x-dev-1280-gc110ae8/orte/mca/ess/hnp/ess_hnp_module.c >>> at line 638 >>> -------------------------------------------------------------------------- >>> It looks like orte_init failed for some reason; your parallel process is >>> likely to abort. There are many reasons that a parallel process can >>> fail during orte_init; some of which are due to configuration or >>> environment problems. This failure appears to be an internal failure; >>> here's some additional information (which may only be relevant to an >>> Open MPI developer): >>> >>> opal_pmix_base_select failed >>> --> Returned value Not found (-13) instead of ORTE_SUCCESS >>> -------------------------------------------------------------------------- >>> tyr hello_1 116 >>> >>> >>> >>> >>> GCC-5.1.0: >>> ========== >>> >>> tyr spawn 121 mpiexec -np 1 --host tyr,sunpc1,linpc1,ruester >>> spawn_multiple_master >>> >>> Parent process 0 running on tyr.informatik.hs-fulda.de >>> I create 3 slave processes. >>> >>> [tyr.informatik.hs-fulda.de:25366] PMIX ERROR: UNPACK-PAST-END in file >>> ../../../../../../openmpi-v2.x-dev-1280-gc110ae8/opal/mca/pmix/pmix112/pmix/src/server/pmix_server_ops.c >>> at line 829 >>> [tyr.informatik.hs-fulda.de:25366] PMIX ERROR: UNPACK-PAST-END in file >>> ../../../../../../openmpi-v2.x-dev-1280-gc110ae8/opal/mca/pmix/pmix112/pmix/src/server/pmix_server.c >>> at line 2176 >>> [tyr:25377] *** An error occurred in MPI_Comm_spawn_multiple >>> [tyr:25377] *** reported by process [3308257281,0] >>> [tyr:25377] *** on communicator MPI_COMM_WORLD >>> [tyr:25377] *** MPI_ERR_SPAWN: could not spawn processes >>> [tyr:25377] *** MPI_ERRORS_ARE_FATAL (processes in this communicator will >>> now abort, >>> [tyr:25377] *** and potentially your MPI job) >>> tyr spawn 122 >>> >>> >>> I would be grateful if somebody can fix the problems. Thank you very >>> much for any help in advance. >>> >>> >>> Kind regards >>> >>> Siegmar >>> <hello_1_mpi.c><spawn_multiple_master.c>_______________________________________________ >>> users mailing list >>> us...@open-mpi.org >>> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >>> Link to this post: >>> http://www.open-mpi.org/community/lists/users/2016/04/28983.php >> >> _______________________________________________ >> users mailing list >> us...@open-mpi.org <mailto:us...@open-mpi.org> >> Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users >> <http://www.open-mpi.org/mailman/listinfo.cgi/users> >> Link to this post: >> http://www.open-mpi.org/community/lists/users/2016/04/28986.php >> <http://www.open-mpi.org/community/lists/users/2016/04/28986.php> >> > > _______________________________________________ > users mailing list > us...@open-mpi.org <mailto:us...@open-mpi.org> > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users > <http://www.open-mpi.org/mailman/listinfo.cgi/users> > Link to this post: > http://www.open-mpi.org/community/lists/users/2016/04/28987.php > <http://www.open-mpi.org/community/lists/users/2016/04/28987.php>