Sorry, I'll try and fill in the background.  I'm attempting to package
openmpi for a number of customers we have, whenever possible on our
clusters we use modules to provide users with a choice of MPI
environment.

I'm using the 1.2.6 stable release and have built the code twice, once
to /opt/openmpi-1.2.6/gnu and once to /opt/openmpi-1.2.6/intel, I have
create two modules environments called openmpi-gnu and openmpi-intel and
am also using a existing one called intel-compiler.  The build was
successful in both cases.

If I load the openmpi-gnu module I can compile and run code using
mpicc/mpirun as expected, if I load openmpi-intel and intel-compiler I
find I can compile code but I get an error about missing libimf.so when
I try to run it (reproduced below).

The application *will* run if I add the line "module load
intel-compiler" to my bashrc as this allows orted to link.  What I think
I want to do is to compile the actual library with icc but to compile
orted with gcc so that I don't need to load the intel environment by
default.  I'm assuming that the link problems only exist with orted and
not with the actual application as the LD_LIBRARY_PATH is set correctly
in the shell which is launching the program.

Ashley Pittman.

sccomp@demo4-sles-10-1-fe:~/benchmarks/IMB_3.0/src> mpirun -H comp00,comp01 
./IMB-MPI1
/opt/openmpi-1.2.6/intel/bin/orted: error while loading shared libraries: 
libimf.so: cannot open shared object file: No such file or directory
/opt/openmpi-1.2.6/intel/bin/orted: error while loading shared libraries: 
libimf.so: cannot open shared object file: No such file or directory
[demo4-sles-10-1-fe:29303] ERROR: A daemon on node comp01 failed to start as 
expected.
[demo4-sles-10-1-fe:29303] ERROR: There may be more information available from
[demo4-sles-10-1-fe:29303] ERROR: the remote shell (see above).
[demo4-sles-10-1-fe:29303] ERROR: The daemon exited unexpectedly with status 
127.
[demo4-sles-10-1-fe:29303] [0,0,0] ORTE_ERROR_LOG: Timeout in file 
base/pls_base_orted_cmds.c at line 275
[demo4-sles-10-1-fe:29303] [0,0,0] ORTE_ERROR_LOG: Timeout in file 
pls_rsh_module.c at line 1166
[demo4-sles-10-1-fe:29303] [0,0,0] ORTE_ERROR_LOG: Timeout in file errmgr_hnp.c 
at line 90
[demo4-sles-10-1-fe:29303] ERROR: A daemon on node comp00 failed to start as 
expected.
[demo4-sles-10-1-fe:29303] ERROR: There may be more information available from
[demo4-sles-10-1-fe:29303] ERROR: the remote shell (see above).
[demo4-sles-10-1-fe:29303] ERROR: The daemon exited unexpectedly with status 
127.
[demo4-sles-10-1-fe:29303] [0,0,0] ORTE_ERROR_LOG: Timeout in file 
base/pls_base_orted_cmds.c at line 188
[demo4-sles-10-1-fe:29303] [0,0,0] ORTE_ERROR_LOG: Timeout in file 
pls_rsh_module.c at line 1198
--------------------------------------------------------------------------
mpirun was unable to cleanly terminate the daemons for this job. Returned value 
Timeout instead of ORTE_SUCCESS.
--------------------------------------------------------------------------

$ ldd /opt/openmpi-1.2.6/intel/bin/orted
        linux-vdso.so.1 =>  (0x00007fff877fe000)
        libopen-rte.so.0 => /opt/openmpi-1.2.6/intel/lib/libopen-rte.so.0 
(0x00007fe97f3ac000)
        libopen-pal.so.0 => /opt/openmpi-1.2.6/intel/lib/libopen-pal.so.0 
(0x00007fe97f239000)
        libdl.so.2 => /lib64/libdl.so.2 (0x00007fe97f135000)
        libnsl.so.1 => /lib64/libnsl.so.1 (0x00007fe97f01f000)
        libutil.so.1 => /lib64/libutil.so.1 (0x00007fe97ef1c000)
        libm.so.6 => /lib64/libm.so.6 (0x00007fe97edc7000)
        libgcc_s.so.1 => /lib64/libgcc_s.so.1 (0x00007fe97ecba000)
        libpthread.so.0 => /lib64/libpthread.so.0 (0x00007fe97eba3000)
        libc.so.6 => /lib64/libc.so.6 (0x00007fe97e972000)
        libimf.so => /opt/intel/compiler_10.1/x86_64/lib/libimf.so 
(0x00007fe97e610000)
        libsvml.so => /opt/intel/compiler_10.1/x86_64/lib/libsvml.so 
(0x00007fe97e489000)
        libintlc.so.5 => /opt/intel/compiler_10.1/x86_64/lib/libintlc.so.5 
(0x00007fe97e350000)
        /lib64/ld-linux-x86-64.so.2 (0x00007fe97f525000)
$ ssh comp00 ldd /opt/openmpi-1.2.6/intel/bin/orted
        libopen-rte.so.0 => /opt/openmpi-1.2.6/intel/lib/libopen-rte.so.0 
(0x00002b1f0c0c5000)
        libopen-pal.so.0 => /opt/openmpi-1.2.6/intel/lib/libopen-pal.so.0 
(0x00002b1f0c23e000)
        libdl.so.2 => /lib64/libdl.so.2 (0x00002b1f0c3bc000)
        libnsl.so.1 => /lib64/libnsl.so.1 (0x00002b1f0c4c0000)
        libutil.so.1 => /lib64/libutil.so.1 (0x00002b1f0c5d7000)
        libm.so.6 => /lib64/libm.so.6 (0x00002b1f0c6da000)
        libgcc_s.so.1 => /lib64/libgcc_s.so.1 (0x00002b1f0c82f000)
        libpthread.so.0 => /lib64/libpthread.so.0 (0x00002b1f0c93d000)
        libc.so.6 => /lib64/libc.so.6 (0x00002b1f0ca54000)
        /lib64/ld-linux-x86-64.so.2 (0x00002b1f0bfa9000)
        libimf.so => not found
        libsvml.so => not found
        libintlc.so.5 => not found
        libimf.so => not found
        libsvml.so => not found
        libintlc.so.5 => not found
$ ldd ./IMB-MPI1
        linux-vdso.so.1 =>  (0x00007fff2cbfe000)
        libmpi.so.0 => /opt/openmpi-1.2.6/intel/lib/libmpi.so.0 
(0x00007f1624821000)
        libopen-rte.so.0 => /opt/openmpi-1.2.6/intel/lib/libopen-rte.so.0 
(0x00007f16246a8000)
        libopen-pal.so.0 => /opt/openmpi-1.2.6/intel/lib/libopen-pal.so.0 
(0x00007f1624535000)
        libdl.so.2 => /lib64/libdl.so.2 (0x00007f1624431000)
        libnsl.so.1 => /lib64/libnsl.so.1 (0x00007f162431b000)
        libutil.so.1 => /lib64/libutil.so.1 (0x00007f1624218000)
        libm.so.6 => /lib64/libm.so.6 (0x00007f16240c3000)
        libgcc_s.so.1 => /lib64/libgcc_s.so.1 (0x00007f1623fb6000)
        libpthread.so.0 => /lib64/libpthread.so.0 (0x00007f1623e9f000)
        libc.so.6 => /lib64/libc.so.6 (0x00007f1623c6e000)
        libimf.so => /opt/intel/compiler_10.1/x86_64/lib/libimf.so 
(0x00007f162390c000)
        libsvml.so => /opt/intel/compiler_10.1/x86_64/lib/libsvml.so 
(0x00007f1623785000)
        libintlc.so.5 => /opt/intel/compiler_10.1/x86_64/lib/libintlc.so.5 
(0x00007f162364c000)
        /lib64/ld-linux-x86-64.so.2 (0x00007f16249e0000)



On Mon, 2008-06-09 at 13:02 -0700, Doug Reeder wrote:
> Ashley,
> 
> I am confused. In your first post you said orted fails, with link  
> errors, when you try to launch a job. From this I inferred that the  
> build and install steps for creating openmpi were successful. Was the  
> build/install step successful. If so what dynamic libraries does ldd  
> say that orted is using.
> 
> Doug Reeder
> On Jun 9, 2008, at 12:54 PM, Ashley Pittman wrote:
> 
> >
> > Putting to side any religious views I might have about static linking
> > how would that help in this case?   It appears to be orted itself that
> > fails to link, I'm assuming that the application would actually run,
> > either because the LD_LIBRARY_PATH is set correctly on the front  
> > end or
> > the --prefix option to mpirun.
> >
> > Or do you mean static linking of the tools?  I could go for that if
> > there is a configure option for it.
> >
> > Ashley Pittman.
> >
> > On Mon, 2008-06-09 at 08:27 -0700, Doug Reeder wrote:
> >> Ashley,
> >>
> >> It could work but I think you would be better off to try and
> >> statically link the intel libraries.
> >>
> >> Doug Reeder
> >> On Jun 9, 2008, at 4:34 AM, Ashley Pittman wrote:
> >>
> >>>
> >>> Is there a way to use a different compiler for the orte component  
> >>> and
> >>> the shared library component when using openmpi?  We are finding
> >>> that if
> >>> we use icc to compile openmpi then orted fails with link errors  
> >>> when I
> >>> try and launch a job as the intel environment isn't loaded by  
> >>> default.
> >>>
> >>> We use the module command heavily and have modules for openmpi- 
> >>> gnu and
> >>> openmpi-intel as well as a intel_compiler module.  To use openmpi-
> >>> intel
> >>> we have to load intel_compiler by default on the compute nodes which
> >>> isn't ideal, is it possible to compile the orte component with  
> >>> gcc and
> >>> the library component with icc?
> >>>
> >>> Yours,
> >>>
> >>> Ashley Pittman,
> >>>
> >>> _______________________________________________
> >>> users mailing list
> >>> us...@open-mpi.org
> >>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >>
> >> _______________________________________________
> >> users mailing list
> >> us...@open-mpi.org
> >> http://www.open-mpi.org/mailman/listinfo.cgi/users
> >
> > _______________________________________________
> > users mailing list
> > us...@open-mpi.org
> > http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users

Reply via email to