Re: [OMPI users] compilation error with pgcc Unknown switch
I tried with 1.5.4 and same compilers, and got the exact same error. I also tried PGI version 11.10.0, and got the same thing. On Wed, Feb 8, 2012 at 3:04 AM, Jeff Squyres (jsquyres) wrote: > Can you try building 1.5.4 with the same compilers? > > Sent from my phone. No type good. > > On Feb 7, 2012, at 3:14 PM, "Abhinav Sarje" wrote: > >> I am trying to build open-mpi 1.4.4 (latest stable from open-mpi.org) >> using PGI compilers on a cray platform. PGI compilers' version is >> 11.9.0. I get the following error while building: >> - >> Making all in tools/wrappers >> make[2]: Entering directory `{my_installation_directory}/opal/tools/wrappers' >> source='opal_wrapper.c' object='opal_wrapper.o' libtool=no \ >> DEPDIR=.deps depmode=none /bin/sh ../../../config/depcomp \ >> cc "-DEXEEXT=\"\"" -I. -I../../../opal/include >> -I../../../orte/include -I../../../ompi/include >> -I../../../opal/mca/paffinity/linux/plpa/src/libplpa -I../../.. >> -D_REENTRANT -O -DNDEBUG -fPIC -c -o opal_wrapper.o opal_wrapper.c >> /bin/sh ../../../libtool --tag=CC --mode=link cc -O -DNDEBUG -fPIC >> -export-dynamic -o opal_wrapper opal_wrapper.o >> ../../../opal/libopen-pal.la -lnsl -lutil >> libtool: link: cc -O -DNDEBUG -fPIC -o .libs/opal_wrapper >> opal_wrapper.o --export-dynamic ../../../opal/.libs/libopen-pal.so >> -ldl -lnsl -lutil -rpath {my_installation_directory}/lib >> pgcc-Error-Unknown switch: --export-dynamic >> make[2]: *** [opal_wrapper] Error 1 >> make[2]: Leaving directory `{my_installation_directory}/opal/tools/wrappers' >> make[1]: *** [all-recursive] Error 1 >> make[1]: Leaving directory `{my_installation_directory}/opal' >> make: *** [all-recursive] Error 1 >> - >> >> I see that the libtool packaged with open-mpi is 2.2.6b >> When I try to compile this particular part with libtool versions 2.2.6 >> or 2.4, I get the following error: >> - >> $ libtool --tag=CC --mode=link cc -O -DNDEBUG -fPIC >> -export-dynamic -o opal_wrapper opal_wrapper.o >> ../../../opal/libopen-pal.la -lnsl -lutil >> libtool: link: cc -O -DNDEBUG -fPIC -o .libs/opal_wrapper >> opal_wrapper.o -Wl,--export-dynamic >> ../../../opal/.libs/libopen-pal.so -ldl -lnsl -lutil -Wl,-rpath >> -Wl,{my_installation_directory}/lib >> /usr/bin/ld: attempted static link of dynamic object >> `../../../opal/.libs/libopen-pal.so' >> - >> >> Looking at earlier posts, apparently there was a bug with libtool a >> couple of years ago because of which the above error occurred. This >> was fixed in newer releases, but I am getting similar errors. >> >> Does anyone have any information on how to fix this, or if I am doing >> something wrong here? >> >> Thanks! >> ___ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users > > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users
Re: [OMPI users] compilation error with pgcc Unknown switch
Hi Abhinav Setting CC='pgcc --noswitcherror', FC='pgfortran --noswitcherror', etc, may prevent the error message [but not sure it will prevent any actual and unreported error]. Check details with 'man pgfortran'. I hope this helps, Gus Corea . On Feb 10, 2012, at 2:03 AM, Abhinav Sarje wrote: > I tried with 1.5.4 and same compilers, and got the exact same error. > I also tried PGI version 11.10.0, and got the same thing. > > > On Wed, Feb 8, 2012 at 3:04 AM, Jeff Squyres (jsquyres) > wrote: >> Can you try building 1.5.4 with the same compilers? >> >> Sent from my phone. No type good. >> >> On Feb 7, 2012, at 3:14 PM, "Abhinav Sarje" wrote: >> >>> I am trying to build open-mpi 1.4.4 (latest stable from open-mpi.org) >>> using PGI compilers on a cray platform. PGI compilers' version is >>> 11.9.0. I get the following error while building: >>> - >>> Making all in tools/wrappers >>> make[2]: Entering directory >>> `{my_installation_directory}/opal/tools/wrappers' >>> source='opal_wrapper.c' object='opal_wrapper.o' libtool=no \ >>>DEPDIR=.deps depmode=none /bin/sh ../../../config/depcomp \ >>>cc "-DEXEEXT=\"\"" -I. -I../../../opal/include >>> -I../../../orte/include -I../../../ompi/include >>> -I../../../opal/mca/paffinity/linux/plpa/src/libplpa -I../../.. >>> -D_REENTRANT -O -DNDEBUG -fPIC -c -o opal_wrapper.o opal_wrapper.c >>> /bin/sh ../../../libtool --tag=CC --mode=link cc -O -DNDEBUG -fPIC >>> -export-dynamic -o opal_wrapper opal_wrapper.o >>> ../../../opal/libopen-pal.la -lnsl -lutil >>> libtool: link: cc -O -DNDEBUG -fPIC -o .libs/opal_wrapper >>> opal_wrapper.o --export-dynamic ../../../opal/.libs/libopen-pal.so >>> -ldl -lnsl -lutil -rpath {my_installation_directory}/lib >>> pgcc-Error-Unknown switch: --export-dynamic >>> make[2]: *** [opal_wrapper] Error 1 >>> make[2]: Leaving directory `{my_installation_directory}/opal/tools/wrappers' >>> make[1]: *** [all-recursive] Error 1 >>> make[1]: Leaving directory `{my_installation_directory}/opal' >>> make: *** [all-recursive] Error 1 >>> - >>> >>> I see that the libtool packaged with open-mpi is 2.2.6b >>> When I try to compile this particular part with libtool versions 2.2.6 >>> or 2.4, I get the following error: >>> - >>> $ libtool --tag=CC --mode=link cc -O -DNDEBUG -fPIC >>> -export-dynamic -o opal_wrapper opal_wrapper.o >>> ../../../opal/libopen-pal.la -lnsl -lutil >>> libtool: link: cc -O -DNDEBUG -fPIC -o .libs/opal_wrapper >>> opal_wrapper.o -Wl,--export-dynamic >>> ../../../opal/.libs/libopen-pal.so -ldl -lnsl -lutil -Wl,-rpath >>> -Wl,{my_installation_directory}/lib >>> /usr/bin/ld: attempted static link of dynamic object >>> `../../../opal/.libs/libopen-pal.so' >>> - >>> >>> Looking at earlier posts, apparently there was a bug with libtool a >>> couple of years ago because of which the above error occurred. This >>> was fixed in newer releases, but I am getting similar errors. >>> >>> Does anyone have any information on how to fix this, or if I am doing >>> something wrong here? >>> >>> Thanks! >>> ___ >>> users mailing list >>> us...@open-mpi.org >>> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> ___ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users > > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users
Re: [OMPI users] compilation error with pgcc Unknown switch
We literally just upgraded the version of autotools on the 1.5 nightly tarballs (i.e., what will become 1.5.5). Could you try the most recent 1.5.5 nightly snapshot? http://www.open-mpi.org/nightly/v1.5/ On Feb 10, 2012, at 8:57 AM, Gustavo Correa wrote: > Hi Abhinav > > Setting CC='pgcc --noswitcherror', FC='pgfortran --noswitcherror', etc, may > prevent the > error message [but not sure it will prevent any actual and unreported error]. > Check details with 'man pgfortran'. > > I hope this helps, > Gus Corea > > . > > On Feb 10, 2012, at 2:03 AM, Abhinav Sarje wrote: > >> I tried with 1.5.4 and same compilers, and got the exact same error. >> I also tried PGI version 11.10.0, and got the same thing. >> >> >> On Wed, Feb 8, 2012 at 3:04 AM, Jeff Squyres (jsquyres) >> wrote: >>> Can you try building 1.5.4 with the same compilers? >>> >>> Sent from my phone. No type good. >>> >>> On Feb 7, 2012, at 3:14 PM, "Abhinav Sarje" wrote: >>> I am trying to build open-mpi 1.4.4 (latest stable from open-mpi.org) using PGI compilers on a cray platform. PGI compilers' version is 11.9.0. I get the following error while building: - Making all in tools/wrappers make[2]: Entering directory `{my_installation_directory}/opal/tools/wrappers' source='opal_wrapper.c' object='opal_wrapper.o' libtool=no \ DEPDIR=.deps depmode=none /bin/sh ../../../config/depcomp \ cc "-DEXEEXT=\"\"" -I. -I../../../opal/include -I../../../orte/include -I../../../ompi/include -I../../../opal/mca/paffinity/linux/plpa/src/libplpa -I../../.. -D_REENTRANT -O -DNDEBUG -fPIC -c -o opal_wrapper.o opal_wrapper.c /bin/sh ../../../libtool --tag=CC --mode=link cc -O -DNDEBUG -fPIC -export-dynamic -o opal_wrapper opal_wrapper.o ../../../opal/libopen-pal.la -lnsl -lutil libtool: link: cc -O -DNDEBUG -fPIC -o .libs/opal_wrapper opal_wrapper.o --export-dynamic ../../../opal/.libs/libopen-pal.so -ldl -lnsl -lutil -rpath {my_installation_directory}/lib pgcc-Error-Unknown switch: --export-dynamic make[2]: *** [opal_wrapper] Error 1 make[2]: Leaving directory `{my_installation_directory}/opal/tools/wrappers' make[1]: *** [all-recursive] Error 1 make[1]: Leaving directory `{my_installation_directory}/opal' make: *** [all-recursive] Error 1 - I see that the libtool packaged with open-mpi is 2.2.6b When I try to compile this particular part with libtool versions 2.2.6 or 2.4, I get the following error: - $ libtool --tag=CC --mode=link cc -O -DNDEBUG -fPIC -export-dynamic -o opal_wrapper opal_wrapper.o ../../../opal/libopen-pal.la -lnsl -lutil libtool: link: cc -O -DNDEBUG -fPIC -o .libs/opal_wrapper opal_wrapper.o -Wl,--export-dynamic ../../../opal/.libs/libopen-pal.so -ldl -lnsl -lutil -Wl,-rpath -Wl,{my_installation_directory}/lib /usr/bin/ld: attempted static link of dynamic object `../../../opal/.libs/libopen-pal.so' - Looking at earlier posts, apparently there was a bug with libtool a couple of years ago because of which the above error occurred. This was fixed in newer releases, but I am getting similar errors. Does anyone have any information on how to fix this, or if I am doing something wrong here? Thanks! ___ users mailing list us...@open-mpi.org http://www.open-mpi.org/mailman/listinfo.cgi/users >>> >>> ___ >>> users mailing list >>> us...@open-mpi.org >>> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> ___ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users > > > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users -- Jeff Squyres jsquy...@cisco.com For corporate legal information go to: http://www.cisco.com/web/about/doing_business/legal/cri/
Re: [OMPI users] Spawn_multiple with tight integration to SGE grid engine
Tom and I talked more about this off list, and I eventually logged in to his cluster to see what I could see. The issue turned out to be not related to SGE or THREAD_MULTIPLE at all. The issue was that RHEL6, by default, activated a virtualization IP interface on all of Tom's nodes. All nodes had a local IP interface in the 192.168.1.x/24 subnet, but that address was only used to communicate to the local Xen interface. But OMPI saw the interface, saw that every MPI process had an address in that IP subnet, and assumed that it could be used for MPI communication. Nope! The simple solution here was to add the RHEL6 Xen virtualization device (virb0) to OMPI's exclude list, like this: mpirun --mca btl_tcp_if_exclude lo,virbr0 \ --mca oob_tcp_if_exclude lo,virbr0 ... Then everything worked fine. On Feb 9, 2012, at 4:19 PM, Reuti wrote: > Am 08.02.2012 um 22:52 schrieb Tom Bryan: > >> >> Yes, this should work across multiple machines. And it's using `qrsh > -inherit > ...` so it's failing somewhere in Open MPI - is it working with 1.4.4? I'm not sure. We no longer have our 1.4 test environment, so I'm in the process of building that now. I'll let you know once I have a chance to run that experiment. >> >> You said that both of these cases worked for you in 1.4. Were you running a >> modified version that did not use THREAD_MULTIPLE? I ask because I'm >> getting worse errors in 1.4. I'm using the same code that was working (in >> some cases) with 1.5.4. >> >> I built 1.4.4 with (among other option) >> --with-threads=posix --enable-mpi-threads > > ./configure --prefix=$HOME/local/openmpi-1.4.4-default-thread --with-sge > --with-threads=posix --enable-mpi-threads > > No problems even with THREAD_MULTIPLE. > > Only as stated in singleton mode one or more additional line (looks like one > per slave host, but not always - race condition?): > > [pc15370:31390] [[24201,0],1] routed:binomial: Connection to lifeline > [[24201,0],0] lost > >> >> ompi_mpi_init: orte_init failed >> --> Returned "Data unpack would read past end of buffer" (-26) instead of >> "Success" (0) >> -- >> *** The MPI_Init_thread() function was called before MPI_INIT was invoked. >> *** This is disallowed by the MPI standard. >> *** Your MPI job will now abort. > > Interesting error message, as it's not true to be disallowed. > > -- Reuti > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users -- Jeff Squyres jsquy...@cisco.com For corporate legal information go to: http://www.cisco.com/web/about/doing_business/legal/cri/
Re: [OMPI users] help: sm btl does not work when I specify the same host twice or more in the node list
Can you provide a specific example? I'm able to do this just fine, for example (with the upcoming OMPI 1.4.5): mpirun --host svbu-mpi001,svbu-mpi001,svbu-mpi002,svbu-mpi002 --mca btl sm,openib,self ring On Feb 9, 2012, at 9:31 AM, ya...@adina.com wrote: > Hi all, > > Good morning! > > I have trouble to communicate through sm btl in open MPI, please > check the attached file for my system information. I am using open > MPI 1.4.3, intel compilers V11.1, on linux RHEL 5.4 with kernel 2.6. > > The tests are the following: > > (1) if I specify the btl to mpirun by "--mca btl self,sm,openib", if I did > not specify any of my computing nodes twice or more in the node > list, my job runs fine. However, if I specify any of the computing > nodes twice or more in the node list, it will hang there forever. > > (2) if I did not specify the sm btl to mpirun as "--mca btl > self,openib", I could run my job smoothly, either put any of the > computing nodes twice or more in the node list, or not. > >> From above 2 tests, apparently something wrong with sm btl > interface on my system. As I checked the user archive, sm btl > issue has been encountered due to the comm_spawned > parent/child processes. But this seems not the case here, if I do > not use any of my MPI based solver, only with MPI initialization and > finalization procedures called, it still has this issue. > > Any comments? > > Thanks, > Yiguang > > The following section of this message contains a file attachment > prepared for transmission using the Internet MIME message format. > If you are using Pegasus Mail, or any another MIME-compliant system, > you should be able to save it or view it from within your mailer. > If you cannot, please ask your system administrator for assistance. > > File information --- > File: ompiinfo-config-uname-output.tgz > Date: 9 Feb 2012, 8:58 > Size: 126316 bytes. > Type: Unknown > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users -- Jeff Squyres jsquy...@cisco.com For corporate legal information go to: http://www.cisco.com/web/about/doing_business/legal/cri/