[OMPI users] mpif90 error with different openmpi editions
my code get this error under openmpi 1.6.4 mpif90 -O2 -m64 -fbounds-check -ffree-line-length-0 -c -o 2dem_mpi.o 2dem_mpi.f90 Fatal Error: Reading module mpi at line 110 column 30: Expected string If I use mpif90: Open MPI 1.3.3 It compiles ok. What the problem for this?
Re: [OMPI users] lsb_launch failed: 0
Hi Jeff Squyres Issue have resolved after resetting environment variables in the user script. Thanks, Bharati Singh -Original Message- From: users-boun...@open-mpi.org [mailto:users-boun...@open-mpi.org] On Behalf Of Jeff Squyres (jsquyres) Sent: Monday, June 17, 2013 7:13 PM To: Open MPI Users Subject: Re: [OMPI users] lsb_launch failed: 0 I'm not an LSF expert, but this usually means that the Open MPI helper executable named "orted" was not able to be found on the remote nodes. Is your PATH set properly, both locally and remotely, such that the Open MPI executables can be found? On Jun 17, 2013, at 7:01 AM, "Singh, Bharati (GE Global Research, consultant)" wrote: > Hi Team, > > Our users jobs are exiting with below error for random nodes. could you please help us to resolve this issue? > > [root@bng1grcdc200 output.228472]# cat user_script.stderr > [bng1grcdc181:08381] [[54933,0],0] ORTE_ERROR_LOG: The specified > application failed to start in file plm_lsf_module.c at line 308 > [bng1grcdc181:08381] lsb_launch failed: 0 > -- > A daemon (pid unknown) died unexpectedly on signal 1 while > attempting to launch so we are aborting. > > There may be more information reported by the environment (see above). > > This may be because the daemon was unable to find all the needed > shared libraries on the remote node. You may set your LD_LIBRARY_PATH > to have the location of the shared libraries on the remote nodes and > this will automatically be forwarded to the remote nodes. > -- > > -- > mpirun noticed that the job aborted, but has no info as to the > process that caused that situation. > -- > > -- > mpirun was unable to cleanly terminate the daemons on the nodes > shown below. Additional manual cleanup may be required - please refer > to the "orte-clean" tool for assistance. > -- > bng1grcdc172 - daemon did not report back when launched > bng1grcdc154 - daemon did not report back when launched > bng1grcdc198 - daemon did not report back when launched > bng1grcdc183 - daemon did not report back when launched > bng1grcdc187 - daemon did not report back when launched > bng1grcdc196 - daemon did not report back when launched > bng1grcdc153 - daemon did not report back when launched > bng1grcdc173 - daemon did not report back when launched > bng1grcdc193 - daemon did not report back when launched > bng1grcdc185 - daemon did not report back when launched > bng1grcdc176 - daemon did not report back when launched > bng1grcdc190 - daemon did not report back when launched > bng1grcdc194 - daemon did not report back when launched > bng1grcdc156 - daemon did not report back when launched > > > Thanks, > Bharati Singh > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users -- Jeff Squyres jsquy...@cisco.com For corporate legal information go to: http://www.cisco.com/web/about/doing_business/legal/cri/ ___ users mailing list us...@open-mpi.org http://www.open-mpi.org/mailman/listinfo.cgi/users
Re: [OMPI users] mpif90 error with different openmpi editions
Can you send all the information for your 1.6.4 installation listed here: http://www.open-mpi.org/community/help/ On Jun 18, 2013, at 12:28 AM, xu wrote: > my code get this error under openmpi 1.6.4 > mpif90 -O2 -m64 -fbounds-check -ffree-line-length-0 -c -o 2dem_mpi.o > 2dem_mpi.f90 Fatal Error: Reading module mpi at line 110 column 30: Expected > string > If I use mpif90: Open MPI 1.3.3 > It compiles ok. What the problem for this? > > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users -- Jeff Squyres jsquy...@cisco.com For corporate legal information go to: http://www.cisco.com/web/about/doing_business/legal/cri/
Re: [OMPI users] Troubles Building OpenMPI on MinGW-w64 (GCC 4.8.0)
Note that the latest binary build we have for Windows is 1.6.2 (after that, as Ralph mentioned, our Windows developer left, and due to persistent lack of interest, Windows support has effectively been dropped from Open MPI). On Jun 17, 2013, at 6:34 PM, Ralph Castain wrote: > You should just be able to pull a native Win version down from our web site - > building it yourself can be an adventure on Windows. > > I'm afraid our Windows supporter has moved on to other pastures, so our > ability to help with such builds is pretty minimal. I don't think he did > anything with MinGW - pretty sure he worked strictly with VisualC > > On Jun 17, 2013, at 3:27 PM, Haroogan wrote: > >>> What version of OMPI are you using? >> Latest, stable: 1.6.4. >> ___ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users > > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users -- Jeff Squyres jsquy...@cisco.com For corporate legal information go to: http://www.cisco.com/web/about/doing_business/legal/cri/
Re: [OMPI users] mpif90 error with different openmpi editions
On 06/18/2013 12:28 AM, xu wrote: my code get this error under openmpi 1.6.4 mpif90 -O2 -m64 -fbounds-check -ffree-line-length-0 -c -o 2dem_mpi.o 2dem_mpi.f90 Fatal Error: Reading module mpi at line 110 column 30: Expected string If I use mpif90: Open MPI 1.3.3 It compiles ok. What the problem for this? Make sure you are not mixing environment variables (PATH and LD_LIBRARY_PATH) of the two OMPI versions you installed.
Re: [OMPI users] MPI_Init_thread hangs in OpenMPI 1.7.1 when using --enable-mpi-thread-multiple
Hi, I now tried using --enable-opal-multi-threads (I added an "s" in the end, since "--enable-opal-multi-thread" was not recognized by the configure script). However, my test code still hangs in the same way as before. To be clear, I now configure like this: ./configure --enable-mpi-thread-multiple --enable-opal-multi-threads but the problem remains: MPI_Init_thread still hangs when trying to run the test case with -np 2. Any other ideas? Are there other configure options that must be used together with --enable-mpi-thread-multiple to make it work? Best regards, Elias Quoting Ralph Castain : Hmmm...well, your code runs fine for me: Ralphs-iMac:mpi rhc$ mpirun -n 2 ./thread_init Calling MPI_Init_thread... Calling MPI_Init_thread... MPI_Init_thread returned, provided = 3 MPI_Init_thread returned, provided = 3 Ralphs-iMac:mpi rhc$ I think the key, however, is that you also have to configure with --enable-opal-multi-thread. On Jun 16, 2013, at 8:54 AM, Elias Rudberg wrote: Hello! I would like to report what seems to be a bug in MPI_Init_thread in OpenMPI 1.7.1. The bug can be reproduced with the following test program (test_mpi_thread_support.c): === #include #include int main(int argc, const char* argv[]) { int provided = -1; printf("Calling MPI_Init_thread...\n"); MPI_Init_thread(NULL, NULL, MPI_THREAD_MULTIPLE, &provided); printf("MPI_Init_thread returned, provided = %d\n", provided); MPI_Finalize(); return 0; } === When trying to run this when OpenMPI was configured with --enable-mpi-thread-multiple, the program hangs when trying to run using anything more than one process. Steps I use to reproduce this in Ubuntu: (1) Download openmpi-1.7.1.tar.gz (2) Configure like this: ./configure --enable-mpi-thread-multiple (3) make (4) Compile test program like this: mpicc test_mpi_thread_support.c (5) Run like this: mpirun -np 2 ./a.out Then you see the following two lines of output: Calling MPI_Init_thread... Calling MPI_Init_thread... And then it hangs. MPI_Init_thread did not hang in earlier OpenMPI versions (for example it worked in 1.5.* and 1.6.*), so it seems like a bug introduced in 1.7. The description above shows how I reproduce this in Ubuntu on my local desktop computer, but the same problem exists for the OpenMPI 1.7.1 installation at the UPPMAX computer center where I wan to run my code in the end. I don't know all details about how they installed it there, but I know they set --enable-mpi-thread-multiple. So maybe it hangs in 1.7.1 on any computer as long as you use MPI_THREAD_MULTIPLE. At least I have not seen it work anywhere. Do you agree that this is a bug, or am I doing something wrong? Best regards, Elias ___ users mailing list us...@open-mpi.org http://www.open-mpi.org/mailman/listinfo.cgi/users ___ users mailing list us...@open-mpi.org http://www.open-mpi.org/mailman/listinfo.cgi/users
[OMPI users] Trouble with Sending Multiple messages to the Same Machine
Hi guys ☺! I'm working with a simple "Hello, World" MPI program that has one master and is sending one message to each worker, receives a message back from each of the workers, and re-sends a new message. This unfortunately is not working :(. When the master only sends one message to each worker, and then receives it, it is working fine, but there are problems with sending more than one message to each worker. When it happens, it prints the error: [[401,1],0][../../../../../openmpi-1.6.3/ompi/mca/btl/tcp/btl_tcp_endpoint.c:638:mca_btl_tcp_endpoint_complete_connect] connect() to 192.168.X.X failed: No route to host (113) I'm wondering how I can go about fixing this. This program is running across multiple Linux nodes, by the way :). BTW, I'm a girl.
Re: [OMPI users] Trouble with Sending Multiple messages to the Same Machine
On 06/18/2013 01:14 PM, Claire Williams wrote: Hi guys ☺! I'm working with a simple "Hello, World" MPI program that has one master and is sending one message to each worker, receives a message back from each of the workers, and re-sends a new message. This unfortunately is not working :(. When the master only sends one message to each worker, and then receives it, it is working fine, but there are problems with sending more than one message to each worker. When it happens, it prints the error: [[401,1],0][../../../../../openmpi-1.6.3/ompi/mca/btl/tcp/btl_tcp_endpoint.c:638:mca_btl_tcp_endpoint_complete_connect] connect() to 192.168.X.X failed: No route to host (113) I'm wondering how I can go about fixing this. This program is running across multiple Linux nodes, by the way :). BTW, I'm a girl. ... and so was Ada. "connect() to 192.168.X.X failed: No route to host" suggests that your 192.168.X.X node cannot be reached by a message sender. Can you ping back and forth across every pair of nodes? http://en.wikipedia.org/wiki/Women_in_computing http://anitaborg.org/news/profiles-of-technical-women/famous-women-in-computer-science/ http://en.wikipedia.org/wiki/Ada_Lovelace
Re: [OMPI users] MPI_Init_thread(..., MPI_THREAD_SERIALIZED) hangs under OSX 10.8.4 if compiled with OpenMPI 1.7.1
Hi! I have tried building with --enable-opal-multi-threads (it was missing in MacPorts), but that did not help. My problem seems very similar to that reported by Elias Rudberg in a separate thread, using Ubuntu. I am using the clang compiler. Best regards, Hans On 18. juni 2013, at 00.08, Ralph Castain wrote: > For 1.7, you also have to configure with --enable-opal-multi-thread. I > suspect MacPorts doesn't do that, so you might want to configure and build > your own version. > > > On Jun 16, 2013, at 5:27 AM, Hans Ekkehard Plesser > wrote: > >> >> On 15. juni 2013, at 01.31, Ralph Castain wrote: >> >>> I have no idea how MacPorts configures OMPI - did you check the output of >>> ompi_info to see if threading was even enabled? >> >> ompi_info shows >> >> Thread support: posix (MPI_THREAD_MULTIPLE: yes, progress: no) >> >> for 1.6.4 and 1.7.1. I could post the full ompi_info if that would help. >> >> Best regards, >> Hans >> >>> >>> >>> On Jun 14, 2013, at 3:12 PM, Hans Ekkehard Plesser >>> wrote: >>> Hi! I use OpenMPI 1.7.1 from MacPorts (+threads +gcc47). When compiling a simple hello world program calling MPI_Init_thread(&argc, &argv, MPI_THREAD_SERIALIZED, &provided); the program hangs if run on more than one process. All works fine if I - either use MPI_THREAD_SINGLE - or use OpenMPI 1.6.4. The actual compiler invocation and compiler version are [Macintosh:~/tmp/hw] hep% openmpicc -o hw_mpi hw_mpi.c --showme /usr/bin/clang -o hw_mpi hw_mpi.c -I/opt/local/include/openmpi -L/opt/local/lib -lmpi [Macintosh:~/tmp/hw] hep% /usr/bin/clang --version Apple clang version 4.1 (tags/Apple/clang-421.11.66) (based on LLVM 3.1svn) Target: x86_64-apple-darwin12.4.0 Thread model: posix The full program code is shown below. I have also reported this to the MacPorts bug tracker. Best regards, Hans E Plesser -- #include #include int main(int argc, char **argv) { int provided; MPI_Init_thread(&argc, &argv, MPI_THREAD_SERIALIZED, &provided); printf("Provided threading level: %d.\n", provided); int rank; MPI_Comm_rank(MPI_COMM_WORLD, &rank); printf("Hello, world! This is rank %d.\n", rank); MPI_Finalize(); return 0; } -- -- Dr. Hans Ekkehard Plesser, Associate Professor Head, Basic Science Section Dept. of Mathematical Sciences and Technology Norwegian University of Life Sciences PO Box 5003, 1432 Aas, Norway Phone +47 6496 5467 Fax +47 6496 5401 Email hans.ekkehard.ples...@umb.no Home http://arken.umb.no/~plesser ___ users mailing list us...@open-mpi.org http://www.open-mpi.org/mailman/listinfo.cgi/users >>> >>> >>> ___ >>> users mailing list >>> us...@open-mpi.org >>> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> -- >> >> Dr. Hans Ekkehard Plesser, Associate Professor >> Head, Basic Science Section >> >> Dept. of Mathematical Sciences and Technology >> Norwegian University of Life Sciences >> PO Box 5003, 1432 Aas, Norway >> >> Phone +47 6496 5467 >> Fax +47 6496 5401 >> Email hans.ekkehard.ples...@umb.no >> Home http://arken.umb.no/~plesser >> >> >> ___ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users > > > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users -- Dr. Hans Ekkehard Plesser, Associate Professor Head, Basic Science Section Dept. of Mathematical Sciences and Technology Norwegian University of Life Sciences PO Box 5003, 1432 Aas, Norway Phone +47 6496 5467 Fax +47 6496 5401 Email hans.ekkehard.ples...@umb.no Home http://arken.umb.no/~plesser
[OMPI users] Application hangs on mpi_waitall
I'm running OpenMPI 1.6.4 and seeing a problem where mpi_waitall never returns. The case runs fine with MVAPICH. The logic associated with the communications has been extensively debugged in the past; we don't think it has errors. Each process posts non-blocking receives, non-blocking sends, and then does waitall on all the outstanding requests. The work is broken down into 960 chunks. If I run with 960 processes (60 nodes of 16 cores each), things seem to work. If I use 160 processes (each process handling 6 chunks of work), then each process is handling 6 times as much communication, and that is the case that hangs with OpenMPI 1.6.4; again, seems to work with MVAPICH. Is there an obvious place to start, diagnostically? We're using the openib btl. Thanks, Ed
Re: [OMPI users] MPI_Init_thread(..., MPI_THREAD_SERIALIZED) hangs under OSX 10.8.4 if compiled with OpenMPI 1.7.1
I'd suggest just pursuing this with her thread then - like I said, with the config I passed her, it works fine for me, and that is on OSX 10.8.4. However, I use the gcc compiler and not clang. On Jun 18, 2013, at 2:23 PM, Hans Ekkehard Plesser wrote: > > Hi! > > I have tried building with --enable-opal-multi-threads (it was missing in > MacPorts), but that did not help. > > My problem seems very similar to that reported by Elias Rudberg in a separate > thread, using Ubuntu. > > I am using the clang compiler. > > Best regards, > Hans > > On 18. juni 2013, at 00.08, Ralph Castain wrote: > >> For 1.7, you also have to configure with --enable-opal-multi-thread. I >> suspect MacPorts doesn't do that, so you might want to configure and build >> your own version. >> >> >> On Jun 16, 2013, at 5:27 AM, Hans Ekkehard Plesser >> wrote: >> >>> >>> On 15. juni 2013, at 01.31, Ralph Castain wrote: >>> I have no idea how MacPorts configures OMPI - did you check the output of ompi_info to see if threading was even enabled? >>> >>> ompi_info shows >>> >>>Thread support: posix (MPI_THREAD_MULTIPLE: yes, progress: no) >>> >>> for 1.6.4 and 1.7.1. I could post the full ompi_info if that would help. >>> >>> Best regards, >>> Hans >>> On Jun 14, 2013, at 3:12 PM, Hans Ekkehard Plesser wrote: > > Hi! > > I use OpenMPI 1.7.1 from MacPorts (+threads +gcc47). When compiling a > simple hello world program calling > > MPI_Init_thread(&argc, &argv, MPI_THREAD_SERIALIZED, &provided); > > the program hangs if run on more than one process. All works fine if I > > - either use MPI_THREAD_SINGLE > - or use OpenMPI 1.6.4. > > The actual compiler invocation and compiler version are > > [Macintosh:~/tmp/hw] hep% openmpicc -o hw_mpi hw_mpi.c --showme > /usr/bin/clang -o hw_mpi hw_mpi.c -I/opt/local/include/openmpi > -L/opt/local/lib -lmpi > > [Macintosh:~/tmp/hw] hep% /usr/bin/clang --version > Apple clang version 4.1 (tags/Apple/clang-421.11.66) (based on LLVM > 3.1svn) > Target: x86_64-apple-darwin12.4.0 > Thread model: posix > > The full program code is shown below. I have also reported this to the > MacPorts bug tracker. > > Best regards, > Hans E Plesser > > -- > #include > > #include > > int main(int argc, char **argv) > { > int provided; > MPI_Init_thread(&argc, &argv, MPI_THREAD_SERIALIZED, &provided); > > printf("Provided threading level: %d.\n", provided); > > int rank; > MPI_Comm_rank(MPI_COMM_WORLD, &rank); > > printf("Hello, world! This is rank %d.\n", rank); > > MPI_Finalize(); > return 0; > } > -- > > > -- > > Dr. Hans Ekkehard Plesser, Associate Professor > Head, Basic Science Section > > Dept. of Mathematical Sciences and Technology > Norwegian University of Life Sciences > PO Box 5003, 1432 Aas, Norway > > Phone +47 6496 5467 > Fax +47 6496 5401 > Email hans.ekkehard.ples...@umb.no > Home http://arken.umb.no/~plesser > > > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users ___ users mailing list us...@open-mpi.org http://www.open-mpi.org/mailman/listinfo.cgi/users >>> >>> -- >>> >>> Dr. Hans Ekkehard Plesser, Associate Professor >>> Head, Basic Science Section >>> >>> Dept. of Mathematical Sciences and Technology >>> Norwegian University of Life Sciences >>> PO Box 5003, 1432 Aas, Norway >>> >>> Phone +47 6496 5467 >>> Fax +47 6496 5401 >>> Email hans.ekkehard.ples...@umb.no >>> Home http://arken.umb.no/~plesser >>> >>> >>> ___ >>> users mailing list >>> us...@open-mpi.org >>> http://www.open-mpi.org/mailman/listinfo.cgi/users >> >> >> ___ >> users mailing list >> us...@open-mpi.org >> http://www.open-mpi.org/mailman/listinfo.cgi/users > > -- > > Dr. Hans Ekkehard Plesser, Associate Professor > Head, Basic Science Section > > Dept. of Mathematical Sciences and Technology > Norwegian University of Life Sciences > PO Box 5003, 1432 Aas, Norway > > Phone +47 6496 5467 > Fax +47 6496 5401 > Email hans.ekkehard.ples...@umb.no > Home http://arken.umb.no/~plesser > > > ___ > users mailing list > us...@open-mpi.org > http://www.open-mpi.org/mailman/listinfo.cgi/users