[OMPI users] mpif90 error with different openmpi editions

2013-06-18 Thread xu
my code get this error under openmpi 1.6.4 
mpif90 -O2 -m64 -fbounds-check -ffree-line-length-0 -c -o 2dem_mpi.o 
2dem_mpi.f90
Fatal Error: Reading module mpi at line 110 column 30: Expected string
If I use mpif90: Open MPI 1.3.3
It compiles ok. What the problem for this?


Re: [OMPI users] lsb_launch failed: 0

2013-06-18 Thread Singh, Bharati (GE Global Research, consultant)
Hi Jeff Squyres

Issue have resolved after resetting environment variables in the user
script.

Thanks,
Bharati Singh

-Original Message-
From: users-boun...@open-mpi.org [mailto:users-boun...@open-mpi.org] On
Behalf Of Jeff Squyres (jsquyres)
Sent: Monday, June 17, 2013 7:13 PM
To: Open MPI Users
Subject: Re: [OMPI users] lsb_launch failed: 0

I'm not an LSF expert, but this usually means that the Open MPI helper
executable named "orted" was not able to be found on the remote nodes.

Is your PATH set properly, both locally and remotely, such that the Open
MPI executables can be found?


On Jun 17, 2013, at 7:01 AM, "Singh, Bharati (GE Global Research,
consultant)"  wrote:

> Hi Team,
>  
> Our users jobs are exiting with below error for random nodes. could
you please help us to resolve this issue?
>  
> [root@bng1grcdc200 output.228472]# cat user_script.stderr 
> [bng1grcdc181:08381] [[54933,0],0] ORTE_ERROR_LOG: The specified 
> application failed to start in file plm_lsf_module.c at line 308 
> [bng1grcdc181:08381] lsb_launch failed: 0
> --
>  A daemon (pid unknown) died unexpectedly on signal 1  while 
> attempting to launch so we are aborting.
>  
> There may be more information reported by the environment (see above).
>  
> This may be because the daemon was unable to find all the needed 
> shared libraries on the remote node. You may set your LD_LIBRARY_PATH 
> to have the location of the shared libraries on the remote nodes and 
> this will automatically be forwarded to the remote nodes.
> --
> 
> --
>  mpirun noticed that the job aborted, but has no info as to the 
> process that caused that situation.
> --
> 
> --
>  mpirun was unable to cleanly terminate the daemons on the nodes 
> shown below. Additional manual cleanup may be required - please refer 
> to the "orte-clean" tool for assistance.
>

--
> bng1grcdc172 - daemon did not report back when launched
> bng1grcdc154 - daemon did not report back when launched
> bng1grcdc198 - daemon did not report back when launched
> bng1grcdc183 - daemon did not report back when launched
> bng1grcdc187 - daemon did not report back when launched
> bng1grcdc196 - daemon did not report back when launched
> bng1grcdc153 - daemon did not report back when launched
> bng1grcdc173 - daemon did not report back when launched
> bng1grcdc193 - daemon did not report back when launched
> bng1grcdc185 - daemon did not report back when launched
> bng1grcdc176 - daemon did not report back when launched
> bng1grcdc190 - daemon did not report back when launched
> bng1grcdc194 - daemon did not report back when launched
> bng1grcdc156 - daemon did not report back when launched
>  
>  
> Thanks,
> Bharati Singh
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users


--
Jeff Squyres
jsquy...@cisco.com
For corporate legal information go to:
http://www.cisco.com/web/about/doing_business/legal/cri/


___
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users



Re: [OMPI users] mpif90 error with different openmpi editions

2013-06-18 Thread Jeff Squyres (jsquyres)
Can you send all the information for your 1.6.4 installation listed here:

http://www.open-mpi.org/community/help/

On Jun 18, 2013, at 12:28 AM, xu  wrote:

> my code get this error under openmpi 1.6.4
> mpif90 -O2 -m64 -fbounds-check -ffree-line-length-0 -c -o 2dem_mpi.o 
> 2dem_mpi.f90 Fatal Error: Reading module mpi at line 110 column 30: Expected 
> string
> If I use mpif90: Open MPI 1.3.3
> It compiles ok. What the problem for this?
> 
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users


-- 
Jeff Squyres
jsquy...@cisco.com
For corporate legal information go to: 
http://www.cisco.com/web/about/doing_business/legal/cri/




Re: [OMPI users] Troubles Building OpenMPI on MinGW-w64 (GCC 4.8.0)

2013-06-18 Thread Jeff Squyres (jsquyres)
Note that the latest binary build we have for Windows is 1.6.2 (after that, as 
Ralph mentioned, our Windows developer left, and due to persistent lack of 
interest, Windows support has effectively been dropped from Open MPI).


On Jun 17, 2013, at 6:34 PM, Ralph Castain  wrote:

> You should just be able to pull a native Win version down from our web site - 
> building it yourself can be an adventure on Windows.
> 
> I'm afraid our Windows supporter has moved on to other pastures, so our 
> ability to help with such builds is pretty minimal. I don't think he did 
> anything with MinGW - pretty sure he worked strictly with VisualC
> 
> On Jun 17, 2013, at 3:27 PM, Haroogan  wrote:
> 
>>> What version of OMPI are you using?
>> Latest, stable: 1.6.4.
>> ___
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users


-- 
Jeff Squyres
jsquy...@cisco.com
For corporate legal information go to: 
http://www.cisco.com/web/about/doing_business/legal/cri/




Re: [OMPI users] mpif90 error with different openmpi editions

2013-06-18 Thread Gus Correa

On 06/18/2013 12:28 AM, xu wrote:

my code get this error under openmpi 1.6.4
mpif90 -O2 -m64 -fbounds-check -ffree-line-length-0 -c -o 2dem_mpi.o
2dem_mpi.f90 Fatal Error: Reading module mpi at line 110 column 30:
Expected string
If I use mpif90: Open MPI 1.3.3
It compiles ok. What the problem for this?


Make sure you are not mixing environment variables (PATH and 
LD_LIBRARY_PATH) of the two OMPI versions you installed.





Re: [OMPI users] MPI_Init_thread hangs in OpenMPI 1.7.1 when using --enable-mpi-thread-multiple

2013-06-18 Thread Elias Rudberg

Hi,

I now tried using --enable-opal-multi-threads (I added an "s" in the  
end, since "--enable-opal-multi-thread" was not recognized by the  
configure script).


However, my test code still hangs in the same way as before.

To be clear, I now configure like this:
./configure --enable-mpi-thread-multiple --enable-opal-multi-threads

but the problem remains: MPI_Init_thread still hangs when trying to  
run the test case with -np 2.


Any other ideas?
Are there other configure options that must be used together with  
--enable-mpi-thread-multiple to make it work?


Best regards,
Elias


Quoting Ralph Castain :


Hmmm...well, your code runs fine for me:

Ralphs-iMac:mpi rhc$ mpirun -n 2 ./thread_init
Calling MPI_Init_thread...
Calling MPI_Init_thread...
MPI_Init_thread returned, provided = 3
MPI_Init_thread returned, provided = 3
Ralphs-iMac:mpi rhc$

I think the key, however, is that you also have to configure with  
--enable-opal-multi-thread.



On Jun 16, 2013, at 8:54 AM, Elias Rudberg  wrote:


Hello!

I would like to report what seems to be a bug in MPI_Init_thread in  
OpenMPI 1.7.1.


The bug can be reproduced with the following test program  
(test_mpi_thread_support.c):

===
#include 
#include 
int main(int argc, const char* argv[]) {
 int provided = -1;
 printf("Calling MPI_Init_thread...\n");
 MPI_Init_thread(NULL, NULL, MPI_THREAD_MULTIPLE, &provided);
 printf("MPI_Init_thread returned, provided = %d\n", provided);
 MPI_Finalize();
 return 0;
}
===

When trying to run this when OpenMPI was configured with  
--enable-mpi-thread-multiple, the program hangs when trying to run  
using anything more than one process.


Steps I use to reproduce this in Ubuntu:

(1) Download openmpi-1.7.1.tar.gz

(2) Configure like this:
./configure --enable-mpi-thread-multiple

(3) make

(4) Compile test program like this:
mpicc test_mpi_thread_support.c

(5) Run like this:
mpirun -np 2 ./a.out
Then you see the following two lines of output:
Calling MPI_Init_thread...
Calling MPI_Init_thread...
And then it hangs.

MPI_Init_thread did not hang in earlier OpenMPI versions (for  
example it worked in 1.5.* and 1.6.*), so it seems like a bug  
introduced in 1.7.


The description above shows how I reproduce this in Ubuntu on my  
local desktop computer, but the same problem exists for the OpenMPI  
1.7.1 installation at the UPPMAX computer center where I wan to run  
my code in the end. I don't know all details about how they  
installed it there, but I know they set  
--enable-mpi-thread-multiple. So maybe it hangs in 1.7.1 on any  
computer as long as you use MPI_THREAD_MULTIPLE. At least I have  
not seen it work anywhere.


Do you agree that this is a bug, or am I doing something wrong?

Best regards,
Elias


___
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users



___
users mailing list
us...@open-mpi.org
http://www.open-mpi.org/mailman/listinfo.cgi/users







[OMPI users] Trouble with Sending Multiple messages to the Same Machine

2013-06-18 Thread Claire Williams
Hi guys ☺!

I'm working with a simple "Hello, World" MPI program that has one master and is 
sending one message to each worker, receives a message back from each of the 
workers, and re-sends a new message. This unfortunately is not working :(. When 
the master only sends one message to each worker, and then receives it, it is 
working fine, but there are problems with sending more than one message to each 
worker. When it happens, it prints the error:

[[401,1],0][../../../../../openmpi-1.6.3/ompi/mca/btl/tcp/btl_tcp_endpoint.c:638:mca_btl_tcp_endpoint_complete_connect]
 connect() to 192.168.X.X failed: No route to host (113)

I'm wondering how I can go about fixing this. This program is running across 
multiple Linux nodes, by the way :). 

BTW, I'm a girl.

Re: [OMPI users] Trouble with Sending Multiple messages to the Same Machine

2013-06-18 Thread Gus Correa

On 06/18/2013 01:14 PM, Claire Williams wrote:

Hi guys ☺!

I'm working with a simple "Hello, World" MPI program that has one master
and is sending one message to each worker, receives a message back from
each of the workers, and re-sends a new message. This unfortunately is
not working :(. When the master only sends one message to each worker,
and then receives it, it is working fine, but there are problems with
sending more than one message to each worker. When it happens, it prints
the error:

[[401,1],0][../../../../../openmpi-1.6.3/ompi/mca/btl/tcp/btl_tcp_endpoint.c:638:mca_btl_tcp_endpoint_complete_connect]
connect() to 192.168.X.X failed: No route to host (113)

I'm wondering how I can go about fixing this. This program is running
across multiple Linux nodes, by the way :).

BTW, I'm a girl.



... and so was Ada.

"connect() to 192.168.X.X failed: No route to host"
suggests that your 192.168.X.X node cannot be
reached by a message sender.
Can you ping back and forth across every pair of nodes?

http://en.wikipedia.org/wiki/Women_in_computing
http://anitaborg.org/news/profiles-of-technical-women/famous-women-in-computer-science/
http://en.wikipedia.org/wiki/Ada_Lovelace




Re: [OMPI users] MPI_Init_thread(..., MPI_THREAD_SERIALIZED) hangs under OSX 10.8.4 if compiled with OpenMPI 1.7.1

2013-06-18 Thread Hans Ekkehard Plesser

Hi!

I have tried building with --enable-opal-multi-threads (it was missing in 
MacPorts), but that did not help.

My problem seems very similar to that reported by Elias Rudberg in a separate 
thread, using Ubuntu.

I am using the clang compiler.

Best regards,
Hans

On 18. juni 2013, at 00.08, Ralph Castain wrote:

> For 1.7, you also have to configure with --enable-opal-multi-thread. I 
> suspect MacPorts doesn't do that, so you might want to configure and build 
> your own version.
> 
> 
> On Jun 16, 2013, at 5:27 AM, Hans Ekkehard Plesser 
>  wrote:
> 
>> 
>> On 15. juni 2013, at 01.31, Ralph Castain wrote:
>> 
>>> I have no idea how MacPorts configures OMPI - did you check the output of 
>>> ompi_info to see if threading was even enabled?
>> 
>> ompi_info shows
>> 
>> Thread support: posix (MPI_THREAD_MULTIPLE: yes, progress: no)
>> 
>> for 1.6.4 and 1.7.1. I could post the full ompi_info if that would help.
>> 
>> Best regards,
>> Hans
>> 
>>> 
>>> 
>>> On Jun 14, 2013, at 3:12 PM, Hans Ekkehard Plesser 
>>>  wrote:
>>> 
 
 Hi!
 
 I use OpenMPI 1.7.1 from MacPorts (+threads +gcc47). When compiling a 
 simple hello world program calling
 
 MPI_Init_thread(&argc, &argv, MPI_THREAD_SERIALIZED, &provided);
 
 the program hangs if run on more than one process. All works fine if I
 
 - either use MPI_THREAD_SINGLE
 - or use OpenMPI 1.6.4.
 
 The actual compiler invocation and compiler version are
 
 [Macintosh:~/tmp/hw] hep% openmpicc -o hw_mpi hw_mpi.c --showme
 /usr/bin/clang -o hw_mpi hw_mpi.c -I/opt/local/include/openmpi 
 -L/opt/local/lib -lmpi
 
 [Macintosh:~/tmp/hw] hep% /usr/bin/clang --version
 Apple clang version 4.1 (tags/Apple/clang-421.11.66) (based on LLVM 3.1svn)
 Target: x86_64-apple-darwin12.4.0
 Thread model: posix
 
 The full program code is shown below. I have also reported this to the 
 MacPorts bug tracker.
 
 Best regards,
 Hans E Plesser
 
 --
 #include 
 
 #include 
 
 int main(int argc, char **argv)
 {
 int provided;
 MPI_Init_thread(&argc, &argv, MPI_THREAD_SERIALIZED, &provided);
 
 printf("Provided threading level: %d.\n", provided);
 
 int rank;
 MPI_Comm_rank(MPI_COMM_WORLD, &rank);
 
 printf("Hello, world! This is rank %d.\n", rank);
 
 MPI_Finalize();
 return 0;
 }
 --
 
 
 --
 
 Dr. Hans Ekkehard Plesser, Associate Professor
 Head, Basic Science Section
 
 Dept. of Mathematical Sciences and Technology
 Norwegian University of Life Sciences
 PO Box 5003, 1432 Aas, Norway
 
 Phone +47 6496 5467
 Fax   +47 6496 5401
 Email hans.ekkehard.ples...@umb.no
 Home  http://arken.umb.no/~plesser
 
 
 ___
 users mailing list
 us...@open-mpi.org
 http://www.open-mpi.org/mailman/listinfo.cgi/users
>>> 
>>> 
>>> ___
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
>> --
>> 
>> Dr. Hans Ekkehard Plesser, Associate Professor
>> Head, Basic Science Section
>> 
>> Dept. of Mathematical Sciences and Technology
>> Norwegian University of Life Sciences
>> PO Box 5003, 1432 Aas, Norway
>> 
>> Phone +47 6496 5467
>> Fax   +47 6496 5401
>> Email hans.ekkehard.ples...@umb.no
>> Home  http://arken.umb.no/~plesser
>> 
>> 
>> ___
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> 
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users

--

Dr. Hans Ekkehard Plesser, Associate Professor
Head, Basic Science Section

Dept. of Mathematical Sciences and Technology
Norwegian University of Life Sciences
PO Box 5003, 1432 Aas, Norway

Phone +47 6496 5467
Fax   +47 6496 5401
Email hans.ekkehard.ples...@umb.no
Home  http://arken.umb.no/~plesser




[OMPI users] Application hangs on mpi_waitall

2013-06-18 Thread Blosch, Edwin L
I'm running OpenMPI 1.6.4 and seeing a problem where mpi_waitall never returns. 
 The case runs fine with MVAPICH.  The logic associated with the communications 
has been extensively debugged in the past; we don't think it has errors.   Each 
process posts non-blocking receives, non-blocking sends, and then does waitall 
on all the outstanding requests.

The work is broken down into 960 chunks. If I run with 960 processes (60 nodes 
of 16 cores each), things seem to work.  If I use 160 processes (each process 
handling 6 chunks of work), then each process is handling 6 times as much 
communication, and that is the case that hangs with OpenMPI 1.6.4; again, seems 
to work with MVAPICH.  Is there an obvious place to start, diagnostically?  
We're using the openib btl.

Thanks,

Ed


Re: [OMPI users] MPI_Init_thread(..., MPI_THREAD_SERIALIZED) hangs under OSX 10.8.4 if compiled with OpenMPI 1.7.1

2013-06-18 Thread Ralph Castain
I'd suggest just pursuing this with her thread then - like I said, with the 
config I passed her, it works fine for me, and that is on OSX 10.8.4. However, 
I use the gcc compiler and not clang.


On Jun 18, 2013, at 2:23 PM, Hans Ekkehard Plesser 
 wrote:

> 
> Hi!
> 
> I have tried building with --enable-opal-multi-threads (it was missing in 
> MacPorts), but that did not help.
> 
> My problem seems very similar to that reported by Elias Rudberg in a separate 
> thread, using Ubuntu.
> 
> I am using the clang compiler.
> 
> Best regards,
> Hans
> 
> On 18. juni 2013, at 00.08, Ralph Castain wrote:
> 
>> For 1.7, you also have to configure with --enable-opal-multi-thread. I 
>> suspect MacPorts doesn't do that, so you might want to configure and build 
>> your own version.
>> 
>> 
>> On Jun 16, 2013, at 5:27 AM, Hans Ekkehard Plesser 
>>  wrote:
>> 
>>> 
>>> On 15. juni 2013, at 01.31, Ralph Castain wrote:
>>> 
 I have no idea how MacPorts configures OMPI - did you check the output of 
 ompi_info to see if threading was even enabled?
>>> 
>>> ompi_info shows
>>> 
>>>Thread support: posix (MPI_THREAD_MULTIPLE: yes, progress: no)
>>> 
>>> for 1.6.4 and 1.7.1. I could post the full ompi_info if that would help.
>>> 
>>> Best regards,
>>> Hans
>>> 
 
 
 On Jun 14, 2013, at 3:12 PM, Hans Ekkehard Plesser 
  wrote:
 
> 
> Hi!
> 
> I use OpenMPI 1.7.1 from MacPorts (+threads +gcc47). When compiling a 
> simple hello world program calling
> 
> MPI_Init_thread(&argc, &argv, MPI_THREAD_SERIALIZED, &provided);
> 
> the program hangs if run on more than one process. All works fine if I
> 
> - either use MPI_THREAD_SINGLE
> - or use OpenMPI 1.6.4.
> 
> The actual compiler invocation and compiler version are
> 
> [Macintosh:~/tmp/hw] hep% openmpicc -o hw_mpi hw_mpi.c --showme
> /usr/bin/clang -o hw_mpi hw_mpi.c -I/opt/local/include/openmpi 
> -L/opt/local/lib -lmpi
> 
> [Macintosh:~/tmp/hw] hep% /usr/bin/clang --version
> Apple clang version 4.1 (tags/Apple/clang-421.11.66) (based on LLVM 
> 3.1svn)
> Target: x86_64-apple-darwin12.4.0
> Thread model: posix
> 
> The full program code is shown below. I have also reported this to the 
> MacPorts bug tracker.
> 
> Best regards,
> Hans E Plesser
> 
> --
> #include 
> 
> #include 
> 
> int main(int argc, char **argv)
> {
> int provided;
> MPI_Init_thread(&argc, &argv, MPI_THREAD_SERIALIZED, &provided);
> 
> printf("Provided threading level: %d.\n", provided);
> 
> int rank;
> MPI_Comm_rank(MPI_COMM_WORLD, &rank);
> 
> printf("Hello, world! This is rank %d.\n", rank);
> 
> MPI_Finalize();
> return 0;
> }
> --
> 
> 
> --
> 
> Dr. Hans Ekkehard Plesser, Associate Professor
> Head, Basic Science Section
> 
> Dept. of Mathematical Sciences and Technology
> Norwegian University of Life Sciences
> PO Box 5003, 1432 Aas, Norway
> 
> Phone +47 6496 5467
> Fax   +47 6496 5401
> Email hans.ekkehard.ples...@umb.no
> Home  http://arken.umb.no/~plesser
> 
> 
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users
 
 
 ___
 users mailing list
 us...@open-mpi.org
 http://www.open-mpi.org/mailman/listinfo.cgi/users
>>> 
>>> --
>>> 
>>> Dr. Hans Ekkehard Plesser, Associate Professor
>>> Head, Basic Science Section
>>> 
>>> Dept. of Mathematical Sciences and Technology
>>> Norwegian University of Life Sciences
>>> PO Box 5003, 1432 Aas, Norway
>>> 
>>> Phone +47 6496 5467
>>> Fax   +47 6496 5401
>>> Email hans.ekkehard.ples...@umb.no
>>> Home  http://arken.umb.no/~plesser
>>> 
>>> 
>>> ___
>>> users mailing list
>>> us...@open-mpi.org
>>> http://www.open-mpi.org/mailman/listinfo.cgi/users
>> 
>> 
>> ___
>> users mailing list
>> us...@open-mpi.org
>> http://www.open-mpi.org/mailman/listinfo.cgi/users
> 
> --
> 
> Dr. Hans Ekkehard Plesser, Associate Professor
> Head, Basic Science Section
> 
> Dept. of Mathematical Sciences and Technology
> Norwegian University of Life Sciences
> PO Box 5003, 1432 Aas, Norway
> 
> Phone +47 6496 5467
> Fax   +47 6496 5401
> Email hans.ekkehard.ples...@umb.no
> Home  http://arken.umb.no/~plesser
> 
> 
> ___
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users