Re: [OMPI users] Building OpenMPI 1.8.7 on XC30

2015-07-28 Thread Gilles Gouaillardet
Erik, the OS X warning (which should not be OS X specific) is fixed in https://github.com/open-mpi/ompi-release/pull/430 it will land into the v2.x series once reviewed in the mean time, feel free to manually apply the patch on the tarball Cheers, Gilles On 7/29/2015 10:35 AM, Erik Schnetter

Re: [OMPI users] Building OpenMPI 1.8.7 on XC30

2015-07-28 Thread Erik Schnetter
Thank you for all the pointers. I was able to build openmpi-v2.x-dev-96-g918650a without problems on Edison, and also on other systems. I'm circumventing the OS X warning by ignoring it via "grep -v"; the other suggestion (--mca oob ^usock) did not work for me. I've tried openmpi-v2.x-dev-100-g26c

Re: [OMPI users] File coherence issues with OpenMPI/torque/NFS (?)

2015-07-28 Thread Gilles Gouaillardet
You are right and I misread your comment. Michael is using ROMIO, which is independent of ompio. Cheers, Gilles On Wednesday, July 29, 2015, Dave Love wrote: > Gilles Gouaillardet > writes: > > > Dave, > > > > On 7/24/2015 1:53 AM, Dave Love wrote: > >> ompio in 1.8 only has pvfs2 (== orangef

Re: [OMPI users] File coherence issues with OpenMPI/torque/NFS (?)

2015-07-28 Thread Dave Love
Gilles Gouaillardet writes: > Dave, > > On 7/24/2015 1:53 AM, Dave Love wrote: >> ompio in 1.8 only has pvfs2 (== orangefs) and ufs support -- which might >> be a good reason to use pvfs2. You'll need an expert to say if you can >> use ufs correctly over an nfs filesystem. (I assume you are act

Re: [OMPI users] SGE segfaulting with OpenMPI 1.8.6

2015-07-28 Thread Dave Love
Ralph Castain writes: > I believe qrsh will execute a simple env command across the allocated nodes - > have you looked into that? qrsh -inherit will run something on any node in the allocation that has a free slot from a tightly integrated parallel environment, but I'm not sure for various rea

Re: [OMPI users] Fatal Error: Cannot read module file 'mpi.mod' opened at (1), because it was created by a different version of GNU Fortran

2015-07-28 Thread Syed Ahsan Ali
Thanks Gilles It solved my issue. Your support is much appreciated. Ahsan On Tue, Jul 28, 2015 at 10:15 AM, Gilles Gouaillardet wrote: > Hi, > > you can run > zcat mpi.mod | head to confirm which gfortran was used to build the > application > > GFORTRAN module version '10' => gcc 4.8.3 > GFORTRA

Re: [OMPI users] Invalid read of size 4 (Valgrind error) with OpenMPI 1.8.7

2015-07-28 Thread Schlottke-Lakemper, Michael
Hi Ralph, That’s what I suspected. Thank you for your confirmation. Michael On 25 Jul 2015, at 16:10 , Ralph Castain mailto:r...@open-mpi.org>> wrote: Looks to me like a false positive - we do malloc some space, and do access different parts of it. However, it looks like we are inside the spa

Re: [OMPI users] strange behavior of MPI_wait() method

2015-07-28 Thread Gilles Gouaillardet
thanks for clarifying there is only one container per host. do you always run 16 tasks per host/container ? or do you always run 16 hosts/containers ? also, do lxc sets iptables when you start a container ? Cheers, Gilles On Tuesday, July 28, 2015, Cristian RUIZ wrote: > Thank you for answe

Re: [OMPI users] strange behavior of MPI_wait() method

2015-07-28 Thread Cristian RUIZ
Thank you for answering. I executed the test with the following command: mpirun --mca btl self,sm,tcp --machinefile machine_file cg.B.128 in both setups. My machine file is composed of 128 lines (each machine hostname is repeated 16 times). There is just one container per machine and the cont

Re: [OMPI users] strange behavior of MPI_wait() method

2015-07-28 Thread Gilles Gouaillardet
Cristian, one more thing... make sure tasks run on the same physical node with and without containers. for example, if in native mode, tasks 0 to 15 run on node 0, then in container mode, tasks 0 to 15 should run on 16 containers hosted by node 0 Cheers, Gilles On Tuesday, July 28, 2015, Gilles

Re: [OMPI users] strange behavior of MPI_wait() method

2015-07-28 Thread Gilles Gouaillardet
Cristian, If the message takes some extra time to land into the receiver, then MPI_Wait will take more time. or even worse, if the sender is late, the receiver will spend even more time in MPI_Wait. First, how do you run 128 tasks on 16 nodes ? if you do a simple mpirun, then you will use sm or v

[OMPI users] strange behavior of MPI_wait() method

2015-07-28 Thread Cristian RUIZ
Hello, I'm measuring the overhead of using Linux container for HPC applications. To do so I was comparing the execution time of NAS parallel benchmarks on two infrastructures: 1) real: 16 real machines 2) container: 16 containers distributed over 16 real machines Each machine used is equippe

Re: [OMPI users] Fatal Error: Cannot read module file 'mpi.mod' opened at (1), because it was created by a different version of GNU Fortran

2015-07-28 Thread Gilles Gouaillardet
Hi, you can run zcat mpi.mod | head to confirm which gfortran was used to build the application GFORTRAN module version '10' => gcc 4.8.3 GFORTRAN module version '12' => gcc 4.9.2 GFORTRAN module version '14' => gcc 5.1.0 i assume the failing command is mpifort ... so you can run mpifort -sho

[OMPI users] Fatal Error: Cannot read module file 'mpi.mod' opened at (1), because it was created by a different version of GNU Fortran

2015-07-28 Thread Syed Ahsan Ali
I am getting this error during installation of an application. Apparently the error seems to be complaining about openmpi being compiled with different version of gnu fortran but I am sure that it was compiled with gcc-4.9.2. The same is also being used for application compilation. I am using open