Erik,
the OS X warning (which should not be OS X specific) is fixed in
https://github.com/open-mpi/ompi-release/pull/430
it will land into the v2.x series once reviewed
in the mean time, feel free to manually apply the patch on the tarball
Cheers,
Gilles
On 7/29/2015 10:35 AM, Erik Schnetter
Thank you for all the pointers. I was able to
build openmpi-v2.x-dev-96-g918650a without problems on Edison, and also on
other systems.
I'm circumventing the OS X warning by ignoring it via "grep -v"; the other
suggestion (--mca oob ^usock) did not work for me. I've
tried openmpi-v2.x-dev-100-g26c
You are right and I misread your comment.
Michael is using ROMIO, which is independent of ompio.
Cheers,
Gilles
On Wednesday, July 29, 2015, Dave Love wrote:
> Gilles Gouaillardet > writes:
>
> > Dave,
> >
> > On 7/24/2015 1:53 AM, Dave Love wrote:
> >> ompio in 1.8 only has pvfs2 (== orangef
Gilles Gouaillardet writes:
> Dave,
>
> On 7/24/2015 1:53 AM, Dave Love wrote:
>> ompio in 1.8 only has pvfs2 (== orangefs) and ufs support -- which might
>> be a good reason to use pvfs2. You'll need an expert to say if you can
>> use ufs correctly over an nfs filesystem. (I assume you are act
Ralph Castain writes:
> I believe qrsh will execute a simple env command across the allocated nodes -
> have you looked into that?
qrsh -inherit will run something on any node in the allocation that has
a free slot from a tightly integrated parallel environment, but I'm not
sure for various rea
Thanks Gilles
It solved my issue. Your support is much appreciated.
Ahsan
On Tue, Jul 28, 2015 at 10:15 AM, Gilles Gouaillardet wrote:
> Hi,
>
> you can run
> zcat mpi.mod | head to confirm which gfortran was used to build the
> application
>
> GFORTRAN module version '10' => gcc 4.8.3
> GFORTRA
Hi Ralph,
That’s what I suspected. Thank you for your confirmation.
Michael
On 25 Jul 2015, at 16:10 , Ralph Castain
mailto:r...@open-mpi.org>> wrote:
Looks to me like a false positive - we do malloc some space, and do access
different parts of it. However, it looks like we are inside the spa
thanks for clarifying there is only one container per host.
do you always run 16 tasks per host/container ?
or do you always run 16 hosts/containers ?
also, do lxc sets iptables when you start a container ?
Cheers,
Gilles
On Tuesday, July 28, 2015, Cristian RUIZ wrote:
> Thank you for answe
Thank you for answering. I executed the test with the following command:
mpirun --mca btl self,sm,tcp --machinefile machine_file cg.B.128 in
both setups. My machine file is composed of 128 lines (each machine
hostname is repeated 16 times). There is just one container per machine
and the cont
Cristian,
one more thing...
make sure tasks run on the same physical node with and without containers.
for example, if in native mode, tasks 0 to 15 run on node 0, then in
container mode, tasks 0 to 15 should run on 16 containers hosted by node 0
Cheers,
Gilles
On Tuesday, July 28, 2015, Gilles
Cristian,
If the message takes some extra time to land into the receiver, then
MPI_Wait will take more time.
or even worse, if the sender is late, the receiver will spend even more
time in MPI_Wait.
First, how do you run 128 tasks on 16 nodes ?
if you do a simple mpirun, then you will use sm or v
Hello,
I'm measuring the overhead of using Linux container for HPC
applications. To do so I was comparing the execution time of NAS
parallel benchmarks on two infrastructures:
1) real: 16 real machines
2) container: 16 containers distributed over 16 real machines
Each machine used is equippe
Hi,
you can run
zcat mpi.mod | head to confirm which gfortran was used to build the
application
GFORTRAN module version '10' => gcc 4.8.3
GFORTRAN module version '12' => gcc 4.9.2
GFORTRAN module version '14' => gcc 5.1.0
i assume the failing command is mpifort ...
so you can run
mpifort -sho
I am getting this error during installation of an application.
Apparently the error seems to be complaining about openmpi being
compiled with different version of gnu fortran but I am sure that it
was compiled with gcc-4.9.2. The same is also being used for
application compilation.
I am using open
14 matches
Mail list logo