Re: [OMPI users] Problem using mpifort(Intel)

2015-10-01 Thread Jeff Squyres (jsquyres)
Excellent; thanks for the feedback. > On Oct 1, 2015, at 2:02 AM, Julien Bodart wrote: > > I have tried the last nightly build and it seems that nobody is complaining > now. > Thanks a lot, > > Julien > >> Date: Fri, 25 Sep 2015 17:55:38 + >> From: "Jeff Squyres (jsquyres)" >> To: "Open

Re: [OMPI users] Setting bind-to none as default via environment?

2015-10-01 Thread Grigory Shamov
Thanks, I guess it be hwloc-base-binding-policy = in the file. Found it. -- Grigory Shamov Westgrid/ComputeCanada Site Lead University of Manitoba E2-588 EITC Building, (204) 474-9625 From: users mailto:users-boun...@open-mpi.org>> on behalf of Nick Papior mailto:nickpap...@gmail.com>> Reply-T

Re: [OMPI users] Setting bind-to none as default via environment?

2015-10-01 Thread Nick Papior
You can define default mca parameters in this file: /etc/openmpi-mca-params.conf 2015-10-01 16:57 GMT+02:00 Grigory Shamov : > Hi All, > > A parhaps naive question: is it possible to set ' mpiexec —bind-to none ' > as a system-wide default in 1.10, like, by setting an OMPI_xxx variable? > > -- >

[OMPI users] Setting bind-to none as default via environment?

2015-10-01 Thread Grigory Shamov
Hi All, A parhaps naive question: is it possible to set ' mpiexec —bind-to none ' as a system-wide default in 1.10, like, by setting an OMPI_xxx variable? -- Grigory Shamov Westgrid/ComputeCanada Site Lead University of Manitoba E2-588 EITC Building, (204) 474-9625

Re: [OMPI users] Using OpenMPI (1.8, 1.10) with Mellanox MXM, ulimits ?

2015-10-01 Thread Mike Dubman
right, it is not attribute of mxm, but general effect. and you are right again - performance engineering will always be needed for best performance in some cases. OMPI, mxm trying to address out of the box performance for any workload, but OS tuning, hw tuning, OMPI or mxm tuning may be needed as

Re: [OMPI users] Using OpenMPI (1.8, 1.10) with Mellanox MXM, ulimits ?

2015-10-01 Thread Mike Dubman
thanks Nathan, you are right, we will fix it. On Wed, Sep 30, 2015 at 7:02 PM, Nathan Hjelm wrote: > > Mike, I see a typo in the mxm warning: > > mxm.c:185 MXM WARN The > 'ulimit -s' on the system is set to 'unlimited'. This may have negative > performance implications. Please set the heap si

Re: [OMPI users] Problem starting jobs

2015-10-01 Thread Emyr James
On 01/10/2015 10:24, Emyr James wrote: "ORTE has lost communication with its daemon located on node: hostname: node123 This is usually due to either a failure of the TCP network connection to the node, or possibly an internal failure of the daemon itself. We cannot recover from this failu

Re: [OMPI users] Using OpenMPI (1.8, 1.10) with Mellanox MXM, ulimits ?

2015-10-01 Thread Dave Love
Mike Dubman writes: > mxm comes with mxm_dump_config utility which provides and explains all > tunables. > Please check HPCX/README file for details. I'm not sure which file that is. I already ran it, according to mxm/share/doc/mxm/README.txt, but it just prints variable names and values. It's

Re: [OMPI users] Using OpenMPI (1.8, 1.10) with Mellanox MXM, ulimits ?

2015-10-01 Thread Dave Love
Mike Dubman writes: > we did not get to the bottom for "why". > Tried different mpi packages (mvapich,intel mpi) and the observation hold > true. Does that mean it's a general effect, unrelated to mxm, or that it is related? > it could be many factors affected by huge heap size (cpu cache misse

[OMPI users] Problem starting jobs

2015-10-01 Thread Emyr James
Hi, I am using openmpi with Platform LSF on our cluster that has 10Gbe connectivity. Sometimes things work fine but we get a lot of occurences of mpi jobs not getting off the ground and the following appears in the log... "ORTE has lost communication with its daemon located on node: hostna

Re: [OMPI users] Problem using mpifort(Intel)

2015-10-01 Thread Julien Bodart
I have tried the last nightly build and it seems that nobody is complaining now. Thanks a lot, Julien > Date: Fri, 25 Sep 2015 17:55:38 + > From: "Jeff Squyres (jsquyres)" > To: "Open MPI User's List" > Subject: Re: [OMPI users] Problem using mpifort(Intel) > Message-ID: <66efd306-64c3-4121