Excellent; thanks for the feedback.
> On Oct 1, 2015, at 2:02 AM, Julien Bodart wrote:
>
> I have tried the last nightly build and it seems that nobody is complaining
> now.
> Thanks a lot,
>
> Julien
>
>> Date: Fri, 25 Sep 2015 17:55:38 +
>> From: "Jeff Squyres (jsquyres)"
>> To: "Open
Thanks, I guess it be hwloc-base-binding-policy = in the file. Found it.
--
Grigory Shamov
Westgrid/ComputeCanada Site Lead
University of Manitoba
E2-588 EITC Building,
(204) 474-9625
From: users mailto:users-boun...@open-mpi.org>> on
behalf of Nick Papior mailto:nickpap...@gmail.com>>
Reply-T
You can define default mca parameters in this file:
/etc/openmpi-mca-params.conf
2015-10-01 16:57 GMT+02:00 Grigory Shamov :
> Hi All,
>
> A parhaps naive question: is it possible to set ' mpiexec —bind-to none '
> as a system-wide default in 1.10, like, by setting an OMPI_xxx variable?
>
> --
>
Hi All,
A parhaps naive question: is it possible to set ' mpiexec —bind-to none ' as a
system-wide default in 1.10, like, by setting an OMPI_xxx variable?
--
Grigory Shamov
Westgrid/ComputeCanada Site Lead
University of Manitoba
E2-588 EITC Building,
(204) 474-9625
right, it is not attribute of mxm, but general effect.
and you are right again - performance engineering will always be needed for
best performance in some cases.
OMPI, mxm trying to address out of the box performance for any workload,
but OS tuning, hw tuning, OMPI or mxm tuning may be needed as
thanks Nathan, you are right, we will fix it.
On Wed, Sep 30, 2015 at 7:02 PM, Nathan Hjelm wrote:
>
> Mike, I see a typo in the mxm warning:
>
> mxm.c:185 MXM WARN The
> 'ulimit -s' on the system is set to 'unlimited'. This may have negative
> performance implications. Please set the heap si
On 01/10/2015 10:24, Emyr James wrote:
"ORTE has lost communication with its daemon located on node:
hostname: node123
This is usually due to either a failure of the TCP network
connection to the node, or possibly an internal failure of
the daemon itself. We cannot recover from this failu
Mike Dubman writes:
> mxm comes with mxm_dump_config utility which provides and explains all
> tunables.
> Please check HPCX/README file for details.
I'm not sure which file that is. I already ran it, according to
mxm/share/doc/mxm/README.txt, but it just prints variable names and
values.
It's
Mike Dubman writes:
> we did not get to the bottom for "why".
> Tried different mpi packages (mvapich,intel mpi) and the observation hold
> true.
Does that mean it's a general effect, unrelated to mxm, or that it is
related?
> it could be many factors affected by huge heap size (cpu cache misse
Hi,
I am using openmpi with Platform LSF on our cluster that has 10Gbe
connectivity.
Sometimes things work fine but we get a lot of occurences of mpi jobs
not getting off the ground and the following appears in the log...
"ORTE has lost communication with its daemon located on node:
hostna
I have tried the last nightly build and it seems that nobody is complaining now.
Thanks a lot,
Julien
> Date: Fri, 25 Sep 2015 17:55:38 +
> From: "Jeff Squyres (jsquyres)"
> To: "Open MPI User's List"
> Subject: Re: [OMPI users] Problem using mpifort(Intel)
> Message-ID: <66efd306-64c3-4121
11 matches
Mail list logo