clue?
Thanks
___
users mailing list
us...@open-mpi.org
Subscription: https://www.open-mpi.org/mailman/listinfo.cgi/users
Link to this post:
http://www.open-mpi.org/community/lists/users/2016/05/29201.php
--
David Shrader
HPC-ENV High Performance Computer Systems
Los Alamos National Lab
so as to be able to better work with my users who are just now
starting to ask questions about it. The FAQ pretty much pushes folks to
the man pages, and the mpirun man page doesn't go in to the reasoning.
Thank you for your time,
David
--
David Shrader
HPC-ENV High Performance Computer Sy
'-mca
plm_base_verbose 100' on the mpirun line, but it didn't mention pmi
specifically. Instead, all I could really find was that it was using the
slurm component. Is there something else I can look at in the output
that would have that detail?
Thank you for your time,
David
--
David
That is really good to know. Thanks!
David
On 10/13/2016 12:27 PM, r...@open-mpi.org wrote:
If you are using mpirun, then neither PMI1 or PMI2 are involved at all. ORTE
has its own internal mechanism for handling wireup.
On Oct 13, 2016, at 10:43 AM, David Shrader wrote:
Hello All,
I
ocaldomain:119319] MCW rank 2 bound to socket 0[core 1[hwt 0]]:
[./B/./././././././././././././././.][./././././././././././././././././.]
...output snipped...
If ranking by slot were somehow following something left over by
mapping, I would have expected rank 2 to end up on a different host. So,
now I don't know what to expect from using "--rank-by s
socket, then you cycle across the procs on a node
by round-robin of sockets, assigning all procs on the node before moving to the
next node. If you then added “span” to that directive, we’d round-robin by
socket across all nodes before circling around to the next proc on this node.
HTH
Ralph
socket of that node.
So you can see that you will indeed get the same relative ranking, even though
the mapping was done using a different algorithm.
HTH
Ralph
On Nov 30, 2016, at 2:16 PM, David Shrader wrote:
Hello Ralph,
I do understand that "slot" is an abstract term and is
members" how it
was configured.
Thank you very much for your time,
David
--
David Shrader
HPC-3 High Performance Computer Systems
Los Alamos National Lab
Email: dshrader lanl.gov
with no CLI options tells you a bunch of stuff; "ompi_info --all"
tells you (a lot) more.
On May 5, 2015, at 2:54 PM, David Shrader wrote:
Hello,
Is there a way to tell what configure line was used in building Open MPI from the
installation itself? That is, not from config.log b
he
configure script while waiting to see if anyone else has run in to this.
Thank you for any and all help,
David
--
David Shrader
HPC-3 High Performance Computer Systems
Los Alamos National Lab
Email: dshrader lanl.gov
sed installation places mxm into /opt/mellanox/mxm and not
into /usr/lib64/libmxm.so.
Do you use HPCx (pack of OMPI and MXM and FCA)?
You can download HPCX, extract it anywhere and compile OMPI pointing
to mxm location under HPCX.
Also, HPCx contains rpms for mxm and fca.
M
On Sa
atch - will fix ompi configure logic with your patch.
mxm can be installed in the system and user spaces - both are valid
and supported logic.
M
On Tue, May 26, 2015 at 5:50 PM, David Shrader <mailto:dshra...@lanl.gov>> wrote:
Hello Mike,
This particular instance of mxm was i
ompi configure logic with
your patch.
>
> mxm can be installed in the system and user spaces - both are
valid and supported logic.
>
> M
>
> On Tue, May 26, 2015 at 5:50 PM, David Shrader
mailto:dshra...@lanl.gov>> wrote:
> Hello M
The config.log goes as an attachment
___
users mailing list
us...@open-mpi.org
Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
Link to this post:
http://www.open-mpi.org/community/lists/users/2015/05/26954.php
--
David Shrader
HPC-3 H
rtland Fortran compiler works? Like pgf90 hello.f ?
Bruno
Em qua, 27 de mai de 2015 às 17:40, David Shrader <mailto:dshra...@lanl.gov>> escreveu:
Looking at the config.log, I see this:
pgi-cc-lin64: LICENSE MANAGER PROBLEM: No such feature exists.
Feature: pgi-cc-l
m
not sure if it hanging when used in this fashion is an actual problem
with orte-clean. If it is unexpected behavior, I'll dig some more.
Thank you very much for your time,
David
--
David Shrader
HPC-3 High Performance Computer Systems
Los Alamos National Lab
Email: dshrader lanl.gov
pi.org
Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
Link to this post:
http://www.open-mpi.org/community/lists/users/2015/07/27300.php
--
David Shrader
HPC-3 High Performance Computer Systems
Los Alamos National Lab
Email: dshrader lanl.gov
ing in specifying --with-hcoll? I would prefer not
to use "--with-hcoll=/usr" as I am pretty sure that spurious linker
flags to that area will work their way in when they shouldn't.
Thanks,
David
--
David Shrader
HPC-3 High Performance Computer Systems
Los Alamos National Lab
Email: dshrader lanl.gov
fix the hcoll test (and code) to be correct.
Any configure test that adds /usr/lib and/or /usr/include to any compile flags
is broken.
+1
Gilles filed https://github.com/open-mpi/ompi/pull/796; I just added some
comments to it.
--
David Shrader
HPC-3 High Performance Computer Systems
Los
lease/pull/492
the patch is attached (it required some back-porting)
Cheers,
Gilles
On 8/12/2015 4:01 AM, David Shrader wrote:
I have cloned Gilles' topic/hcoll_config branch and, after running
autogen.pl, have found that './configure --with-hcoll' does indeed
work now. I used
wrote:
Hi David,
This issue is from hcoll library. This could be because of symbol
conflict with ml module. This is fixed recently in HCOLL. Can you
try with "-mca coll ^ml" and see if this workaround works in your setup?
-Devendar
On Wed, Aug 12, 2015 at 9:30 AM, David Shrader <m
n Wed, Aug 12, 2015 at 9:47 AM, David Shrader <mailto:dshra...@lanl.gov>> wrote:
Hey Devendar,
It looks like I still get the error:
[dshrader@zo-fe1 tests]$ mpirun -n 2 -mca coll ^ml ./a.out
App launch reported: 1 (out of 1) daemons - 2 (out of 2) procs
[1439397957.3
.php
http://www.open-mpi.org/community/lists/devel/2015/06/17562.php
-Devendar
On Wed, Aug 12, 2015 at 2:52 PM, David Shrader <mailto:dshra...@lanl.gov>> wrote:
Interesting... the seg faults went away:
[dshrader@zo-fe1 tests]$ export LD_PRELOAD=/usr/lib64/libhcoll.so
[dshr
figure option? This might
force coll_ml to be loaded first even with -mca coll ^ml.
next HPCX is expected to release by end of Aug.
-Devendar
On Wed, Aug 12, 2015 at 3:30 PM, David Shrader <mailto:dshra...@lanl.gov>> wrote:
I remember seeing those, but forgot about them. I am cu
ml/.ompi_ignore
./autogen.pl
./configure ...
make && make install
so the ml component is not even built
Cheers,
Gilles
On 8/13/2015 7:30 AM, David Shrader wrote:
I remember seeing those, but forgot about them. I am curious, though, why using
'-mca coll ^ml' wouldn't work
./configure --enable-mca-no-build=coll-ml ...
This will disable the build of the coll/ml component altogether.
On Aug 13, 2015, at 11:23 AM, David Shrader wrote:
Hey Jeff,
I'm actually not able to find coll_ml related files at that location. All I see
are the following files:
[dsh
x27; in to LDFLAGS/LIBS? Using LDFLAGS/LIBS will
link mxm in to everything, which I would prefer not to do.
Thanks in advance!
David
--
David Shrader
HPC-3 High Performance Computer Systems
Los Alamos National Lab
Email: dshrader lanl.gov
52000)
Both hcoll and mxm where installed using the rpms provided by Mellanox.
Thanks again,
David
On 10/21/2015 09:34 AM, David Shrader wrote:
Hello All,
I'm currently trying to install 1.10.0 with hcoll and mxm, and am
getting an error during configure:
--- MCA component coll:hcoll (
fe3196c2000)
libnsl.so.1 => /lib64/libnsl.so.1 (0x7fe3194a8000)
libutil.so.1 => /lib64/libutil.so.1 (0x7fe3192a5000)
libnl.so.1 => /lib64/libnl.so.1 (0x7fe319052000)
Both hcoll and mxm where installed using the rpms provided by
Mellanox.
rote:
could you please check if you have file /etc/ld.so.conf.d/mxm.conf on
your system?
it will help us understand why hcoll did not detect libmxm.so at the
1st attempt.
Thanks
On Wed, Oct 21, 2015 at 7:19 PM, David Shrader <mailto:dshra...@lanl.gov>> wrote:
We're using T
ks without having to specify LD_LIBRARY_PATH.
So, not an Open MPI issue, but I am very grateful for all the help!
David
On 10/21/2015 12:00 PM, David Shrader wrote:
I'm sorry I missed reporting on that. I do not have
/etc/ld.so.conf.d/mxm.conf.
Interestingly enough, the rpm reports that
e compilers, is there a way to fix my above situation or
prevent it from happening at build time?
Thanks,
David
--
David Shrader
HPC-3 High Performance Computer Systems
Los Alamos National Lab
Email: dshrader lanl.gov
installation in the
/lib/*.la 1.10.1 files and would like to ask if I need to
worry at all? It seems the way files were named and organized in
/lib changed in 1.7 which may be why 1.10.1 is working.
Thank you very much for your time,
David
On 02/10/2016 10:58 AM, David Shrader wrote:
Hello
ight?
___
users mailing list
us...@open-mpi.org
Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/users
Link to this post:
http://www.open-mpi.org/community/lists/users/2016/02/28590.php
--
David Shrader
HPC-3 High Performance Computer Systems
Los Alamos National Lab
Email: dshrader lanl.gov
I forgot to include a link to the official announcement of the change,
and that info might be helpful in navigating the different versions and
backwards compatibility:
https://www.open-mpi.org/community/lists/announce/2015/06/0069.php
Thanks,
David
On 02/26/2016 10:43 AM, David Shrader wrote
35 matches
Mail list logo