Re: [OMPI users] OpenMPI-5.0.6: -x LD_LIBRARY_PATH not able to load shared objects

2025-02-14 Thread Patrick Begou
Bad answer, sorry I did not managed prted was part of OpenMPI stack. Le 14/02/2025 à 19:19, Patrick Begou a écrit : Hi Sangam could you check that the install location of the library is the same on all the nodes ?  May be checking LD_LIBRARY_PATH after sourcing the intel vars.sh file ? I&#

Re: [OMPI users] OpenMPI-5.0.6: -x LD_LIBRARY_PATH not able to load shared objects

2025-02-14 Thread Patrick Begou
On Fri, Feb 14, 2025 at 6:30 PM Patrick Begou wrote: Le 14/02/2025 à 13:22, Sangam B a écrit : > Hi, > > OpenMPI-5.0.6 is compiled with ucx-1.18 and Intel 1api 2024 v2.1 > compilers. An mpi program is compiled with this openmpi-5.0.6. > > While submi

Re: [OMPI users] OpenMPI-5.0.6: -x LD_LIBRARY_PATH not able to load shared objects

2025-02-14 Thread Patrick Begou
Le 14/02/2025 à 13:22, Sangam B a écrit : Hi, OpenMPI-5.0.6 is compiled with ucx-1.18 and Intel 1api 2024 v2.1 compilers. An mpi program is compiled with this openmpi-5.0.6. While submitting job thru PBS on a Linux cluster, the intel compilers is sourced and the same is passed thru OpenMPI's

Re: [OMPI users] Performances problems with OpenMPI 5.0.5 and UCX 1.17.0 with Qlogiq infiniband

2024-10-02 Thread Patrick Begou via users
e network is welcome. Patrick Le 30/09/2024 à 18:41, Patrick Begou via users a écrit : Hi Nathan thanks for this suggestion. I have understood that now all is managed by the UCX layer. Am I wrong ? These options do not seams to work with my openMPI 5.0.5 build. But I've built OpenMPI on

Re: [OMPI users] Performances problems with OpenMPI 5.0.5 and UCX 1.17.0 with Qlogiq infiniband

2024-09-30 Thread Patrick Begou via users
On Sep 30, 2024, at 10:18 AM, Patrick Begou via users wrote: Hi, I'm working on refreshing an old cluster with Almalinux 9 (instead of CentOS6 😕) and building a fresh OpenMPI 5.0.5 environment. I've reached the step where OpenMPI begins to work with ucx 1.17 and Pmix 5.0.3 but

[OMPI users] Performances problems with OpenMPI 5.0.5 and UCX 1.17.0 with Qlogiq infiniband

2024-09-30 Thread Patrick Begou via users
Hi, I'm working on refreshing an old cluster with Almalinux 9 (instead of CentOS6 😕) and building a fresh OpenMPI 5.0.5 environment. I've reached the step where OpenMPI begins to work with ucx 1.17 and Pmix 5.0.3 but not totally. Nodes are using a Qlogic QDR HBA with a managed Qlogic switch (

Re: [OMPI users] OpenMPI and names of the nodes in a cluster

2022-06-21 Thread Patrick Begou via users
by default, but it has certain expectations about the format of hostnames. Try using the "naive" regex component, instead. -- Jeff Squyres jsquy...@cisco.com ________ From: Patrick Begou Sent: Thursday, June 16, 2022 9:48 AM To: Jeff Squyres (jsquyres);

Re: [OMPI users] OpenMPI and names of the nodes in a cluster

2022-06-16 Thread Patrick Begou via users
occurring? -- Jeff Squyres jsquy...@cisco.com From: users on behalf of Patrick Begou via users Sent: Thursday, June 16, 2022 3:21 AM To: Open MPI Users Cc: Patrick Begou Subject: [OMPI users] OpenMPI and names of the nodes in a cluster Hi all, we are

[OMPI users] OpenMPI and names of the nodes in a cluster

2022-06-16 Thread Patrick Begou via users
Hi all, we are facing a serious problem with OpenMPI (4.0.2) that we have deployed on a cluster. We do not manage this large cluster and the names of the nodes do not agree with Internet standards for protocols: they contain a "_" (underscore) character. So OpenMPI complains about this and d

Re: [OMPI users] Need help for troubleshooting OpenMPI performances

2022-04-07 Thread Patrick Begou via users
e legacy openib btl? If the former, is it built with multi threading support? If the latter, I suggest you give UCX - built with multi threading support - a try and see how it goes Cheers, Gilles On Thu, Mar 24, 2022 at 5:43 PM Patrick Begou via users wrote: Le 28/02/2022 à 17:56, Pa

Re: [OMPI users] Need help for troubleshooting OpenMPI performances

2022-03-24 Thread Patrick Begou via users
Le 28/02/2022 à 17:56, Patrick Begou via users a écrit : Hi, I meet a performance problem with OpenMPI on my cluster. In some situation my parallel code is really slow (same binary running on a different mesh). To investigate, the fortran code code is built with profiling option (mpifort

[OMPI users] Need help for troubleshooting OpenMPI performances

2022-02-28 Thread Patrick Begou via users
Hi, I meet a performance problem with OpenMPI on my cluster. In some situation my parallel code is really slow (same binary running on a different mesh). To investigate, the fortran code code is built with profiling option (mpifort -p -O3.) and launched on 91 cores. One mon.out file pe

Re: [OMPI users] [EXTERNAL] Re: OpenMPI 4.0.5 error with Omni-path

2021-02-08 Thread Patrick Begou via users
debian, so i can't be much > more help > > if i had to guess totally pulling junk from the air, there's probably > something incompatible with PSM and OPA when running specifically on debian > (likely due to library versioning). i don't know how common that is, so

Re: [OMPI users] [EXTERNAL] Re: OpenMPI 4.0.5 error with Omni-path

2021-01-27 Thread Patrick Begou via users
ot > sure if it's supposed to stop at some point > > i'm running rhel7, gcc 10.1, openmpi 4.0.5rc2, with-ofi, > without-{psm,ucx,verbs} > > On Tue, Jan 26, 2021 at 3:44 PM Patrick Begou via users > wrote: > > > >

Re: [OMPI users] OpenMPI 4.0.5 error with Omni-path

2021-01-26 Thread Patrick Begou via users
MPI app that reproduces > the problem? I can’t think of another way I can give you more help > without being able to see what’s going on. It’s always possible > there’s a bug in the PSM2 MTL but it would be surprising at this point. > > Sent from my iPad > >> On Jan 26, 20

Re: [OMPI users] OpenMPI 4.0.5 error with Omni-path

2021-01-26 Thread Patrick Begou via users
Hi all, I ran many tests today. I saw that an older 4.0.2 version of OpenMPI packaged with Nix was running using openib. So I add the --with-verbs option to setup this module. That I can see now is that: mpirun -hostfile $OAR_NODEFILE *--mca mtl psm -mca btl_openib_allow_ib true* - the

Re: [OMPI users] OpenMPI 4.0.5 error with Omni-path

2021-01-26 Thread Patrick Begou via users
07 but expect 4007 but it fails too. Patrick Le 25/01/2021 à 19:34, Ralph Castain via users a écrit : > I think you mean add "--mca mtl ofi" to the mpirun cmd line > > >> On Jan 25, 2021, at 10:18 AM, Heinz, Michael William via users >> wrote: >> >>

Re: [OMPI users] OpenMPI 4.0.5 error with Omni-path

2021-01-25 Thread Patrick Begou via users
Hi Howard and Michael, thanks for your feedback. I did not want to write a toot long mail with non pertinent information so I just show how the two different builds give different result. I'm using a small test case based on my large code, the same used to show the memory leak with mpi_Alltoallv c

[OMPI users] OpenMPI 4.0.5 error with Omni-path

2021-01-25 Thread Patrick Begou via users
Hi, I'm trying to deploy OpenMPI 4.0.5 on the university's supercomputer: * Debian GNU/Linux 9 (stretch) * Intel Corporation Omni-Path HFI Silicon 100 Series [discrete] (rev 11) and for several days I have a bug (wrong results using MPI_AllToAllW) on this server when using OmniPath. Running

Re: [OMPI users] [version 2.1.5] invalid memory reference

2018-10-11 Thread Patrick Begou
ver. All prerequisites compile successfully with 3.1.2. Patrick -- === | Equipe M.O.S.T. | | | Patrick BEGOU | mailto:patrick.be...@grenoble-inp.fr | | LEGI| | | BP 53 X | Tel 04 76

Re: [OMPI users] [version 2.1.5] invalid memory reference

2018-09-19 Thread Patrick Begou
at 3:50 AM Patrick Begou <mailto:patrick.be...@legi.grenoble-inp.fr>> wrote: Hi I'm moving a large CFD code from Gcc 4.8.5/OpenMPI 1.7.3 to Gcc 7.3.0/OpenMPI 2.1.5 and with this latest config I have random segfaults. Same binary, same server, same number of proces

[OMPI users] [version 2.1.5] invalid memory reference

2018-09-18 Thread Patrick Begou
cxx --without-slurm --enable-mpi-thread-multiple --enable-debug --enable-mem-debug Any help appreciated Patrick -- === | Equipe M.O.S.T. | | | Patrick

Re: [OMPI users] openmpi-3.1.2 libgfortran conflict

2018-09-03 Thread Patrick Begou
Solved. Strange conflict (not explained) after several compilation test of OpenMPI with gcc7. Solved by removing the destination directory before any new "make install" command. Patrick Patrick Begou wrote: I am compiling openmpi-3.1.2 on CentOS 7 with GCC 7.3 installed in /opt

Re: [OMPI users] MPI advantages over PBS

2018-09-01 Thread Patrick Begou
ttps://lists.open-mpi.org/mailman/listinfo/users -- === | Equipe M.O.S.T. | | | Patrick BEGOU | mailto:patrick.be...@grenoble-inp.fr | | LEGI|

[OMPI users] openmpi-3.1.2 libgfortran conflict

2018-09-01 Thread Patrick Begou
7f5e7ae1b000)     libc.so.6 => /lib64/libc.so.6 (0x7f5e7aa4e000)     /lib64/ld-linux-x86-64.so.2 (0x7f5e7b945000) -- ======= | Equipe M.O.S.T. | | | Patrick BEGOU | mailto:patrick.be...@grenoble-inp.fr | | LEGI

Re: [OMPI users] OpenMPI-1.10.0 bind-to core error

2015-09-24 Thread Patrick Begou
o-not-exist in (/home/PROJECTS/... Patrick -- === | Equipe M.O.S.T. | | | Patrick BEGOU | mailto:patrick.be...@grenoble-inp.fr | | LEGI|

Re: [OMPI users] OpenMPI-1.10.0 bind-to core error

2015-09-21 Thread Patrick Begou
;frog" ;-) */ Cheers, Gilles On Friday, September 18, 2015, Patrick Begou <mailto:patrick.be...@legi.grenoble-inp.fr>> wrote: Gilles Gouaillardet wrote: Patrick, by the way, this will work when running on a single node. i do not know what will happen when you

Re: [OMPI users] OpenMPI-1.10.0 bind-to core error

2015-09-18 Thread Patrick Begou
binding more processes than cpus on a resource: Bind to: CORE Node:frog5 #processes: 2 #cpus: 1 You can override this protection by adding the "overload-allowed" option to your binding directive. Cheers, Gilles On 9/18/2015 4:54 PM, Patrick Begou w

Re: [OMPI users] OpenMPI-1.10.0 bind-to core error

2015-09-18 Thread Patrick Begou
STS]$ *cat frog7.txt* frog7 slots=4 Patrick On Sep 16, 2015, at 1:00 AM, Patrick Begou <mailto:patrick.be...@legi.grenoble-inp.fr>> wrote: Thanks all for your answers, I've added some details about the tests I have run. See below. Ralph Castain wrote: Not precisely corre

Re: [OMPI users] OpenMPI-1.10.0 bind-to core error

2015-09-16 Thread Patrick Begou
option implicitly sets "max_slots" equal to the "slots" value for each node. It also looks like -map-by has a way to implement it as well (see man page). Thanks for letting me/us know about this. On a system of mine I sort of depend on the -nooversubscribe behavio

[OMPI users] OpenMPI-1.10.0 bind-to core error

2015-09-15 Thread Patrick Begou
== | Equipe M.O.S.T. | | | Patrick BEGOU | mailto:patrick.be...@grenoble-inp.fr | | LEGI| | | BP 53 X | Tel 04 76 82 51 35 | | 38041 GRENOBLE CEDEX

Re: [OMPI users] slowdown with infiniband and latest CentOS kernel

2014-02-27 Thread Patrick Begou
member that OpenFOAM does this yet. Patrick -- === | Equipe M.O.S.T. | | | Patrick BEGOU | mailto:patrick.be...@

Re: [OMPI users] Compiling OpenMPI 1.7.x with core afinity

2013-10-21 Thread Patrick Begou
Jeff Squyres (jsquyres) wrote: Can you manually install a recent version of hwloc (http://www.open-mpi.org/projects/hwloc/) on kareline, and run lstopo on it? Send the output here. What kind of machine is kareline? On Oct 21, 2013, at 11:09 AM, Patrick Begou kareline (front-end) is a

Re: [OMPI users] Compiling OpenMPI 1.7.x with core afinity

2013-10-21 Thread Patrick Begou
to bind by default, then you need to add that directive to your default MCA param file: /etc/openmpi-mca-params.conf On Oct 21, 2013, at 3:17 AM, Patrick Begou wrote: I am compiling OpenMPI 1.7.3 and 1.7.2 with GCC 4.8.1 but I'm unable to activate some binding policy at compi

[OMPI users] Compiling OpenMPI 1.7.x with core afinity

2013-10-21 Thread Patrick Begou
====== | Equipe M.O.S.T. | | | Patrick BEGOU | mailto:patrick.be...@grenoble-inp.fr | | LEGI| | | BP 53 X | Tel 04 76 82 51 35 | | 38041 GRENOBLE CEDEX| Fax 04 76 82 52 71 | ===

Re: [OMPI users] MPI_BCAST and fortran subarrays

2011-12-15 Thread Patrick Begou
Patrick -- === | Equipe M.O.S.T. | http://most.hmg.inpg.fr | | Patrick BEGOU | | | LEGI| mailto:patrick.be...@hmg.inpg.fr | | BP 53 X | Tel 04 76 82 51 35 | | 38041 GRENOBLE CEDEX

Re: [OMPI users] MPI_BCAST and fortran subarrays

2011-12-14 Thread Patrick Begou
Q. 0) print*,tab CALL MPI_FINALIZE(ierr) END PROGRAM bide -- ======= | Equipe M.O.S.T. | http://most.hmg.inpg.fr | | Patrick BEGOU | | | LEGI | mailto:patrick.be...@hmg.inpg.fr | | BP 53 X | Tel 04 76 82 51 35 | | 38041 GRENOBLE CEDEX | Fax 04 76 82 52 71 | ===

[OMPI users] MPI_BCAST and fortran subarrays

2011-12-12 Thread Patrick Begou
about this behavior ? Patrick -- === | Equipe M.O.S.T. | http://most.hmg.inpg.fr | | Patrick BEGOU | | | LEGI| mailto:patrick.be...@hmg.inpg.fr | | BP 53 X

Re: [OMPI users] Changing plm_rsh_agent system wide

2011-10-26 Thread Patrick Begou
r the suggestion, I was fixed on a syntax error in my config... Patrick On Oct 26, 2011, at 3:11 AM, Patrick Begou wrote: I need to change system wide how OpenMPI launch the jobs on the nodes of my cluster. Setting: export OMPI_MCA_plm_rsh_agent=oarsh works fine but I would like this config to b

[OMPI users] Changing plm_rsh_agent system wide

2011-10-26 Thread Patrick Begou
.3 Open MPI SVN revision: r23834 Open MPI release date: Oct 05, 2010 Thanks Patrick -- === | Equipe M.O.S.T. | http://most.hmg.inpg.fr | | Patrick BEGOU | | | LEGI