Re: [OMPI users] Error build Open MPI 4.1.5 with GCC 11.3

2023-07-18 Thread Tom Kacvinsky via users
ok.  But deleting the whole source tree and re-configuring + re-building also works.  🙂  From: Jeffrey Layton <layto...@gmail.com> Sent: Tuesday, July 18, 2023 11:38 AM To: Jeff Squyres (jsquyres) <jsquy...@cisco.com> Cc: Open MPI Users <users@lists.open-mpi.org> Subject: Re: [OMPI u

Re: [OMPI users] Error build Open MPI 4.1.5 with GCC 11.3

2023-07-18 Thread Bernstein, Noam CIV USN NRL (6393) Washington DC (USA) via users
-ran configure, it probably would have built ok. But deleting the whole source tree and re-configuring + re-building also works. 🙂 From: Jeffrey Layton mailto:layto...@gmail.com>> Sent: Tuesday, July 18, 2023 11:38 AM To: Jeff Squyres (jsquyres) mailto:jsquy..

Re: [OMPI users] Error build Open MPI 4.1.5 with GCC 11.3

2023-07-18 Thread Jeff Squyres (jsquyres) via users
n Sent: Tuesday, July 18, 2023 12:51 PM To: Jeff Squyres (jsquyres) Cc: Open MPI Users Subject: Re: [OMPI users] Error build Open MPI 4.1.5 with GCC 11.3 As soon as you pointed out /usr/lib/gcc/x86_64-linux-gnu/9/include/float.h that made me think of the previous build. I did "make clean"

Re: [OMPI users] Error build Open MPI 4.1.5 with GCC 11.3

2023-07-18 Thread Jeffrey Layton via users
source tree and re-configuring > + re-building also works. 🙂 > -- > *From:* Jeffrey Layton > *Sent:* Tuesday, July 18, 2023 11:38 AM > *To:* Jeff Squyres (jsquyres) > *Cc:* Open MPI Users > *Subject:* Re: [OMPI users] Error build Open MPI 4.1.5 with GCC 11.3 > > Jeff, >

Re: [OMPI users] Error handling

2023-07-18 Thread George Bosilca via users
Alex, How are your values "random" if you provide correct values ? Even for negative values you could use MIN to pick one value and return it. What is the problem with `MPI_Abort` ? it does seem to do what you want. George. On Tue, Jul 18, 2023 at 4:38 AM Alexander Stadik via users < users@li

Re: [OMPI users] Error build Open MPI 4.1.5 with GCC 11.3

2023-07-18 Thread Jeff Squyres (jsquyres) via users
n MPI Users Subject: Re: [OMPI users] Error build Open MPI 4.1.5 with GCC 11.3 Jeff, Thanks for the tip - it started me thinking a bit. I was using a directory in my /home account with 4.1.5 that I had previously built using GCC 9.4 (Ubuntu 20.04). I rebuilt the system with Ubuntu-22.04 but I d

Re: [OMPI users] Error build Open MPI 4.1.5 with GCC 11.3

2023-07-18 Thread Jeffrey Layton via users
Jeff, Thanks for the tip - it started me thinking a bit. I was using a directory in my /home account with 4.1.5 that I had previously built using GCC 9.4 (Ubuntu 20.04). I rebuilt the system with Ubuntu-22.04 but I did a backup of /home. Then I copied the 4.1.5 directory to /home again. I checke

Re: [OMPI users] Error build Open MPI 4.1.5 with GCC 11.3

2023-07-17 Thread Jeff Squyres (jsquyres) via users
That's a little odd. Usually, the specific .h files that are listed as dependencies came from somewhere​ -- usually either part of the GNU Autotools dependency analysis. I'm guessing that /usr/lib/gcc/x86_64-linux-gnu/9/include/float.h doesn't actually exist on your system -- but then how did

Re: [OMPI users] Error using rankfile to bind multiple cores on the same node for threaded OpenMPI application (users Digest, Vol 4715, Issue 1)

2022-02-07 Thread David Perozzi via users
Hi Bernd, Thanks for your valuable input! Your suggested approach indeed seems like the correct one and is actually what I've always wanted to do. In the past, I've also asked our cluster support if there was this possibility, but they always suggested the following approach: export OMP_NUM

Re: [OMPI users] Error using rankfile to bind multiple cores on the same node for threaded OpenMPI application

2022-02-03 Thread David Perozzi via users
, 2019 [1,0]:opal:version:full:4.0.2 [1,0]:opal:version:repo:v4.0.2 [1,0]:opal:version:release_date:Oct 07, 2019 [1,0]:mpi-api:version:full:3.1.0 [1,0]:ident:4.0.2 Best Christoph - Original Message - From: "Open MPI Users" To: "Open MPI Users" Cc: "Ralph Castain&q

Re: [OMPI users] Error using rankfile to bind multiple cores on the same node for threaded OpenMPI application

2022-02-03 Thread Ralph Castain via users
release_date:Oct 07, 2019 [1,0]:opal:version:full:4.0.2 [1,0]:opal:version:repo:v4.0.2 [1,0]:opal:version:release_date:Oct 07, 2019 [1,0]:mpi-api:version:full:3.1.0 [1,0]:ident:4.0.2 Best Christoph - Original Message - From: "Open MPI Users" To: "Open MPI Users" Cc:

Re: [OMPI users] Error using rankfile to bind multiple cores on the same node for threaded OpenMPI application (users Digest, Vol 4715, Issue 1)

2022-02-03 Thread Bernd Dammann via users
Hi David, On 03/02/2022 00:03 , David Perozzi wrote: Helo, I'm trying to run a code implemented with OpenMPI and OpenMP (for threading) on a large cluster that uses LSF for the job scheduling and dispatch. The problem with LSF is that it is not very straightforward to allocate and bind the r

Re: [OMPI users] Error using rankfile to bind multiple cores on the same node for threaded OpenMPI application

2022-02-03 Thread David Perozzi via users
pi:version:repo:v4.0.2 [1,0]:ompi:version:release_date:Oct 07, 2019 [1,0]:orte:version:full:4.0.2 [1,0]:orte:version:repo:v4.0.2 [1,0]:orte:version:release_date:Oct 07, 2019 [1,0]:opal:version:full:4.0.2 [1,0]:opal:version:repo:v4.0.2 [1,0]:opal:version:release_date:Oct 07, 2019 [1,0]:mpi-api:versio

Re: [OMPI users] Error using rankfile to bind multiple cores on the same node for threaded OpenMPI application

2022-02-03 Thread Ralph Castain via users
ackage:Open MPI spackapps@eu-c7-042-03 Distribution >>> [1,0]:ompi:version:full:4.0.2 >>> [1,0]:ompi:version:repo:v4.0.2 >>> [1,0]:ompi:version:release_date:Oct 07, 2019 >>> [1,0]:orte:version:full:4.0.2 >>> [1,0]:orte:version:repo:v4.0.2 >>> [1,0]

Re: [OMPI users] Error using rankfile to bind multiple cores on the same node for threaded OpenMPI application

2022-02-03 Thread David Perozzi via users
2 [1,0]:opal:version:repo:v4.0.2 [1,0]:opal:version:release_date:Oct 07, 2019 [1,0]:mpi-api:version:full:3.1.0 [1,0]:ident:4.0.2 Best Christoph - Original Message - From: "Open MPI Users" To: "Open MPI Users" Cc: "Ralph Castain" Sent: Thursday, 3 February,

Re: [OMPI users] Error using rankfile to bind multiple cores on the same node for threaded OpenMPI application

2022-02-02 Thread Ralph Castain via users
al:version:full:4.0.2 > [1,0]:opal:version:repo:v4.0.2 > [1,0]:opal:version:release_date:Oct 07, 2019 > [1,0]:mpi-api:version:full:3.1.0 > [1,0]:ident:4.0.2 > > Best > Christoph > > - Original Message - > From: "Open MPI Users" > To: "Open MPI Users

Re: [OMPI users] Error using rankfile to bind multiple cores on the same node for threaded OpenMPI application

2022-02-02 Thread Christoph Niethammer via users
rs" Cc: "Ralph Castain" Sent: Thursday, 3 February, 2022 00:22:30 Subject: Re: [OMPI users] Error using rankfile to bind multiple cores on the same node for threaded OpenMPI application Errr...what version OMPI are you using? > On Feb 2, 2022, at 3:03 PM, David Perozzi via use

Re: [OMPI users] Error using rankfile to bind multiple cores on the same node for threaded OpenMPI application

2022-02-02 Thread Ralph Castain via users
Errr...what version OMPI are you using? > On Feb 2, 2022, at 3:03 PM, David Perozzi via users > wrote: > > Helo, > > I'm trying to run a code implemented with OpenMPI and OpenMP (for threading) > on a large cluster that uses LSF for the job scheduling and dispatch. The > problem with LSF is

Re: [OMPI users] Error code for I/O operations

2021-07-01 Thread Aurelien Bouteiller via users
Eric, Error codes are **not** mandated to be consistent across ranks in the MPI standard. Although file errors are not expected to cause the program to abort, there is no special case for I/O in that regard in the MPI standard. The Open MPI implementation acts accordingly and will not automati

Re: [OMPI users] Error intialising an OpenFabrics device.

2021-03-13 Thread Heinz, Michael William via users
I’ve begun getting this annoyingly generic warning, too. It appears to be coming from the openib provider. If you disable it with -mtl ^openib the warning goes away. Sent from my iPad > On Mar 13, 2021, at 3:28 PM, Bob Beattie via users > wrote: > > Hi everyone, > > To be honest, as an MPI

Re: [OMPI users] Error with building OMPI with PGI

2021-01-19 Thread Passant A. Hafez via users
". make[2]: *** [param.o] Error 2 make[2]: Leaving directory `/tmp/openmpi-4.0.3/orte/tools/orte-info' make[1]: *** [install-recursive] Error 1 make[1]: Leaving directory `/tmp/openmpi-4.0.3/orte' make: *** [install-recursive] Error 1 Best, Passant _____________

Re: [OMPI users] Error with building OMPI with PGI

2021-01-19 Thread Gilles Gouaillardet via users
n switch: +03 > > > Please advise. > > All the best, > Passant > > From: users on behalf of Gus Correa via > users > Sent: Friday, January 15, 2021 2:36 AM > To: Open MPI Users > Cc: Gus Correa > Subject: Re: [OMPI users] Err

Re: [OMPI users] Error with building OMPI with PGI

2021-01-19 Thread Passant A. Hafez via users
__ From: users on behalf of Gus Correa via users Sent: Friday, January 15, 2021 2:36 AM To: Open MPI Users Cc: Gus Correa Subject: Re: [OMPI users] Error with building OMPI with PGI Hi Passant, list This is an old problem with PGI. There are many threads in the OpenMPI mailing list archiv

Re: [OMPI users] Error with building OMPI with PGI

2021-01-14 Thread Gus Correa via users
Hi Passant, list This is an old problem with PGI. There are many threads in the OpenMPI mailing list archives about this, with workarounds. The simplest is to use FC="pgf90 -noswitcherror". Here are two out of many threads ... well, not pthreads! :) https://www.mail-archive.com/users@lists.open

Re: [OMPI users] error while running mpirun

2020-07-13 Thread Srijan Chatterjee via users
Hi Gilles, Thanks for the suggestion. If I do exclude sm in the command line in the following way it seems to work. @chimera examples]$ mpirun -np 1 --mca btl ^sm hello_c libibverbs: Warning: no userspace device-specific driver found for /sys/class/infiniband_verbs/uverbs0 libibverbs: Warning: n

Re: [OMPI users] error while running mpirun

2020-07-12 Thread Gilles Gouaillardet via users
Srijan, The logs suggest you explicitly request the btl/sm component, and this typically occurs via a openmpi-mca-params.conf (that contains a line such as btl = sm,openib,self), or the OMPI_MCA_btl environment variable Cheers, Gilles On Mon, Jul 13, 2020 at 1:50 AM Srijan Chatterjee via users

Re: [OMPI users] Error with MPI_GET_ADDRESS and MPI_TYPE_CREATE_RESIZED?

2020-05-18 Thread Jeff Squyres (jsquyres) via users
Was your Open MPI compiled with -r8? We definitely recommend using the same compiler flags to compile Open MPI as your application. As George noted, -r8 can definitely cause issues if one was compiled with it and the other was not. On May 18, 2020, at 12:24 AM, George Bosilca via users mail

Re: [OMPI users] Error with MPI_GET_ADDRESS and MPI_TYPE_CREATE_RESIZED?

2020-05-17 Thread George Bosilca via users
Diego, I see nothing wrong with the way you create the datatype. In fact this is the perfect example on how to almost do it right in FORTRAN. The almost is because your code is highly dependent on the -r8 compiler option (otherwise the REAL in your type will not match the MPI_DOUBLE_PRECISION you

Re: [OMPI users] Error with MPI_GET_ADDRESS and MPI_TYPE_CREATE_RESIZED?

2020-05-17 Thread Diego Avesani via users
Dear Gilles, dear All, as far I remember no. The compiler is the same as the options which I use. Maybe, the error is some other places of my code. However, the results look errors in allocation of sent and received vector of datatype. The import is that at least mydata type definitions in corre

Re: [OMPI users] Error with MPI_GET_ADDRESS and MPI_TYPE_CREATE_RESIZED?

2020-05-17 Thread Gilles Gouaillardet via users
Diego, Did you change your compiler options? Cheers, Gilles - Original Message - > Dear all, > > I would like to share with what I have done in oder to create my own MPI > data type. The strange thing is that it worked until some day ago and then > it stopped working. This because pr

Re: [OMPI users] Error with OpenMPI: Could not resolve generic procedure mpi_irecv

2019-08-21 Thread Sangam B via users
Thanks Jeff for the resolution. On Mon, Aug 19, 2019 at 7:45 PM Jeff Squyres (jsquyres) wrote: > On Aug 19, 2019, at 6:15 AM, Sangam B via users > wrote: > > > > subroutine recv(this,lmb) > > class(some__example6), intent(inout) :: this > > integer, intent(in) :: lmb(2,2) > > > > i

Re: [OMPI users] Error with OpenMPI: Could not resolve generic procedure mpi_irecv

2019-08-19 Thread Jeff Squyres (jsquyres) via users
On Aug 19, 2019, at 6:15 AM, Sangam B via users wrote: > > subroutine recv(this,lmb) > class(some__example6), intent(inout) :: this > integer, intent(in) :: lmb(2,2) > > integer :: cs3, ierr > integer(kind=C_LONG) :: size This ^^ is your problem. More below. > ! receive

Re: [OMPI users] Error with OpenMPI: Could not resolve generic procedure mpi_irecv

2019-08-19 Thread Sangam B via users
Hi, Evan after recompiling OpenMPI with -fdefault-real-8, it fails with same error. It seems to me that, it's an issue with OpenMPI itself, because: Intel MPI + gnu compiler --- works Intel MPI + intel compiler --- works Open MPI + gnu compiler --- fails Open MPI + AOCC compiler --- fails Wil

Re: [OMPI users] Error with OpenMPI: Could not resolve generic procedure mpi_irecv

2019-08-19 Thread Gilles Gouaillardet via users
Thanks, but this is not really helping. Could you please build a Minimal, Reproducible Example as described at https://stackoverflow.com/help/minimal-reproducible-example ? Cheers, Gilles On Mon, Aug 19, 2019 at 7:19 PM Sangam B via users wrote: > > Hi, > > Here is the sample program snippet:

Re: [OMPI users] Error with OpenMPI: Could not resolve generic procedure mpi_irecv

2019-08-19 Thread Sangam B via users
Hi, Here is the sample program snippet: #include "intrinsic_sizes.h" #include "redef.h" module module1_m use mod1_m, only: some__example2 use mod2_m, only: some__example3 use mod3_m, only: some__example4 use mpi use, intrinsic :: iso_c_binding implicit none private

Re: [OMPI users] Error with OpenMPI: Could not resolve generic procedure mpi_irecv

2019-08-19 Thread Gilles Gouaillardet via users
I am not questioning whether you are facing an issue with Open MPI or not. I am just asking for "the same application" (read minimal source code) so I can reproduce the issue, investigate it and hopefully help you. Meanwhile, try rebuilding Open MPI with '-fdefault-real-8' in your FCFLAGS (since t

Re: [OMPI users] Error with OpenMPI: Could not resolve generic procedure mpi_irecv

2019-08-19 Thread Sangam B via users
Hi, I've tried both gcc-8.1.0 and AOCC-2.0 compilers with openmpi-3.1.1. It fails for both the compilers. Posted error message was from OpenMPI-3.1.1 + AOCC-2.0 compiler. To cross-check whether it is problem with OpenMPI or the base compiler, compiled the same application with Intel MPI using ba

Re: [OMPI users] Error with OpenMPI: Could not resolve generic procedure mpi_irecv

2019-08-19 Thread Gilles Gouaillardet via users
One more thing ... Your initial message mentioned a failure with gcc 8.2.0, but your follow-up message mentions LLVM compiler. So which compiler did you use to build Open MPI that fails to build your test ? Cheers, Gilles On Mon, Aug 19, 2019 at 6:49 PM Gilles Gouaillardet wrote: > > Thanks,

Re: [OMPI users] Error with OpenMPI: Could not resolve generic procedure mpi_irecv

2019-08-19 Thread Gilles Gouaillardet via users
Thanks, and your reproducer is ? Cheers, Gilles On Mon, Aug 19, 2019 at 6:42 PM Sangam B via users wrote: > > Hi, > > OpenMPI is configured as follows: > > export CC=`which clang` > export CXX=`which clang++` > export FC=`which flang` > export F90=`which flang` > > ../configure --prefix=/sw/op

Re: [OMPI users] Error with OpenMPI: Could not resolve generic procedure mpi_irecv

2019-08-19 Thread Sangam B via users
Hi, OpenMPI is configured as follows: export CC=`which clang` export CXX=`which clang++` export FC=`which flang` export F90=`which flang` ../configure --prefix=/sw/openmpi/3.1.1/aocc20hpcx210-mpifort --enable-mpi-fortran --enable-mpi-cxx --without-psm --without-psm2 --without-knem --without-libf

Re: [OMPI users] Error with OpenMPI: Could not resolve generic procedure mpi_irecv

2019-08-19 Thread Gilles Gouaillardet via users
Hi, Can you please post a full but minimal example that evidences the issue? Also please post your Open MPI configure command line. Cheers, Gilles Sent from my iPod > On Aug 19, 2019, at 18:13, Sangam B via users > wrote: > > Hi, > > I get following error if the application is compiled

Re: [OMPI users] error running mpirun command

2019-05-16 Thread Eric F. Alemany via users
Hi Jeff, Thank you for picking up on the thread. I followed your suggestion and I also did some research. I think I fixed the issue. My problem was two things: 1) From the mater node I made sure i could do a password-less ssh connection to the other nodes via; IP address of the nodes, host

Re: [OMPI users] error running mpirun command

2019-05-14 Thread Jeff Squyres (jsquyres) via users
Looks like this thread accidentally got dropped; sorry! More below. > On May 4, 2019, at 10:40 AM, Eric F. Alemany via users > wrote: > > Hi Gilles, > > Thank you for your message and your suggestion. As you suggested i tried > mpirun -np 84 - -hostfile hostsfile --mca routed direct ./open

Re: [OMPI users] error running mpirun command

2019-05-04 Thread Eric F. Alemany via users
Hi Gilles, Thank you for your message and your suggestion. As you suggested i tried mpirun -np 84 - -hostfile hostsfile --mca routed direct ./openmpi_hello.c The command hangs with no message or error message until i hit "control + z". Then i have the same error message as before. To answer

Re: [OMPI users] error running mpirun command

2019-05-03 Thread Gilles Gouaillardet via users
Eric, which version of Open MPI are you using ? how many hosts in your hostsfile ? The error message suggests this could be a bug within Open MPI, and a potential workaround for you would be to try mpirun -np 84 - -hostfile hostsfile --mca routed direct ./openmpi_hello.c You might also want to

Re: [OMPI users] Error when Building an MPI Java program

2019-04-11 Thread Kawashima, Takahiro
Benson, I think OpenJDK 9 is enough. I tested Open MPI Java bindings with OpenJDK 8, 9, 10, and 11 several months ago. Compiling your MPI Java program requires mpi.jar, which should have been installed in $OMPI_INSTALL_DIR/lib. Probably your compilation error message indicates mpi.jar is not f

Re: [OMPI users] Error when Building an MPI Java program

2019-04-10 Thread Benson Muite
Thanks. The code is here: https://github.com/mboysan/ping-pong-mpi-tcp OpenMPI is built using: ./configure --prefix=/PATH_TO_HOME_DIRECTORY/openmpi-4.0.1-install \     --enable-mpi-java The code builds on a number of other machines. Are there any minimum java build requirements for Op

Re: [OMPI users] Error when Building an MPI Java program

2019-04-09 Thread Jeff Squyres (jsquyres) via users
Can you provide a small, standalone example + recipe to show the problem? > On Apr 8, 2019, at 6:45 AM, Benson Muite wrote: > > Hi > > Am trying to build an MPI Java program using OpenMPI 4.0.1: > > I get the following error: > > Compilation failure > /MY_DIRECTORY/ping-pong-mpi-tcp/src/mai

Re: [OMPI users] error "unacceptable operand for unary &" for openmpi-master-201903260242-dfbc144 on Linux with Sun C

2019-03-26 Thread Nathan Hjelm via users
This really looks like a compiler bug. There is no & @ osc_pt2pt.h line 579. There is one at line 577 but there is no “unacceptable operand” on that line. If I have time this week I will try to find a workaround but it might be worth filing a bug with Oracle and see what they say. -Nathan > On

Re: [OMPI users] Error initializing an UCX / OpenFabrics device. #6300

2019-03-22 Thread Jeff Squyres (jsquyres) via users
Greetings Charlie. Yes, it looks like you replied to a closed issue on Github -- would you mind opening a new issue about it? You can certainly refer to the old issue for context. But replying to closed issues is a bit dangerous: if we miss the initial email from GitHub (and all of us have ov

Re: [OMPI users] error while using open mpi to compile parallel-netcdf-1.8.1

2019-03-07 Thread Jeff Squyres (jsquyres) via users
FWIW, the error message is telling you exactly what is wrong and provides a link to the FAQ item on how to fix it. It's a bit inelegant that it's segv'ing after that, but the real issue is what is described in the help message. > On Mar 6, 2019, at 3:07 PM, Zhifeng Yang wrote: > > Hi > > I

Re: [OMPI users] error building openmpi-master-201810310352-a1e85b0 on Linux with Sun C

2018-11-01 Thread Siegmar Gross
Hi Takahiro, thank you very much for your help. Adding "-xatomic=studio" to CFLAGS solved the problem. Now I get a new error which I will report in a separate email. Kind regards Siegmar Am 01.11.2018 um 04:45 schrieb Kawashima, Takahiro: Hi Siegmar, According to the man page of cc in Oracl

Re: [OMPI users] error building openmpi-master-201810310352-a1e85b0 on Linux with Sun C

2018-10-31 Thread Kawashima, Takahiro
Hi Siegmar, According to the man page of cc in Oracle Developer Studio 12.6, any atomics support runtime library is not linked by default. Probably you'll need -xatomic=studio, -xatomic=gcc, or -latomic option. I'm not sure whether the option should be added by Open MPI configure or by users.

Re: [OMPI users] error building Java api for openmpi-v4.0.x-201810090241-2124192 and openmpi-master-201810090329-e9e4d2a

2018-10-09 Thread Kawashima, Takahiro
I confirmed the patch resolved the issue with OpenJDK 11. I've created a PR for the master branch. I'll also create PRs for release branches. Thanks to your bug report! https://github.com/open-mpi/ompi/pull/5870 Takahiro Kawashima, Fujitsu > Siegmar, > > I think you are using Java 11 and it c

Re: [OMPI users] error building Java api for openmpi-v4.0.x-201810090241-2124192 and openmpi-master-201810090329-e9e4d2a

2018-10-09 Thread Kawashima, Takahiro
Siegmar, I think you are using Java 11 and it changed the default output HTML version. I'll take a look. But downloading OpenJDK 11 takes time... Probably the following patch will resolve the issue. diff --git ompi/mpi/java/java/Com

Re: [OMPI users] error building Java api for openmpi-v4.0.x-201810090241-2124192 and openmpi-master-201810090329-e9e4d2a

2018-10-09 Thread Kawashima, Takahiro
What version of Java are you using? Could you type "java -version" and show the output? Takahiro Kawashima, Fujitsu > today I've tried to build openmpi-v4.0.x-201810090241-2124192 and > openmpi-master-201810090329-e9e4d2a on my "SUSE Linux Enterprise Server > 12.3 (x86_64)" with Sun C 5.15, gcc 6

Re: [OMPI users] error building openmpi-master-201809290304-73075b8 with Sun C 5.15

2018-10-01 Thread Jeff Squyres (jsquyres) via users
Siegmar -- I created a GitHub issue for this: https://github.com/open-mpi/ompi/issues/5814 Nathan posted a test program on there for you to try; can you try it and reply on the issue? Thanks. > On Oct 1, 2018, at 9:23 AM, Siegmar Gross > wrote: > > Hi, > > I've tried to install openm

Re: [OMPI users] *** Error in `orted': double free or corruption (out): 0x00002aaab4001680 ***, in some node combos.

2018-09-13 Thread Jeff Squyres (jsquyres) via users
On Sep 12, 2018, at 4:54 AM, Balázs Hajgató wrote: > > Setting mca oob to tcp works. I will stick to this solution in our production > environment. Great! > I am not sure that it is relevant, but I also tried the patch on a > non-procduction OpenMPI 3.1.1, and "mpirun -host nic114,nic151 host

Re: [OMPI users] *** Error in `orted': double free or corruption (out): 0x00002aaab4001680 ***, in some node combos.

2018-09-12 Thread Balázs Hajgató
Dear Jeff, Setting mca oob to tcp works. I will stick to this solution in our production environment. I am not sure that it is relevant, but I also tried the patch on a non-procduction OpenMPI 3.1.1, and "mpirun -host nic114,nic151 hostname" works without any parameters, but issuing the libi

Re: [OMPI users] *** Error in `orted': double free or corruption (out): 0x00002aaab4001680 ***, in some node combos.

2018-09-11 Thread Jeff Squyres (jsquyres) via users
Thanks for reporting the issue. First, you can workaround the issue by using: mpirun --mca oob tcp ... This uses a different out-of-band plugin (TCP) instead of verbs unreliable datagrams. Second, I just filed a fix for our current release branches (v2.1.x, v3.0.x, and v3.1.x): https

Re: [OMPI users] Error in file base/plm_base_launch_support.c: OPAL_HWLOC_TOPO

2018-07-28 Thread Lovelace III, Henry
:18 PM To: Open MPI Users Subject: Re: [OMPI users] Error in file base/plm_base_launch_support.c: OPAL_HWLOC_TOPO Hi Gilles, If I understand the instruction correctly by invoking the -showme command to give addition information. I did use the -showme command, however the files that are being

Re: [OMPI users] Error in file base/plm_base_launch_support.c: OPAL_HWLOC_TOPO

2018-07-27 Thread Lovelace III, Henry
updated that is the reason behind the directory change. Locally I have no problem with the build. Thanks, Henry From: users on behalf of gil...@rist.or.jp Sent: Tuesday, July 24, 2018 10:00:51 PM To: Open MPI Users Subject: Re: [OMPI users] Error in file base/plm_

Re: [OMPI users] Error in file base/plm_base_launch_support.c: OPAL_HWLOC_TOPO

2018-07-24 Thread gilles
or from the MPIfort? Thanks, Henry From: users on behalf of Brian Smith Sent: Monday, July 23, 2018 2:06:14 PM To: Open MPI Users Subject: Re: [OMPI users] Error in file base/plm_base_launch_support. c: OPAL_HWLOC_TOPO On Sat, Jul 21, 2018 at 9:13 PM, Gill

Re: [OMPI users] Error in file base/plm_base_launch_support.c: OPAL_HWLOC_TOPO

2018-07-24 Thread Lovelace III, Henry
. The build is not seeing the mpi libraries even though before the build I defined export LD_LIBRARY_PATH=${LD_LIBRARY_PATH}:$HOME/BMAD/bmad_dist_2018_0717/production/lib Is this error from the MPIfort? Thanks, Henry From: users on behalf of Brian Smith Sent: M

Re: [OMPI users] Error in file base/plm_base_launch_support.c: OPAL_HWLOC_TOPO

2018-07-23 Thread Brian Smith
On Sat, Jul 21, 2018 at 9:13 PM, Gilles Gouaillardet wrote: > Brian, > > As Ralph already stated, this is likely a hwloc API issue. > From debian9, you can > lstopo --of xml | ssh debian8 lstopo --if xml -i - > > that will likely confirm the API error. > > If you are willing to get a bit more deta

Re: [OMPI users] Error in file base/plm_base_launch_support.c: OPAL_HWLOC_TOPO

2018-07-21 Thread Gilles Gouaillardet
Brian, As Ralph already stated, this is likely a hwloc API issue. From debian9, you can lstopo --of xml | ssh debian8 lstopo --if xml -i - that will likely confirm the API error. If you are willing to get a bit more details, you can add some printf in opal_hwloc_unpack (from opal/mca/hwloc/base/

Re: [OMPI users] Error in file base/plm_base_launch_support.c: OPAL_HWLOC_TOPO

2018-07-21 Thread r...@open-mpi.org
More than likely the problem is the difference in hwloc versions - sounds like the topology to/from xml is different between the two versions, and the older one doesn’t understand the new one. > On Jul 21, 2018, at 12:04 PM, Brian Smith > wrote: > > Greetings, > > I'm having trouble getting

Re: [OMPI users] error building openmpi-master-201806060243-64a5baa on Linux with Sun C

2018-06-07 Thread Siegmar Gross
Hi, it seems that the problem results from a compiler bug, because today I was able to build the package with the compilers from "Intel Parallel Studio XE 2018" and "Portland Group Community Edition 2018". Unfortunately, I have no contact person at Oracle to report the bug, so that it can be fixe

Re: [OMPI users] error building openmpi-master-201806060243-64a5baa on Linux with Sun C

2018-06-06 Thread Nathan Hjelm
The bindings in v3.1.0 are incorrect. They are missing the asynchronous attribute. That will be fixed in v3.1.1. > On Jun 6, 2018, at 12:06 PM, Siegmar Gross > wrote: > > Hi Jeff, > >> I asked some Fortran gurus, and they don't think that there >> is any restriction on having ASYNCHRONOUS

Re: [OMPI users] error building openmpi-master-201806060243-64a5baa on Linux with Sun C

2018-06-06 Thread Siegmar Gross
Hi Jeff, I asked some Fortran gurus, and they don't think that there is any restriction on having ASYNCHRONOUS and INTENT on the same line. Indeed, Open MPI's definition of MPI_ACCUMULATE seems to agree with what is in MPI-3.1. Is this a new version of a Fortran compiler that you're using, per

Re: [OMPI users] error building openmpi-master-201806060243-64a5baa on Linux with Sun C

2018-06-06 Thread Jeff Squyres (jsquyres) via users
Siegmar -- I asked some Fortran gurus, and they don't think that there is any restriction on having ASYNCHRONOUS and INTENT on the same line. Indeed, Open MPI's definition of MPI_ACCUMULATE seems to agree with what is in MPI-3.1. Is this a new version of a Fortran compiler that you're using, p

Re: [OMPI users] error building openmpi-master-201806060243-64a5baa on Linux with Sun C

2018-06-06 Thread Nathan Hjelm
I put in a PR to "fix" this but if you look at the standard it has both intent(in) and asynchronous. Might be a compiler problem? -Nathan > On Jun 6, 2018, at 5:11 AM, Siegmar Gross > wrote: > > Hi, > > I've tried to install openmpi-master-201806060243-64a5baa on my "SUSE Linux > Enterprise

Re: [OMPI users] Error in hello_cxx.cc

2018-04-23 Thread r...@open-mpi.org
and 3) we no longer support Windows. You could try using the cygwin port instead. > On Apr 23, 2018, at 7:52 PM, Nathan Hjelm wrote: > > Two things. 1) 1.4 is extremely old and you will not likely get much help > with it, and 2) the c++ bindings were deprecated in MPI-2.2 (2009) and > removed

Re: [OMPI users] Error in hello_cxx.cc

2018-04-23 Thread Nathan Hjelm
Two things. 1) 1.4 is extremely old and you will not likely get much help with it, and 2) the c++ bindings were deprecated in MPI-2.2 (2009) and removed in MPI-3.0 (2012) so you probably want to use the C bindings instead. -Nathan > On Apr 23, 2018, at 8:14 PM, Amir via users wrote: > > Yes,

Re: [OMPI users] Error in hello_cxx.cc

2018-04-23 Thread r...@open-mpi.org
Also, I note from the screenshot that you appear to be running on Windows with a Windows binary. Correct? > On Apr 23, 2018, at 7:08 AM, Jeff Squyres (jsquyres) > wrote: > > Can you send all the information listed here: > >https://www.open-mpi.org/community/help/ > > > >> On Apr 22, 2

Re: [OMPI users] Error in hello_cxx.cc

2018-04-23 Thread Jeff Squyres (jsquyres)
Can you send all the information listed here: https://www.open-mpi.org/community/help/ > On Apr 22, 2018, at 2:28 PM, Amir via users wrote: > > Hi everybody, > > After having some problems with setting up the debugging environment for > Visual Studio 10, I am trying to debug the first O

Re: [OMPI users] error building openmpi-3.0.1 on Linux with gcc or Sun C and Java-10

2018-04-12 Thread Kawashima, Takahiro
Siegmar, Thanks for your report. But it is a known issue and will be fixed in Open MPI v3.0.2. https://github.com/open-mpi/ompi/pull/5029 If you want it now in v3.0.x series, try the latest nightly snapshot. https://www.open-mpi.org/nightly/v3.0.x/ Regards, Takahiro Kawashima, MPI develop

Re: [OMPI users] error in mpi processes: openmpi-3.0

2018-04-07 Thread Jeff Squyres (jsquyres)
Are you 100% sure that you are not accidentally mixing and matching multiple versions of Open MPI in the same job? This type of error (PMIX bad param) is typical when you accidentally use Open MPI vXYZ on one node and Open MPI vABC on a different node. > On Apr 4, 2018, at 12:20 AM, abhisek M

Re: [OMPI users] Error when using Einstein toolkit

2018-04-04 Thread Jeff Squyres (jsquyres)
Greetings, and welcome to the wonderful world of MPI. :-) First thing to note is that there are multiple different software packages that implement the MPI specification. Open MPI -- the mailing list that you sent to -- is one of them. MPICH, from Argonne National Labs, is another. From the

Re: [OMPI users] Error building openmpi on Raspberry pi 2

2017-10-03 Thread Pavel Shamis
I'm building on ARMv8 (64bit kernel, ompi master) and so far no problems. On Wed, Sep 27, 2017 at 7:34 AM, Jeff Layton wrote: > I could never get OpenMPI < 2.x to build on a Pi 2. I ended up using the > binary from the repos. Pi 3 is a different matter - I got that to build > after a little expe

Re: [OMPI users] Error building openmpi on Raspberry pi 2

2017-09-27 Thread Jeff Layton
I could never get OpenMPI < 2.x to build on a Pi 2. I ended up using the binary from the repos. Pi 3 is a different matter - I got that to build after a little experimentation :) Jeff On Wednesday, September 27, 2017 8:03 AM, Nathan Hjelm wrote: Open MPI does not officially support A

Re: [OMPI users] Error building openmpi on Raspberry pi 2

2017-09-27 Thread Nathan Hjelm
Open MPI does not officially support ARM in the v2.1 series. Can you download a nightly tarball from https://www.open-mpi.org/nightly/master/ and see if it works for you? -Nathan > On Sep 26, 2017, at 7:32 PM, Faraz Hussain wrote: > > I am receiving the make errors below on my pi 2: > > pi@p

Re: [OMPI users] Error building openmpi on Raspberry pi 2

2017-09-26 Thread John Hearns via users
This might be of interest for ARM users: https://developer.arm.com/products/software-development-tools/hpc/arm-compiler-for-hpc On 27 September 2017 at 06:58, Gilles Gouaillardet < gilles.gouaillar...@gmail.com> wrote: > Faraz, > > which OS are you running ? > > iirc, i faced similar issues, an

Re: [OMPI users] Error building openmpi on Raspberry pi 2

2017-09-26 Thread Gilles Gouaillardet
Faraz, which OS are you running ? iirc, i faced similar issues, and the root cause is that though ARMv7 does support these instructions, the compiler only generate ARMv6 code and hence failed to build Open MPI Cheers, Gilles On Wed, Sep 27, 2017 at 10:32 AM, Faraz Hussain wrote: > I am receiv

Re: [OMPI users] error building openmpi-v2.* with SUN C 5.15 on SuSE Linux

2017-08-08 Thread r...@open-mpi.org
Should be fixed for 2.x here: https://github.com/open-mpi/ompi/pull/4054 > On Jul 31, 2017, at 5:56 AM, Siegmar Gross > wrote: > > Hi, > > I've been able to install openmpi-v2.0.x-201707270322-239c439 and > openmpi-v2.x-201707271804-3b1e9fe on my

Re: [OMPI users] Error "Bad parameter" in mpirun

2017-02-16 Thread Jeff Squyres (jsquyres)
Are you running on a Mac, perchance? If so, see question 8: https://www.open-mpi.org/faq/?category=osx#startup-errors-with-open-mpi-2.0.x > On Feb 16, 2017, at 5:10 AM, Alessandra Bonazzi > wrote: > > Goodmorning, > I’m a beginner and I’m trying to run a simple “HelloWorld.f90”, but I g

Re: [OMPI users] Error during installation

2017-02-14 Thread Jeff Squyres (jsquyres)
+1 on what Gilles said. >From config.log: configure:18397: gcpc --version >&5 ./configure: line 18399: gcpc: command not found configure:18408: $? = 127 That's the failure. Sent from my phone. No type good. On Feb 13, 2017, at 9:23 PM, Gilles Gouaillardet mailto:gil...@rist.or.jp>> wrote:

Re: [OMPI users] Error during installation

2017-02-13 Thread Gilles Gouaillardet
Alessandra, your configure command line is ./configure CC=gcc CXX=gcpc F77=gfortran FC=gfortran --prefix=/usr/local i guess gcpc is a typo and you meant g++ instead. fwiw, the default prefix is /usr/local, so you do not really need '--prefix=/usr/local' Cheers, Gilles On 2/14/2017 12

Re: [OMPI users] Error during installation

2017-02-13 Thread Jeff Squyres (jsquyres)
Can you send all the information listed here: https://www.open-mpi.org/community/help/ > On Feb 13, 2017, at 3:57 AM, Alessandra Bonazzi > wrote: > > Goodmorning, > I'm facing a problem during the installation of Open MPI. > The error appears on terminal at a certain point during the com

Re: [OMPI users] Error using hpcc benchmark

2017-02-02 Thread wodel youchi
de, nodes, and memory per node to have an idea? A quick test could be to > start with very small number of ranks to see if it runs. > > > > Thanks, > > Regards, > > > > _MAC > > > > *From:* users [mailto:users-boun...@lists.open-mpi.org] *On Behalf O

Re: [OMPI users] Error using hpcc benchmark

2017-02-01 Thread Cabral, Matias A
pi.org] On Behalf Of wodel youchi Sent: Wednesday, February 01, 2017 3:36 AM To: Open MPI Users Subject: Re: [OMPI users] Error using hpcc benchmark Hi, Thank you for you replies, but :-) it didn't work for me. Using hpcc compiled with OpenMPI 2.0.1 : I tried to use export PSM_MQ_RECVREQS_

Re: [OMPI users] Error using hpcc benchmark

2017-02-01 Thread wodel youchi
EQS_MAX) or increase both with > PSM_MEMORY=max. Note that the psm library will allocate more system memory > for the queues. > > > > Thanks, > > > > _MAC > > > > *From:* users [mailto:users-boun...@lists.open-mpi.org] *On Behalf Of *Howard > Pritchar

Re: [OMPI users] Error using hpcc benchmark

2017-01-31 Thread Cabral, Matias A
PSM_MEMORY=max. Note that the psm library will allocate more system memory for the queues. Thanks, _MAC From: users [mailto:users-boun...@lists.open-mpi.org] On Behalf Of Howard Pritchard Sent: Tuesday, January 31, 2017 6:38 AM To: Open MPI Users Subject: Re: [OMPI users] Error using hpcc

Re: [OMPI users] Error using hpcc benchmark

2017-01-31 Thread Howard Pritchard
Hi Wodel Randomaccess part of HPCC is probably causing this. Perhaps set PSM env. variable - Export PSM_MQ_REVCREQ_MAX=1000 or something like that. Alternatively launch the job using mpirun --mca plm ob1 --host to avoid use of psm. Performance will probably suffer with this option

Re: [OMPI users] Error bash: /usr/mpi/gcc/openmpi-1.8.8/bin/orted: No such file or directory

2016-11-20 Thread Sebastian Antunez N.
Hello Thank for your comment Only the frontend was updated directly via install.sh fron ofed 2.4.3 to ofed 3.1.1.0 and contains openmpi 1.8.8. Now the compute node have a older version of ofed 2.4 with openmpi 1.6.4 My question; if is possible update ofed directly in the compute node executing

Re: [OMPI users] Error bash: /usr/mpi/gcc/openmpi-1.8.8/bin/orted: No such file or directory

2016-11-20 Thread Gilles Gouaillardet
Sebastian, The error message is pretty self-explanatory /usr/mpi/gcc/openmpi-1.8.8/bin/orted is missing on your compute nodes. it seems you are using /usr/mpi/gcc/openmpi-1.8.8/bin/mpirun on your frontend node (e.g. the node on which mpirun is invoked) but Open MPI was not updated on some nodes l

Re: [OMPI users] error on dlopen

2016-11-04 Thread Mahmood Naderan
>​What problems are you referring to? I mean errors that are saying failed to load X.so. Then the user has to add some paths to LD_LIBRARY_PATH. Although such problem can be fixed by adding an export to the .bashrc, but I prefer to avoid that. >We might need a bit more detail than that; I use "--

Re: [OMPI users] error on dlopen

2016-11-04 Thread Jeff Squyres (jsquyres)
On Nov 4, 2016, at 12:14 PM, Mahmood Naderan wrote: > > >​If there's a reason you did --enable-static --disable-shared​ > Basically, I want to prevent dynamic library problems (ldd) on a distributed > environment. What problems are you referring to? > $ mpifort --showme > gfortran -I/opt/openm

Re: [OMPI users] error on dlopen

2016-11-04 Thread Mahmood Naderan
>​If there's a reason you did --enable-static --disable-shared​ Basically, I want to prevent dynamic library problems (ldd) on a distributed environment. ​$ mpifort --showme gfortran -I/opt/openmpi-2.0.1/include -pthread -I/opt/openmpi-2.0.1/lib -Wl,-rpath -Wl,/opt/openmpi-2.0.1/lib -Wl,--enable-

Re: [OMPI users] error on dlopen

2016-11-04 Thread Jeff Squyres (jsquyres)
> On Nov 4, 2016, at 7:07 AM, Mahmood Naderan wrote: > > > You might have to remove -ldl from the scalapack makefile > I removed that before... I will try one more time > > Actually, using --disable-dlopen fixed the error. To clarify: 1. Using --enable-static causes all the plugins in Open MPI

  1   2   3   4   5   6   >