[OMPI users] OpenMPI against multiple, evolving SLURM versions

2016-01-29 Thread William Law
Hi, Our group can't find anyway to do this and it'd be helpful. We use slurm and keep upgrading the slurm environment. OpenMPI bombs out against PMI each time the libslurm stuff changes, which seems to be fairly regularly. Is there a way to compile against slurm but insulate ourselves from t

Re: [OMPI users] OpenMPI against multiple, evolving SLURM versions

2016-01-29 Thread Ralph Castain
It makes sense - but isn’t it slurm that is linking libpmi against libslurm? I don’t think we are making that connection, so it would be a slurm issue to change it. > On Jan 28, 2016, at 10:12 PM, William Law wrote: > > Hi, > > Our group can't find anyway to do this and it'd be helpful. > >

[OMPI users] difference between OpenMPI - intel MPI mpi_waitall

2016-01-29 Thread Diego Avesani
Dear all, I have created a program in fortran and OpenMPI, I test it on my laptop and it works. I would like to use it on a cluster that has, unfortunately, intel MPI. The program crushes on the cluster and I get the following error: *Fatal error in MPI_Waitall: Invalid MPI_Request, error stack:

Re: [OMPI users] difference between OpenMPI - intel MPI mpi_waitall

2016-01-29 Thread Gilles Gouaillardet
Diego, your code snippet does MPI_Waitall(2,...) but the error is about MPI_Waitall(3,...) Cheers, Gilles On Friday, January 29, 2016, Diego Avesani wrote: > Dear all, > > I have created a program in fortran and OpenMPI, I test it on my laptop > and it works. > I would like to use it on a clu

Re: [OMPI users] OpenMPI against multiple, evolving SLURM versions

2016-01-29 Thread Gilles Gouaillardet
Is openmpi linked with a static libpmi.a that requires a dynamic libslurm ? that can be checked with ldd mca_ess_pmi.so btw, do slurm folks increase the libpmi.so version each time slurm is upgraded ? that could be a part of the issue ... but if they increase lib version because of abi changes, it

Re: [OMPI users] difference between OpenMPI - intel MPI mpi_waitall

2016-01-29 Thread Diego Avesani
Dear all, Dear Gilles, I do not understand, I am sorry. I did a "grep" on my code and I find only "MPI_WAITALL(2", so I am not able to find the error. Thanks a lot Diego On 29 January 2016 at 11:58, Gilles Gouaillardet < gilles.gouaillar...@gmail.com> wrote: > Diego, > > your code snippet

Re: [OMPI users] difference between OpenMPI - intel MPI mpi_waitall

2016-01-29 Thread Jeff Squyres (jsquyres)
You must have an error elsewhere in your code; as Gilles pointed, the error message states that you are calling MPI_WAITALL with a first argument of 3: -- MPI_Waitall(271): MPI_Waitall(count=3, req_array=0x7445f0, status_array=0x744600) failed -- We can't really help you with problems w

Re: [OMPI users] OpenMPI against multiple, evolving SLURM versions

2016-01-29 Thread Gilles Gouaillardet
on second thought, is there any chance your sysadmin removed the old libslurm.so.x but kept the old libpmix.so.y ? in this case, the real issue would be hidden your sysadmin "broke" the old libpmi, but you want to use the new one indeed. Cheers, Gilles On Friday, January 29, 2016, Gilles Gouaill

Re: [OMPI users] difference between OpenMPI - intel MPI mpi_waitall

2016-01-29 Thread Diego Avesani
Dear all, Dear Jeff, Dear Gilles, I am sorry, porblably I am a stubborn. In all my code I have CALL MPI_WAITALL(2,REQUEST,send_status_list,MPIdata%iErr) how can it became "3"? the only thing that I can think is that MPI starts to allocate the vector from "0", while fortran starts from 1. Indee

Re: [OMPI users] difference between OpenMPI - intel MPI mpi_waitall

2016-01-29 Thread Jeff Squyres (jsquyres)
On Jan 29, 2016, at 7:55 AM, Diego Avesani wrote: > > Dear all, Dear Jeff, Dear Gilles, > > I am sorry, porblably I am a stubborn. > > In all my code I have > > CALL MPI_WAITALL(2,REQUEST,send_status_list,MPIdata%iErr) > > how can it became "3"? I don't know. You'll need to check your code

Re: [OMPI users] difference between OpenMPI - intel MPI mpi_waitall

2016-01-29 Thread Gilles Gouaillardet
Diego, First, you can double check the program you are running has been compiled from your sources. then you can run your program under a debugger, and browse the stack when it crashes. there could be a bug in intelmpi, that incorrectly translates 2 in Fortran to 3 in C... but as far as I am con

Re: [OMPI users] difference between OpenMPI - intel MPI mpi_waitall

2016-01-29 Thread Diego Avesani
Dear all, I am really sorry for the time that you dedicated to me. this is what I found: REQUEST = MPI_REQUEST_NULL !send data share with UP IF(MPIdata%rank.NE.0)THEN MsgLength = MPIdata%imaxN DO icount=1,MPIdata%imaxN iNode = MPIdata%nodeFromUp(icount) send_me

Re: [OMPI users] difference between OpenMPI - intel MPI mpi_waitall

2016-01-29 Thread Jeff Squyres (jsquyres)
> On Jan 29, 2016, at 9:43 AM, Diego Avesani wrote: > > Dear all, > > I am really sorry for the time that you dedicated to me. > > this is what I found: > > REQUEST = MPI_REQUEST_NULL I'm not enough of a fortran expert to know -- does this assign MPI_REQUEST_NULL to every entry in the REQU

Re: [OMPI users] difference between OpenMPI - intel MPI mpi_waitall

2016-01-29 Thread Jeff Hammond
On Fri, Jan 29, 2016 at 2:45 AM, Diego Avesani wrote: > Dear all, > > I have created a program in fortran and OpenMPI, I test it on my laptop > and it works. > I would like to use it on a cluster that has, unfortunately, intel MPI. > > You can install any open-source MPI implementation from user