Re: [OMPI users] Porting MPI-3 C-program to Fortran

2016-04-18 Thread Jeff Hammond
MPI uses void** arguments to pass pointer by reference so it can be updated. In Fortran, you always pass by reference so you don't need this. Just pass your Fortran pointer argument. There are MPI-3 shared memory examples in Fortran somewhere. Try Using Advanced MPI (latest edition) or MPI Trac

Re: [OMPI users] OMPI users] Possible bug in MPI_Barrier() ?

2016-04-18 Thread Gilles Gouaillardet
could you also test the nightly tarball of the v2.x and v1.10 branches ? when a process sends a message to an other process for the first time, it establish a TCP connection if not already present. so i A sends to B first, then A connects to B. when B wants to send to A, it uses the previously es

[OMPI users] Porting MPI-3 C-program to Fortran

2016-04-18 Thread Tom Rosmond
Hello, I am trying to port a simple halo exchange program from C to fortran. It is designed to demonstrate the shared memory features of MPI-3. The original C program was download from an Intel site, and I have modified it to simplify the port. A tarfile of a directory with each program and

[OMPI users] Fw: LSF's LSB_PJL_TASK_GEOMETRY + OpenMPI 1.10.2

2016-04-18 Thread Farid Parpia
Hi Ralph, Thank you for looking into this! The job #BSUB -J "task_geometry" #BSUB -n 9 #BSUB -R "span[ptile=3]" #BSUB -m "p10a30 p10a33 p10a35 p10a55 p10a58" #BSUB -R "affinity[core]" #BSUB -e "task_geometry.stderr.%J" #BSUB -o "task_geometry.stdout.%J" #BSUB -q "normal" #BSUB -M "800" #BSUB -R "

Re: [OMPI users] LSF's LSB_PJL_TASK_GEOMETRY + OpenMPI 1.10.2

2016-04-18 Thread Ralph Castain
Hi Farid I’m not sure I understand what you are asking here. If your point is that OMPI isn’t placing and binding procs per the LSF directives, then you are quite correct. The LSF folks never provided that level of integration, nor the info by which we might have derived it (e.g., how the patte

Re: [OMPI users] OMPI users] Possible bug in MPI_Barrier() ?

2016-04-18 Thread dpchoudh .
Dear developers Thank you all for jumping in to help. Here is what I have found so far: 1. Running Netpipe (NPmpi) between the two nodes (in either order) was successful, but following this test, my original code still hung. 2. Following Gilles's advice, I then added an MPI_Barrier() at the end o

[OMPI users] openib MTL not working via slurm after update

2016-04-18 Thread Nathan Smith
We recently updated and rebooted Infiniband-attached nodes, and now when trying to schedule MPI jobs with slurm, we are seeing the following: -- No OpenFabrics connection schemes reported that they were able to be used on a s

[OMPI users] LSF's LSB_PJL_TASK_GEOMETRY + OpenMPI 1.10.2

2016-04-18 Thread Farid Parpia
Greetings! The following batch script will successfully demo the use of LSF's task geometry feature using IBM Parallel Environment: #BUB -J "task_geometry" #BSUB -n 9 #BSUB -R "span[ptile=3]" #BSUB -network "type=sn_single:mode=us" #BSUB -R "affinity[core]" #BSUB -e "task_geometry.stderr.%J" #BSU

Re: [OMPI users] OMPI users] Possible bug in MPI_Barrier() ?

2016-04-18 Thread George Bosilca
Durga, Can you run a simple netpipe over TCP using any of the two interfaces you mentioned? George On Apr 18, 2016 11:08 AM, "Gilles Gouaillardet" < gilles.gouaillar...@gmail.com> wrote: > An other test is to swap the hostnames. > If the single barrier test fails, this can hint to a firewall. >

Re: [OMPI users] OMPI users] Possible bug in MPI_Barrier() ?

2016-04-18 Thread Gilles Gouaillardet
An other test is to swap the hostnames. If the single barrier test fails, this can hint to a firewall. Cheers, Gilles Gilles Gouaillardet wrote: >sudo make uninstall >will not remove modules that are no more built >sudo rm -rf /usr/local/lib/openmpi >is safe thought > >i confirm i did not see a

Re: [OMPI users] Possible bug in MPI_Barrier() ?

2016-04-18 Thread Gilles Gouaillardet
sudo make uninstall will not remove modules that are no more built sudo rm -rf /usr/local/lib/openmpi is safe thought i confirm i did not see any issue on a system with two networks Cheers, Gilles On 4/18/2016 2:53 PM, dpchoudh . wrote: Hello Gilles I did a sudo make uninstall followed by a

Re: [OMPI users] Possible bug in MPI_Barrier() ?

2016-04-18 Thread dpchoudh .
Hello Gilles I did a sudo make uninstall followed by a sudo make install on both nodes. But that did not make a difference. I will try your tarball build suggestion a bit later. What I find a bit strange is that only I seem to be getting into this issue. What could I be doing wrong? Or am I disco

Re: [OMPI users] Possible bug in MPI_Barrier() ?

2016-04-18 Thread Gilles Gouaillardet
so you might want to rm -rf /usr/local/lib/openmpi and run make install again, just to make sure old stuff does not get in the way Cheers, Gilles On 4/18/2016 2:12 PM, dpchoudh . wrote: Hello Gilles Thank you very much for your feedback. You are right that my original stack trace was on code

Re: [OMPI users] Possible bug in MPI_Barrier() ?

2016-04-18 Thread dpchoudh .
Hello Gilles Thank you very much for your feedback. You are right that my original stack trace was on code that was several weeks behind, but updating it just now did not seem to make a difference: I am copying the stack from the latest code below: On the master node: (gdb) bt #0 0x7fc0524c

Re: [OMPI users] Possible bug in MPI_Barrier() ?

2016-04-18 Thread Gilles Gouaillardet
here is your stack trace #6 0x7f72a0d09cd5 in mca_pml_ob1_send (buf=0x7fff81057db0, count=4, datatype=0x601080 , dst=1, tag=1, sendmode=MCA_PML_BASE_SEND_STANDARD, comm=0x601280 ) at line 251 that would be line 259 in current master, and this file was updated 21 days ago and t

Re: [OMPI users] Possible bug in MPI_Barrier() ?

2016-04-18 Thread dpchoudh .
An update: Building the latest from master did not make any difference; the code still hangs with identical stack trace as before. This should be a simple case to reproduce (positively or negatively). Would somebody in the developer community mind giving it a quick try? Thank you Durga 1% of th

Re: [OMPI users] Possible bug in MPI_Barrier() ?

2016-04-18 Thread dpchoudh .
Thank you for your suggestion, Ralph. But it did not make any difference. Let me say that my code is about a week stale. I just did a git pull and am building it right now. The build takes quite a bit of time, so I avoid doing that unless there is a reason. But what I am trying out is the most bas