MPI uses void** arguments to pass pointer by reference so it can be updated. In
Fortran, you always pass by reference so you don't need this. Just pass your
Fortran pointer argument.
There are MPI-3 shared memory examples in Fortran somewhere. Try Using Advanced
MPI (latest edition) or MPI Trac
could you also test the nightly tarball of the v2.x and v1.10 branches ?
when a process sends a message to an other process for the first time,
it establish a TCP connection if not already present.
so i A sends to B first, then A connects to B.
when B wants to send to A, it uses the previously es
Hello,
I am trying to port a simple halo exchange program from C to fortran.
It is designed to demonstrate the shared memory features of MPI-3. The
original C program was download from an Intel site, and I have modified
it to simplify the port. A tarfile of a directory with each program and
Hi Ralph,
Thank you for looking into this!
The job
#BSUB -J "task_geometry"
#BSUB -n 9
#BSUB -R "span[ptile=3]"
#BSUB -m "p10a30 p10a33 p10a35 p10a55 p10a58"
#BSUB -R "affinity[core]"
#BSUB -e "task_geometry.stderr.%J"
#BSUB -o "task_geometry.stdout.%J"
#BSUB -q "normal"
#BSUB -M "800"
#BSUB -R "
Hi Farid
I’m not sure I understand what you are asking here. If your point is that OMPI
isn’t placing and binding procs per the LSF directives, then you are quite
correct. The LSF folks never provided that level of integration, nor the info
by which we might have derived it (e.g., how the patte
Dear developers
Thank you all for jumping in to help. Here is what I have found so far:
1. Running Netpipe (NPmpi) between the two nodes (in either order) was
successful, but following this test, my original code still hung.
2. Following Gilles's advice, I then added an MPI_Barrier() at the end o
We recently updated and rebooted Infiniband-attached nodes, and now when
trying to schedule MPI jobs with slurm, we are seeing the following:
--
No OpenFabrics connection schemes reported that they were able to be
used on a s
Greetings!
The following batch script will successfully demo the use of LSF's task
geometry feature using IBM Parallel Environment:
#BUB -J "task_geometry"
#BSUB -n 9
#BSUB -R "span[ptile=3]"
#BSUB -network "type=sn_single:mode=us"
#BSUB -R "affinity[core]"
#BSUB -e "task_geometry.stderr.%J"
#BSU
Durga,
Can you run a simple netpipe over TCP using any of the two interfaces you
mentioned?
George
On Apr 18, 2016 11:08 AM, "Gilles Gouaillardet" <
gilles.gouaillar...@gmail.com> wrote:
> An other test is to swap the hostnames.
> If the single barrier test fails, this can hint to a firewall.
>
An other test is to swap the hostnames.
If the single barrier test fails, this can hint to a firewall.
Cheers,
Gilles
Gilles Gouaillardet wrote:
>sudo make uninstall
>will not remove modules that are no more built
>sudo rm -rf /usr/local/lib/openmpi
>is safe thought
>
>i confirm i did not see a
sudo make uninstall
will not remove modules that are no more built
sudo rm -rf /usr/local/lib/openmpi
is safe thought
i confirm i did not see any issue on a system with two networks
Cheers,
Gilles
On 4/18/2016 2:53 PM, dpchoudh . wrote:
Hello Gilles
I did a
sudo make uninstall
followed by a
Hello Gilles
I did a
sudo make uninstall
followed by a
sudo make install
on both nodes. But that did not make a difference. I will try your tarball
build suggestion a bit later.
What I find a bit strange is that only I seem to be getting into this
issue. What could I be doing wrong? Or am I disco
so you might want to
rm -rf /usr/local/lib/openmpi
and run
make install
again, just to make sure old stuff does not get in the way
Cheers,
Gilles
On 4/18/2016 2:12 PM, dpchoudh . wrote:
Hello Gilles
Thank you very much for your feedback. You are right that my original
stack trace was on code
Hello Gilles
Thank you very much for your feedback. You are right that my original stack
trace was on code that was several weeks behind, but updating it just now
did not seem to make a difference: I am copying the stack from the latest
code below:
On the master node:
(gdb) bt
#0 0x7fc0524c
here is your stack trace
#6 0x7f72a0d09cd5 in mca_pml_ob1_send (buf=0x7fff81057db0, count=4,
datatype=0x601080 , dst=1, tag=1,
sendmode=MCA_PML_BASE_SEND_STANDARD, comm=0x601280
)
at line 251
that would be line 259 in current master, and this file was updated 21
days ago
and t
An update:
Building the latest from master did not make any difference; the code still
hangs with identical stack trace as before.
This should be a simple case to reproduce (positively or negatively). Would
somebody in the developer community mind giving it a quick try?
Thank you
Durga
1% of th
Thank you for your suggestion, Ralph. But it did not make any difference.
Let me say that my code is about a week stale. I just did a git pull and am
building it right now. The build takes quite a bit of time, so I avoid
doing that unless there is a reason. But what I am trying out is the most
bas
17 matches
Mail list logo