Hello again ;),
after getting acceptable latency on our Chelsio S320E-CXA adapters we now want
to check if we can
also tune the bandwidth. On TCP level (measured via iperf) we get 1.15 GB/s, on
MPI level (measured
via MPI-Ping-Pong) just 930 MB/s. We already set btl_tcp_sndbuf and
btl_tcp_rcvb
With OpenMPI 1.3 / iWARP you should get around 8us latency using mpi
pingpong tests.
Andy Georgi wrote:
Thanks again for all the answers. It seems that were was a bug in the
driver in combination with
Suse Linux Enterprise Server 10. It was fixed with version 1.0.146.
Now we have 12us with NP
Andy Georgi wrote:
Hello again ;),
after getting acceptable latency on our Chelsio S320E-CXA adapters we
now want to check if we can
also tune the bandwidth. On TCP level (measured via iperf) we get 1.15
GB/s, on MPI level (measured
via MPI-Ping-Pong) just 930 MB/s. We already set btl_tcp_sndb
Thanks.
I had also posted the bug on the MPICH2 list, and received an
aswer from the ROMIO maintainers: the issue seems to be related to
NFS file locking bugs. I had been testing on an NFS system, and
when I re-tested under a local (ext3) file system, I did not reproduce
the bug.
I had been exper
Jitendra,
There is a problem with the addresses you provide to MPI_Type_struct.
For all arrays instead of giving the pointer to the array, you provide
a pointer to the pointer in the individual struct.
Try the following
MPI_Get_address(&parentPop[0].indiv[0].intvar[0], &disp[0]);
instead of
Tom,
I did the same modification as you on the osu_latency and the
resulting application run to completion. I don't get any TRUNCATE
error messages. I'm using the latest version of Open MPI (1.4a1r19313).
There was a bug that might be related to your problem but our commit
log shows it wa
Brian,
Some of these files are used for startup, while others are used during
application execution (such as the back-end for shared memory files).
Over the years we had a lot of discussions about this topic, and so
far we have two ways to help people deal with such situations.
However, f
Dave,
Thanks for your report. As you discovered we had a memory leak in the
MPI_Alltoallw. A very small one, but it was there. Basically, we
didn't release two internal arrays of data-types, used to convert from
the Fortran data-types (as supplied by the user) to their C version
(as requi
Mostyn,
There was a problem with the SM BTL. Please upgrade to at least 19315
and [hopefully] your application will run to completion.
Thanks,
george.
On Jul 24, 2008, at 3:39 AM, Mostyn Lewis wrote:
Hello,
Using a very recent svn version (1.4a1r18899) I'm getting a non-
terminati