Nathan,
Thank you, I can confirm that it works as expected with master on our
system. I will stick to this version then until 3.1.1 is out.
Joseph
On 05/08/2018 05:34 PM, Nathan Hjelm wrote:
Looks like it doesn't fail with master so at some point I fixed this
bug. The current plan is to br
I am using ompi -3.1.0 version in my program and compiler is mpicc
its a parallel program which uses multiple nodes with 16 cores in each
node.
but its not working and generates a error file . i Have attached the error
file below.
can anyone please tell what is the issue actually
bicgstab_Test
Ankita, looks like your program is not launching correctly.
I would try the following:
define two hosts in a machinefile. Use mpirun -np 2 machinefile date
Ie can you use mpirun just to run the command 'date'
Secondly compile up and try to run an MPI 'Hello World' program
On 9 May 2018 at 12:
MPI "Hello World" program is also working
please see this error file attached below. its of a different program
On Wed, May 9, 2018 at 4:10 PM, John Hearns via users <
users@lists.open-mpi.org> wrote:
> Ankita, looks like your program is not launching correctly.
> I would try the following:
> de
Ankita,
Do you have any reason to suspect the root cause of the crash is Open MPI ?
Cheers,
Gilles
On Wednesday, May 9, 2018, Ankita m wrote:
> MPI "Hello World" program is also working
>
> please see this error file attached below. its of a different program
>
> On Wed, May 9, 2018 at 4:10 P
yes. Because previously i was using intel-mpi. That time the program was
running perfectly. Now when i use openmpi this shows this error
files...Though i am not quite sure. I just thought if the issue will be for
Openmpi then i could get some help here.
On Wed, May 9, 2018 at 6:47 PM, Gilles Gouai
It looks like you're getting a segv when calling MPI_Comm_rank().
This is quite unusual -- MPI_Comm_rank() is just a local lookup / return of an
integer. If MPI_Comm_rank() is seg faulting, it usually indicates that there's
some other kind of memory error in the application, and this seg fault
before i debug ucx further (cause it's totally not working for me), i
figured i'd check to see if it's *really* required to use shmem inside
of openmpi. i'm pretty sure the answer is yes, but i wanted to double
check.
___
users mailing list
users@lists.o
I'm trying to play with oshmem on a single node (just to have a way to
do some simple
experimentation and playing around) and having spectacular problems:
CentOS 6.9 (gcc 4.4.7)
built and installed ucx 1.3.0
built and installed openmpi-3.1.0
[cfreese]$ cat oshmem.c
#include
int
Thanks for confirming that it works for you as well. I have a PR open on v3.1.x
that brings osc/rdma up to date with master. I will also be bringing some code
that greatly improves the multi-threaded RMA performance on Aries systems (at
least with benchmarks— github.com/hpc/rma-mt). That will no
Hi Craig,
You are experiencing problems because you don't have a transport installed
that UCX can use for oshmem.
You either need to go and buy a connectx4/5 HCA from mellanox (and maybe a
switch), and install that
on your system, or else install xpmem (https://github.com/hjelmn/xpmem).
Note ther
11 matches
Mail list logo