Yes I had checked running mpirun on all nodes one by one to see the
problematic one. I had already mentioned that compute-01-01 is causing
problem, when I remove it from the hostlist mpirun works fine. Here is
ibstatus of compute-01-01.
Infiniband device 'mlx4_0' port 1 status:
default gid
In making the leap from 1.6 to 1.8, how can I check whether or not
process/memory affinity is supported?
I've built OpenMPI on a system where the numactl-devel package was not
installed, and another where it was, but I can't see anything in the output of
ompi_info that suggests any difference b
Rob,
Well, if I understand what you are saying, maybe my concerns about
sub-optimality are unfounded. Yes, the 'buffer' array is contiguous in
memory for each subdomain, and that is what the successful call to
'mpi_file_read_at_all' sees. The call to 'mpi_type_indexed' creates my
'datatype' and
On 07/20/2014 04:23 PM, Tom Rosmond wrote:
Hello,
For several years I have successfully used MPIIO in a Fortran global
atmospheric ensemble data assimilation system. However, I always
wondered if I was fully exploiting the power of MPIIO, specifically by
using derived data types to better des
Hi Rolf,
yes this is exactly what I was looking for, I just hoped that there is
also a way to manually control this behavior.
But in most cases that would be the best setting.
Thanks!
Tobias
On 07/21/2014 05:01 PM, Rolf vandeVaart wrote:
With Open MPI 1.8.1, the library will use the NIC that
With Open MPI 1.8.1, the library will use the NIC that is "closest" to the CPU.
There was a bug in earlier versions of Open MPI 1.8 so that did not happen.
You can see this by running with some verbosity using the "btl_base_verbose"
flag. For example, this is what I observed on a two node clus
You have to check the ports states on *all* nodes in the run/job/submission.
Checking on a single node is not enough.
My guess is the 01-00 tries to connect 01-01 and the ports are down on 01-01.
You may disable support for infiniband by adding --mca btl ^openib.
Best,
Pavel (Pasha) Shamis
---
Hello Ralph,
thanks for your answer.
> I can look to see if there is something generic we can do (perhaps
enclosing each param in quotes to avoid any special character issues) -
will see if something like that might help. Best that will happen,
however, is that we launch the app and then have
Dear All
I need your help to solve this cluster related issue causing mpirun
malfunction. I get following warning for some of the nodes and then the
route failure message comes causing failure to mpirun.
*WARNING: There is at least one OpenFabrics device found but there are no
active ports dete
Please see:
http://bugs.centos.org/view.php?id=5812
From: users [users-boun...@open-mpi.org] on behalf of Ralph Castain
[r...@open-mpi.org]
Sent: Sunday, July 20, 2014 9:30 AM
To: Open MPI Users
Subject: Re: [OMPI users] Mpirun 1.5.4 problems when request > 28 sl
NIC = network interface controller?
There is QDR Infiniband 4x/10G Ethernet/Gigabit Ethernet.
I want to use QDR Infiniband.
Here is a new output:
$ mpirun -mca mca_base_env_list 'LD_PRELOAD' --debug-daemons --mca
plm_base_verbose 5 -mca oob_base_verbose 10 -mca rml_base_verbose 10 -np 2
hello
11 matches
Mail list logo