Yes, it looks like you have a heterogeneous system (i.e., a binary compiled on
one server doesn't necessarily run properly on another server).
In this case, you should see the heterogeneous section of the FAQ.
Fair warning, though -- heterogeneous systems are more difficult to
manage/maintain/u
It may be because the other system is running upgraded version of linux
which is not having infiniband drivers. Any solution?
On Tue, Mar 26, 2013 at 12:42 PM, Syed Ahsan Ali wrote:
> Tried this but mpirun exits with this error
>
> mpirun -np 40 /home/MET/hrm/bin/hrm
> librdmacm: couldn't read A
Tried this but mpirun exits with this error
mpirun -np 40 /home/MET/hrm/bin/hrm
librdmacm: couldn't read ABI version.
librdmacm: assuming: 4
librdmacm: couldn't read ABI version.
librdmacm: assuming: 4
librdmacm: couldn't read ABI version.
librdmacm: assuming: 4
librdmacm: couldn't read ABI versio
On Mar 22, 2013, at 3:42 AM, Syed Ahsan Ali wrote:
> Actually due to some data base corruption I am not able to add any new node
> to cluster from the installer node. So I want to run parallel job on more
> nodes without adding them to existing cluster.
> You are right the binaries must be pre
Actually due to some data base corruption I am not able to add any new node
to cluster from the installer node. So I want to run parallel job on more
nodes without adding them to existing cluster.
You are right the binaries must be present on the remote node as well.
Is this possible throught nfs?
Am 22.03.2013 um 10:14 schrieb Syed Ahsan Ali:
> I have a very basic question. If we want to run mpirun job on two systems
> which are not part of cluster, then how we can make it possible. Can the host
> be specifiend on mpirun which is not compute node, rather a stand alone
> system.
Sure, t