Dear All
I have an application which is run using openmpi and uses infiniband flags.
The application is a forecast model simulation. A frequent problem arises
that the Infiniband mezzanine cards of servers become faulty (don't know
the reason why it happens so frequent), the model simulation becom
Short answer. Run ibstats or ibstatus.
Look also at the logs of your subnet manager.
You can try downloading and installing a fresher version of MXM from
mellanox web site. There was a thread on the list with the same problem,
you can search for it.
2012/11/28 Joseph Farran
> Howdy.
>
> I a have a stock CentOS 6.3 OS and a Mellanox MT26428 card.
>
> I installed the Mellanox OFE
ibstats comes with some other distribution? I don't have this command
available right now
On Wed, Nov 28, 2012 at 1:14 PM, John Hearns wrote:
> Short answer. Run ibstats or ibstatus.
> Look also at the logs of your subnet manager.
>
> ___
> users mail
Those diagnostics are from Openfabrics.
What type of infiniband card do you have?
What drivers are you using?
I'm trying to build openmpi with support for BLCR plus qlogic infiniband
(plus grid engine). Everything seems to compile OK and checkpoints are
taken but whenever I try to restore a checkpoint I get the following error:
- do_mmap(, 2aaab18c7000, 1000, ...) failed:
ffea
On 11/28/2012 10:52 AM, Pavel Mezentsev wrote:
> You can try downloading and installing a fresher version of MXM from mellanox
> web site. There was a thread on the list with the same problem, you can
> search for it.
Indeed, that OFED version comes with older version of MXM.
You can get the new
I know java MPI bindings are in the development tree, and the FAQ states they
are derived from the HLRS bindings (which I can't appear to find online).
Is it possible to take the bindings from the dev tree and build them against
the 1.6 stable? If not what mpiJava bindings do you recommend?
B
On Nov 28, 2012, at 10:06 AM, Brock Palen wrote:
> I know java MPI bindings are in the development tree, and the FAQ states they
> are derived from the HLRS bindings (which I can't appear to find online).
I don't remember exactly where we got them; the repository might not have been
public. I
Perfect and Thanks!
I had searched the Mellanox web site for the mxm package to no avail.
I will try recompiling later today.
Best,
Joseph
On 11/28/2012 06:29 AM, Yevgeny Kliteynik wrote:
On 11/28/2012 10:52 AM, Pavel Mezentsev wrote:
You can try downloading and installing a fresher version
Question on the version to use. The tar files contains the following RPMS:
# ls | grep -v debug
mxm-1.1.3a5e745-1.x86_64-centos5u6.rpm
mxm-1.1.3a5e745-1.x86_64-centos5u7.rpm
mxm-1.1.3a5e745-1.x86_64-centos6u0.rpm
mxm-1.1.3a5e745-1.x86_64-rhel5u5.rpm
mxm-1.1.3a5e745-1.x86_64-rhel6u1.rpm
mxm-1
Our case is a single user expressing interest, and maybe long term mixing as we
explore hadoop options, and would mix MPI+hadoop.
I would not go into the effort if it is non-trivial to add it to 1.6.
Brock Palen
www.umich.edu/~brockp
CAEN Advanced Computing
bro...@umich.edu
(734)936-1985
On
FWIW: Greenplum is about to release exactly that solution based on OMPI, though
it will require the OMPI trunk. Not back portable to v1.6 I'm afraid, but
allows you to run Hadoop on any cluster as opposed to requiring dedicated
hardware.
I can get an early-adopter release for you, if it helps.
You need mxm-1.1.3a5e745-1.x86_64-**rhel6u3.rpm
On Wed, Nov 28, 2012 at 7:44 PM, Joseph Farran wrote:
> mxm-1.1.3a5e745-1.x86_64-**rhel6u3.rpm
>
Hello,
I am trying to build OpenMPI 1.6.3 on an IBM/Intel RHEL-6 cluster.
I tried building with variations (meaning enable-...=no, disable-,
changing switch order, etc.) of this:
./configure CC=icc CFLAGS="-xT -O2" F77=ifort FFLAGS="-xT -O2"
FC=ifort FCFLAGS="-xT -O2" --enable-mpi-cxx=no -
Hi openmpi'users
I now trying to install mrmpi in a cluster to use it with openmpi, I installed
locally openmpi with the option --disable-dlopen , I install also mrmpi without
apparent problems but when testing with the basic commands in a python27 shell
from mrmpi import mrmpi
mr=mrmpi()
I ge
I'll bet we're not disabling the C++ test properly when you disable the C++
bindings. Bummer. I'll file a bug, but I don't know when that will be fixed.
However, this kind of error typically only occurs when your C++ compiler fails
altogether (e.g., it's broken). Check the config.log file and
Hi Jeff,
Thanks. I am just running the Intel 13.0.1 compiler on the Quarry
cluster at IU. It would be very odd to have a serious issue without
users complaining. I tried running it again with C++ turned on with:
./configure CC=icc CFLAGS="-xT -O2" F77=ifort FFLAGS="-xT -O2"
FC=ifort FCFL
According to config.log, your icpc is broken -- it won't compile a trivial C++
program. Try it yourself -- try compiling
-
#include
#include
using namespace std;
int main(int argc, char* argv[]) {
cout << "Hello, world" << endl;
return 0;
}
-
Do you need to set some environmen
I was looking for some info about MPI port on iOS or Android devices.
I have some old devices that may result useful, if I could be able to include
them in my computation scheme.
OpenCL runs on iOS and Android, so I was wondering if there is any way to have
an old iPhone/phone or iPad/tablet to
You might want to post in beowulf mailing list see cc
and you want to install linux of course.
OpenFabrics releases openmpi, yet it only works at a limited number
of distributions - most important is having
the correct kernel (usually old kernel).
I'm gonna try get it to work at debian soon.
shameless plug:
http://www.mathematik.tu-dortmund.de/~goeddeke/pubs/pdf/Goeddeke_2012_EEV.pdf
In the MontBlanc project (www.montblanc-project.eu), a lot of folks from
all around Europe look into exactly this. Together with a few
colleagues, we have been honoured to get access to an early proto
I am not sure about drivers because those were installed by someone else
during cluster setup. I see following information about infiniband card.
The card is DDR InfiniBand Mellanox ConnectX.
On Wed, Nov 28, 2012 at 3:17 PM, John Hearns wrote:
> Those diagnostics are from Openfabrics.
> What ty
On Nov 28, 2012, at 12:21 PM, Mariana Vargas Magana
wrote:
> Hi openmpi'users
>
> I now trying to install mrmpi in a cluster to use it with openmpi, I
> installed locally openmpi with the option --disable-dlopen , I install also
> mrmpi without apparent problems but when testing with the bas
I seem to vaguely recall someone porting OMPI to the iPad at one time as part
of a large-scale, impromptu cluster demo at some conference - everyone was
supposed to bring a computer, network them all into a large "cluster", and then
run a benchmark to see how fast it would work. I can't find or
25 matches
Mail list logo