[OMPI users] OpenFabrics (openib)

2013-02-27 Thread Khapare Joshi
Hello all, First time poster, I recently installed openmpi 1.6.4 in my cluster with resource manager support as : ./configure --with-tm --prefix=/opt/openmpi/1.6.2/ it works well, but I always get some error saying : [[58551,1],0]: A high-performance Open MPI point-to-point messaging module was

Re: [OMPI users] OpenFabrics (openib)

2013-02-27 Thread Ralph Castain
--without-openib will do the trick On Feb 27, 2013, at 7:24 AM, Khapare Joshi wrote: > Hello all, > > First time poster, I recently installed openmpi 1.6.4 in my cluster with > resource manager support as : > > ./configure --with-tm --prefix=/opt/openmpi/1.6.2/ > > it works well, but I alwa

Re: [OMPI users] rcu_sched stalls on CPU

2013-02-27 Thread Simon DeDeo
We've resolved this issue, which appears to have been an early warning of a large-scale hardware failure. Twelve hours later the machine was unable to power-on or self-test. We are now running on a new machine, and the same jobs are finishing normally -- without having to worry about Send/Ssen

Re: [OMPI users] rcu_sched stalls on CPU

2013-02-27 Thread Jeff Squyres (jsquyres)
I'm glad you figured this out. Your mail was on my to-do list to reply to today; I didn't reply earlier simply because I had no idea what the problem could have been. I'm also kinda glad it wasn't related to MPI. ;-) On Feb 27, 2013, at 11:20 AM, Simon DeDeo wrote: > We've resolved this

Re: [OMPI users] OpenFabrics (openib)

2013-02-27 Thread Jeff Squyres (jsquyres)
You can also just disable/unload the OpenFabrics drivers in your systems. Open MPI is reacting to the fact that it could the drivers loaded (even though there is no OpenFabrics-based hardware active, apparently). If you unload the drivers, this message should go away. On Feb 27, 2013, at 10:

[OMPI users] Option -cpus-per-proc 2 not working with given machinefile?

2013-02-27 Thread Reuti
Hi, I have an issue using the option -cpus-per-proc 2. As I have Bulldozer machines and I want only one process per FP core, I thought using -cpus-per-proc 2 would be the way to go. Initially I had this issue inside GridEngine but then tried it outside any queuingsystem and face exactly the sam