[OMPI users] Cluster with IB hosts and Ethernet hosts

2009-01-22 Thread Sangamesh B
Hello all, We''ve a cluster with 23 nodes connected to IB switch and 8 nodes have connected to ethernet switch. Master node is also connected to IB switch. SGE(with tight integration, -pe orte) is used for parallel/serial job submission. Open MPI-1.3 is installed on master node with IB suppo

[OMPI users] MPIAbort and raise(SIGTERM)

2009-01-22 Thread Gijsbert Wiesenekker
I am running OpenMPI on Fedora Core 10. MPI_Abort does not seem to abort the processes on FC10. I found out that when you send a signal all processes seem to abort as expected, so I am now using raise(SIGTERM) instead of MPI_Abort(). Are there any consequences to not using MPI_Abort? Regards,

Re: [OMPI users] dead lock in MPI_Finalize

2009-01-22 Thread Jeff Squyres
Per this note on the "getting help" page, we still need the following: "A _detailed_ description of what is failing. The more details that you provide, the better. E-mails saying "My application doesn't work!" will inevitably be answered with requests for more information about exactly what

Re: [OMPI users] 1.3 and --preload-files and --preload-binary

2009-01-22 Thread Doug Reeder
Josh, It sounds like . is not in your path. That would prevent mpirun from seeing the binary in the current directory. Doug Reeder On Jan 22, 2009, at 10:48 AM, Josh Hursey wrote: As a followup. I can confirm that --preload-files is not working as it should. I was able to use --preload-bi

Re: [OMPI users] 1.3 and --preload-files and --preload-binary

2009-01-22 Thread Josh Hursey
As a followup. I can confirm that --preload-files is not working as it should. I was able to use --preload-binary with a full path to the binary without a problem though. The following commands worked fine (where / tmp is not mounted on all machines): shell$ mpirun -np 2 --preload-binary

Re: [OMPI users] Handling output of processes

2009-01-22 Thread Ralph Castain
Hmmmit works fine for me. Note that orte-iof must be executed from the same node as mpirun. Also, ensure that rank=0 is actually the MPI rank that is printing something! Otherwise, nothing will come out. It should also be noted that it does take time for you to execute orte- iof, and tha

Re: [OMPI users] 1.3 hangs running 2 exes with different names

2009-01-22 Thread Ralph Castain
I can't replicate that behavior - it all seems to be working just fine. I can launch apps of different name, we correctly detect and respond to missing executables, etc. Can you provide more info as to how this was built? Also, be sure to check that the remote hosts are using the same versi

Re: [OMPI users] 1.3 and --preload-files and --preload-binary

2009-01-22 Thread Josh Hursey
The warning is to be expected if the file already exists on the remote side. Open MPI has a policy not to replace the file if it already exists. The segv is concerning. :/ I will take a look and see if I can diagnose what is going on here. Probably in the next day or two. Thanks for the

[OMPI users] 1.3 and --preload-files and --preload-binary

2009-01-22 Thread Geoffroy Pignot
Hello, As you can notice , I am trying the work done on this new release. preload-files and preload-binary options are very interesting to me because I work on a cluster without any shared space between nodes. I tried those basically , but no success . You will find below the error messages. If I

Re: [OMPI users] Handling output of processes

2009-01-22 Thread jody
Hi I downloaded 1.3 the day its release was announced :) I now wanted to try it, so i pieced it together from orte-iof --help: - started an mpirun with 4 process - opened a new terminal - did ps to find the mpirun's pid - then: orte-iof --pid 8449 --rank 0 --stdout but nothing happens... So did i

Re: [OMPI users] dead lock in MPI_Finalize

2009-01-22 Thread Bernard Secher - SFME/LGLS
Hello Tim, I send you the information in join files. Bernard Tim Mattox a écrit : Can you send all the information listed here: http://www.open-mpi.org/community/help/ On Wed, Jan 21, 2009 at 8:58 AM, Bernard Secher - SFME/LGLS wrote: Hello, I have a case wher i have a dead lock in

Re: [OMPI users] Handling output of processes

2009-01-22 Thread Ralph Castain
If you are willing/able to upgrade to 1.3, you will find a new utility called "ompi-iof" (or "orte-iof") that allows you to capture the output from any specified rank or combination of ranks. Note that the output is copied to the tool, so it will also still arrive at the output of mpirun.

[OMPI users] 1.3 hangs running 2 exes with different names

2009-01-22 Thread Geoffroy Pignot
Hello , still a bug ??? compil03% /tmp/openmpi-1.3/bin/mpirun -n 1 --wdir /tmp --host compil03 a.out : -n 1 --host compil02 a.out Hello world from process 0 of 2 Hello world from process 1 of 2 compil03% mv a.out a.out_32 compil03% /tmp/openmpi-1.3/bin/mpirun -n 1 --wdir /tmp --host compil03 a.ou

[OMPI users] Handling output of processes

2009-01-22 Thread jody
Hi I have a small cluster consisting of 9 computers (8x2 CPUs, 1x4 CPUs). I would like to be able to observe the output of the processes separately during an mpirun. What i currently do is to apply the mpirun to a shell script which opens a xterm for each process, which then starts the actual appl