Re: [OMPI users] signal 15 (terminated)

2009-02-03 Thread Hana Milani
sorry if I didn't answer: Have you checked to ensure that the job manager is not killing your job? I am not quite sure what you mean by job manager, but, this is the personal computer of mine. Much to my surprise, I have also open suse on my laptop, took the similar procedure then the same mess

Re: [OMPI users] OpenMPI hangs across multiple nodes.

2009-02-03 Thread Ralph Castain
Could you tell us which version of OpenMPI you are using, and how it was configured? Did you install the OMPI libraries and binaries on both nodes? Are they in the same absolute path locations? Thanks Ralph On Feb 3, 2009, at 3:46 PM, Robertson Burgess wrote: Dear users, I am quite new

[OMPI users] OpenMPI hangs across multiple nodes.

2009-02-03 Thread Robertson Burgess
Dear users, I am quite new to OpenMPI, I have compiled it on two nodes, each node with 8 CPU cores. The two nodes are identical. The code I am using works in parallel across the 8 cores on a single node. However, whenever I try to run across both nodes, OpenMPI simply hangs. There is no output w

Re: [OMPI users] Handling output of processes

2009-02-03 Thread Ralph Castain
Hi Jody Well, the problem with both the output filename and the xterm option was that I wasn't passing them back to the remote daemons under the ssh launch environment. I should have that corrected now - things will hopefully work with any tarball of r20407 or above. Let me know... Ralph

Re: [OMPI users] Handling output of processes

2009-02-03 Thread Ralph Castain
Ah! I know the problem - forgot you are running under ssh, so the environment doesn't get passed. I'll have to find a way to pass the output filename to the backend nodes...should have it later today. On Feb 3, 2009, at 11:09 AM, jody wrote: Hi Ralph --output-filename It creates files,

Re: [OMPI users] v1.3: mca_common_sm_mmap_init error

2009-02-03 Thread Eugene Loh
Prentice Bisbal wrote: Jeff Squyres wrote: On Feb 2, 2009, at 4:48 PM, Prentice Bisbal wrote No. I was running just a simple "Hello, world" program to test v1.3 when these errors occured. And as soon as I reverted to 1.2.8, the errors disappeared. FW

Re: [OMPI users] v1.3: mca_common_sm_mmap_init error

2009-02-03 Thread Jeff Squyres
By peers I mean the number of MPI processes on the same host. So if you run with 4 processes on a single host, OMPI sets up shared memory for those 4 processes during MPI_INIT, regardless of whether you call MPI send/receive functions or not. On Feb 3, 2009, at 1:15 PM, Prentice Bisbal wr

Re: [OMPI users] v1.3: mca_common_sm_mmap_init error

2009-02-03 Thread Prentice Bisbal
Jeff Squyres wrote: > On Feb 2, 2009, at 4:48 PM, Prentice Bisbal wrote: > >> No. I was running just a simple "Hello, world" program to test v1.3 when >> these errors occured. And as soon as I reverted to 1.2.8, the errors >> disappeared. > > FWIW, OMPI allocates shared memory based on the number

Re: [OMPI users] Handling output of processes

2009-02-03 Thread jody
Hi Ralph >> >> --output-filename >> It creates files, but only for the local processes: >> [jody@localhost neander]$ mpirun -np 8 -hostfile testhosts >> --output-filename gnana ./MPITest >> ... output ... >> [jody@localhost neander]$ ls -l gna* >> -rw-r--r-- 1 jody morpho 549 2009-02-03 18:02 gna

Re: [OMPI users] Handling output of processes

2009-02-03 Thread Ralph Castain
On Feb 3, 2009, at 10:16 AM, jody wrote: Hi Ralph Thanks for the fixes and the "!". --xterm: The "!" works, but i still don't have any xterms from my remote nodes even with all my xhost+ and -x DISPLAY tricks explained below :( --output-filename It creates files, but only for the local proces

Re: [OMPI users] Handling output of processes

2009-02-03 Thread jody
Hi Ralph Thanks for the fixes and the "!". --xterm: The "!" works, but i still don't have any xterms from my remote nodes even with all my xhost+ and -x DISPLAY tricks explained below :( --output-filename It creates files, but only for the local processes: [jody@localhost neander]$ mpirun -np 8

Re: [OMPI users] signal 15 (terminated)

2009-02-03 Thread Jeff Squyres
Not to be pedantic, but you haven't answered the questions that I asked from the last time I tried to help you. :-( http://www.open-mpi.org/community/lists/users/2009/01/7710.php http://www.open-mpi.org/community/lists/users/2009/01/7717.php On Feb 3, 2009, at 6:52 AM, Hana Milani wrote: I

Re: [OMPI users] v1.3: mca_common_sm_mmap_init error

2009-02-03 Thread Jeff Squyres
On Feb 2, 2009, at 4:48 PM, Prentice Bisbal wrote: No. I was running just a simple "Hello, world" program to test v1.3 when these errors occured. And as soon as I reverted to 1.2.8, the errors disappeared. FWIW, OMPI allocates shared memory based on the number of peers on the host. This a

Re: [OMPI users] Open MPI error

2009-02-03 Thread Jeff Squyres
Have you tried this: http://www.open-mpi.org/faq/?category=openfabrics#v1.2-use-early-completion On Feb 2, 2009, at 2:52 PM, c.j@exxonmobil.com wrote: I am using openmpi to run a job on 4 nodes, 2 processors per node. Seems like 5 out of the 8 processors executed the app successfu

Re: [OMPI users] Joe Griffin: "Re: Cannot compile onLinuxItanium system"

2009-02-03 Thread Jeff Squyres
On Feb 2, 2009, at 1:27 PM, Joe Griffin wrote: Hi Jeff and Tony, I just tried 1.2.2 on my Itanium and got the same error. Our code uses 1.2.2, but I built it back in 2007 with V8 of the compiler. Now I am using V10.1. So, perhaps the code is incompatible with Intel V10.1. (on Itanium) Ah

[OMPI users] signal 15 (terminated)

2009-02-03 Thread Hana Milani
Dear Users, I have posted this problem before, but still I am trying to solve it so I put it again so that maybe you can help me. I run my simulating code in a system with Intel Core 2 Quad, 4x2.4 GHz cpu, with 7.9 GiB memopry. My linux is opensuse 11.0 with gfortran compiler. I have installed