sorry if I didn't answer:
Have you checked to ensure that the job manager is not killing your job?
I am not quite sure what you mean by job manager, but, this is the personal
computer of mine. Much to my surprise, I have also open suse on my laptop, took
the similar procedure then the same mess
Could you tell us which version of OpenMPI you are using, and how it
was configured?
Did you install the OMPI libraries and binaries on both nodes? Are
they in the same absolute path locations?
Thanks
Ralph
On Feb 3, 2009, at 3:46 PM, Robertson Burgess wrote:
Dear users,
I am quite new
Dear users,
I am quite new to OpenMPI, I have compiled it on two nodes, each node with 8
CPU cores. The two nodes are identical. The code I am using works in parallel
across the 8 cores on a single node. However, whenever I try to run across both
nodes, OpenMPI simply hangs. There is no output w
Hi Jody
Well, the problem with both the output filename and the xterm option
was that I wasn't passing them back to the remote daemons under the
ssh launch environment. I should have that corrected now - things will
hopefully work with any tarball of r20407 or above.
Let me know...
Ralph
Ah! I know the problem - forgot you are running under ssh, so the
environment doesn't get passed.
I'll have to find a way to pass the output filename to the backend
nodes...should have it later today.
On Feb 3, 2009, at 11:09 AM, jody wrote:
Hi Ralph
--output-filename
It creates files,
Prentice Bisbal wrote:
Jeff Squyres wrote:
On Feb 2, 2009, at 4:48 PM, Prentice Bisbal wrote
No. I was running just a simple "Hello, world" program to test v1.3 when
these errors occured. And as soon as I reverted to 1.2.8, the errors
disappeared.
FW
By peers I mean the number of MPI processes on the same host. So if
you run with 4 processes on a single host, OMPI sets up shared memory
for those 4 processes during MPI_INIT, regardless of whether you call
MPI send/receive functions or not.
On Feb 3, 2009, at 1:15 PM, Prentice Bisbal wr
Jeff Squyres wrote:
> On Feb 2, 2009, at 4:48 PM, Prentice Bisbal wrote:
>
>> No. I was running just a simple "Hello, world" program to test v1.3 when
>> these errors occured. And as soon as I reverted to 1.2.8, the errors
>> disappeared.
>
> FWIW, OMPI allocates shared memory based on the number
Hi Ralph
>>
>> --output-filename
>> It creates files, but only for the local processes:
>> [jody@localhost neander]$ mpirun -np 8 -hostfile testhosts
>> --output-filename gnana ./MPITest
>> ... output ...
>> [jody@localhost neander]$ ls -l gna*
>> -rw-r--r-- 1 jody morpho 549 2009-02-03 18:02 gna
On Feb 3, 2009, at 10:16 AM, jody wrote:
Hi Ralph
Thanks for the fixes and the "!".
--xterm:
The "!" works, but i still don't have any xterms from my remote nodes
even with all my xhost+ and -x DISPLAY tricks explained below :(
--output-filename
It creates files, but only for the local proces
Hi Ralph
Thanks for the fixes and the "!".
--xterm:
The "!" works, but i still don't have any xterms from my remote nodes
even with all my xhost+ and -x DISPLAY tricks explained below :(
--output-filename
It creates files, but only for the local processes:
[jody@localhost neander]$ mpirun -np 8
Not to be pedantic, but you haven't answered the questions that I
asked from the last time I tried to help you. :-(
http://www.open-mpi.org/community/lists/users/2009/01/7710.php
http://www.open-mpi.org/community/lists/users/2009/01/7717.php
On Feb 3, 2009, at 6:52 AM, Hana Milani wrote:
I
On Feb 2, 2009, at 4:48 PM, Prentice Bisbal wrote:
No. I was running just a simple "Hello, world" program to test v1.3
when
these errors occured. And as soon as I reverted to 1.2.8, the errors
disappeared.
FWIW, OMPI allocates shared memory based on the number of peers on the
host. This a
Have you tried this:
http://www.open-mpi.org/faq/?category=openfabrics#v1.2-use-early-completion
On Feb 2, 2009, at 2:52 PM, c.j@exxonmobil.com wrote:
I am using openmpi to run a job on 4 nodes, 2 processors per node.
Seems
like 5 out of the 8 processors executed the app successfu
On Feb 2, 2009, at 1:27 PM, Joe Griffin wrote:
Hi Jeff and Tony,
I just tried 1.2.2 on my Itanium and got the same error.
Our code uses 1.2.2, but I built it back in 2007 with
V8 of the compiler. Now I am using V10.1. So, perhaps
the code is incompatible with Intel V10.1.
(on Itanium)
Ah
Dear Users,
I have posted this problem before, but still I am trying to solve it so I put
it again so that maybe you can help me. I run my simulating code in a system
with Intel Core 2 Quad, 4x2.4 GHz cpu, with 7.9 GiB memopry. My linux is
opensuse 11.0 with gfortran compiler.
I have installed
16 matches
Mail list logo