e receiver gets the message.
>
> On Oct 4, 2010, at 11:13 AM, "Andrei Fokau"
> wrote:
>
> Does OMPI have shared memory capabilities (as it is mentioned in MPI-2)?
> How can I use them?
>
> On Sat, Sep 25, 2010 at 23:19, Andrei Fokau <
> andrei.fo...@neutron.kth.s
Does OMPI have shared memory capabilities (as it is mentioned in MPI-2)?
How can I use them?
Andrei
On Sat, Sep 25, 2010 at 23:19, Andrei Fokau wrote:
> Here are some more details about our problem. We use a dozen of 4-processor
> nodes with 8 GB memory on each node. The code we run
We have a a program which is normally (running w/o MPI) controlled by Ctrl+C
and then a number of options. Is it possible to set STDIN for a command
executed via mpirun? The following page describes parameter -stdin, but it
seems to be not supported by OMPI's mpirun.
http://w3.ualg.pt/~dubuf/calhau
to constrain how processes are
> > mapped to nodes, this could be easy. (E.g., "every 4 processes are
> > collocal".) If you want to discover dynamically at run time which are
> > collocal, it would be harder. The mmap stuff could be in a stand-alone
> > function
! exit program
On Fri, Sep 24, 2010 at 13:44, Reuti wrote:
> Am 24.09.2010 um 13:26 schrieb John Hearns:
>
> > On 24 September 2010 08:46, Andrei Fokau
> wrote:
> >> We use a C-program which consumes a lot of memory per process (up to few
> >> GB), 9
We use a C-program which consumes a lot of memory per process (up to few
GB), 99% of the data being the same for each process. So for us it would be
quite reasonable to put that part of data in a shared memory.
In the source code, the memory is allocated via malloc() function. What
would it requir
yet.
>
> Specifically, the code that does what you specify below is done and works.
> It is recovery of the MPI job itself (collectives, lost messages, etc.) that
> remains to be completed.
>
>
> On Thu, Sep 23, 2010 at 7:22 AM, Andrei Fokau <
> andrei.fo...@neutron.kth.s
Dear users,
Our cluster has a number of nodes which have high probability to crash, so
it happens quite often that calculations stop due to one node getting down.
May be you know if it is possible to block the crashed nodes during run-time
when running with OpenMPI? I am asking about principal pos