may be because of insufficient storage space...???( I mean hard
disk space.)
On Thu, Aug 6, 2009 at 11:23 PM, Jeff Squyres wrote:
> Any chance you could re-try the experiment with Open MPI 1.3.3?
>
>
> On Aug 4, 2009, at 11:10 AM, Hoelzlwimmer Andreas - S0810595005 wrote:
>
> Hello,
Hi Craig,
WRF has pattern of talking to nearest neighbours like p+1, p-1,
p+3 and p-3, where p is the particular process. But in addition to that,
it also uses collective calls like MPI_Bcast, MPI_AlltoAllv,
MPI_Allgather, MPI_Gather, MPI_Gatherv, MPI_Scatterv.
Apparently openm
There's been quite some discussion here lately about the effect of OMPI
tuning parameters, particularly w.r.t. collectives.
Is there some tool to probe performance on any given network/collection
of nodes to aid optimisation of these parameters?
(I'm thinking something along the philosophy of ATL
Craig,
Did your affinity script bind the processes per socket or linearly to cores.
If the former you'll want to look at using rankfiles and place the ranks based
on sockets. TWe have found this especially useful if you are not running fully
subscribed on your machines.
Also, if you think t
Hi Terry,
I feel hierarchical collectives are slower compare to tuned one. I
had done some benchmark in the past specific to collectives, and this is
what i feel based on my observation.
Regards
Neeraj Chourasia (MTS)
Computational Research Laboratories Ltd.
(A wholly Owned Subsidiary
Hi Craig,
How was the nodefile selected for execution? Whether it was
provided by scheduler say LSF/SGE/PBS or you manually gave it?
With WRF, we observed giving sequential nodes (Blades which are in the
same order as in enclosure) gave us some performance benefit.
Regards
Neeraj Chour
Hi Neeraj,
Were there specific collectives that were slower? Also what kind of cluster
were you running on? How many nodes and cores per node?
thanks,
--td
Message: 3
Date: Fri, 7 Aug 2009 16:51:05 +0530
From: nee...@crlindia.com
Subject: Re: [OMPI users] Performance question about OpenMPI
Hi Terry,
I had tested mostly MPI_Bcast, MPI_Reduce, MPI_Gather kind of MPI
with openmpi-1.3 and hierarchical option enabled.In all these, i found
results slower than regular tuned collectives.
We have HP Blade with intel clovertown processor(two quad core)
connected with DDR i
Terry Dontje wrote:
Craig,
Did your affinity script bind the processes per socket or linearly to
cores. If the former you'll want to look at using rankfiles and place
the ranks based on sockets. TWe have found this especially useful if
you are not running fully subscribed on your machines.
nee...@crlindia.com wrote:
Hi Craig,
How was the nodefile selected for execution? Whether it was
provided by scheduler say LSF/SGE/PBS or you manually gave it?
With WRF, we observed giving sequential nodes (Blades which are in the
same order as in enclosure) gave us some performance be
I'm way behind on my mail; apologies for the delay in replying.
Did you figure this out?
As a pure guess, it sounds like you have a heterogeneous setup --
nodes have different distros and/or versions. As such, your glibc's
may be different, etc. In such situations, it is definitely
recom
Date: Fri, 07 Aug 2009 07:12:45 -0600
From: Craig Tierney
Subject: Re: [OMPI users] Performance question about OpenMPI and
MVAPICH2 on IB
To: Open MPI Users
Message-ID: <4a7c284d.3040...@noaa.gov>
Content-Type: text/plain; charset=ISO-8859-1; format=flowed
Terry Dontje wrote:
>
Terry Dontje wrote:
Craig,
Did your affinity script bind the processes per socket or linearly to
cores. If the former you'll want to look at using rankfiles and place
the ranks based on sockets. TWe have found this especially useful if
you are not running fully subscribed on your machines.
To use tuned collectives, do all I have to do is add --mca coll tuned?
I am trying to run with:
# mpirun -np 8 --mca coll tuned --mca orte_base_help_aggregate 0 ./wrf.exe
But all the processes fail with the folling message:
--
Hello all,
I'm using hpc8.2 :
Lidia-jobic% ompi_info
Displaying Open MPI information for 32-bit ...
Package: ClusterTools 8.2
Open MPI: 1.3.3r21324-ct8.2-b09j-r40
[...]
And i've got a X4600 machine (8*4 cores).
When i'm trying to run a 32 processor jobs, i've got
This message is telling you that you have run out of file descriptors.
I am surprised that the -mca parameter setting did not fix the problem.
Can you run limit or ulimit on your shell and send the information? I
typically set my limit to 65536 assuming the system allows it.
burl-16 58 =>limit
Rolf Vandevaart wrote:
This message is telling you that you have run out of file descriptors.
I am surprised that the -mca parameter setting did not fix the problem.
Can you run limit or ulimit on your shell and send the information? I
typically set my limit to 65536 assuming the system allows
Hi Craig, Terry, Neeraj, list
Craig: A fellow here runs WRF.
I grep the code and there is plenty of collectives there:
MPI_[All]Gather[v], MPI_[All]Reduce, etc.
Domain decomposition code like WRF, MITgcm, and other atmosphere
and ocean codes has point-to-point communication to exchange
subdomain
Hello,
I have three sets of nodes, each with openmpi installed in
a different location. I am getting an error related to orted:
/users/kenneth/info/openmpi/install/bin/orted: Command not found.
I think it's looking for orted in the wrong place on some of
the nodes. Is there an easy way to
Terry Frankcombe wrote:
There's been quite some discussion here lately about the effect of OMPI
tuning parameters, particularly w.r.t. collectives.
Is there some tool to probe performance on any given network/collection
of nodes to aid optimisation of these parameters?
(I'm thinking something a
I'm not sure that there is a general "best set" of parameters, given the
dependence of that set on comms patterns, etc...
Still, this *is* a classic parameter sweep and optimization problem
(unlike ATLAS), with a small number of parameters, and is the sort of
thing one should be able to hook u
I have several Sun x4100 with Infiniband which appear to be running at
400MB/sec instead of 800MB/sec. It a freshly reformatted cluster
converting from solaris to linux. We also reset the bios settings
with "load optimal defaults". Does anyone know which bios setting i
changed to dump the BW?
x4
Gus Correa wrote:
Terry Frankcombe wrote:
There's been quite some discussion here lately about the effect of OMPI
tuning parameters, particularly w.r.t. collectives.
Is there some tool to probe performance on any given network/collection
of nodes to aid optimisation of these parameters?
(I'm t
Not that I know of - I don't think we currently have any way for you to
specify a location for individual nodes.
Is there some reason why you installed it this way?
On Fri, Aug 7, 2009 at 11:27 AM, Kenneth Yoshimoto wrote:
>
> Hello,
>
> I have three sets of nodes, each with openmpi installed
24 matches
Mail list logo