> I was hoping that someone might have some examples of real application
> behaviour rather than micro benchmarks. It can be crazy hard to get that
> information from users.
[Tom] 
I don't have direct performance information on knem, but with Intel's (formerly 
QLogic's) PSM layer as delivered in our software stack (Intel True Scale Fabric 
Suite) known as IFS, there is a kcopy module that assists shared memory MPI 
bandwidth in a way similar to knem.

We ran SPEC MPI2007 benchmarks quite a while ago and kcopy showed about a 2% 
advantage on average over the 13 applications that make up the suite. -- There 
were codes which did not benefit, but no downside.  This was run over 16 nodes 
at 8 cores per node, so not very fat nodes.

More interestingly, on one of our software revs. a few years ago, a bug crept 
in which disabled kcopy.  A customer filed an issue that one of their apps 
slowed down by 30%.  Fixing that bug restored the previous performance.  The 
application was proprietary, so I don't even know what it did in general.  It 
was run over multiple nodes, so this was not a single-node performance 
comparison.

More recently, some customers with large memory nodes, and > 40 cores per node 
found kcopy was important to the performance of their most important app, a 
finite element code (I don't have a percentage figure).

kcopy works with Open MPI over PSM , so using knem instead of kcopy is not 
likely to speed up that configuration much (unless you get your PSM from OFED 
or a Linux distro, then that won't include kcopy; we weren't able to get kcopy 
accepted upstream).  Recent PSM (from OFED 3.5 say) can be built to use knem 
for kernel-assisted copies.     kcopy also works with the other MPIs that 
support PSM.

Hope these anecdotes are relevant to Open MPI users considering knem.

-Tom Elken  





> 
> Unusually for us, we're putting in a second cluster with the same
> architecture, CPUs, memory and OS as the last one. I might be able to use
> this as a bigger stick to get some better feedback. If so, I'll pass it
> on.
> 
> > Darius Buntinas, Brice Goglin, et al. wrote an excellent paper about
> > exactly this set of issues; see http://runtime.bordeaux.inria.fr/knem/.
> ...
> 
> I'll definitely take a look - thanks again.
> 
> All the best,
> 
> Mark
> --
> -----------------------------------------------------------------
> Mark Dixon                       Email    : m.c.di...@leeds.ac.uk
> HPC/Grid Systems Support         Tel (int): 35429
> Information Systems Services     Tel (ext): +44(0)113 343 5429
> University of Leeds, LS2 9JT, UK
> -----------------------------------------------------------------
> _______________________________________________
> users mailing list
> us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/users

Reply via email to