On Fri, Mar 17, 2017 at 3:33 PM, Joshua Mora <joshua_m...@usa.net> wrote:

> Thanks for the quick reply.
> This test is between 2 cores that are on different CPUs. Say data has to
> traverse coherent fabric (eg. QPI,UPI, cHT).
> It has to go to main memory independently of cache size. Wrong assumption ?
>

Depends on the usage pattern. Some benchmarks have options to clean/flush
the cache before each round of tests.


> Can data be evicted from cache and put into cache of second core on
> different
> CPU without placing it first in main memory ?
>

It would depend on the memory coherency protocol. Usually it gets marked as
shared, and as a result it might not need to be pushed into main memory
right away.


> I am more thinking that there is a parameter that splits large messages in
> smaller ones at 64k or 128k ?
>

Pipelining is not the answer to all situations. Once your messages are
larger than the caches, you already built memory pressure (by getting
outside the cache size) so the pipelining is bound by the memory bandwidth.



> This seems (wrong assumption ?) like the kind of parameter I would need for
> large messages on a NIC. Coalescing data / large MTU,...


Sure, but there are hard limits imposed by the hardware, especially with
regards to intranode communications. Once you saturate the memory bus, you
hit a pretty hard limit.

  George.



>
> Joshua
>
>
>
>
>
>
>
>
> ------ Original Message ------
> Received: 02:15 PM CDT, 03/17/2017
> From: George Bosilca <bosi...@icl.utk.edu>
> To: Open MPI Users <users@lists.open-mpi.org>
>
> Subject: Re: [OMPI users] tuning sm/vader for large messages
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
> > Joshua,
> >
> > In shared memory the bandwidth depends on many parameters, including the
> > process placement and the size of the different cache levels. In your
> > particular case I guess after 128k you are outside the L2 cache (1/2 of
> the
> > cache in fact) and the bandwidth will drop as the data need to be flushed
> > to main memory.
> >
> >   George.
> >
> >
> >
> > On Fri, Mar 17, 2017 at 1:47 PM, Joshua Mora <joshua_m...@usa.net>
> wrote:
> >
> > > Hello,
> > > I am trying to get the max bw for shared memory communications using
> > > osu_[bw,bibw,mbw_mr] benchmarks.
> > > I am observing a peak at ~64k/128K msg size and then drops instead of
> > > sustaining it.
> > > What parameters or linux config do I need to add to default openmpi
> > > settings
> > > to get this improved ?
> > > I am already using vader and knem.
> > >
> > > See below one way bandwidth with peak at 64k.
> > >
> > > # Size      Bandwidth (MB/s)
> > > 1                       1.02
> > > 2                       2.13
> > > 4                       4.03
> > > 8                       8.48
> > > 16                     11.90
> > > 32                     23.29
> > > 64                     47.33
> > > 128                    88.08
> > > 256                   136.77
> > > 512                   245.06
> > > 1024                  263.79
> > > 2048                  405.49
> > > 4096                 1040.46
> > > 8192                 1964.81
> > > 16384                2983.71
> > > 32768                5705.11
> > > 65536                7181.11
> > > 131072               6490.55
> > > 262144               4449.59
> > > 524288               4898.14
> > > 1048576              5324.45
> > > 2097152              5539.79
> > > 4194304              5669.76
> > >
> > > Thanks,
> > > Joshua
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > >
> > > _______________________________________________
> > > users mailing list
> > > users@lists.open-mpi.org
> > > https://rfd.newmexicoconsortium.org/mailman/listinfo/users
> > >
> >
>
>
>
>
>
>
>
>
> > _______________________________________________
> > users mailing list
> > users@lists.open-mpi.org
> > https://rfd.newmexicoconsortium.org/mailman/listinfo/users
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
>
_______________________________________________
users mailing list
users@lists.open-mpi.org
https://rfd.newmexicoconsortium.org/mailman/listinfo/users

Reply via email to