Ron,

please post actual figures:

1. Cpu, mem, disk, network utilisation
2. Response times
3. Load
4. Hardware config of server
5. Software config of server


On Thu, 25 May 2023, 18:51 Joel Bernstein, <joels...@gmail.com> wrote:

> One thing to understand about the topLevelDv approach is you'll need to
> warm both sides of the join. You can do this by adding a  static warming
> query that facets on 'group_id_mv'  and 'group_member_id' in both
> collections.
>
>
> Joel Bernstein
> http://joelsolr.blogspot.com/
>
>
> On Thu, May 25, 2023 at 9:17 AM Joel Bernstein <joels...@gmail.com> wrote:
>
> > If you are using a recent version of Solr try adding the parameter
> >
> > method=topLevelDV
> >
> > Let us know how this effects performance in your use case.
> >
> > What matters most here is the number of documents the from side of join
> > matches.
> >
> >
> >
> >
> >
> > Joel Bernstein
> > http://joelsolr.blogspot.com/
> >
> >
> > On Thu, May 25, 2023 at 8:52 AM Ron Haines <mickr...@gmail.com> wrote:
> >
> >> I've been using the 'join' query parser to 'filter out' related
> documents
> >> that should not be part of the result set.  Functionally, it is working
> >> fine.  However, when we throw a 'real' level of customer traffic at it,
> it
> >> pretty much brings Solr to its knees.  CPU increases ALOT.  Close to 3X,
> >> when I enable this feature in our system.  Solr response times shoot up,
> >> and thread counts shoot up.  Before I 'give up' on the join query
> parser,
> >> I
> >> thought I'd seek some advice here.
> >>
> >> So, when this feature is enabled, this negative &fq gets added:
> >> -{!join fromIndex=primary_rollup from=group_id_mv to=group_member_id
> >> score=none}${q}
> >>
> >> The 'local' collection size is about 27 million docs, but the number of
> >> docs that actually contain a 'group_member_id' is only about 125k.  And,
> >> in
> >> the 'fromIndex' collection, there are only 80k documents in that
> >> collection, and they all have the 'group_id_mv' field.  The 'fromIndex'
> >> collection is a single shard, with a replica on each shard of the local
> >> collection.  The local collection only has about 300k docs per shard, at
> >> 96
> >> shards.
> >>
> >> I guess I'm just trying to understand why this appears to be causing
> such
> >> problems for Solr, as the amount of work (the # of documents involved)
> >> seems relatively small.
> >>
> >> I hope I'm missing something...
> >> Thanks for any input.
> >>
> >
>

Reply via email to