A reason for sharding on a single server is the 2.1b  max docs per core
limitation.

On Thu, Oct 6, 2022, 12:51 PM Dave <hastings.recurs...@gmail.com> wrote:

> I know these machines. Sharding is kind of useless. Set the ssd tb drives
> up in fastest raid read available, 31 xms xmx, one solr instance. Buy back
> up ssd drives when you burn one out and it fails over to the master server.
> Multiple solr instances on one machine makes little sense unless they have
> different purposes like a ml instance and a text highlighting instance but
> even then you get no performance improvement
>
>
> > On Oct 6, 2022, at 12:21 PM, Shawn Heisey <apa...@elyograg.org> wrote:
> >
> > On 10/6/22 01:57, Dominique Bejean wrote:
> >> One of our customer have huge servers
> >>
> >>    - Bar-metal
> >>    - 64 CPU
> >>    - 512 Gb RAM
> >>    - 6x2Tb disk in RAID 6 (so 2Tb disk space available)
> >>
> >>
> >> I think the best way to optimize resources usage of these servers is to
> >> install several Solr instances.
> >
> > That is not what I would do.
> >
> >> Do not configure disks in RAID 6 but, leave 6 standard volumes (more
> space
> >> disk, more I/O available)
> >> Install 3 or 6 solr instances each one using 1 ou 2 disk volumes
> >
> > RAID10 will get you the best performance.  Six 2TB drives in RAID10 has
> 6TB of total space.  The ONLY disadvantage that RAID10 has is that you pay
> for twice the usable storage.  Disks are relatively cheap, though hard to
> get in quantity these days.  I would recommend going with the largest
> stripe size your hardware can support.  1MB is typically where that maxes
> out.
> >
> > Any use of RAID5 or RAID6 has two major issues:  1) A serious
> performance problem that also affects reads if there are ANY writes
> happening.  2) If a disk fails, performance across the board is terrible.
> When the bad disk is replaced, performance is REALLY terrible as long as a
> rebuild is happening, and I have seen a RAID5/6 rebuild take 24 to 48 hours
> with 2TB disks on a busy array.  It would take even longer with larger
> disks.
> >
> >> What I am not sure is how MMapDirectory will work with several Solr
> >> instances. Will off heap memory correctly managed and shared between
> >> several Solr instances ?
> >
> > With symlinks or multiple mount points in the solr home, you can have a
> single instance handle indexes on multiple storage devices.  One instance
> has less overhead, particularly in memory, than multiple instances. Off
> heap memory for the disk cache should function as expected with multiple
> instances or one instances.
> >
> > Thanks,
> > Shawn
> >
>

Reply via email to