On 2022-10-07 9:34 AM, Dominique Bejean wrote:
Hi Dima,
About ZFS, I read this *"In an effort to maximize read/write performance,
ZFS uses all available space in RAM to create a huge cache"*. How does it
work with MMapDirectory ? No conflict ?
ZFS uses half of the available RAM for L1 cache (c
The ideal jvm size will be influenced by the latency sensitivity of the
application. Large VM's are mostly good if you need to hold large data
objects in memory, otherwise they fill up with large numbers of small
objects and that leads to long GC pauses (GC time relates to the number,
not the size
Hi Dima,
About ZFS, I read this *"In an effort to maximize read/write performance,
ZFS uses all available space in RAM to create a huge cache"*. How does it
work with MMapDirectory ? No conflict ?
Regards
Dominique
Le jeu. 6 oct. 2022 à 17:44, dmitri maziuk a
écrit :
> On 2022-10-06 2:57 AM
On 10/6/22 15:54, Dominique Bejean wrote:
We are starting to investigate performance issues with a new customer.
There are several bad practices (commit, sharding, replicas count and
types, heap size, ...), that can explain these issues and we will work on
it in the next few days. I agree we need
Hi Dave,
Are you suggesting to use historical Solr master/slave architecture ?
In Sorlcloud / SolrJ architecture this can be achieved by creating only
TLOG replicas then FORCELEADER located on a specific server (then indexing
server) and search only on TLOG replicas with the parameter
"shards.pre
A reason for sharding on a single server is the 2.1b max docs per core
limitation.
On Thu, Oct 6, 2022, 12:51 PM Dave wrote:
> I know these machines. Sharding is kind of useless. Set the ssd tb drives
> up in fastest raid read available, 31 xms xmx, one solr instance. Buy back
> up ssd drives w
Run a GC analyzer on that JVM. I cannot imagine that they need 80 GB of heap.
I’ve never run with more than 16 GB, even for a collection with 70 million
documents.
Look at the amount of heap used after full collections. Add a safety factor to
that, then use that heap size.
wunder
Walter Underw
On 2022-10-06 5:52 PM, Dominique Bejean wrote:
Thank you Dima,
Updates are highly multi-threaded batch processes at any time.
We won't have all index in RAM cache
Disks are SSD
You'd have to benchmark, pref. with you real jobs, on RAID-10 (as per my
previous e-mail) vs JBOD. I suspect you wo
You should never index directly into your query servers by the way. Index to
the indexing server and replicate out to you query servers and tune each as
needed
> On Oct 6, 2022, at 6:52 PM, Dominique Bejean
> wrote:
>
> Thank you Dima,
>
> Updates are highly multi-threaded batch processes a
Thank you Dima,
Updates are highly multi-threaded batch processes at any time.
We won't have all index in RAM cache
Disks are SSD
Dominique
Le ven. 7 oct. 2022 à 00:28, dmitri maziuk a
écrit :
> On 2022-10-06 4:54 PM, Dominique Bejean wrote:
>
> > Storage configuration is the second point tha
On 2022-10-06 4:54 PM, Dominique Bejean wrote:
Storage configuration is the second point that I would like to investigate
in order to better share disk resources.
Instead have one single RAID 6 volume, isn't it better to have one distinct
not RAID volume per Solr node (if multiple Solr nodes are
Hi,
Thank you all for your responses. I will try to answer your questions in
one single message.
We are starting to investigate performance issues with a new customer.
There are several bad practices (commit, sharding, replicas count and
types, heap size, ...), that can explain these issues and w
We have kept a 72 CPU machine busy with a single Solr process, so I doubt that
multiple processes are needed.
The big question is the size of the index. If it is too big to fit in RAM (OS
file buffers), then the system is IO bound and CPU doesn’t really matter.
Everything will depend on the spe
I know these machines. Sharding is kind of useless. Set the ssd tb drives up in
fastest raid read available, 31 xms xmx, one solr instance. Buy back up ssd
drives when you burn one out and it fails over to the master server. Multiple
solr instances on one machine makes little sense unless they h
On 10/6/22 01:57, Dominique Bejean wrote:
One of our customer have huge servers
- Bar-metal
- 64 CPU
- 512 Gb RAM
- 6x2Tb disk in RAID 6 (so 2Tb disk space available)
I think the best way to optimize resources usage of these servers is to
install several Solr instances.
That
On 2022-10-06 2:57 AM, Dominique Bejean wrote:
Do not configure disks in RAID 6 but, leave 6 standard volumes (more space
disk, more I/O available)
If they're running linux: throw out the raid controller, replace with
ZFS on 2 SSDs and 4 spinning rust drives. You're not going to have more
i/
What would the iops look like?
Deepak
"The greatness of a nation can be judged by the way its animals are treated
- Mahatma Gandhi"
+91 73500 12833
deic...@gmail.com
Facebook: https://www.facebook.com/deicool
LinkedIn: www.linkedin.com/in/deicool
"Plant a Tree, Go Green"
Make In India : http:
It depends... on your data, on your usage, etc. The best answers are
obtained by testing various configurations, if possible by replaying
captured query load from production. There is (for all java programs) an
advantage to staying under 32 GB RAM, but without an idea of the number of
machines you
Why do you want to split it up at all?
On Thu, Oct 6, 2022 at 3:58 AM Dominique Bejean
wrote:
>
> Hi,
>
> One of our customer have huge servers
>
>- Bar-metal
>- 64 CPU
>- 512 Gb RAM
>- 6x2Tb disk in RAID 6 (so 2Tb disk space available)
>
>
> I think the best way to optimize resou
Hi,
One of our customer have huge servers
- Bar-metal
- 64 CPU
- 512 Gb RAM
- 6x2Tb disk in RAID 6 (so 2Tb disk space available)
I think the best way to optimize resources usage of these servers is to
install several Solr instances.
I imagine 2 scenarios to be tested according to d
20 matches
Mail list logo