* Go multi threaded for each core as Shawn says. Try e.g. 2, 3 and 4 threads
* Experiment with different batch sizes, e.g. try 500 and 2000 - depends on 
your docs what is optimal
* Do NOT commit after each batch of 1000 docs. Instead, commit as seldom as 
your requirements allows, e.g. try commitWithin=60000 to commit every minute

Tip: Try to push Solr metrics to DataDog or some other service, where you can 
see a dashboard with stats on requests/sec, RAM, CPU, threads, GC etc which may 
answer your last question.

Jan

> 8. jun. 2022 kl. 14:06 skrev Shawn Heisey <apa...@elyograg.org>:
> 
> On 6/8/2022 3:35 AM, Marius Grigaitis wrote:
>> * 9 different cores. Each weighs around ~100 MB on disk and has
>> approximately 90k documents inside each.
>> * Updating is performed using update method in batches of 1000, around 9
>> processes in parallel (split by core)
> 
> This means that indexing within each Solr core is single-threaded.  The way 
> to increase indexing speed is to index in parallel with multiple threads or 
> processes per index.  If you can increase the CPU power available on the Solr 
> server when you increase the number of processes/threads sending data to 
> Solr, that might help.
> 
> Thanks,
> Shawn
> 

Reply via email to