Hi Nikhilesh,

Try hard-committing more often. This way you'll have smaller tlog files and
there will be less data to recover. My suggestion is to add a maxSize
constraint to autoCommit. 100MB is a good rule of thumb, makes sure you
don't replay more than 100MB worth of data (even if you have an indexing
spike).

Best regards,
Radu
--
Elasticsearch/OpenSearch & Solr Consulting, Production Support & Training
Sematext Cloud - Full Stack Observability
http://sematext.com/


On Wed, Jun 29, 2022 at 9:10 AM Nikhilesh Jannu <nikhil...@predictspring.com>
wrote:

> Dear Users,
>
> We are using the SOLR TRA collection for capturing the logs.  We  are
> writing the logs to SOLR using the Rest API in a batch of 100 and also we
> are using SOFT commit interval of 15000 and Hard commit interval of 60000.
>
> Solr Version : 8.11.1.
>
> When we restart the SOLR node in the cloud the current day's collection
> goes in recovery mode and we see the following logs. It takes a long time
> for the recovery process to complete. Not sure how to avoid it. Any
> suggestions ?
>
> Sample of the logs below.
>
> 2022-06-29 06:03:27.500 INFO
>  (recoveryExecutor-67-thread-1-processing-n:10.0.42.157:8983_solr
> x:logs__TRA__2022-06-29_shard1_replica_n1 c:logs__TRA__2022-06-29 s:shard1
> r:core_node2) [c:logs__TRA__2022-06-29 s:shard1 r:core_node2
> x:logs__TRA__2022-06-29_shard1_replica_n1] o.a.s.u.UpdateLog log replay
> status
>
> tlog{file=/var/solr/data/logs__TRA__2022-06-29_shard1_replica_n1/data/tlog/tlog.0000000000000000243
> refcount=3} active=false starting pos=0 current pos=1119002110 current
> size=3287152529 % read=34.0
>
> Regards,
> Nikhilesh Jannu
>

Reply via email to