Hi,

We are experiencing heavy slowness on updates for SolrCloud implementation.
We are using it as 1 shard with 2 clusters. Also we have 3 zookeeper nodes.
The *solr* version is *8.7.0 *and *ZK* version is *3.6.2*.

Everyday we have some heavy updates (100,000 to 500,000 updates
processed parallel)
which includes delete, add and update.
We have a total 2.2 M records indexed.

Every time when this updates/deletes happens we are seeing a lot of
'*Reordered
DBQs detected*’ messages and finally the processing becomes very very slow
and an update/delete request is increasing from 100ms to 30 minutes to
complete.
At the same time we started to see error messages with "*Task queue
processing has stalled for 115231 ms with 100 remaining elements to process*
”  and "*Idle timeout expired: 120000/120000 ms*” , “*cancel_stream_error*
” etc.
Sometimes one node goes to a recovery state and recovers after some time.


2021-04-21 18:21:08.219 INFO  (qtp938463537-5550) [c:datacore s:shard1
r:core_node4 x:datacore_shard1_replica_n2]
o.a.s.u.DirectUpdateHandler2 *Reordered
DBQs detected*.
Update=add{_version_=1697674936592629760,id=S-5942167-P-108089342-F-800102562-E-180866483}
DBQs=[DBQ{version=1697674943496454144,q=store_element_id:395699},
DBQ{version=1697674943408373760,q=store_element_id:395698},
DBQ{version=1697674943311904768,q=store_element_id:395678},
DBQ{version=1697674943221727232,q=store_element_id:395649},
DBQ{version=1697674943143084032,q=store_element_id:395642},
DBQ{version=1697674943049760768,q=store_element_id:395612},
DBQ{version=1697674942964826112,q=store_element_id:395602},
DBQ{version=1697674942871502848,q=store_element_id:395587},
DBQ{version=1697674942790762496,q=store_element_id:395582},
DBQ{version=1697674942711070720,q=store_element_id:395578},
DBQ{version=1697674942622990336,q=store_element_id:199511},
DBQ{version=1697674942541201408,q=store_element_id:199508},
DBQ{version=1697674942452072448,q=store_element_id:397242},
DBQ{version=1697674942356652032,q=store_element_id:397194},
DBQ{version=1697674942268571648,q=store_element_id:397166},
DBQ{version=1697674942178394112,q=store_element_id:397164},
DBQ{version=1697674942014816256,q=store_element_id:397149},
DBQ{version=1697674941901570048,q=store_element_id:395758},
DBQ{version=1697674941790420992,q=store_element_id:395725},
DBQ{version=1697674941723312128,q=store_element_id:395630},

2021-04-21 18:30:23.636 INFO
 (recoveryExecutor-11-thread-5-processing-n:solr-1.solrcluster:8983_solr
x:datacore_shard1_replica_n2 c:datacore s:shard1 r:core_node4) [c:datacore
s:shard1 r:core_node4 x:datacore_shard1_replica_n2] o.a.s.c.*RecoveryStrategy
PeerSync Recovery was not successful - trying replication.*
2021-04-21 18:30:23.636 INFO
 (recoveryExecutor-11-thread-5-processing-n:solr-1.solrcluster:8983_solr
x:datacore_shard1_replica_n2 c:datacore s:shard1 r:core_node4) [c:datacore
s:shard1 r:core_node4 x:datacore_shard1_replica_n2] o.a.s.c.*RecoveryStrategy
Starting Replication Recovery.*


Below given the  autoCommit and autoSoftCommit values used.

<*autoCommit*>
<maxTime>${solr.autoCommit.maxTime:*90000*}</maxTime>
<openSearcher>*false*</openSearcher>
</autoCommit>
<*autoSoftCommit*>
<maxTime>${solr.autoSoftCommit.maxTime:*15000*}</maxTime>
</*autoSoftCommit*>

Here is the GC logs for reference:

[2021-04-21T18:28:32.296+0000][408073.843s] GC(541) User=0.07s Sys=0.01s
Real=0.04s
[2021-04-21T18:28:32.296+0000][408073.843s] GC(541) Pause Young (Normal)
(G1 Evacuation Pause) 7849M->1725M(10240M) 39.247ms
[2021-04-21T18:28:32.296+0000][408073.843s] GC(541) Metaspace:
85486K->85486K(1126400K)
[2021-04-21T18:28:32.296+0000][408073.843s] GC(541) Humongous regions:
15->15
[2021-04-21T18:28:32.296+0000][408073.843s] GC(541) Old regions: 412->412
[2021-04-21T18:28:32.296+0000][408073.843s] GC(541) Survivor regions:
5->5(192)
[2021-04-21T18:28:32.296+0000][408073.843s] GC(541) Eden regions:
1531->0(1531)
[2021-04-21T18:28:32.296+0000][408073.843s] GC(541)   Other: 0.5ms
[2021-04-21T18:28:32.296+0000][408073.843s] GC(541)   Post Evacuate
Collection Set: 6.1ms
[2021-04-21T18:28:32.296+0000][408073.843s] GC(541)   Evacuate Collection
Set: 32.5ms
[2021-04-21T18:28:32.296+0000][408073.843s] GC(541)   Pre Evacuate
Collection Set: 0.1ms
[2021-04-21T18:28:32.257+0000][408073.804s] GC(541) Using 2 workers of 2
for evacuation
[2021-04-21T18:28:32.257+0000][408073.804s] GC(541) Pause Young (Normal)
(G1 Evacuation Pause)


Would really appreciate any help on this.

Thanks,
Rekha

Reply via email to