Hi Team,


I am trying to setup solr(3 nodes) and zk(3 nodes) in GCP environment. All
the setup looks okay but except while indexing the data which is taking a
lot of time.



Below is the machine configuration and I already configured and loaded the
data into the local disk of one of the solr nodes before starting the
indexing  of the data into Solr cluster. The home directory of the solr is
in the pd-ssd persistent disk. And the index is also stored in the
Persistent disk.



*Solr nodes [3 nodes]:*

1.       c3d-highmem-16, local disk – 100GB [PD-SSD, Ephemeral], Persistent
disk [PD-SSD, 2000 GB]

2.       Solr version – 9.6.0, Java version -11, LuceneMatchVersion 9.9,

1.       9 shards,

2.       2 replication factor,

3.       JVM memory – 12G

4.       Autocommit/maxtime =  300000

5.       Autosoftcommit/maxtime = 120000

6.       Each document of the collection may be ~ 2kb size in json format



*ZK nodes [3 nodes]:*

1.       c3d-highcpu -4, local disk – 100GB [PD-SSD, Ephemeral], Persistent
disk [PD-SSD, 2000 GB]

2.       Zk version – 3.9.2, Java version -11


While ingesting the data into solr index noticed the CPU utilization &
memory is normal



Already tried below options from disk side:



1.  
https://cloud.google.com/compute/docs/disks/optimizing-pd-performance#no-ext3

2.  
https://cloud.google.com/compute/docs/disks/optimizing-pd-performance#formatting_parameters



Please let me know the way to increase the performance of solr indexing and
reduce the indexing hours.



Regards

Sathish P

Sathish Ponnusamy

Chennai
m: + 91 9962331981
e: sathishrp...@gmail.com

GitHub - github.com/sathishrpt24
<http://github%20-%20github.com/sathishrpt24> [image: Youtbue icon]
<https://www.youtube.com/channel/UCyPZ_DNpq2zOsTW9Txod4tg>

Reply via email to