Hi Team,
I am trying to setup solr(3 nodes) and zk(3 nodes) in GCP environment. All the setup looks okay but except while indexing the data which is taking a lot of time. Below is the machine configuration and I already configured and loaded the data into the local disk of one of the solr nodes before starting the indexing of the data into Solr cluster. The home directory of the solr is in the pd-ssd persistent disk. And the index is also stored in the Persistent disk. *Solr nodes [3 nodes]:* 1. c3d-highmem-16, local disk – 100GB [PD-SSD, Ephemeral], Persistent disk [PD-SSD, 2000 GB] 2. Solr version – 9.6.0, Java version -11, LuceneMatchVersion 9.9, 1. 9 shards, 2. 2 replication factor, 3. JVM memory – 12G 4. Autocommit/maxtime = 300000 5. Autosoftcommit/maxtime = 120000 6. Each document of the collection may be ~ 2kb size in json format *ZK nodes [3 nodes]:* 1. c3d-highcpu -4, local disk – 100GB [PD-SSD, Ephemeral], Persistent disk [PD-SSD, 2000 GB] 2. Zk version – 3.9.2, Java version -11 While ingesting the data into solr index noticed the CPU utilization & memory is normal Already tried below options from disk side: 1. https://cloud.google.com/compute/docs/disks/optimizing-pd-performance#no-ext3 2. https://cloud.google.com/compute/docs/disks/optimizing-pd-performance#formatting_parameters Please let me know the way to increase the performance of solr indexing and reduce the indexing hours. Regards Sathish P Sathish Ponnusamy Chennai m: + 91 9962331981 e: sathishrp...@gmail.com GitHub - github.com/sathishrpt24 <http://github%20-%20github.com/sathishrpt24> [image: Youtbue icon] <https://www.youtube.com/channel/UCyPZ_DNpq2zOsTW9Txod4tg>