Hi, I have a sharded collection distributed over several solr nodes. Each solr node hosts one shard and one replica of another shard. shards are huge (100 millions documents). Queries are using several filterQuery. filterCache for this number of documents can use high amount of heap memory.
Is it a good idea to split shards by 2 or 4 in order to have shards with 50 or 25 millions documents ? With a split by 4, a Solr node will host 8 replicas instead of 2, but with smaller filterCache for each replica. I don't expect to have better search performances, but I expect to have faster warming and mainly less impacted heap memory by open searcher during sofcommit. For instance, instead of having one large filterCache warming up once each minute, 4 smaller filterCaches will warm up not at the same time (hopefully). So shard splitting or not on the same Solr nodes ? Thank you Dominique
