For 200 values you need neither spark nor Solr. A plain Java in mem filter is much simpler 😉
Sorry, you cannot benchmark like this. You have to select a real use case and then select technology base on the requirements at hand. And to benchmark you must use a realistic data set. Jan Høydahl > 28. jan. 2023 kl. 23:11 skrev marc nicole <mk1853...@gmail.com>: > > Hello guys, > > I have been playing with Solr lately, and I tested it over a csv file of > about 200 rows (that I indexed in Solr). I also read the file in Spark and > perform filtering over an attribute value and compute time of processing > when the dataset is loaded from File System vs. Solr. > > I find the time of execution longer when the dataset is loaded from Solr. > Any explanation? > Maybe the dataset is small to reflect improved performance for Solr? > Thanks for clarifying.