For 200 values you need neither spark nor Solr. A plain Java in mem filter is 
much simpler 😉

Sorry, you cannot benchmark like this. You have to select a real use case and 
then select technology base on the requirements at hand. And to benchmark you 
must use a realistic data set.

Jan Høydahl

> 28. jan. 2023 kl. 23:11 skrev marc nicole <mk1853...@gmail.com>:
> 
> Hello guys,
> 
> I have been playing with Solr lately, and I tested it over a csv file of
> about 200 rows (that I indexed in Solr). I also read the file in Spark and
> perform filtering over an attribute value and compute time of processing
> when the dataset is loaded from File System vs. Solr.
> 
> I find the time of execution longer when the dataset is loaded from Solr.
> Any explanation?
> Maybe the dataset is small to reflect improved performance for Solr?
> Thanks for clarifying.

Reply via email to