Let's say you're right about the 200 rows being too few. From which row
count I can see the difference reflected in the results as it should (Solr
faster)?

Le dim. 29 janv. 2023 à 00:34, Jan Høydahl <jan....@cominvent.com> a écrit :

> For 200 values you need neither spark nor Solr. A plain Java in mem filter
> is much simpler 😉
>
> Sorry, you cannot benchmark like this. You have to select a real use case
> and then select technology base on the requirements at hand. And to
> benchmark you must use a realistic data set.
>
> Jan Høydahl
>
> > 28. jan. 2023 kl. 23:11 skrev marc nicole <mk1853...@gmail.com>:
> >
> > Hello guys,
> >
> > I have been playing with Solr lately, and I tested it over a csv file of
> > about 200 rows (that I indexed in Solr). I also read the file in Spark
> and
> > perform filtering over an attribute value and compute time of processing
> > when the dataset is loaded from File System vs. Solr.
> >
> > I find the time of execution longer when the dataset is loaded from Solr.
> > Any explanation?
> > Maybe the dataset is small to reflect improved performance for Solr?
> > Thanks for clarifying.
>

Reply via email to