Generally you go over the least frequent to most frequent required terms as
the pivot. As you get more and more complicated queries, the ordering of
the query results tends to dominate. This also means there are two types
of measurement. One for the running query and one for query + results in
o
If you have a 'small' application and you want search, you can embed a
lucent index. This is a lower impact dependency graph.
Solr is a service. I would suspect the only dependencies *required* are
client dependencies. And even those might boil down to a shell equivalent
curl + json parser.
As
Hi Doug,
I don't know the specific limits. But the document limits are going to be
around an int, probably signed. This comes out to mean about 2 billion
documents per lucene index. This is fairly embedded into the lucene code.
The way the collective we have solved this is through forms of shar