Re: limitation on token-length for KeywordAnalyzer?

2014-01-28 Thread Andreas Brandl
Hi, - Original Message - > On Mon, Jan 27, 2014 at 3:48 AM, Andreas Brandl > wrote: > > Is there some limitation on the length of fields? How do I get > > around this? > [cut] > > My overall goal is to index (arbitrary sized) text files and run a > > re

limitation on token-length for KeywordAnalyzer?

2014-01-26 Thread Andreas Brandl
Hi, I'm trying to get a RegexpQuery to work properly with Lucene 4.6. However, it fails consistently when the document gets bigger than 32kb (this document will never show up in search results, even if it is a match). Is there some limitation on the length of fields? How do I get around this?

Re: expensive post filtering of a query's result

2013-11-26 Thread Andreas Brandl
Uwe, > Lucene Filters are always executed before on the full index. This is > done inside getDocIdSet(), which is similar to scorer() in Querys. > Most filters return a bitset in this method, so they calculate the > whole bitset on the full index - this is what your filter is doing. > The strategy

expensive post filtering of a query's result

2013-11-25 Thread Andreas Brandl
Hi, I have a Query that is fast and cheap to answer compared to a Filter implementation that is quite expensive (* for a background see below). I was under the impression that when combining Query and Filter, lucene is able to calculate matches based on the query and *for these matches* applies