> I'd like to use lucene to search text
> documents for the existence of a large
> list of search terms. I have a file that contains thousands
> of entries, one
> word per line. I was thinking about to writing a
> specialized analyzer
> that tokenizes the document by looking up each token
> in th
I'd like to use lucene to search text documents for the existence of a large
list of search terms. I have a file that contains thousands of entries, one
word per line. I was thinking about to writing a specialized analyzer
that tokenizes the document by looking up each token in the source document
Hi,
I'm building a BooleanQuery that may contain a NumericRangeQuery. The
NRQ may be one of several sub-queries in the parent BooleanQuery. I
wasn't able to make the NRQ function properly by extending the
QueryParser and overriding the getRangeQuery method. So I'm building
the entire Boolea
Are you using the same analyzer for searching, in your unshown
getHitCount() method?
There is lots of good advice in the FAQ under "Why am I getting no
hits / incorrect hits?". And/or write the index to disk and use Luke
to check that the correct content is being indexed.
--
Ian.
On Tue, Apr
I am encountering a strange issue. I have a CustomStopAnalyzer. If I
do this (supporting code taken from AnalyzerUtils in LIA3 source code
Mike uploaded):
Analyzer customStopAnalyzer = new CustomStopAnalyzer();
AnalyzerUtils.displayTokensWithFullDetails(customStopAnalyzer,
"mail77")
<<>>
You've hit one of my personal hot buttons . Never, never, never
until
1> you're reasonably sure you actually *have* a performance issue.
2> your performance issues can't be fixed more cheaply. Memory is
wy cheap compared to all the time you'll spend programming
your caching solut
Hi Ariel,
Please ask on solr-u...@lucene.apache.org. Also, please provide more info
about your index, etc. when sending that mail.
-Grant
On Apr 19, 2010, at 11:54 AM, Ariel wrote:
> Hi everybody:
>
> I have a big problem with solr in a server with the memory size it is using,
> I would want
You have to open an IndexReader in order to walk the postings of newly
indexed docs.
EG after indexing the doc, ask for a new near-real-time reader
(IndexWriter.getReader()), then use the postings API
(TermDocs/TermPositions in 3.0.x, but in trunk, with the flex API,
DocsEnum/DocsAndPositionsEnum)
Generally best to let the OS do it. Having plenty of spare RAM is good.
--
Ian.
On Tue, Apr 20, 2010 at 6:16 AM, Chris B wrote:
> Mike cheers for the reply.
>
> Is it worth setting up your own caching or letting the OS do it? I've setup
> a caching system, but if the OS is doing it it seems p
Hi all,
I have a question about usage of lucene, I want to figure out how I
can get one or all posting lists, after adding a document to the
index, but without materializing it in files. So after I add a
document (let's say "to be or not to be"), I am asking if I can get
something like this:
to -
10 matches
Mail list logo