> Have you tried setting the termInfosIndexDivisor when opening the
> IndexReader? EG a setting of 2 would load every 256th term (instead
> of every 128th term) into RAM, halving RAM usage, with the downside
> being that looking up a term will generally take longer since it'll
> require more scann
Sorry, LUCENE-1458 is "continuing" under LUCENE-2111 (ie, flexible
indexing is not yet committed). I've just added a comment to
LUCENE-1458 to that effect.
Lucene, even with flexible indexing, loads the terms index entirely
into RAM (it's just that the terms index in flexible indexing has less
RA
> This (very large number of unique terms) is a problem for Lucene currently.
>
> There are some simple improvements we could make to the terms dict
> format to not require so much RAM per term in the terms index...
> LUCENE-1458 (flexible indexing) has these improvements, but
> unfortunately tied
how to run lucene on eclipse
Great! If I understand correctly it looks like RAM savings? Will
there be an improvement in lookup speed? (We're using binary
search here?).
Is there a precedence in database systems for what was mentioned
about placing the term dict, delDocs, and filters onto disk and
reading them from there (wit
Roughly, the current approach for the default terms dict codec in
LUCENE-1458 is:
* Create a separate class per-field (the String field in each Term
is redundant). This is a big change over Lucene today
* That class has String[] indexText and long[] indexPointer, each
length = th
> LUCENE-1458 (flexible indexing) has these improvements,
Mike, can you explain how it's different? I looked through the code once
but yeah, it's in with a lot of other changes.
On Wed, Jun 10, 2009 at 5:40 AM, Michael McCandless <
luc...@mikemccandless.com> wrote:
> This (very large number of
Message-
> From: Michael McCandless [mailto:luc...@mikemccandless.com]
> Sent: Wednesday, June 10, 2009 5:40 AM
> To: java-user@lucene.apache.org
> Subject: Re: Lucene memory usage
>
>
> This (very large number of unique terms) is a problem for Lucene currently.
>
> There are so
isheng
-Original Message-
From: Michael McCandless [mailto:luc...@mikemccandless.com]
Sent: Wednesday, June 10, 2009 5:40 AM
To: java-user@lucene.apache.org
Subject: Re: Lucene memory usage
This (very large number of unique terms) is a problem for Lucene currently.
There are some simple improv
This (very large number of unique terms) is a problem for Lucene currently.
There are some simple improvements we could make to the terms dict
format to not require so much RAM per term in the terms index...
LUCENE-1458 (flexible indexing) has these improvements, but
unfortunately tied in w/ lots
Hej hej,
i have a question regarding lucenes memory usage
when launching a query. When i execute my query
lucene eats up over 1gig of heap-memory even
when my result-set is only a single hit. I
found out that this is due to the "ensureIndexIsRead()"
method-call in the "TermInfosReader" class, wh
On 5/31/06, Heng Mei <[EMAIL PROTECTED]> wrote:
Does Lucene do any caching of Document fields during a search?
No... but it wouldn't be too difficult to make your own cache, or use
Solr which does have a Document cache (among other types of caches).
-Yonik
http://incubator.apache.org/solr Solr
Hi experts,
Does Lucene do any caching of Document fields during a search?
If I perform a search and retrieve some fields in the Document hits,
then I repeat the same search, are those fields cached in memory? It
doesn't seem to be -- I'm performing several thousand unique search
and retrievely
13 matches
Mail list logo