I believe so. Java's synchronized' keyword is what makes it possible to
use multiple threads with a single IndexWriter instance, and I believe
that means threads are processed in a FIFO manner.
Otis
--- Chris Fraschetti <[EMAIL PROTECTED]> wrote:
> If i'm using multiple threads to add documents
If i'm using multiple threads to add documents to the index, can it be
assumed that they will be added to the index in the order they are
presented to the indexwriter? and thus keeping my local doc id count
would hold true?
-Chris Fraschetti
On 7/29/05, Erik Hatcher <[EMAIL PROTECTED]> wrote:
>
Does anyone know why lucene would throw the following stacktrace
when a search runs while the system is in the process of merging
documents? Based on the book, this should be a legal operation.
I've seen this stacktrace mentioned several times in the archives but
I never seen a solut
May I suggest:
Don't call optimize. You don't need it. Here is my approach:
Keep each one of your 250,000 document indexes separate - so run your
batch, build the index, and then just close it. Don't try to optimize
it. For each 250,000 document batch, just put it into a different folder.
Hi all,
I have a Punctuation Filter that filters out certain criteria
(ex. - and /). This filter is called whenever I search for normal
items.. Search 56-TXT
The filter does not seem to be called when I search for 56-TXT* even
though I use my extended analyzer which has the filter in it s
I am going to play it safe.
I'm going to wipe the index files, and start over (only put about 2 days
of processing time into it so far).
This time with a maximum memory of over 512MB
Yonik Seeley wrote:
If all segments were flushed to the disk (no adds since the last time
the index writer was
If all segments were flushed to the disk (no adds since the last time
the index writer was opened), then it seems like the index should be
fine.
The big question I have is what happens when there are in-memory
segments in the case of an OOM exception during an optimize? Is data
loss possible?
-Y
Hi,
If you want highlighting like this:
http://simpy.com/simpy/User.do?username=otis&q=lucene
or like this:
http://lucenebook.com/search?query=highlighter
Then you'll want to pull Lucene directly from SVN, and look in the
contrib/ directory, where you'll find classes for highlighter. You can
I think you can achieve what you want using Span queries with slop
of 1 for ? or MAX_INT for * -- but you'll probably need to make some
nested queries in order to have 0 slop between "of" and "america"
: Date: Mon, 1 Aug 2005 09:15:02 -0700
: From: Rajesh Munavalli <[EMAIL PROTECTED]>
: Reply-To:
Hi Chris,
The method you suggested is definitely a good solution. However there
is one more reason I would like to do n-gram generation at indexing time. The
examples below are a text search equivalent of what I am trying to do for a
different kind of data. Anyway the example should con
Hy all,
which is the way to get a snippet of a query?
Does Lucene offer some explicit methods?
If is not so, it's possible to know the position of the matched
keyword in the query into the field of the doc, so that I can take
some words before and after the keyword to create the snippet?
Or there
Your index should be fine. You could use "luke" if you want to remove any
dangling
files not in use. Just run optimize again to fix it all up... You might want
to
allocate more memory that 175 though. Depending on your document sizes and how
efficiently and fast you want lucene to work to ind
Oh!!! So *that's* what the documentation is for! ;)
Thanks for pointing that out. Sometimes you get so frustrated with a
problem you fail to notice the obvious solution.
Doug
-Original Message-
From: [EMAIL PROTECTED]
[mailto:[EMAIL PROTECTED] On Behalf Of Chris Hostetter
Sent: Monday,
13 matches
Mail list logo