Hello! I am using Lucene 1.4.3
I'm building a Lucene index, that will have about 25 million documents
when it is done.
I'm adding 250,000 at a time.
Currently there is about 1.2Million in there, and I ran into a problem.
After I had added a batch of 250,000 I go a 'java.lang.outOfMemory'
threw by writer.optimize(); (a standard IndexWriter)
The exception caused my program to quit out, and it didn't call
'writer.close();'
First, with it dying in the middle of an .optimize() is there any chance
my index is corrupted?
Second, I know I can remove the /tmp/lucene*.lock file to remove the
lock in order to add more, but is it safe to do that?
I've since figured out that I can pass -Xmx to the 'java' program in
order to increase the maximum amount of RAM.
It was using the default of 64M, I plan on increasing that to 175M to
start with.
That should solve the memory problems (I can allocate more if necessary
down the line).
Lastly, when I go back, open it again, and add another 250,000 and then
call optimize again, will a failed previous optimize hurt the index at all?
---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]