Note that large stored fields do not use up any RAM in IndexWriter's
RAM buffer because these stored fields are immediately written to the
directory and not stored in RAM for very long.
Aditi, I'd love to see the full stack trace of the OOME that was
originally hit if you still have it...
Mike
Ganesh wrote:
Single document of 16 MB seems to be big. I think you are trying to
store the entire document content. If it is so drop the stored field
and store its reference information in the database, which could
help to retreive the content later.
Regards
Ganesh
----- Original Message ----- From: "Aditi Goyal" <[EMAIL PROTECTED]
>
To: <java-user@lucene.apache.org>
Sent: Friday, October 03, 2008 3:03 PM
Subject: Re: Document larger than setRAMBufferSizeMB()
Thanks Anshum.
Although it raises another query, committing the current buffer
will commit
the docs before and what will happen to the current doc which threw
an error
while adding a field to it, will that also get committed in the
half??
Thanks a lot
Aditi
On Fri, Oct 3, 2008 at 2:12 PM, Anshum <[EMAIL PROTECTED]> wrote:
Hi Aditi,
I guess increasing the buffer size would be a solution here, but
in case
you
wouldn't know the expected max doc size. I guess the best way to
handle
that
would be a regular try catch block in which you could commit the
current
buffer. At the least you could just continue the loop after doing
whatever
you wish to do using an exception handling block.
--
Anshum Gupta
Naukri Labs!
http://ai-cafe.blogspot.com
The facts expressed here belong to everybody, the opinions to me.
The
distinction is yours to draw............
On Fri, Oct 3, 2008 at 1:56 PM, Aditi Goyal <[EMAIL PROTECTED]>
wrote:
> Hi Everyone,
>
> I have an index which I am opening at one time only. I keep
adding the
> documents to it until I reach a limit of 500.
> After this, I close the index and open it again. (This is done
in order
to
> save time taken by opening and closing the index)
> Also, I have set setRAMBufferSizeMB to 16MB.
>
> If the document size itself is greater than 16MB what will
happen in > this
> case??
> It is throwing
> java.lang.OutOfMemoryError: Java heap space
> Now, my query is,
> Can we change something in the way we parse/index to make it
more > memory
> friendly so that it doesnt throw this exception.
> And, Can it be caught and overcome gracefully?
>
>
> Thanks a lot
> Aditi
>
Send instant messages to your online friends http://in.messenger.yahoo.com
---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]
---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]