hey chris, i will check and let you know just to make sure, basically i see the OS allocating memory (up to about 4GB) while loading the indexes to memory and then crashing on the TermInfosReader class. what i noticed was that the crash occured when lucene tried to create a Term array with the following code new Term[indexSize] i assume, since this is an array java was trying to allocate consecutive blocks in memory and this is hard to find , even in a 16 GB RAM machine, especially since (if im not mistaken) indexSize here is the termEnum size (which in my case is rather large) i will get back to you about the one liner, if you have any other thoughts id be extremely happy to hear them as this problem is a Major road block thanks a million
Chris Hostetter <[EMAIL PROTECTED]> wrote: : i am recieving the following stack trace: : : JVMDUMP013I Processed Dump Event "uncaught", detail "java/lang/OutOfMemoryError". : Exception in thread "main" java.lang.OutOfMemoryError : at org.apache.lucene.index.TermInfosReader.readIndex(TermInfosReader.java:82) is it possible that parts of your application are eating up all of the heap in your JVM before this exception is encountered? Possibly by opening a the index many times without closing it? More specifically, if you write a 4 line app that does nothing by open your index and then close it again, do you get an OOM? ... public class Main { public static void main(String[] args) throws Exception { Searcher s = new IndexSearcher("/your/index/path"); s.close(); } } -Hoss --------------------------------------------------------------------- To unsubscribe, e-mail: [EMAIL PROTECTED] For additional commands, e-mail: [EMAIL PROTECTED] --------------------------------- Yahoo! Mail Bring photos to life! New PhotoMail makes sharing a breeze.