Hi,

That mergeFactor is too high.  I suggest going back to default (10).
maxBufferedDocs is an old and not very accurate setting (imagine what happens 
with the JVM heap if your indexer hits a SUPER LARGE document).  Use 
setRamBufferSizeMB instead.

Otis
--
Sematext is hiring -- http://sematext.com/about/jobs.html?mls
Lucene, Solr, Nutch, Katta, Hadoop, HBase, UIMA, NLP, NER, IR



----- Original Message ----
> From: Laxmilal Menariya <lmena...@chambal.com>
> To: java-user@lucene.apache.org
> Sent: Monday, August 10, 2009 3:23:17 AM
> Subject: Taking too much time in optimization
> 
> Hello everyone,
> 
> I have created a sample application & indexing files properties, have index
> appx 107K files.
> 
> I am getting OutofMemoryError after 100K while indexing,  got the cause from
> MaxBuffereddocs=100K, but after that I am calling optimize() method, this is
> taking too much time appx 12-HRS, and index size is more than 500GB, its too
> large.
> 
> I am using Lucene 2.4.0. Could some one please let me know what wrong with
> my configuration.
> 
> My Configuration is :
> 
>       lucWriter = new IndexWriter("C:\\Laxmilal", new KeywordAnalyzer(),
> true);
>       lucWriter.setMergeFactor((int) 1000);
>       lucWriter.setMaxMergeDocs((int) 2147483647);
>       lucWriter.setMaxBufferedDocs((int) 100000);
> 
> 
> -- 
> Thanks,
> Laxmilal Menariya
> 
> http://www.bucketexplorer.com/
> http://www.sdbexplorer.com/
> http://www.chambal.com/


---------------------------------------------------------------------
To unsubscribe, e-mail: java-user-unsubscr...@lucene.apache.org
For additional commands, e-mail: java-user-h...@lucene.apache.org

Reply via email to