I used the following settings for speeding up indexing on a similarly sized
db table
If you have enough ram it might help you.

IndexWriter writer = *new* IndexWriter(fdDir,*new* StandardAnalyzer(), *true
*);

writer.setMergeFactor(100);

writer.setMaxMergeDocs(999999);

writer.setMaxBufferedDocs(1500);




On 1/11/07, Rangarirayi Muvavarirwa <[EMAIL PROTECTED]> wrote:

One option: https://cool-apps-distributedindex.dev.java.net/

caveat: you would have to setup an account (you get 10CPUhr & 10GB account
upon signup)

On 1/11/07, Alice <[EMAIL PROTECTED]> wrote:
>
> Unfortunately I can't use multiple machines.
>
> And I cannot start lots of threads because the server crashes.
>
> -----Original Message-----
> From: Russ [mailto:[EMAIL PROTECTED]
> Sent: quinta-feira, 11 de janeiro de 2007 14:33
> To: java-user@lucene.apache.org
> Subject: Re: Huge Index
>
> Can you use multiple threads/machines to index the data into separate
> indexes, and then combine them?
>
> Russ
> Sent wirelessly via BlackBerry from T-Mobile.
>
> -----Original Message-----
> From: "Alice" <[EMAIL PROTECTED]>
> Date: Thu, 11 Jan 2007 13:47:36
> To:<java-user@lucene.apache.org>
> Subject: Huge Index
>
> Hello!
>
>
>
> I have to index 37million documents retrieved from the database.
>
>
>
> I was trying to do by loading intervals of 10000 records but it is too
> slow.
>
>
>
> Anybody could sugest a better way to get all the data indexed in a
> reasonable time?
>
>
>
> Thanks
>
>
>
>
>
>
>
>
>
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: [EMAIL PROTECTED]
> For additional commands, e-mail: [EMAIL PROTECTED]
>
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: [EMAIL PROTECTED]
> For additional commands, e-mail: [EMAIL PROTECTED]
>
>


Reply via email to