I'm trying to come to terms with http://lucene.apache.org/java/docs/api/org/apache/lucene/index/IndexWriter.h tml#setMaxFieldLength(int) too. I've been attempting to index large text files as single Lucene documents, passing them as java.io.Reader to cope with RAM. I was assuming (like - I suspect - manu mohedano) that an unstored field could be of any length and that maxFieldLength was only applicable to stored fields. Do we in fact need to break the document into manageable parts?
-----Original Message----- From: Pasha Bizhan [mailto:[EMAIL PROTECTED] Sent: 09 June 2006 21:35 To: java-user@lucene.apache.org Subject: RE: Problems indexing large documents Hi, > From: manu mohedano [mailto:[EMAIL PROTECTED] > Hi All! I have a trouble... When I index text documents in > english, there is no problem, buy when I index Spanish text > documents (And they're big), a lot of information from the > document don't become indexed (I suppose it is due to the > Analyzer, but if the documents is less tahn 400kb it works > perfectly). Howewer I want to Index ALL the strings in the > document with no StopWords. Is this possible?? Read javadoc about DEFAULT_MAX_FIELD_LENGTH at http://lucene.apache.org/java/docs/api/org/apache/lucene/index/IndexWriter.h tml#setMaxFieldLength(int) Pasha Bizhan --------------------------------------------------------------------- To unsubscribe, e-mail: [EMAIL PROTECTED] For additional commands, e-mail: [EMAIL PROTECTED] --------------------------------------------------------------------- To unsubscribe, e-mail: [EMAIL PROTECTED] For additional commands, e-mail: [EMAIL PROTECTED]