No other exceptions I could spot. OS: win2003 32bits, with NTFS. This is a vm running on vmware fusion on a mac.
jvm: I made sure, java version "1.6.0_14" IndexWriter settings: writer.setMaxFieldLength(maxFieldLength); writer.setMergeFactor(10); writer.setRAMBufferSizeMB(10); writer.setUseCompoundFile(true); writer.setTermIndexInterval(512); In this production environment only a single thread was indexing. No addIndexes() is used. In some ocassions yes write.lock might be deleted by code. I have added infoStream to stdout in my code so in the future we would run this with debug on. javier On Fri, Nov 27, 2009 at 4:00 PM, Michael McCandless <luc...@mikemccandless.com> wrote: > Any Lucene-related exceptions hit in your env? What OS (looks like > Windows, but which one?), filesystem are you on? > > And are you really certain about the java version being used in your > production env? Don't just trust which java your interactive shell > finds on its PATH -- double check how your app server is actually > launched. > > Have you changed any of IndexWriter's defaults? Are you adding docs > with many threads? Using addIndexes? Do you ever forcefully remove > the write.lock, or use Lucene's API to unlock the index? The more > info you can give about how your app uses Lucene, the better. > > And if possible run with infoStream enabled so we can maybe capture > the corruption happening. > > Mike > > On Fri, Nov 27, 2009 at 7:09 AM, jm <jmugur...@gmail.com> wrote: >> I manually did CheckIndex in all indexes and found two with issues: >> >> first >> Segments file=segments_42w numSegments=21 version=FORMAT_HAS_PROX [Lucene >> 2.4] >> 1 of 21: name=_109 docCount=10410 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=55,789 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...OK [170764 terms; 6214606 terms/docs >> pairs; 29694242 tokens] >> test: stored fields.......OK [10410 total field count; avg 1 fields per >> doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> >> 2 of 21: name=_160 docCount=2178 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=11,676 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...OK [92275 terms; 1264593 terms/docs >> pairs; 5807908 tokens] >> test: stored fields.......OK [2178 total field count; avg 1 fields per >> doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> >> 3 of 21: name=_1i9 docCount=2225 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=10,872 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...OK [92284 terms; 1193562 terms/docs >> pairs; 5253891 tokens] >> test: stored fields.......OK [2225 total field count; avg 1 fields per >> doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> >> 4 of 21: name=_1qx docCount=2654 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=14,374 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...OK [95153 terms; 1623058 terms/docs >> pairs; 7151784 tokens] >> test: stored fields.......OK [2654 total field count; avg 1 fields per >> doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> >> 5 of 21: name=_248 docCount=2165 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=11,446 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...OK [92177 terms; 1243801 terms/docs >> pairs; 5632974 tokens] >> test: stored fields.......OK [2165 total field count; avg 1 fields per >> doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> >> 6 of 21: name=_2a4 docCount=2508 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=13,917 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...OK [94222 terms; 1566966 terms/docs >> pairs; 6967989 tokens] >> test: stored fields.......OK [2508 total field count; avg 1 fields per >> doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> >> 7 of 21: name=_2pu docCount=2092 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=10,217 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...OK [91592 terms; 1105091 terms/docs >> pairs; 4865932 tokens] >> test: stored fields.......OK [2092 total field count; avg 1 fields per >> doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> >> 8 of 21: name=_35q docCount=3285 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=16,378 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...ERROR [term >> attachments:c7a0e2a47ebad4214e933621f6c9dac7: doc 426 <= lastDoc 426] >> java.lang.RuntimeException: term >> attachments:c7a0e2a47ebad4214e933621f6c9dac7: doc 426 <= lastDoc 426 >> at >> org.apache.lucene.index.CheckIndex.testTermIndex(CheckIndex.java:644) >> at org.apache.lucene.index.CheckIndex.checkIndex(CheckIndex.java:530) >> at org.apache.lucene.index.CheckIndex.checkIndex(CheckIndex.java:319) >> at org.getopt.luke.Luke$6.run(Unknown Source) >> test: stored fields.......OK [3285 total field count; avg 1 fields per >> doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> FAILED >> WARNING: fixIndex() would remove reference to this segment; full >> exception: >> java.lang.RuntimeException: Term Index test failed >> at org.apache.lucene.index.CheckIndex.checkIndex(CheckIndex.java:543) >> at org.apache.lucene.index.CheckIndex.checkIndex(CheckIndex.java:319) >> at org.getopt.luke.Luke$6.run(Unknown Source) >> >> 9 of 21: name=_3g7 docCount=2016 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=11,813 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...OK [91094 terms; 1286115 terms/docs >> pairs; 5942307 tokens] >> test: stored fields.......OK [2016 total field count; avg 1 fields per >> doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> >> 10 of 21: name=_3o8 docCount=1702 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=10,008 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...OK [89179 terms; 1081620 terms/docs >> pairs; 4960310 tokens] >> test: stored fields.......OK [1702 total field count; avg 1 fields per >> doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> >> 11 of 21: name=_3wq docCount=2249 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=13,441 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...OK [92829 terms; 1491635 terms/docs >> pairs; 6851263 tokens] >> test: stored fields.......OK [2249 total field count; avg 1 fields per >> doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> >> 12 of 21: name=_3zd docCount=2264 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=14,153 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...OK [92747 terms; 1602790 terms/docs >> pairs; 7248272 tokens] >> test: stored fields.......OK [2264 total field count; avg 1 fields per >> doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> >> 13 of 21: name=_44r docCount=2743 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=16,56 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...OK [95135 terms; 1817602 terms/docs >> pairs; 8534778 tokens] >> test: stored fields.......OK [2743 total field count; avg 1 fields per >> doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> >> 14 of 21: name=_4fl docCount=2105 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=13,041 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...OK [91940 terms; 1447522 terms/docs >> pairs; 6565862 tokens] >> test: stored fields.......OK [2105 total field count; avg 1 fields per >> doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> >> 15 of 21: name=_4gf docCount=1891 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=10,88 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...OK [90339 terms; 1186203 terms/docs >> pairs; 5392055 tokens] >> test: stored fields.......OK [1891 total field count; avg 1 fields per >> doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> >> 16 of 21: name=_54z docCount=2007 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=10,848 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...OK [91154 terms; 1199551 terms/docs >> pairs; 5284819 tokens] >> test: stored fields.......OK [2007 total field count; avg 1 fields per >> doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> >> 17 of 21: name=_5ao docCount=530 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=3,743 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...OK [76964 terms; 311259 terms/docs >> pairs; 1569162 tokens] >> test: stored fields.......OK [530 total field count; avg 1 fields per doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> >> 18 of 21: name=_5am docCount=113 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=2,118 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...OK [66266 terms; 142915 terms/docs >> pairs; 711223 tokens] >> test: stored fields.......OK [113 total field count; avg 1 fields per doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> >> 19 of 21: name=_5ap docCount=117 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=0,525 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...OK [15930 terms; 50518 terms/docs pairs; >> 203306 tokens] >> test: stored fields.......OK [117 total field count; avg 1 fields per doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> >> 20 of 21: name=_5ar docCount=165 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=0,73 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...OK [22756 terms; 67573 terms/docs pairs; >> 273697 tokens] >> test: stored fields.......OK [165 total field count; avg 1 fields per doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> >> 21 of 21: name=_5at docCount=115 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=0,409 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...OK [14805 terms; 40657 terms/docs pairs; >> 139527 tokens] >> test: stored fields.......OK [115 total field count; avg 1 fields per doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> >> WARNING: 1 broken segments (containing 3285 documents) detected >> >> >> =============================================================================================== >> second: >> >> Segments file=segments_7s4 numSegments=12 version=FORMAT_HAS_PROX [Lucene >> 2.4] >> 1 of 12: name=_3hm docCount=17119 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=88,715 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...ERROR [term body:mas: doc 13174 <= lastDoc >> 13174] >> java.lang.RuntimeException: term body:mas: doc 13174 <= lastDoc 13174 >> at >> org.apache.lucene.index.CheckIndex.testTermIndex(CheckIndex.java:644) >> at org.apache.lucene.index.CheckIndex.checkIndex(CheckIndex.java:530) >> at org.apache.lucene.index.CheckIndex.checkIndex(CheckIndex.java:319) >> at org.getopt.luke.Luke$6.run(Unknown Source) >> test: stored fields.......OK [17119 total field count; avg 1 fields per >> doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> FAILED >> WARNING: fixIndex() would remove reference to this segment; full >> exception: >> java.lang.RuntimeException: Term Index test failed >> at org.apache.lucene.index.CheckIndex.checkIndex(CheckIndex.java:543) >> at org.apache.lucene.index.CheckIndex.checkIndex(CheckIndex.java:319) >> at org.getopt.luke.Luke$6.run(Unknown Source) >> >> 2 of 12: name=_4bl docCount=3302 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=19,256 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...OK [98534 terms; 2136125 terms/docs >> pairs; 9970629 tokens] >> test: stored fields.......OK [3302 total field count; avg 1 fields per >> doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> >> 3 of 12: name=_4w9 docCount=4380 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=27,348 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...OK [103306 terms; 3081204 terms/docs >> pairs; 14235295 tokens] >> test: stored fields.......OK [4380 total field count; avg 1 fields per >> doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> >> 4 of 12: name=_5jh docCount=4372 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=28,867 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...OK [103363 terms; 3187068 terms/docs >> pairs; 15176139 tokens] >> test: stored fields.......OK [4372 total field count; avg 1 fields per >> doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> >> 5 of 12: name=_6nn docCount=6526 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=32,316 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...OK [111787 terms; 3720002 terms/docs >> pairs; 16710886 tokens] >> test: stored fields.......OK [6526 total field count; avg 1 fields per >> doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> >> 6 of 12: name=_82m docCount=6889 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=36,753 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...OK [113120 terms; 4284180 terms/docs >> pairs; 19188993 tokens] >> test: stored fields.......OK [6889 total field count; avg 1 fields per >> doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> >> 7 of 12: name=_8kv docCount=2286 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=13,367 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...OK [93101 terms; 1437797 terms/docs >> pairs; 6844358 tokens] >> test: stored fields.......OK [2286 total field count; avg 1 fields per >> doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> >> 8 of 12: name=_8i0 docCount=141 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=1,618 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...OK [63287 terms; 85330 terms/docs pairs; >> 414225 tokens] >> test: stored fields.......OK [141 total field count; avg 1 fields per doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> >> 9 of 12: name=_8kt docCount=275 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=2,549 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...OK [73391 terms; 173116 terms/docs >> pairs; 847099 tokens] >> test: stored fields.......OK [275 total field count; avg 1 fields per doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> >> 10 of 12: name=_8n3 docCount=103 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=0,618 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...OK [18925 terms; 50696 terms/docs pairs; >> 248460 tokens] >> test: stored fields.......OK [103 total field count; avg 1 fields per doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> >> 11 of 12: name=_8n4 docCount=3 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=1,209 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...OK [49084 terms; 49234 terms/docs pairs; >> 267700 tokens] >> test: stored fields.......OK [3 total field count; avg 1 fields per doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> >> 12 of 12: name=_8n5 docCount=122 >> compound=true >> hasProx=true >> numFiles=1 >> size (MB)=0,66 >> no deletions >> test: open reader.........OK >> test: fields..............OK [6 fields] >> test: field norms.........OK [6 fields] >> test: terms, freq, prox...OK [20211 terms; 54518 terms/docs pairs; >> 267390 tokens] >> test: stored fields.......OK [122 total field count; avg 1 fields per doc] >> test: term vectors........OK [0 total vector count; avg 0 >> term/freq vector fields per doc] >> >> WARNING: 1 broken segments (containing 17119 documents) detected >> >> >> I have not been able to reproduce this in my env. >> >> javier >> >> On Fri, Nov 27, 2009 at 12:23 PM, jm <jmugur...@gmail.com> wrote: >>> Ok, I got the index from the production machine, but I am having some >>> problem to find the index..., our process deals with multiple indexes, >>> in the current exception I cannot see any indication about the index >>> having the issue. I opened all my indexes with luke and old opened >>> succesfully, some had around 4000 docs (so I can discard those I >>> think), and some had around 40.000 docs, I was hoping one that had >>> around 24658 docs, but I guess this is the number of docs in one >>> segment, not the whole index right? >>> >>> Any idea to help find the index? >>> >>> In the codebase it would be nice to give any hint about the index >>> (path in case of a fsdirectory etc) when found an exception, dont >>> assume there is a single index in the application. >>> >>> javier >>> >>> On Fri, Nov 27, 2009 at 11:39 AM, Michael McCandless >>> <luc...@mikemccandless.com> wrote: >>>> Also, if you're able to reproduce this, can you call >>>> writer.setInfoStream and capture & post the resulting output leading >>>> up to the exception? >>>> >>>> Mike >>>> >>>> On Thu, Nov 26, 2009 at 7:12 AM, jm <jmugur...@gmail.com> wrote: >>>>> The process is still running and ops dont want to stop it. As soon as >>>>> stops I'll try checkindex. >>>>> >>>>> Its created brand new with 2.4.1 >>>>> >>>>> On Thu, Nov 26, 2009 at 12:42 PM, Michael McCandless >>>>> <luc...@mikemccandless.com> wrote: >>>>>> I think you're using a JRE that has the fix for the issue found in >>>>>> LUCENE-1282. >>>>>> >>>>>> Can you run CheckIndex on your index and post the output? >>>>>> >>>>>> Was this index created from scratch on Lucene 2.4.1? Or, created from >>>>>> an earlier Lucene version? >>>>>> >>>>>> Mike >>>>>> >>>>>> On Thu, Nov 26, 2009 at 6:03 AM, jm <jmugur...@gmail.com> wrote: >>>>>>> or are we really? I think we are on 1.6 update 14 right?? >>>>>>> >>>>>>> sorry Im lost right now on jdk version numbering >>>>>>> >>>>>>> On Thu, Nov 26, 2009 at 12:01 PM, jm <jmugur...@gmail.com> wrote: >>>>>>>> on second thought...I hadnt noticed the jdk numbers properly, we are >>>>>>>> using using b28, and JDK 6 Update 10 (b28) is the one fixing this... >>>>>>>> >>>>>>>> ok forget this then >>>>>>>> thanks! >>>>>>>> >>>>>>>> On Thu, Nov 26, 2009 at 11:55 AM, jm <jmugur...@gmail.com> wrote: >>>>>>>>> Hi, >>>>>>>>> >>>>>>>>> Dont know if this should be here or in java-dev, posting to this one >>>>>>>>> first. In one of our installations, we have encountered an exception: >>>>>>>>> >>>>>>>>> Exception in thread "Lucene Merge Thread #0" >>>>>>>>> org.apache.lucene.index.MergePolicy$MergeException: >>>>>>>>> org.apache.lucene.index.CorruptIndexException: docs out of order >>>>>>>>> (24658 <= 24658 ) >>>>>>>>> at >>>>>>>>> org.apache.lucene.index.ConcurrentMergeScheduler.handleMergeException(ConcurrentMergeScheduler.java:309) >>>>>>>>> at >>>>>>>>> org.apache.lucene.index.ConcurrentMergeScheduler$MergeThread.run(ConcurrentMergeScheduler.java:286) >>>>>>>>> Caused by: org.apache.lucene.index.CorruptIndexException: docs out of >>>>>>>>> order (24658 <= 24658 ) >>>>>>>>> at >>>>>>>>> org.apache.lucene.index.SegmentMerger.appendPostings(SegmentMerger.java:641) >>>>>>>>> at >>>>>>>>> org.apache.lucene.index.SegmentMerger.mergeTermInfo(SegmentMerger.java:586) >>>>>>>>> at >>>>>>>>> org.apache.lucene.index.SegmentMerger.mergeTermInfos(SegmentMerger.java:547) >>>>>>>>> at >>>>>>>>> org.apache.lucene.index.SegmentMerger.mergeTerms(SegmentMerger.java:500) >>>>>>>>> at >>>>>>>>> org.apache.lucene.index.SegmentMerger.merge(SegmentMerger.java:140) >>>>>>>>> at >>>>>>>>> org.apache.lucene.index.IndexWriter.mergeMiddle(IndexWriter.java:4231) >>>>>>>>> at >>>>>>>>> org.apache.lucene.index.IndexWriter.merge(IndexWriter.java:3884) >>>>>>>>> at >>>>>>>>> org.apache.lucene.index.ConcurrentMergeScheduler.doMerge(ConcurrentMergeScheduler.java:205) >>>>>>>>> at >>>>>>>>> org.apache.lucene.index.ConcurrentMergeScheduler$MergeThread.run(ConcurrentMergeScheduler.java:260) >>>>>>>>> >>>>>>>>> I have investigated in the list and it looked like >>>>>>>>> https://issues.apache.org/jira/browse/LUCENE-1282. But we are using >>>>>>>>> 2.4.1, and >>>>>>>>> C:\Documents and Settings\Administrator>java -version >>>>>>>>> java version "1.6.0_14" >>>>>>>>> Java(TM) SE Runtime Environment (build 1.6.0_14-b08) >>>>>>>>> Java HotSpot(TM) Client VM (build 14.0-b16, mixed mode) >>>>>>>>> >>>>>>>>> java process launched like this: >>>>>>>>> java -server -Xmx1536m ... >>>>>>>>> >>>>>>>>> So I understand this bug should not be happening?? >>>>>>>>> >>>>>>>>> any idea? >>>>>>>>> thanks >>>>>>>>> javi >>>>>>>>> >>>>>>>> >>>>>>> >>>>>>> --------------------------------------------------------------------- >>>>>>> To unsubscribe, e-mail: java-user-unsubscr...@lucene.apache.org >>>>>>> For additional commands, e-mail: java-user-h...@lucene.apache.org >>>>>>> >>>>>>> >>>>>> >>>>>> --------------------------------------------------------------------- >>>>>> To unsubscribe, e-mail: java-user-unsubscr...@lucene.apache.org >>>>>> For additional commands, e-mail: java-user-h...@lucene.apache.org >>>>>> >>>>>> >>>>> >>>>> --------------------------------------------------------------------- >>>>> To unsubscribe, e-mail: java-user-unsubscr...@lucene.apache.org >>>>> For additional commands, e-mail: java-user-h...@lucene.apache.org >>>>> >>>>> >>>> >>>> --------------------------------------------------------------------- >>>> To unsubscribe, e-mail: java-user-unsubscr...@lucene.apache.org >>>> For additional commands, e-mail: java-user-h...@lucene.apache.org >>>> >>>> >>> >> >> --------------------------------------------------------------------- >> To unsubscribe, e-mail: java-user-unsubscr...@lucene.apache.org >> For additional commands, e-mail: java-user-h...@lucene.apache.org >> >> > > --------------------------------------------------------------------- > To unsubscribe, e-mail: java-user-unsubscr...@lucene.apache.org > For additional commands, e-mail: java-user-h...@lucene.apache.org > > --------------------------------------------------------------------- To unsubscribe, e-mail: java-user-unsubscr...@lucene.apache.org For additional commands, e-mail: java-user-h...@lucene.apache.org