Hello all, We have an extremely large number of terms in our indexes. I want to be able to extract a sample of the terms, say something like every 128th term. If I use code based on org.apache.lucene.misc.HighFreqTerms or org.apache.lucene.index.CheckIndex I would get a TermsEnum, call termEnum.next() 128 times, grab the term and then call next another 128 times. termEnum = reader.terms(); while (termEnum.next() { }
Since the tii file contains every 128th (or IndexInterval ) term and it is loaded into memory, is there some programmatic way (in the public API) to read that data structure in memory rather than having to force Lucene to actually read the entire tis file by using termEnum.next() ? Tom Burton-West http://www.hathitrust.org/blogs/large-scale-search