Hi, 

> You could even store all of the page offsets in your 
> meta-data document
> in a special field if you wanted, then lazy-load that field 
> rather than
> dynamically counting. 

How can I lazy load a field?

> You'd have to be careful that your offsets
> corresponded to the data *after* it was analyzed, but that shouldn't
> be too hard. 

The TermPosition is the position after analyzing?

> You'd have to read this field before deleting the doc
> and make sure it was stored with the replacement.....

Don't understand...

> And, since I'm getting random ideas anyway, here's another.
> The PositionIncrementGap is the "distance" (measured in
> terms) between two tokens. Let's claim that you have no
> page with more than 10,000 (or whatever) tokens. Just
> bump the positionincrementgap to the next 10,000 at the
> start of each page. So, the first term on the first page
> has an offset of 0. the first term on the second page
> has an offset of 10,000. The first term on the third
> page has an offset of 20,000.
> 
> Now, with the SpanNearQuery trick from above, your
> term position modulo 10,000 is also your page. This would
> also NOT match across pages. Hmmmm, I kind of like that
> idea.

But I have to know, how many tokens each page has!?

Thank you!


---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]

Reply via email to