The contrib/analyzers has several n-gram based tokenization and token filter options.

On Jun 18, 2009, at 10:15 PM, Neha Gupta wrote:

Hey,

I was wondering if there is a way to read the index and generate n- grams of words for a document in lucene? I am quite new to it and am using pylucene.

Thanks,
Neha

--------------------------
Grant Ingersoll
http://www.lucidimagination.com/

Search the Lucene ecosystem (Lucene/Solr/Nutch/Mahout/Tika/Droids) using Solr/Lucene:
http://www.lucidimagination.com/search


---------------------------------------------------------------------
To unsubscribe, e-mail: java-user-unsubscr...@lucene.apache.org
For additional commands, e-mail: java-user-h...@lucene.apache.org

Reply via email to