nalysis page), you
> can mechanically translate to raw Lucene API calls.
>
> Look at the standard tokenizer, it should do a better job with punctuation.
>
> -- Jack Krupansky
>
> -Original Message- From: Todd Hunt
> Sent: Thursday, June 27, 2013 1:14 PM
> To: java
nize numeric tokens and generate extra tokens at the same position
>for each token variant that you want to search for.
>
>-- Jack Krupansky
>
>-Original Message-
>From: Todd Hunt
>Sent: Friday, June 28, 2013 2:18 PM
>To: java-user@lucene.apache.org
>Subject: How to
I have an application that is indexing the text from various reports and forms
that are generated from our core system. The reports will contain dollar
amounts and various indexes that contain all numbers, but have leading zeros.
If a document contains that following text that is stored in one
I am working on an application that is using Tika to index text based documents
and store the text results in Lucene. These documents can range anywhere from
1 page to thousands of pages.
We are currently using Lucene 3.0.3. I am currently using the StandarAnalyzer
to index and search for the