Whitespace analyzer does preserve those symbols, but not as tokens. It simply leaves them attached to the original term.

As an example of what I'm talking about, consider a document that contains (without the quotes) "foo, ".

Now, using WhitespaceAnalyzer, I could only get that document by searching for "foo,". Using StandardAnalyzer or any analyzer that removes punctuation, I could only find it by searching for "foo".

I want an analyzer that will allow me to find it if I build a phrase query with the term "foo" followed immediately by ",". After all, the comma may be relevant to the search, but is definitely not part of the word.

Extending StandardAnalyer is what I had in mind, but I don't know where to start. I also wonder why no-one seems to have done it before- it makes me suspect that there's some reason I haven't seen yet that makes it impossible ot impractical.



Karl Wettin wrote:

1 okt 2007 kl. 15.33 skrev John Byrne:

Has anyone written an analyzer that preserves puncuation and
synmbols ("£", "$", "%" etc.) as tokens?

WhitespaceAnalyzer?

You could also extend the lexical rules of StandardAnalyzer.




---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]

Reply via email to