I'm using default analyzer. Actually one that is set by default by Compass
framework but I assume it is the same that would be used in Lucene by default.
Which one should I use?
-Original Message-
From: Robert Muir [mailto:rcm...@gmail.com]
Sent: Friday, July 10, 2009 6:13 PM
To: java-us
on, Jun 15, 2009 at 10:30 PM, OBender
Hotmail wrote:
> That's the thing there is no actual requirement.
> I've been presented with all the languages that company theoretically
> provides.
> My guess is that what I'm going to end up with is all western languages, good
>
h?
I think you might have larger problems!
On Mon, Jun 15, 2009 at 9:18 PM, OBender Hotmail wrote:
> Here is the list of possible languages. Don't laugh :) I know those are
> almost all world languages but it is a true requirement. Well, actual number
> will be closer to 70 not 100
return ts;
}
}
can you give a better idea as to what languages you have and what your
search requirements are (accent marks, punctuation, etc etc) ?
On Mon, Jun 15, 2009 at 5:39 PM, OBender Hotmail wrote:
> I've looked over SolR quickly, it is a bit too heavy for my project.
> So w
-user@lucene.apache.org
Subject: Re: Lucene and multi-lingual Unicode - advice needed
Well just reply back if SolR is inappropriate for your needs.
In that case, you will need to build a custom analyzer (its not too
bad), so that you can use compass.
On Mon, Jun 15, 2009 at 4:19 PM, OBender Ho
ework, but recently there has been some improvements added to SolR
so that the default type 'text' is pretty good for multilingual
processing.
In fact I hope in the future it will be improved in lucene so that
your decision is really based upon other application needs...
On Mon, Jun 15, 2009
Hi All!
I'm new to Lucene so forgive me if this question was asked before.
I have a database with records in the same table in many different languages
(up to 70) it includes all W-European, Arabic, Eastern, CJK, Cyrillic, etc.
you name it.
I've looked at what people say about Lucene and it l