:47 AM, Ashley Ryan
> wrote:
> > Obviously, our work around of escaping the angle brackets works as we
> need
> > it to, but it seems to me that your documentation is incorrect. Am I
> > misunderstanding the documentation or conflating the issue I'm seeing
> with
> the topic of special characters?
Maybe the documentation you're reading is for the older QueryParser
and not StandardQueryParser? Neither StandardQueryParser and
StandardSyntaxParser appear to say anything about special
characters... or indeed very much about the syntax at all, which
Hi,
I'm working on a project that uses
the org.apache.lucene.queryparser.flexible.standard.parser.StandardSyntaxParser.
I found in the documentation online that the list of special characters is
as follows:
+ - && || ! ( ) { } [ ] ^ " ~ * ? : \
This list does not incl
Thank you.
Can you please share me code snippet to deal with these chars.
I tried but couldn't achieve.
On Tue, Sep 6, 2016 at 10:59 PM, Iker Huerga wrote:
> here is the thing, you are probably using the StandardAnalyzer so those
> special characters are going to be removed at in
here is the thing, you are probably using the StandardAnalyzer so those
special characters are going to be removed at indexing time
If you don't want that to happen, if you don't want that to happen you can
try with KeywordAnalyzer or just create your own Analyzer
You can test with the
.ietf.org/html/rfc3986 see section 2.2 so you would have to
> URL encode them
>
> My 2 cents
>
> 2016-09-06 10:20 GMT-04:00 Chaitanya Kumar Ch :
>
> > Thanks for the reply.
> > I have tried that but didn't work.
> > Also please note that *@,# are not part of c
t *@,# are not part of current special characters
> list*.
>
> On Tue, Sep 6, 2016 at 7:47 PM, Iker Huerga wrote:
>
> > I'd try scaping the characters as in
> > https://lucene.apache.org/core/2_9_4/queryparsersyntax.
> > html#Escaping%20Special%20Characters
>
Thanks for the reply.
I have tried that but didn't work.
Also please note that *@,# are not part of current special characters list*.
On Tue, Sep 6, 2016 at 7:47 PM, Iker Huerga wrote:
> I'd try scaping the characters as in
> https://lucene.apache.org/core/2_9_4/querypars
d but I am not getting results.
>
> Below query is generated If i am remove ignoreFieldBridge()
>
> +(+body:johndaly +body:baby)
>
>
> Stack overflow link
> <http://stackoverflow.com/questions/39350676/hibernate-
> search-lucene-search-text-with-special-characters-like>
> --
> Thank You,
> Chaitanya Kumar Ch,
> +91 9550837582
>
--
Iker Huerga
http://www.ikerhuerga.com/
Bridge() to the
field but I am not getting results.
Below query is generated If i am remove ignoreFieldBridge()
+(+body:johndaly +body:baby)
Stack overflow link
<http://stackoverflow.com/questions/39350676/hibernate-search-lucene-search-text-with-special-characters-like>
--
Thank You,
Chaitanya Kumar Ch,
+91 9550837582
.
-- Jack Krupansky
-Original Message-
From: saisantoshi
Sent: Sunday, October 20, 2013 7:43 PM
To: java-user@lucene.apache.org
Subject: Re: Handling special characters in Lucene 4.0
what about other characters like '&,'( quote) characters. We have a
requirement that a text
View this message in context:
http://lucene.472066.n3.nabble.com/Handling-special-characters-in-Lucene-4-0-tp4096674p4096732.html
Sent from the Lucene - Java Users mailing list archive at Nabble.com.
-
To unsubscrib
Right, the "Escaping Special Characters" is simply to escape query operators
like "&&" (means "AND") and "+" (which means "AND" or "MUST"). Yes, the
white space analyzer could be used, or a custom analyzer that uses the
dAnalyzer wont work for the following
> below as it strips out the special characters and does search only on
> searchText ( in this case).
>
> queryText = *&&searchText*
>
> If we want to do a search like "*&&**" then we need to use
> WhiteSpaceAna
Thanks.
So, if I understand correctly, StandardAnalyzer wont work for the following
below as it strips out the special characters and does search only on
searchText ( in this case).
queryText = *&&searchText*
If we want to do a search like "*&&**" then we need to use
The standard analyzer should remove those ampersands and pluses, so the core
alpha terms should be matched. You would need to use the white space
analyzer or a custom analyzer to preserve such special characters.
Please give a specific indexed text string and a specific query that fails
StandardAnalyzer both at index and search time. We use the default one and
don't have any custom analyzers.
Thanks,
Sai
--
View this message in context:
http://lucene.472066.n3.nabble.com/Handling-special-characters-in-Lucene-4-0-tp4096674p4096710.html
Sent from the Lucene - Java
using at query time?
-- Jack Krupansky
-Original Message-
From: saisantoshi
Sent: Sunday, October 20, 2013 12:47 PM
To: java-user@lucene.apache.org
Subject: Handling special characters in Lucene 4.0
I have created strings like the below
&&searchtext
+sampletext
and when
I have created strings like the below
&&searchtext
+sampletext
and when I try to search the following using *&&** or *+** it does not give
any result.
I am using QueryParser.escape(String s) method to handle the special
characters but does not look like it did anything.
Als
Ok..The problem seems to be solved.. Using the escape method of
QueryParser, I am able to parse the given input with special characters
also.
But if this leads to another pitfall someone may please guide..
On 9/10/2013 12:19 PM, Ankit Murarka wrote:
Hello,
I am using StandardAnalyzer
special characters in the code.
When I supply the same phrase in PhraseQuery, I get the required hit..
Understood that PhraseQuery and QueryParser handles query differently.
How can I achieve this. Accepting special characters also in user
supplied inputs. I cannot make any changes in the
uffer[0]) == -1" to test if
> termBuffer[0] is equal to null, but apparently the special characters
> return -1 as well when given as parameter.
>
> Thank you for your help.
>
> 2013/1/14 Hankyu Kim
>
>> I did intend to ignore all the spaces, so that's
I just found the cause of error and you were right about my code being the
source.
I used "Character.getNumericValue(termBuffer[0]) == -1" to test if
termBuffer[0] is equal to null, but apparently the special characters
return -1 as well when given as parameter.
Thank you for your hel
also realised the issue wasn't with querying, but it was indexing
> whihch
> >> left the terms with leading special character out.
> >>
> >> My goal was to do a fuzzymatch by creating a trigram index. The idea is
> to
> >> tokenize the documents into trigra
ing, but it was indexing whihch
>> left the terms with leading special character out.
>>
>> My goal was to do a fuzzymatch by creating a trigram index. The idea is to
>> tokenize the documents into trigrams, not by words during indexing and
>> searching so lucene ca
t; creating trigrmas. However, when I check the index with lukeall, all the
> other trigrams are indexed correctly except for the terms ':)a' and ')an'.
> Since the missing indexes are related to lucene's special characters, I
> don't think it's got to do
ting a trigram index. The idea is to
tokenize the documents into trigrams, not by words during indexing and
searching so lucene can search for part of a word or phrase.
Say the original text in the document said : "Sample text with special
characters :) and such"
It's tokenized int
on of something, that
custom version has to be the prime suspect for any problems.
--
Ian.
On Thu, Jan 10, 2013 at 12:08 PM, Hankyu Kim wrote:
> Hi.
>
> I've created a custom analyzer that treats special characters just like any
> other. The index works fine all the time even w
If
> > the user is entering the query, it's his responsibility to escape the
> > query.
> >
> > On Mon, Jul 4, 2011 at 4:21 AM, govind bhardwaj
> > wrote:
> >
> > > Hi,
> > >
> > > I am using Lucene version 3.1
> > > Previousl
on, Jul 4, 2011 at 4:21 AM, govind bhardwaj
> wrote:
>
> > Hi,
> >
> > I am using Lucene version 3.1
> > Previously I had trouble regarding special characters as when I entered
> > "---" as my input, it gave the following error
> >
> >
> >
sly I had trouble regarding special characters as when I entered
> "---" as my input, it gave the following error
>
>
> *Caused by: org.apache.lucene.queryParser.ParseException: Encountered " "-&quo
Hi,
I am using Lucene version 3.1
Previously I had trouble regarding special characters as when I entered
"---" as my input, it gave the following error
*Caused by: org.apache.lucene.queryParser.ParseException: Encountered " "-"
"- "" a
: Wed, 14 Apr 2010 08:24:38 -0400
> Von: Steven A Rowe
> An: "java-user@lucene.apache.org"
> Betreff: RE: PrefixQuery and special characters
> Hi Franz,
>
> The likely problem is that you're using an index-time analyzer that strips
> out the paren
n the case of synonyms, most people will want
their index and query analyzers to be the same.
Steve
From: Franz Roth [franzr...@gmx.de]
Sent: Wednesday, April 14, 2010 7:42 AM
To: java-user@lucene.apache.org
Subject: PrefixQuery and special characters
H
Hi all,
say I have an Index with one field named "category". There are two documents
one with value "(testvalue)" and one with value "test value".
Now somone search with "test". My Searchenine uses the
org.apache.lucene.search.PrefixQuery and finds 2 documents. Maybe he estimated
only one hit;
Hi
I am looking at handling special characters in the query as using certain
characters cause an exception. I looked at QueryParser.escape(..) to handle
this. It works to a certain extent for example using '!' doesn't cause an
exception. However when I use a wildcard then the wild
On Fri, Jul 31, 2009 at 5:00 PM, wrote:
> Hi Ahmet,
>
> Thanks for the clarification and information! That was exactly what I was
> looking for.
>
> Jim
>
>
> AHMET ARSLAN wrote:
>>
>> > I guess that the obvious question is "Which characters are
>> > considered 'punctuation characters'?".
Hi Ahmet,
Thanks for the clarification and information! That was exactly what I was
looking for.
Jim
AHMET ARSLAN wrote:
>
> > I guess that the obvious question is "Which characters are
> > considered 'punctuation characters'?".
>
> Punctuation = ("_"|"-"|"/"|"."|",")
>
> > In part
> I guess that the obvious question is "Which characters are
> considered 'punctuation characters'?".
Punctuation = ("_"|"-"|"/"|"."|",")
> In particular, does the analyzer consider "=" (equal) and
> ":" (colon) to be punctuation characters?
":" is special character at QueryParser (if you are
Phil Whelan wrote:
> On Thu, Jul 30, 2009 at 7:12 PM, wrote:
> > I was wonder if there is a list of special characters for the standard
> > analyzer?
> >
> > What I mean by "special" is characters that the analyzer considers break
> > chara
On Thu, Jul 30, 2009 at 7:12 PM, wrote:
> I was wonder if there is a list of special characters for the standard
> analyzer?
>
> What I mean by "special" is characters that the analyzer considers break
> characters.
> For example, if I have something like &qu
Hi,
I was wonder if there is a list of special characters for the standard
analyzer?
What I mean by "special" is characters that the analyzer considers break
characters. For example, if I have something like "foo=something", apparently
the analyzer considers this as
Thanks Ian!
You were right, the problem was simply the InputStreamReader.
(Sorry this reply is a bit off-topic now, I didn't have time to work
on this specific problem earlier).
Laura
On Apr 27, 2009, at 3:24 PM, Ian Lea wrote:
Hi
The problem may well lie with the reading of the queries
ed a standalone test case that demonstrates the current
> behavior. I've attached it to this email. It should just require junit and
> lucene. It might actually be useful in general for anyone trying to figure
> out various Lucene behaviors.
>
> At a high level, am I correctly
rs.
At a high level, am I correctly understanding that Lucene doesn't support
searching on indexed special characters without significant additional
machinations? If so, has anyone gone through those machinations and posted
a link :)? Given the test case, is this worthy of a JIRA issue?
On
tokens indexed.
Warning: I haven't used the ngram tokenizers, so I know just enough to
be dangerous. That said, you could tokenize these as ngrams. I'm not sure
what the base ngram tokenizer does with your special characters, but you
could pretty easily create your own analyzer that spits out,
that the escaped special characters seem to be turned into spaces, not
used literally.
Up to this point, even attempting to directly create an appropriate query
(PrefixQuery, PhraseQuery, TermQuery, etc.), I've been unable to come up
with a query that will match the text with special characters a
Hi
The problem may well lie with the reading of the queries from disk
into your program.
Using an InputStreamReader with correct encoding (UTF-8?) should solve it.
--
Ian.
On Mon, Apr 27, 2009 at 12:32 PM, Laura Hollink wrote:
> Hi,
>
> I have built an index using the standardAnalyser, and a
Hi,
I have built an index using the standardAnalyser, and am now querying
that index, also with standardAnalyser. However, I get results I don't
understand.
I know there are a few documents about Brazil in my corpus. My corpus
is in Dutch, and the Dutch term used is "Brazilië".
I query
t;>
>> After gone threw with the site which you given... i understood that
>> StandardAnalyzer is enough to handle these special characters.
>>
>> i'm attaching one class called AnalysisDemo.java. By executing that class
>> i'm able to say the above sentance
derstood that
> StandardAnalyzer is enough to handle these special characters.
>
> i'm attaching one class called AnalysisDemo.java. By executing that class
> i'm able to say the above sentance(i.e StandardAnalyzer is enough).
>
> Here is the out put when i ran the abov
Hi Thanks for your reply.
After gone threw with the site which you given... i understood that
StandardAnalyzer is enough to handle these special characters.
i'm attaching one class called AnalysisDemo.java. By executing that class
i'm able to say the above sentance(i.e StandardA
Please help me in this. please..
> >>
> >> Erick Erickson wrote:
> >> >
> >> > Take a look at DutchAnalyzer. The problem you'll have is if you're
> >> > indexing
> >> > this document along with a bunch of documents from other langua
t;> > Take a look at DutchAnalyzer. The problem you'll have is if you're
>> > indexing
>> > this document along with a bunch of documents from other languages.
>> > You could search the mail archive for extensive discussions of
>> indexing/
>> &
t;> > Take a look at DutchAnalyzer. The problem you'll have is if you're
>> > indexing
>> > this document along with a bunch of documents from other languages.
>> > You could search the mail archive for extensive discussions of
>> indexing/
>> &
indexing/
> > searching documents from several languages.
> >
> > Best
> > Erick
> >
> > On Tue, Apr 21, 2009 at 2:40 AM, Uday Kumar Maddigatla
> > wrote:
> >
> >> HI,
> >>
> >>
> >>
> >> I'm new to the lu
n Tue, Apr 21, 2009 at 2:40 AM, Uday Kumar Maddigatla
> wrote:
>
>> HI,
>>
>>
>>
>> I'm new to the lucene. I downloaded lucene 2.4.1.
>>
>>
>>
>> I have one xml file which contains few special characters like 'å', '
1, 2009 at 2:40 AM, Uday Kumar Maddigatla wrote:
> HI,
>
>
>
> I'm new to the lucene. I downloaded lucene 2.4.1.
>
>
>
> I have one xml file which contains few special characters like 'å', 'ø,' °'
> etc.(these are Danish language elements)
HI,
I'm new to the lucene. I downloaded lucene 2.4.1.
I have one xml file which contains few special characters like 'å', 'ø,' °'
etc.(these are Danish language elements).
How can I search these things.
Uday Kumar Reddy Maddigatla
Software Engin
h...@all
I'd like to do a fuzzy search with german special characters. For instance I
want to query for "müller", but also terms like "mueller" should be respected,
as ü can also be written as ue. How could this be done? At index creation time,
I could convert ü t
ds in the hibernate POJO-classes. But as far as I know
> this shouldn't change things that much because I can use the core Lucene.
>
> And I've used Luke already and the indexed special characters are
> represented as "¡"(¡) and "¿" (¿) in the index.
>
>
hings that much because I can use the core
Lucene.
And I've used Luke already and the indexed special characters are
represented as "¡"(¡) and "¿" (¿) in the index.
But the analyzer should have nothing to do with the problem currently
because the problem is that, thos
d be helped a lot be getting a copy of Luke and examining your
index
to see exactly what's been indexed, it'll reveal a lot.
Best
Erick
On Tue, Nov 18, 2008 at 10:05 AM, Pekka Nykyri <[EMAIL PROTECTED]>wrote:
> Hi!
>
> I'm having problems with entities including specia
Hi!
I'm having problems with entities including special characters (Spanish
language) not getting indexed.
I haven't been able to find the the reason why some entities get indexed
while some don't.
I have 3 fields that (currently) hold the same value. The value for the
fi
out escape char //NOT OK
> Q5: key:"hatti ghoda (e-mail)" //without escape char //OK
> Q6: key:"hatti ghoda \(e\-mail\)" //with escape char //OK
>
> And with .NET API the behaviour is different.
> Q1 doesnt work
> Q2 works
> Q3 works
> Q4 throws except
Q2 works
Q3 works
Q4 throws exception
Q5 works
Q6 doesnt work
Any suggestions?
--
View this message in context:
http://www.nabble.com/Issues-with-Special-Characters-tp19511428p19515241.html
Sent from the Lucene - Java Users mailing list archive at Nabble.com.
t WhitespaceTokenizer
> > produces by getting a copy of Luke and looking at your index. Or writing
> > a really simple test program that prints out tokens.
> >
> > At the bottom of this page is a list of special characters for escaping:
> > http://lucene.apache.o
r the questions about what WhitespaceTokenizer
> produces by getting a copy of Luke and looking at your index. Or writing
> a really simple test program that prints out tokens.
>
> At the bottom of this page is a list of special characters for escaping:
> http://lucene.apache.org/j
You can easily answer the questions about what WhitespaceTokenizer
produces by getting a copy of Luke and looking at your index. Or writing
a really simple test program that prints out tokens.
At the bottom of this page is a list of special characters for escaping:
http://lucene.apache.org/java
Hi there,
I am using WhiteSpaceAnalyser to index documents. I have used this because i
need to split tokens based on space only. Also Tokensized=true
While indexing what does it do with special characters like + - && || ! ( )
{ } [ ] ^ " ~ * ? : \, will these characters be index
It was, after all an XML issue, the servlets creating the content that was
being indexed were not sending UTF but the XML declaration stated the code
WAS UTF, so it really was not a Lucene issue after all. Thanks for all the
help.
On Thu, Aug 21, 2008 at 6:18 PM, Juan Pablo Morales
<[EMAIL PROTECT
You are right, it does work. I'll look into my example to see where the
difference is.
On Thu, Aug 21, 2008 at 5:30 PM, Grant Ingersoll <[EMAIL PROTECTED]>wrote:
> Here's a unit test:
> import junit.framework.TestCase;
> import org.apache.lucene.analysis.snowball.SnowballAnalyzer;
> import org.ap
ablo Morales wrote:
I have an index in Spanish and I use Snowball to stem and analyze
and it
works perfectly. However, I am running into trouble storing (not
indexing,
only storing) words that have special characters.
That is, I store the special character but the it comes garbled when
I am running into
> > trouble storing (not indexing, only storing) words that
> > have special characters.
> >
> > That is, I store the special character but the it comes
> > garbled when I read it back.
> > To provide an example:
> >
> > String content =
Hola Juan,
On 08/21/2008 at 1:16 PM, Juan Pablo Morales wrote:
> I have an index in Spanish and I use Snowball to stem and
> analyze and it works perfectly. However, I am running into
> trouble storing (not indexing, only storing) words that
> have special characters.
>
> Th
I have an index in Spanish and I use Snowball to stem and analyze and it
works perfectly. However, I am running into trouble storing (not indexing,
only storing) words that have special characters.
That is, I store the special character but the it comes garbled when I read
it back.
To provide an
Steven A Rowe wrote:
On 08/11/2008 at 2:14 PM, Chris Hostetter wrote:
Aravind R Yarram wrote:
can i escape built in lucene keywords like OR, AND aswell?
as of the last time i checked: no, they're baked into the grammer.
I have not tested this, but I've read somewhere on t
You can simply change your input string to lowercase before passing it
to the analyzers, which will give you the effect of escaping the boolean
operators. (I.E you will now search on and or and not) Remember
however that these are extremely common words, and chances are high that
you are remo
On 08/11/2008 at 2:14 PM, Chris Hostetter wrote:
> Aravind R Yarram wrote:
> > can i escape built in lucene keywords like OR, AND aswell?
>
> as of the last time i checked: no, they're baked into the grammer.
I have not tested this, but I've read somewhere on this list that enclosing OR
and AND
: can i escape built in lucene keywords like OR, AND aswell?
as of the last time i checked: no, they're baked into the grammer.
(that may have changed when it switchedfrom a javac to a flex grammer
though, so i'm not 100% positive)
-Hoss
-
can i escape built in lucene keywords like OR, AND aswell?
Regards,
Aravind R Yarram
Chris Hostetter <[EMAIL PROTECTED]>
08/06/2008 07:05 PM
Please respond to
java-user@lucene.apache.org
To
java-user@lucene.apache.org
cc
Subject
Re: escaping special characters
:
: String escapedKeywords = QueryParser.escape(keywords);
: Query query = new QueryParser("content", new
: StandardAnalyzer()).parse(escapedKeywords);
:
: this works with most of the special characters like * and ~ except \ . I
: can't do a search for a keyword like "ho\w"
Hi,
I followed the following procedure to escape special characteres.
String escapedKeywords = QueryParser.escape(keywords);
Query query = new QueryParser("content", new
StandardAnalyzer()).parse(escapedKeywords);
this works with most of the special characters like * and ~ except \ .
; inserted between 'L' and ''article'.
>
> Thanks,
>
> Leire
>
> -Message d'origine-
> De: Leire Urcelay [mailto:[EMAIL PROTECTED]
> Envoyé: lundi, 5. novembre 2007 13:02
> À: java-user@lucene.apache.org
> Objet: blank space bef
ECTED]
Envoyé : lundi, 5. novembre 2007 13:02
À : java-user@lucene.apache.org
Objet : blank space before special characters
Hello,
I have the following problem with my lucene index.
When indexing fields containing special characters (like &), a blank space
is inserted before the special
Hello,
I have the following problem with my lucene index.
When indexing fields containing special characters (like &), a blank space
is inserted before the special character. For example: the content
"L'article" is indexed as "L '" (with a blank space b
Hi,
In case this is of help to others:
Crux of problem:
I wanted numbers and characters such as # and + to be considered.
Solution:
implement a LowercaseWhitespaceAnalyzer and a
LowercaseWhitespaceTokenizer.
Tom
===
Diagnostic
ed
>
> Enter Querystring:
> (c++ and c#)
> Searching for: c c
>
> Enter Querystring:
> (c\+\+ and c\#)
> Searching for: c c
>
> -Original Message-
> From: Erick Erickson [mailto:[EMAIL PROTECTED]
> Sent: 23 September 2007 14:47
> To: java-user@lucene.a
ailto:[EMAIL PROTECTED]
Sent: 23 September 2007 14:47
To: java-user@lucene.apache.org
Subject: Re: Escaping special characters
Default as in StandardAnalyzer? "Stuff happens" with StandardAnalyzer.
I'd recommend that you use something really simple like
WhitespaceAnalyzer for something li
7;ll try.
>
> Tom
>
> -Original Message-
> From: Karl Wettin [mailto:[EMAIL PROTECTED]
> Sent: 23 September 2007 10:39
> To: java-user@lucene.apache.org
> Subject: Re: Escaping special characters
>
>
> 23 sep 2007 kl. 10.53 skrev Tom Conlon:
>
> >
---Original Message-
From: Karl Wettin [mailto:[EMAIL PROTECTED]
Sent: 23 September 2007 10:39
To: java-user@lucene.apache.org
Subject: Re: Escaping special characters
23 sep 2007 kl. 10.53 skrev Tom Conlon:
>
> Unless I'm missing something, according to:
>
> http:/
23 sep 2007 kl. 10.53 skrev Tom Conlon:
Unless I'm missing something, according to:
http://lucene.apache.org/java/docs/queryparsersyntax.html#Escaping%
20Spe
cial%20Characters
I should be able to search for C++ and C# using something like:
C\+\+ and C\#.
That is correct.
This doesn't
Hi,
Unless I'm missing something, according to:
http://lucene.apache.org/java/docs/queryparsersyntax.html#Escaping%20Spe
cial%20Characters
I should be able to search for C++ and C# using something like:
C\+\+ and C\#.
This doesn't work.
I apologise if this sort of question has been asked man
direct writing term
> (CONTENT:"emp-id") will not.
>
> Andy
>
> -Original Message-
> From: SK R [mailto:[EMAIL PROTECTED]
> Sent: Wednesday, June 20, 2007 5:24 PM
> To: java-user@lucene.apache.org
> Subject: zero termfreq for some search strings with
ED]
Sent: Wednesday, June 20, 2007 5:24 PM
To: java-user@lucene.apache.org
Subject: zero termfreq for some search strings with special characters
Hi,
I'm using standard tokenizer for both indexing and searching
process.Myindexed value is like "emp-id Aq234 kaith creating document
for search"
Sent: Wednesday, June 20, 2007 5:24 PM
To: java-user@lucene.apache.org
Subject: zero termfreq for some search strings with special characters
Hi,
I'm using standard tokenizer for both indexing and searching
process.Myindexed value is like "emp-id Aq234 kaith creating document
for s
Hi,
I'm using standard tokenizer for both indexing and searching
process.Myindexed value is like "emp-id Aq234 kaith creating document
for search".
I can get search results for the query CONTENT:"emp-id" by using hits =
indexSearcher.search(*query*).
But if I try to get termfrequency of t
Thanks for the replies on my question.
In the end I've taken the StandardAnalyser grammar, modified it and
generated a new analyser with JavaCC. Seems to be working a treat!
Adrian
On 11 Aug 2006, at 14:32, Erik Hatcher wrote:
On Aug 11, 2006, at 1:23 AM, Martin Braun wrote:
Hello Adrian
On Aug 11, 2006, at 1:23 AM, Martin Braun wrote:
Hello Adrian,
I am indexing some text in a java object that is "%772B" with the
standard analyser and Lucene 2.
Should I be able to search for this with the same text as the
query, or
do I need to do any escaping of characters?
Besides Luk
Hello Adrian,
>> I am indexing some text in a java object that is "%772B" with the
>> standard analyser and Lucene 2.
>>
>> Should I be able to search for this with the same text as the query, or
>> do I need to do any escaping of characters?
Besides Luke there are the AnalyzerUtils from the LIA
1 - 100 of 130 matches
Mail list logo