Hi all.
This is something I have been wondering for a while but can't find a
good answer by reading the code myself.
If you have a query like this:
( field:Value1 OR
field:Value2 OR
field:Value3 OR
... )
How many TermEnum / TermDocs scans should this execute?
(a) One per clau
We actually have that implementation thinking of contributing.
-John
On Sat, Apr 4, 2009 at 3:00 PM, Grant Ingersoll wrote:
> Note that I believe with some work (marking the "zones" during analysis),
> one can accomplish this with Spans without the field creation problem that
> John mentions.
>
If you have luke, you should be able to submit your query and use
the explain functionality to gain some insights into what the query
actually looks like as well
Best
Erick
On Mon, Apr 6, 2009 at 5:24 PM, Ariel wrote:
> Well I have luke lucene, the index has been build fine.
> The field whe
Well I have luke lucene, the index has been build fine.
The field where I am searching is the content field.
I am using the same analyzer in query and indexing time: SnowBall English
Analyzer.
I am going to submit later the snippet code.
Regards
Ariel
On Mon, Apr 6, 2009 at 4:37 PM, Erick Eric
I am sorry Nittin, I may have injected you the doubt about this...
semantic-vectors is a project based on Lucene:
http://code.google.com/p/semanticvectors/
you probably want to look there and ask questions on the forum there.
paul
Le 06-avr.-09 à 22:45, Richard Marr a écrit :
Hi Nitin,
Hi Nitin,
I'm assuming you're asking about Latent Semantic Indexing and similar.
This may not be the best place to ask about this. Not sure where else
to suggest though.
If I understand your quesion correctly, the basic idea is that you
take a document (usually text but it could also be an image,
We really need some more data. First, I *strongly* recommend you
get a copy of Luke and examine your index to see what is
*actually* there. Google "lucene luke". That often answers
many questions.
Second, query.toString is your friend. For instance, if the query
you provided below is all that you'
Hi every body:
Why when I make a query with this search query : "the fool of the hill"
doesn't appear documents in the search results that contains the entire
phrase "the fool of the hill" and it does exist documents that contain that
phrase, I am using snowball analyzer for English ???
Could you
Hey there,
Does de function doc.setBoost(x.y) accept negative values or values minor
than 1?? I mean... it compile and doesn't give errors but the behabiour is
not exactly what I was expecting.
In my use case I have the field title... I want to give very very low
relevance to the documents witch t
6 apr 2009 kl. 15.47 skrev Lebiram:
I am thinking of adding search filters to my application thinking
that they would more efficient.
Can anyone explain what lucene does with search filters?
Like, what generally happens when calling search()
A filter is a bitset, one bit per document in t
6 apr 2009 kl. 14.59 skrev Glyn Darkin:
Hi Glyn,
to be able to spell check phrases
E.g
"Harry Poter" is converted to "Harry Potter"
We have a fixed dataset so can build indexes/ dictionaries from our
own data.
the most obvious solution is index your contrib/spell checker with
shingles. T
Hi all,
Just to let everyone know... I'm organising (if you can call it that)
an informal London meet-up in the presence of alcohol to talk about
Lucene, and search in general. We're not planning on having any formal
talks or presentations.
We've just finished a big-ish search project that uses L
... erm.. I'm still not quite sure what you are talking about.
But what you are trying to do, really isn't that hard. Here's some
sample code that should get you to where you want to be:
During document creation time do something like this:
doc.add(new Field("data",
/da
What I've done is to put copies of the same field, built with different
analyzers, and later use a MultiFieldQueryParser matching all fields.
eg: "name", "name_phonetic", "name_keyword", ad nauseum
To define which analyzer will go to which field, use PerFieldAnalyzerWrapper
On Mon, Apr 6, 2009 a
This really doesn't make sense. KeywordAnalyzer will NOT
tokenize the input stream. StandardAnalyzer WILL tokenize
the input stream. I can't imagine what it means to do both at
the same time.
Perhaps you could give us some examples of what your desired
inputs and outputs are we could steer you in
Hi,
I want to add multiple Analyzer on single field. I want properties of
KeywordAnalyzer, SimpleAnalyzer, StandardAnalyzer, WhiteSpaceAnalyzer. Is there
any easy way to have all analyzer bundled on single field.
Regards,
Allahbaksh
CAUTION - Disclaimer *
T
Hi All,
I am thinking of adding search filters to my application thinking that they
would more efficient.
Can anyone explain what lucene does with search filters?
Like, what generally happens when calling search()
This problem is filed at:
https://issues.apache.org/jira/browse/LUCENE-1489
You may want to take a look at LUCENE-1522 for highlighting N-gram tokens:
https://issues.apache.org/jira/browse/LUCENE-1522
Koji
ito hayato wrote:
> Hi All,
> My name is Hayato.
>
> I have a question for Highlighter
Hi Guys,
We are using Lucene for our search and have implemented "Did you
mean?" with the built in Spell Checker. However what we really need is
to be able to spell check phrases
E.g
"Harry Poter" is converted to "Harry Potter"
We have a fixed dataset so can build indexes/ dictionaries from our
Hi All,
My name is Hayato.
I have a question for Highlighter.
I indexed following text and use Tokenizer.
text : abracadabra
Tokenizer: NGramAnalyzer
and requested following query
query: ab
expected result and actual result is following
expect : abracadabra
actual : abracadabra
20 matches
Mail list logo