? (Luke has a "reconstruct & edit" button which does not seem
to work on deleted documents, if it ever was designed to undelete a
document).
Thank you!
Regards,
Martin
pgpBd0XCCpK5b.pgp
Description: PGP signature
I need to return all of the keys for a certain field, essentially
"fieldName:*".This causes a ParseException / lexical error
Encountered: "*" (42), after : ""
I understand why this fails, WildCard prevent this to keep the results
manageble. In my case the number of results will always be man
non-zero.
Is there any relation between the value of the explanation and the score?
Thanks!
Regards,
Martin Kobele
-
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]
Hi Chris,
thank you very much. Let's hope the developers can get this bug fixed soon :)
Regards,
Martin
On Wednesday 03 May 2006 17:17, Chris Hostetter wrote:
> : Now we want to use Search.search(Query, HitCollection).
> : HitCollection adds the document to a BitSet as soon as t
so the score is always correct, just the Explain can return values greater
than 0 although the query did not match the document due to a bug in Explain?
Are all version affected of that? I use lucene 1.9.1.
Thanks!
Regards,
Martin
On Thursday 04 May 2006 09:27, Yonik Seeley wrote:
> On 5/4
awesome, thank you very much!
Regards,
Martin
On Thursday 04 May 2006 17:15, Chris Hostetter wrote:
> : so the score is always correct, just the Explain can return values
> : greater than 0 although the query did not match the document due to a
> : bug in Explain? Are all version af
Hi,
is writing into an index with several threads safe?
I have a static IndexWriter and all the threads use that one to do an
addDocument(Document).
Thanks!
Martin
--
Martin Kobele
Software Engineer
-
To unsubscribe, e-mail
ouch, I missed it there. I always thought that the FAQ is really good! thank
you! sorry for the noise.
Martin
On Friday 05 May 2006 16:04, Otis Gospodnetic wrote:
> Yes, and I believe it's in the FAQ.
>
> Otis
>
> - Original Message
> From: Martin Kobele <[EM
!
Regards,
Martin
-
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]
On Wednesday 17 May 2006 09:09, karl wettin wrote:
> On Wed, 2006-05-17 at 09:02 -0400, Martin Kobele wrote:
> > How can I check whether a document was actually added?
> > Is there another way to monitor whether the doc was added?
>
> Luke is your friend when it comes
On Wednesday 17 May 2006 12:03, Yonik Seeley wrote:
> On 5/17/06, Martin Kobele <[EMAIL PROTECTED]> wrote:
> > several threads add document to the same index.
> > As noted in the FAQ, adding documents is thread safe.
> > (I even synchronize my static index writer wh
with the create flag = true;
Do you want me to delete the directory every time?
Martin
>
>
> -
> To unsubscribe, e-mail: [EMAIL PROTECTED]
> For additional commands,
I've deployed 1.9.1 and 2.0.0 to the ASF Maven 2 repo and requested a sync
with ibiblio, so hopefully they'll be available soon.
--
Martin Cooper
On 6/10/06, Otis Gospodnetic <[EMAIL PROTECTED]> wrote:
It's really just a matter of putting the Jars in the appropr
get two words as a synonym at the same increment
position during indexing? What will happen with a phrase search?
Does anybody know existing approaches for this?
thanks in advance,
Martin
-
To unsubscribe, e-mail: [EMAIL PR
tion of the
> index"? If you haven't encountered a bottleneck that's forcing you into
Perhaps for privacy reasons? that only specific users should be able to
search the whole index.
Is there a best practice approach to realize this?
thx,
martin
--
Universitaetsbibliothek Heidelb
800 ms.
Do you use a kind of sorting? I had this issue with SortField.STRING, it
takes a lot of time and memory to build the sort cache of all terms in
the field (of type String). So if the searcher needs to rebuild this
with each query it takes a while...
hth,
martin
> Here also i noticed
[EMAIL PROTECTED] schrieb:
> hi,
>
> my problem is that i am using mysql db in which one table is
> present and i want index each row in the table and then search
>
> plz reply
>
> how this can be done?
http://wiki.apache.org/jakarta-lucene/LuceneFAQ
How can I use Lucene to index a database?
Co
check-contrib from the
SVN-repository, but it seems to me that there are no ant-targets for that?
thanks,
martin
--
Universitaetsbibliothek Heidelberg Tel: +49 6221 54-2580
Ploeck 107-109, D-69117 Heidelberg Fax: +49 6221 54
Hi,
since I use many wildcards, I get the exception, that the number of boolean
queries exceeds the default value (1024).
I could simply increase the value to like 10,000 or something
What would be the trade-off of using a high max value?
Thanks!
Martin
thank you!
Martin
On Friday 30 June 2006 14:21, Erick Erickson wrote:
> The tradeoff is that it'll blow up eventually . I have a really hard
> time trusting increasing the clause count, since eventually, more
> data/terms/something will blow my limit again.
>
> You pr
XXVIII, 307, |<180>| S. :
zahlr. Ill.
ISBN 0-8240-6868-8
Any Ideas?
thanks in advance,
martin
Davide schrieb:
> mark harwood wrote:
>> Does your index have only the one document?
>>
>> MoreLikeThis will only generate queries with terms that occur in more than
>&
he example doc which may produce tokens that do not match those of the
> indexed content. Use setAnalyzer() to ensure they are in sync.
>
>
>
>
> - Original Message
> From: Martin Braun <[EMAIL PROTECTED]>
> To: java-user@lucene.apache.org
> Sent: Frid
ve found in the List-Archives so far is that I have to use
Filters for my search.
Does anybody knows where to find some Source Code, to get an Idea how to
implement this?
I think that's a useful property for a search engine, so are there any
contributions for Lucene
hi miles,
thanks for the response.
I think I didn't explain my Problem good enough.
The harder problem for me is how to get the proposals for the
refinement? I have a date-range of 16xx to now, for about 4 bn. docs.
So the number of found documents could be quite large. But the
distribution of t
hi herbert,
>> WhitespaceAnalyzer looks brutal. Is it possible that I keep
>> StandardAnalyzer and at the same time to tell the parser to keep a
>> list of chars during indexing?
Perhaps it would be sufficient to use the WhitespaceAnalyzer and keep
StandardAnalyzer for the other fields by using a
Yonik Seeley schrieb:
> On 7/23/06, karl wettin <[EMAIL PROTECTED]> wrote:
>> I'm want to filter words with a dash in them.
>>
>> ["x-men"]
>> ["xmen"]
>> ["x", "men"]
>>
>> All of above should be synonyms. The problem is ["x", "men"] requiring a
>> distance between the terms and thus also matching
drawbacks with this?
1: [the]
2: [x] [xmen]
3: [men] [xmen]
4: [story]
thanks,
martin
-
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]
t; So, be prepared to suffer quite a bit. ;)
>
regards,
martin
-
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]
quot;west (berlin westberlin)"
with +arbeiterjugend +westberlin I get the result.
It seems that the synonyms don't work with the query. How do you solve
this in Solr? Do I have to build a TermQuery?
thanks in advance,
martin
ot;title");
}
});
hth,
martin
-
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]
andard.StandardAnalyzer:
[772b]
1: [772b:1->5:]
1: [772b]
Analzying "%772B"
org.apache.lucene.analysis.KeywordAnalyzer:
[%772B]
1: [%772B:0->5:word]
1: [%772B]
hth,
martin
---
hello ould,
sid'ahmed schrieb:
> Hello,
> I indexed my document but, Can I search for an address web, it returns
> me no result,
> and when I search the same address with a query like "http*" it returns
> me a result,
It depends on which analyzer you use:
the StandardAnalyzer will do this with
the deleteDocuments function (normally you won't
know the lucene's index document-id for the specific doc). The downside
of this is that the lucene-index document-id will change.
hth,
martin
-
To unsubscribe, e-mail: [EM
er Problem? How do I get all tokens?
tia,
martin
-
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]
Hi,
I have read it several times that sometimes it is practical to warm up a
searcher. What exactly does that mean? Would I simply perform a search
for "foo bar" or something like that? What happens inside the searcher so
that it becomes 'warmed up'?
Thank y
document, would cause many redundancies, e.g. thousands of docs with
test in it.
Does anybody have programmed something similar, or is it just not a
reasonable idea?
thanks,
martin
-
To unsubscribe, e-mail: [EMAIL PROTECTED]
For
nce won't cut it, because there are other
relevance factors affecting the ranking. Is there any good (easy &
efficient) way to do this?
Thanks,
Martin
-
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-
only entries that the user may view? Should the returned list
of results be post processed to filter out the invalid entries, and if
so how?
-Is 2 TB of data to large to be handled by Lucene?
-Has anybody implemented a similar type of application?
Thanks in advance
Martin
er - Anchor Men [mailto:[EMAIL PROTECTED]
Sent: 13 April 2005 13:56
To: java-user@lucene.apache.org
Subject: RE: Searching an NTFS File Server
You can use a filter with the IndexSearcher so that it removes all
'unwanted' results.
-Original Message-
From: Maher Martin
Sent: 13
the
field, and if so boosting the document by 5000. However, it doesn't seem
to have the desired effect.
Any ideas?
Thanks,
Martin
-
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]
I've got the book (which is great, btw). I used Luke to get explanations
of the results, but I don't see any boosts in the explanations.
Martin
On Thu, 2005-04-14 at 13:24 -0700, Otis Gospodnetic wrote:
> I'd look a the output of Explain to see how ranking score is calculated
eems
that this is not currently possible in Lucene (and please, correct me if
I'm wrong).
Thanks,
Martin
Gabe wrote:
I don't think you can do this continuously, but this
might be possible to do by setting up "regions" of how
far away the person is. I'm not using real lats
Thanks Mark for the tip. I've tried out JCIFS and it's just what the
doctor ordered. I've used NTLM SSO to authenticate the User against the
NTFS file server and then I can filter out the search results returned
based on the user's access rights.
I've completed the authentication and authorization
Thanks Mark for the tip. I've tried out JCIFS and it's just what the
doctor ordered. I've used NTLM SSO to authenticate the User against the
NTFS file server and then I can filter out the search results returned
based on the user's access rights.
I've completed the authentication and authorization
if anybody has
had the same problems and found already a solution for that!
Thanks for your help!
Take Care,
Martin
-
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]
pers can at least at an option, so that
reader.terms(new Term("henrie")) yields "henriette".
Anyway... I need a comletition which includes the Umlauts, so I can just
emit two tokens and be happy. That should do it, shouldn't it?
Best Regards,
Martin
P.S.: When I am do
Erlend,
try this code:
QueryParser q = new QueryParser("text", analyzer);
q.setOperator(QueryParser.DEFAULT_OPERATOR_AND);
Query query = q.parse(search);
Hits hits = isearcher.search(query);
Best,
Martin
[EMAIL PROTECTED] wrote:
Hi there!
I'm fairly ne
u then search, the "text" will give you access to all your data.
Anyway you should definitely get "luke". Browse the index you created!
Best,
Martin
[EMAIL PROTECTED] wrote:
Hi there Martin!
try this code:
QueryParser q = new QueryParser("text", analyzer);
in advance,
martin
-
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]
You can also use the
PerFieldAnalyzerWrapper to search in KeywordFields using QueryParser.
Best,
Martin
Vanlerberghe, Luc wrote:
Is your Analyzer aware that that particular field does not need to be
tokenized?
During indexation, if a field is passed that is passed as
tokenize=false, the
Hi everybody,
Has anyone tried to code a solution like Google's "Did you mean?" in
Lucene?
I would be very happy to hear your ideas, approaches, suggestions.
Best,
Martin
-
To unsubscribe, e-mail: [EMAIL
optimize() on the writer, most of the time I get OutOfMemory
(even with 800MB for the JVM).
What do I do wrong?
Best,
Martin
code snip start -
// optimize index
IndexReader[] readers = new IndexReader[1];
readers[0] = IndexReader.open(diri
elves often contain lots of information between the data
you want to match against.
Or is there some way of telling a proximity query to not cross field
boundaries?
And no, I have no idea what an opattor is either ... ;)
martin
-
/is regenerated from the url value
input.removeField("score");
server.add(input);
}
--
Ole-Martin Mørk
On Mon, Oct 5, 2009 at 11:15 AM, Simon Willnauer <
simon.willna...@googlemail.com> wrote:
> Did you change any boost values for URL field or document whil
this time. It might be that the index was
really small the first time the document was added. Could that affect the
fieldNorm value?
--
Ole-Martin Mørk
On Mon, Oct 5, 2009 at 11:39 AM, Simon Willnauer <
simon.willna...@googlemail.com> wrote:
>
>
> On Mon, Oct 5, 2009 at 11:31
I did not change the url. The length of the title was increased by 1, from
41 to 42 characters.
--
Ole-Martin Mørk
On Mon, Oct 5, 2009 at 12:39 PM, Karl Wettin wrote:
> sorry, I ment title.
>
> 5 okt 2009 kl. 11.57 skrev Simon Willnauer:
>
>
> Ole-Martin, did you mentio
That might be true. The document boost did not change, but maybe the field
boost changed. Is it possible to retrieve the field boost from solr?
--
Ole-Martin Mørk
On Mon, Oct 5, 2009 at 2:01 PM, Simon Willnauer <
simon.willna...@googlemail.com> wrote:
> I still guess that the document
Thanks. It might be that Nutch sets some values. I am not able to find
anything in the config files though.
We are using nutch' solrindex.
--
Ole-Martin Mørk
http://twitter.com/olemartin
http://flickr.com/olemartin
On Mon, Oct 5, 2009 at 2:28 PM, Simon Willnauer <
simon.willna...@google
d" would match
nothing, which is what I want.
Is this possible?
Cheers,
martin
-
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]
Hi,
I have few questions related to updateDocValues methods and usages, it
would be great if I can be helped
1) Is it possible to update a stored numeric field? I saw from the java-doc
that updateNumericDocValue is to update NumericDocValues. The fields
NumericDocValuesField aren't stored, if I w
101 - 160 of 160 matches
Mail list logo