: up into 1 segment of 1000 docs. So far so good. However if I add 900 docs to
: the index and then add 100 docs in a second pass (closing the IndexWriter
: between the two runs), I get 10 segments on disk but they are not merged.
: Also, if I add 10 docs with mergeFactor=1, I get 10 segments.
:
:
It's not easy for us to have the installer team do that (it's a big
product).
We actually found IndexWriter.AddIndexes to do almost what we wanted:
just pass in empty list of IndexReader's, and customize the mergedName
to what we want it to be; I made extended IndexWriter to have this new
method.
Look at PerFieldAnalyzerWrapper. It allows you to use different
analyzers on different fields during the query parsing phase.
But I wouldn't go there if you don't have to. I suspect you'll spend a
LOT of time tracking down errors in your use of a mixed case index.
If for no other reason than your
Hi All,
I'm running Lucene 1.4.3 and have noticed that the mergeFactor setting is
not working as I would expect. To provide a little background - this all
started as an attempt to reduce the number of segments in the index
directory. After several months of use in a high update environment, I end
I have an index with both fields that are case sensitive and insensitive. I
am trying to use a QueryParser to accept query from end users for
searching. The default behavior of QueryParser is to lowercase the prefix
text to create the PrefixQuery. So wildcard search on the case sensitive
fields
If I where going to do this I would create special autocomplete index. I
would add each name that I wanted to be able to autocomplete on to this
index. You can tokenize if you want (the lowercase filter) but do not use
something that will split up words...you want "Appleton Bakery Café" to go
in a
Hi. I'm trying to design a proper index and query mechanism for looking up a
business listing using an Ajax-style autocompleting text box. While I have
gotten "versions" to work, I'm wondering what the optimal approach is.
Someone may be looking for "Appleton Café." That listing might be
Is there an example on how to do this (or can someone provide a sample code)?
I've been looking at the Lucene API, particularly at IndexCommitPoint and
IndexDeletionPolicy interfaces, but have no idea what I need to do in order
to implement a live backup. Any help would be appreciated.
LH
M
Thank you, I got the natch plugin, and it is working great
-Original Message-
From: Otis Gospodnetic [mailto:[EMAIL PROTECTED]
Sent: Thursday, May 03, 2007 4:17 PM
To: java-user@lucene.apache.org
Subject: Re: Language detection library
LingPipe - commercial unless your data/product/serv
A minor clarification. What I'm measuring is the size of the index
before I index a doc, and the size after (this involves "expert" usage
of my IndexWriter subclass). Using those two figures, I get
a "bloat" factor. That is, how much the memory was consumed by
this document in the index. Say i
First, are you sure you're free memory calculation is OK? Why not
just use freeMemory?
I _think_ my calculation is ok - my reasoning:
Runtime.maxMemory() - Amount of memory that can be given to the JVM -
based on -Xmx
Runtime.totalMemory() - Amount of memory currently owned by the JVM
Runtime.
Thanks, youve sparked a thought Ive now realised I was calling the wrong
method i had another method (simpleSearch) that used the Query parser to
parse the search string which I should have been using
cheers Paul
Koji Sekiguchi wrote:
Hi Pail,
Try WildcardQuery instead of TermQuery as follow
Hi Pail,
Try WildcardQuery instead of TermQuery as follows:
//Search on column columnId for value searchstring
WildcardQuery query = new WildcardQuery(new
Term(String.valueOf(columnId),searchstring));
Thank you,
Koji
-
I have indexed some records in a JTable , and I am trying to return all
records where the value in a particular column starts with a particular
value(http://musicbrainz.org:), but I get no matches. If I match for an
exact values it works fine, Im stumped as to why this would be the case
//Work
14 matches
Mail list logo