Walt,
I am no expert, but it sounds like you need to associate many
dates to a single record. Can this be handled as you would a synonym?
Basically add a token at the same offset as the row itself? i.e. you
would have a record that would also have a date field that has 3 offsets
that woul
is.search(query, hc);
-Original Message-
From: Aigner, Thomas [mailto:[EMAIL PROTECTED]
Sent: Thursday, December 07, 2006 1:36 PM
To: java-user@lucene.apache.org
Subject: RE: Reading Performance
Thanks Grant and Erik for your suggestions. I will try both of them and
let you know i
have it in the loop?
On Dec 7, 2006, at 1:14 PM, Aigner, Thomas wrote:
>
>
>
>
> Howdy all,
>
>
>
> I have a question on reading many documents and time to do this.
> I have a loop on the hits object reading a record, then writing it
> to a
> file
Howdy all,
I have a question on reading many documents and time to do this.
I have a loop on the hits object reading a record, then writing it to a
file. When there is only 1 user on the Index Searcher, this process to
read say 100,000 takes around 3 seconds. This is slow, but can
I have seen numerous posts on warming up a searcher, but was wondering
if someone could post their code that would spin off another thread to
warm up a searcher, then switch to the new one when it is warmed up?
-
To unsubscri
objects aren't
properly disposed.
-drj
On 10/26/06, Aigner, Thomas < [EMAIL PROTECTED]> wrote:
>
> Howdy all,
>
> I have a issue with java running out of memory after the
search
> has been running for a while. We are using 1.9.1 release and I check
> the indexrea
Howdy all,
I have a issue with java running out of memory after the search
has been running for a while. We are using 1.9.1 release and I check
the indexreader's version to determine if I need to get a new searcher
for searches (so I pick up any changes to the index). I am seeing jumps
i
Howdy all,
I am having a performance issue. When I do a search for items,
getting more information takes a long time.
Ex. If there are 1M hits (I know, why look for that many or even allow
it, but let's say we return 1M hits). When the user wants to see the
last 25, it takes a LONG time
I believe that the files are actually deleted from lucene when the
optimize is run.
-Original Message-
From: Dan Armbrust [mailto:[EMAIL PROTECTED]
Sent: Monday, February 13, 2006 12:27 PM
To: java-user@lucene.apache.org
Subject: When do files in 'deleteable' get deleted?
If I am using l
Anyone have a comment on the below message?
-Original Message-
From: Aigner, Thomas
Sent: Wednesday, February 08, 2006 11:50 AM
To: java-user@lucene.apache.org
Subject: 1.9 lucene version
Hello all,
I have a couple of questions for the community about the 1.9
Lucene version
Hello all,
I have a couple of questions for the community about the 1.9
Lucene version. As I understand it, this has not been released and I
can't find an approximate date for release (I know you can download the
development version and compile it). I see a nightly build going on
(http:/
deas from that, you can impliment a much faster version of
PrefixQuery that doesn't score documents based on term frequency ...
which
may be ok depending on your needs.
: Date: Mon, 30 Jan 2006 13:51:15 -0500
: From: "Aigner, Thomas" <[EMAIL PROTECTED]>
: Reply-To: java-
I am curious what would be the difference between searching for a number
verses a character.
I have a large index consisting of a few fields (So index would look
something like: " 123123123 my description my catalog"
Searching for 12* is much slower than searching for de*
I don't have a
I'm still a bit new to Lucene, but the way I think you can handle it is
using a Boolean query.
//Create a bool Query
BooleanQuery boolQuery = new BooleanQuery();
Loop through each query and add it to the boolQuery
//Addend the boolean queries together Using the And syntax
boolQuery.add(queryX, t
s that the thread has X Mb of memory,
it's really the same physical block as all the others.
You see this all the time in a Tomcat app server box, where each Http
Connector is a thread, and appears as it's own process.
cheers,
Paul Smith
On 17/01/2006, at 7:11 AM, Aigner, Thomas wrote:
Hi all,
Is anyone experiencing possible memory problems on LINUX with
Lucene search? Here is our scenario, we have a service that lives on
LINUX that takes all incoming request through a port and does the
search. Only 1 IndexSearcher is instantiated to do this from our
service. When I r
Is there an error list for lucene errors? I am creating a project that
will do something specific for certain lucene errors that occur (Lock
file out there, etc).
Thanks,
Tom
-
To unsubscribe, e-mail: [EMAIL PROTECTED]
For
the last time you checked. The index's version number
changes whenever the index is updated.
Peter
> -Original Message-
> From: Aigner, Thomas [mailto:[EMAIL PROTECTED]
> Sent: Monday, November 21, 2005 3:48 PM
> To: java-user@lucene.apache.org
> Subject: Lucene
Hi all,
Is there an index changed event that I can jump on that will
tell me when my index has been updated so I can close and reopen my
searcher to get the new changes?
I can't seem to find the event, but see some tools that might
accomplish this (DLESE DPC software components?).
ber of segments, and hence speed searching.
-Yonik
Now hiring -- http://forms.cnet.com/slink?231706
On 11/16/05, Aigner, Thomas <[EMAIL PROTECTED]> wrote:
> Howdy all, have a quick question for you...
>
> I am seeing quite a difference between optimized index and one that is
>
Howdy all, have a quick question for you...
I am seeing quite a difference between optimized index and one that is
not optimized. I have read a few papers that say that it shouldn't
matter, but I am seeing 7X speed or better when the index is optimized.
Is it possible that I am creating the origi
wapped to a reader for reads. By keeping this
index
small speeds up the constant opening and closing. Searching is
straightforward using the MultiSearcher.
I don't know anything about the lock problem - it's not something I've
ever
seen (I'm using 1.4.3).
Regards
Pa
Howdy all,
I am having a problem with inserting/updating records into my
index. I have approximately 1.5M records in the index taking about 2.5G
space when optimized.
If I want to update 1000 records, I delete the old item and insert the
new one. This is taking a LONG time to accomplis
I have seen quite a few posts on using the 1.9 dev version for
production uses. How stable is it? Is it really ready for production?
I would like to use it.. but I never ever put beta packages in
procution.. but then again.. I'm always dealing with Microsoft :)
Tom
-Original Message-
Fro
Another way around it is to increase the max clause count.
//Setting the clause Count
BooleanQuery.setMaxClauseCount(int);
Can use maxint or some number smaller.. When I set this high, I have had
to set the java pool higher for memory as well.
Tom
-Original Message-
From: Sharma, Siddh
Hi all,
Was just wondering if anyone has come across this or if I'm doing
something wrong here. On initial load of my index, I can close the
writer and delete an entry and then update an entry, then open the
writer again and go on to the next entry etc. Then while searching,
everything t
Tokens created:
mailto:[EMAIL PROTECTED]
Sent: Wednesday, October 05, 2005 9:27 AM
To: java-user@lucene.apache.org
Subject: Re: Optimization
On Oct 5, 2005, at 9:05 AM, Aigner, Thomas wrote:
> Have a question.. Is there any obvious things that can be done
> to help speed up
Howdy all,
Have a question.. Is there any obvious things that can be done
to help speed up query lookups especially wildcard searches (i.e.
*lamps).
We have created a server application on a linux box that listens
to a socket and processes searches as they come in. We thought th
What seems to be working for me is a punctuation filter that removes / -
_ etc and makes the token without them. Then "most" of the time the
word XYZZZY_DE_SA0001 will be tokenized as XYZZZYDESA0001. For this to
work, you will have to use the same punctuation filter on the strings
before you sear
Please forgive my jumping on this thread, but I have a similar issue. I
have a server process on Linux that creates the java process (java
-Xms256m -Xmx512m -jar Suchmaschine.jar). The problem is that after the
processing is done, the memory is retained. Is there a collection
argument that would
Howdy all,
I am running some tests against our search. I have created a
script that will run 30 simultaneous executions of 25 searches each.
The 25 searches are randomized but the same data is used. I am
consistently getting "Lock obtain timed out:
Lock@/tmp/lucene-77053e9cdda0a530f729e7
[mailto:[EMAIL PROTECTED]
Sent: Tuesday, August 02, 2005 1:00 PM
To: java-user@lucene.apache.org
Subject: Re: Question on Wildcard
On Monday 01 August 2005 22:22, Aigner, Thomas wrote:
> The filter does not seem to be called when I search for 56-TXT* even
> though I use my extended analyzer whi
Hi all,
I have a Punctuation Filter that filters out certain criteria
(ex. - and /). This filter is called whenever I search for normal
items.. Search 56-TXT
The filter does not seem to be called when I search for 56-TXT* even
though I use my extended analyzer which has the filter in it s
Hi all,
I am creating a punctuation filter to filter certain punctuation
out of the token stream. I am getting a "The field t.termText is not
visible" error. I'm not sure what I would need to include to make this
property visible (I am still new to Lucene and Java for that matter). I
copi
#x27;t matter.
Two other issues are code/data size (ICU can be big) and the
performance hit while indexing documents.
-- Ken
>Aigner, Thomas wrote:
>
>>Hello all,
>>
>> I am VERY new to Lucene and we are trying out Lucene to see if
>>it will accomplish the v
g my SynonymnAnalyzer for some aliases to build the
index and the SnowballAnalyzer to query the index (nice stemming in it)
Tom
-Original Message-
From: Chris D [mailto:[EMAIL PROTECTED]
Sent: Tuesday, June 28, 2005 2:41 PM
To: java-user@lucene.apache.org
Subject: Re: Indexing puncutati
Hello all,
I am VERY new to Lucene and we are trying out Lucene to see if
it will accomplish the vast majority of our search functions.
I have a question about a good way to index some of our product
description codes. We have description codes like 21-MA-GAB and other
punctuatio
37 matches
Mail list logo