Thank you.
> From: dawid.we...@gmail.com
> Date: Thu, 19 Jul 2012 13:34:26 +0200
> Subject: Re: RAM or SSD...
> To: java-user@lucene.apache.org
>
> Read this:
> http://blog.thetaphi.de/2012/07/use-lucenes-mmapdirectory-on-64bit.html
>
> Dawid
>
> On Thu, Ju
address
> > space allocated by the OS but no memory.
> >
> > On Wed, Jul 18, 2012 at 10:39 PM, Toke Eskildsen
> wrote:
> >> On Wed, 2012-07-18 at 17:50 +0200, Dragon Fly wrote:
> >>> If I want to improve performance, which of the following is better and
Hi,
If I want to improve performance, which of the following is better and why?
1. Buy a machine with a lot of RAM and use a RAMDirectory for the index.
2. Put the index on a solid state drive.
By the way, my index is about 30 GB. Thank you.
the values
> are read from the inverted index.
>
> Best
> Erick
>
> On Mon, Feb 27, 2012 at 8:04 AM, Dragon Fly wrote:
> >
> > Erick, what if the search returns 100,000 hits? I'm trying to avoid loading
> > a large number of documents from disk (i.e. a s
ooked at the Searcher.search variant
> that takes a Sort parameter?
>
> Best
> Erick
>
> On Sun, Feb 26, 2012 at 8:30 AM, Dragon Fly wrote:
> >
> > Hi,
> >
> > Let's say I have 6 documents and each document has 2 fields (i.e.
> > CustomerName
Hi,
Let's say I have 6 documents and each document has 2 fields (i.e. CustomerName
and OrderDate). For example:
Doc 1John20120115
Doc 2Mary20120113
Doc 3Peter 20120117
Doc 4Kate20120208
Doc 5John20120211
Doc 6Alan20110423
Is there a way to execute
uggest you upgrade to something more recent.
>
> In the 2.9 both IndexReader and IndexWriter have commit() methods.
>
> Best
> Erick
>
> On Tue, Jan 3, 2012 at 8:35 AM, Dragon Fly wrote:
> >
> > Hi, I'm using Lucene 2.0 and was wondering how to flush/commit
Hi, I'm using Lucene 2.0 and was wondering how to flush/commit index data to
disk. It doesn't look like there is a flush() or commit() method in the 2.0
IndexWriter. Is there a way to flush the data without calling close()? Thank
you.
you can directly load fields, do
> lazy loading or stop the loading process once a certain field was
> loaded.
> If you do not load a field due to the fieldSelector a call to
> Document.get("fieldname") will return null.
>
> Simon
>
> On Tue, Oct 6, 2009 at 2:38 PM, Drag
Hi,
Which of the following method actually loads the document from disk?
(1) Document document = searcher.doc (docId);
OR
(2) string value = document.get ("FirstNameField");
It's probably searcher.doc but I just want to be sure. Thank you.
___
t; > PANGAEA - Publishing Network for Geoscientific and Environmental Data
> > MARUM - University of Bremen
> > Room 2500, Leobener Str., D-28359 Bremen
> > Tel.: +49 421 218 65595
> > Fax: +49 421 218 65505
> > http://www.pangaea.de/
> > E-mail: uschind
Hi,
I have a question regarding RAMDirectory. I have a 5 GB index on disk and it
is opened like the following:
searcher = new IndexSearcher (new RAMDirectory (indexDirectory));
Approximately how much memory is needed to load the index? 5GB of memory or
10GB because of Unicode? Does the ent
Let's say I have 3 fields in a document (Type, FirstName, and LastName). For
example:
Document 0
--
Type: Public
FirstName: John
LastName: Deere
If I execute the following boolean query, document 0 is returned.
Type:Public OR FirstName:Candy OR LastName:Deere
about a URL inside a single
> document, by appending the user id to the field name. So I have, for
> example, bookmark_title_29:"My Bookmark", and bookmark_title_35:"Another
> bookmark" in the same document, and I can search bookmark titles by specific
> users.
>
> On Wed, Feb
:06:28 -0500
> From: sar...@syr.edu
> To: java-user@lucene.apache.org
>
> Hi Dragon Fly,
>
> You could split the original document into multiple Lucene Documents,
> one for each array index, all sharing the same "DocID" field value.
>
> Then your queries "
Hi,
Let's say I have a single document with 2 fields (namely Field1 and Field2). 2
values are added to each field like below.
// Add 2 values to Field1.
doc.Add (new Field ("Field1", "A", Field.Store.YES, Field.Index.UN_TOKENIZED));
doc.Add (new Field ("Field1", "B", Field.Store.YES, Field.Ind
http://www.gossamer-threads.com/lists/lucene/java-user/
> Date: Fri, 7 Nov 2008 14:27:38 -0700
> From: [EMAIL PROTECTED]
> To: java-user@lucene.apache.org
> Subject: searchable archives
>
> Hey,
>
> Is this list available somewhere that you can search the entire archives at
> one time?
>
> Tha
Thank you both for your help.
> Date: Fri, 31 Oct 2008 09:06:50 +0100
> From: [EMAIL PROTECTED]
> To: java-user@lucene.apache.org
> Subject: Re: Read all the data from an index
>
> Erick Erickson wrote:
> > I'm not sure what *could* be easier than looping with IndexSearcher.doc(),
> > looping fro
Best
> Erick
>
> On Thu, Oct 30, 2008 at 3:24 PM, Dragon Fly <[EMAIL PROTECTED]>wrote:
>
> > Hi,
> >
> > I have an old index that was built a few months ago. The data that I used
> > to build the index has been deleted from the database. I'd like
Hi,
I have an old index that was built a few months ago. The data that I used to
build the index has been deleted from the database. I'd like to read all the
data from the old index to build a new index. Which Lucene API calls should I
use to read all the data from the old index? Thank you i
Let's say my index has two fields (Type and Description). Type is either 0 or
1 and Description is a string (up to 250 characters). I'd like to execute the
following search:
+Description:Honda* +Type:0
Would the query run faster if I specify the Type first:
+Type:0 +Description:Honda
ing to close/open IndexWriter &
> IndexReader back and forth to do the deletions.
>
> IndexModifier is deprecated.
>
> Mike
>
> Dragon Fly wrote:
>
> > I'd like to delete some documents from my index. Should I use the
> > DeleteDocument method
I'd like to delete some documents from my index. Should I use the
DeleteDocument method in the IndexReader class or the IndexModifier class? Does
it make a difference which one I use? Thank you.
_
Get more from your digital life.
ontrib/benchmark program helpful for running experiments, although it
> isn't a substitute for your actual data.
>
> -Grant
>
>
> On Jul 30, 2008, at 9:46 AM, Dragon Fly wrote:
>
> > Perhaps I didn't explain myself clearly so please let me try it
> &g
Jul 2008 15:03:37 +0100
> From: [EMAIL PROTECTED]
> To: java-user@lucene.apache.org
> Subject: Re: Index optimization ...
>
> OK, but why do you need to optimize before every swap? Have you tried
> with less frequent optimizes?
>
> --
> Ian.
>
>
> On Wed, Jul
re the inactive copy is made active.
> Date: Wed, 30 Jul 2008 14:54:03 +0100
> From: [EMAIL PROTECTED]
> To: java-user@lucene.apache.org
> Subject: Re: Index optimization ...
>
> Why do you run an optimize every 4 hours?
>
>
> --
> Ian.
>
>
> On Wed, Ju
Perhaps I didn't explain myself clearly so please let me try it again. I'm
happy with the search/indexing performance. However, my index gets fully
optimized every 4 hours and the time it takes to fully optimize the index is
longer than I like. Is there anything that I can do to speed up the
I'd like to shorten the time it takes to optimize my index and am willing to
sacrifice search and indexing performance. Which parameters (e.g. merge
factor) should I change? Thank you.
_
Stay in touch when you're away with Windows
Thanks.
> Date: Mon, 24 Mar 2008 21:03:13 -0700
> From: [EMAIL PROTECTED]
> To: java-user@lucene.apache.org
> Subject: RE: Field values ...
>
>
> : The Id and Phone fields are stored. So I can just do a MatchAllQuery as
> : you suggested. I have read about field selectors on this mailing list
The Id and Phone fields are stored. So I can just do a MatchAllQuery as you
suggested. I have read about field selectors on this mailing list but have
never used it. Does anyone know where I can find some sample code? Thank you.
> Date: Sat, 22 Mar 2008 16:03:54 -0700
> From: [EMAIL PROTECTED
that'll iterate them all.
>
> Best
> Erick
>
> On Thu, Mar 20, 2008 at 9:55 AM, Dragon Fly <[EMAIL PROTECTED]>
> wrote:
>
> > What's the easiest way to extract the values of 2 fields from each
> > document in the index. For example, each docume
'll iterate them all.
>
> Best
> Erick
>
> On Thu, Mar 20, 2008 at 9:55 AM, Dragon Fly <[EMAIL PROTECTED]>
> wrote:
>
> > What's the easiest way to extract the values of 2 fields from each
> > document in the index. For example, each document has
What's the easiest way to extract the values of 2 fields from each document in
the index. For example, each document has 5 fields:
Id Name Address Phone Preference
I'd like to extract the values for the Id and Phone fields for each document in
the index. Thank you.
Hi Robert,
Did you run into any performance issues (because multiple searchers accessed a
single index on a shared directory)? Also, did you employ some redundancy
scheme to ensure that the shared directory is always "available"? Thank you.
> To: java-user@lucene.apache.org
> Subject: Re: Lucen
amp; test search performance.
>
> Mike
>
> Dragon Fly wrote:
>
> > Hi,
> >
> > I'd like to find out if I can do the following with Lucene (on
> > Windows).
> >
> > On server A:
> > - An index writer creates/updates the index. The index
Hi,
I'd like to find out if I can do the following with Lucene (on Windows).
On server A:
- An index writer creates/updates the index. The index is physically stored on
server A.
- An index searcher searches against the index.
On server B:
- Maps to the index directory.
- An index searcher sea
%). In lucene, the returing documents is set to 100 for the sake of
speed.
I am not quite sure my way of pagination is best: but it works fine under
test preasure: Just keep the first search result in cache and fetch the
snippet when the document is presented in current page.
2007/12/26, Drag
Any advice on this? Thanks.
> From: [EMAIL PROTECTED]
> To: java-user@lucene.apache.org
> Subject: Pagination ...
> Date: Sat, 22 Dec 2007 10:19:30 -0500
>
>
> Hi,
>
> What is the most efficient way to do pagination in Lucene? I have always done
> the following because this "flavor" of the se
Hi,
What is the most efficient way to do pagination in Lucene? I have always done
the following because this "flavor" of the search call allows me to specify the
top N hits (e.g. 1000) and a Sort object:
TopFieldDocs topFieldDocs = searcher.search(query, null, 1000,
SORT_BY_DATE);
Is it
Hi,
My application needs to close/open the index searcher periodically so that
newly added documents are visible. Is there a way to determine if there are
any pending searches running against an index searcher or do I have to do my
own reference counting? Thank you.
_
t; To: java-user@lucene.apache.org
> Subject: Re: Sort by date with Lucene 2.2.0 ...
> Date: Fri, 19 Oct 2007 19:50:46 +0200
>
> On Thursday 18 October 2007 21:35, Dragon Fly wrote:
>
> > I'm am trying to sort a date field in my index but I'm seeing strange
> > resul
t;
> Or is it just late and I'm missing the obvious (a specialty of mine)...
>
> Erick
>
> On 10/18/07, Dragon Fly <[EMAIL PROTECTED]> wrote:
> >
> >
> > Hi,
> >
> > I'm am trying to sort a date field in my index but I'm seeing stran
Hi,
I'm am trying to sort a date field in my index but I'm seeing strange results.
I have searched the Lucene user mail archive for Datetools but still couldn't
figure out the problem.
The date field is indexed as follows (i.e. DateTools is used, date field is
stored and untokenized):
Strin
Mike, which version of Lucene supports lazy loading? Thanks.
From: Michael McCandless <[EMAIL PROTECTED]>
Reply-To: java-user@lucene.apache.org
To: java-user@lucene.apache.org
Subject: Re: Field compression too slow
Date: Fri, 11 Aug 2006 06:59:58 -0400
I can share the data.. but it would be q
Thank you very much.
From: "Erick Erickson" <[EMAIL PROTECTED]>
Reply-To: java-user@lucene.apache.org
To: java-user@lucene.apache.org
Subject: Re: Empty fields ...
Date: Wed, 19 Jul 2006 09:48:04 -0400
Try something like
TermDocs termDocs = reader.termDocs();
termDocs.seek(new Term("",
My index gets rebuilt every night so I probably can afford
to construct the filters right after the index is rebuilt. How
do I check each document (for empty fields) though? Would
I use an IndexReader to loop through the documents? If so,
which method(s) in the IndexReader class should I use?
ter
Thanks for the quick reply, Erick. A couple of follow-up questions though.
If I had 200 fields (instead of 10), would you still recommend the same
approach? Would 200 filters use up too much memory? Would 200 filters
be too slow to construct/search?
From: "Erick Erickson" <[EMAIL PROTECTED]>
Re
Hi,
I have 10 fields in my index and some of the fields can be empty.
I'd like to be able do something like "IS NOT NULL" in SQL.
For example:
Name:Jane AND Addr IS NOT NULL AND Zip IS NOT NULL
Zip:90210 AND Name IS NOT NULL
Is there an easy way to do this? Thank you.
_
Hi,
I have an index that contains 3 fields: Book Id, Book Title, and Related
Book Ids.
For example:
=
Book Id Book Title Related Book Ids
A0001 Title 1 A0003, A0004
A0002 Title 2
A0003 Title 3 A0001, A0002
A0004 Title 4
I'll give it a try, thanks.
From: "Yonik Seeley" <[EMAIL PROTECTED]>
Reply-To: java-user@lucene.apache.org
To: java-user@lucene.apache.org
Subject: Re: Performance ...
Date: Mon, 22 May 2006 11:40:46 -0400
On 5/22/06, Dragon Fly <[EMAIL PROTECTED]> wrote:
The
Hi,
The search results of my Lucene application are always sorted
alphabetically.
Therefore, score and relevance are not needed. With that said, is there
anything that I can "disable" to:
(a) Improve the search performance
(b) Reduce the size of the index
(c) Shorten the indexing time
Thank
Hi,
What is the best way to implement the following?
Document 1 contains the following text:
"THE CZECH REPUBLIC ORGANIZATION"
Document 2 contains the following text:
"THE CZE ORGANISATION"
Synonym rules:
(1) CZECH REPUBLIC --> CZE
(2) CZE --> CZECH REPUBLIC
(3) ORGANIZATION --> ORG, ORGA
52 matches
Mail list logo