RE: Searching an Index on Another Machine

2008-08-07 Thread John Griffin
Dana, RMI. We use it exclusively and we are able to cluster for fail-over. The clustering complicates things quite a bit so you may not need or want it. John G. -Original Message- From: DanaWhite [mailto:[EMAIL PROTECTED] Sent: Thursday, August 07, 2008 9:54 AM To: java-user@lucene.apac

RE: Searching an Index on Another Machine

2008-08-07 Thread John Griffin
Alexander, RMI. We use it exclusively and we are able to cluster for fail-over. The clustering complicates things quite a bit so you may not need or want it. John G. -Original Message- From: Alexander Aristov [mailto:[EMAIL PROTECTED] Sent: Thursday, August 07, 2008 10:35 AM To: java-us

Re: Searching Tokenized x Un_tokenized

2008-08-07 Thread Otis Gospodnetic
Hi, Perhaps you can give some examples. Yes, untokenized means "full string" - it requires an "exact match". Otis -- Sematext -- http://sematext.com/ -- Lucene - Solr - Nutch - Original Message > From: Andre Rubin <[EMAIL PROTECTED]> > To: java-user@lucene.apache.org > Sent: Thursda

Re: merge Index

2008-08-07 Thread 장용석
Thank you for your reply. :) You are correct. I did run addIndexes method after the ramWriter colse. And It is good working and create correct index file. Thank you very much. Jang. 08. 8. 8, Grant Ingersoll <[EMAIL PROTECTED]>님이 작성: > > Lucene does this RAM/FS stuff seamlessly for you in the

Re: merge Index

2008-08-07 Thread Grant Ingersoll
Lucene does this RAM/FS stuff seamlessly for you in the background. Just open the FSDirectory and do your indexing. If I had to guess, though, from a quick glance, I think you should do the addIndexes after the ramWriter close, but that's just a guess, as I haven't tried it. -Grant On

merge Index

2008-08-07 Thread 장용석
hi, I am using RamDirectory and FSDirectory for indexing documents. I use RamDirectory as Buffer. For example, --- String indexDir = "d:/lucene_data/merge_test_index"; Analyzer analyzer = new StopAnalyzer(); RAMDirectory ramDir= new RAMDirectory(); IndexWriter

Searching Tokenized x Un_tokenized

2008-08-07 Thread Andre Rubin
Hi all, When I switched a String field from tokenized to untokenized, some searches started not returning some obvious values. Am I missing something on querying untokenized fields? Another question is, do I need an Analyzer if my search is on an Untokenized field, wouldn't the search be based on

Re: updating existing field values

2008-08-07 Thread Michael McCandless
It's also possible to create your own code that uses FieldsReader and FieldsWriter (both in org.apache.lucene.index) directly to create new fdx/fdt files, with the date fixed on each document. You'd have to enumerate over all segments in your index, create a FieldsReader instance on each

Re: updating existing field values

2008-08-07 Thread Andrzej Bialecki
Robert Stewart wrote: Given an existing index, is there any way to update the value of a particular field across all documents, without deleting/re-indexing documents? For instance, if I have a date field, and I need to offset all dates based on change to stored time zone (subtract 12 hours from

updating existing field values

2008-08-07 Thread Robert Stewart
Given an existing index, is there any way to update the value of a particular field across all documents, without deleting/re-indexing documents? For instance, if I have a date field, and I need to offset all dates based on change to stored time zone (subtract 12 hours from each value). The so

Re : Stop search process when a given number of hits is reached

2008-08-07 Thread renou oki
Thanks a lot for your responses... I have tried the HitCollector and throw an exception when the limit of hits is reached... It works fine and the search time is really reduce when there is a lot of docs which are matching the query... I did that : public class CountCollector extends HitCollec

Re: Paging & Sorting

2008-08-07 Thread Yonik Seeley
On Thu, Aug 7, 2008 at 11:57 AM, Neeraj Gupta <[EMAIL PROTECTED]> wrote: > A Gentle Reminder! Any Solution around this problem? Lucky you, it's already been solved :-) Seriously though, I think you are making some incorrect assumptions about how Lucene works. Lucene does not retrieve all 50K doc

Re: Searching an Index on Another Machine

2008-08-07 Thread Alexander Aristov
Hi Mount the file system of that machine to another one as a network share. But I don;t think it would be optimal. Alexander 2008/8/7 DanaWhite <[EMAIL PROTECTED]> > > Hello, > > I have two machines on the same network, but I want to use one machine to > search an index located on the file syst

Re: Searching an Index on Another Machine

2008-08-07 Thread ನಾಗೇಶ್ ಸುಬ್ರಹ್ಮಣ್ಯ (Nagesh S)
I guess the starting point would be to have the IndexSearcher to take a URI as the path to index (I don't think such a constructor exists) than the ones currently available. This might also call for some security policy to be in place to restrict access. Otherwise, J2EE perhaps ? Good question !

Re: Paging & Sorting

2008-08-07 Thread Neeraj Gupta
A Gentle Reminder! Any Solution around this problem? Regards -Neeraj Thank you for the reply, As you said Sure, just iterate over the first 100 entries in your Hits object It means before Iteration Lucene has already spent time and memory in finding all the 50k documents and sorting them,

Searching an Index on Another Machine

2008-08-07 Thread DanaWhite
Hello, I have two machines on the same network, but I want to use one machine to search an index located on the file system of the other machine. Any ideas on how to achieve this? Thanks Dana -- View this message in context: http://www.nabble.com/Searching-an-Index-on-Another-Machine-tp188736

RE: Lucene Concurrency Issue

2008-08-07 Thread Alex Wang
Thanks Mark and Jason for your responses and your contrib to Lucene. I will try to dig into them and incorporate the ideas into my app. Thanks again! Alex >-Original Message- >From: Jason Rutherglen [mailto:[EMAIL PROTECTED] >Sent: Thursday, August 07, 2008 10:07 AM >To: java-user@lucen

Re: Tree search

2008-08-07 Thread Sergey Kabashnyuk
Thanks for you reply Glen But I saw this solution before. And as I said before, moving is very often operation and storing full path can cost additional unwanted operations and therefore it's not a desirable solution. Sergey Kabashnyuk eXo Platform SAS There are a number of ways to do this.

Re: Tree search

2008-08-07 Thread Glen Newton
There are a number of ways to do this. Here is one: Lose the parentid field (unless you have other reasons to keep it). Add a field fullName, and a field called depth : doc1 fullName: state depth: 0 doc2 fullName: state/department depth:1 doc3 fullName: state/department/Boston depth: 2 doc4 ful

Re: Lucene Concurrency Issue

2008-08-07 Thread Jason Rutherglen
The contrib realtime search patch enables the functionality you described. https://issues.apache.org/jira/browse/LUCENE-1313 On Wed, Aug 6, 2008 at 7:45 PM, Alex Wang <[EMAIL PROTECTED]> wrote: > > Hi all, > > To allow mutilple users concurrently add, delete docs and at the same time > search the

Tree search

2008-08-07 Thread Sergey Kabashnyuk
Hello I have such document structure doc1 id - 1 parentid - 0 name -state doc2 id - 2 parentid - 1 name - department doc3 id - 3 parentid - 2 name – Boston doc4 id - 4 parentid - 2 name – Opera

Efficient delete

2008-08-07 Thread Michael Zehrer
Dear List, I have a rather big index around 20gb. My documents have a unique id that I store in in an untokenized field. Using an IndexReader I delete documents by term using the id. The applications tries to batch as many delete operations as possible for this. The applications runs on a 8Core

Re: Stop search process when a given number of hits is reached

2008-08-07 Thread Andrzej Bialecki
Doron Cohen wrote: Nothing built in that I'm aware of will do this, but it can be done by searching with your own HitCollector. There is a related feature - stop search after a specified time - using TimeLimitedCollector. It is not released yet, see issue LUCENE-997. In short, the collector's col

Re: search with accent not match

2008-08-07 Thread Christophe from paris
Yes markrmiller,the order is important then TokenStream result = new StandardTokenizer(reader); result = new StandardFilter(result); result = new StopFilter(result, stoptable); result = new ISOLatin1AccentFilter(result); result = new FrenchStemFilter(result, excltable);

Re: Stop search process when a given number of hits is reached

2008-08-07 Thread Doron Cohen
Nothing built in that I'm aware of will do this, but it can be done by searching with your own HitCollector. There is a related feature - stop search after a specified time - using TimeLimitedCollector. It is not released yet, see issue LUCENE-997. In short, the collector's collect() method is invo

Re: CustomScoreQuery and BooleanQuery

2008-08-07 Thread Doron Cohen
When combining any sub queries, a scorer has at least two things to decide: which docs to match, and once matched, how to score. Boolean Queries applies specific logics for this, and some queries allow some control of the way to score. For current CustomScoreQuery things are more straightforward -

Re: LineDocMaker usage

2008-08-07 Thread Michael McCandless
Also, since the docs are so regular from a line file, to gain indexing throughput you can re-use the Document & Field instances, and inside the while loop just use Field.setValue(...) to change the value per document. Mike Anshum wrote: Hi, How about just opening a file and parsing thr

Stop search process when a given number of hits is reached

2008-08-07 Thread renou oki
Hello Is there a way to stop the search process when a given number of hits is reached? I have a counter feature which displays how many docs match a query. This counter is blocked; I mean that if there are more than 500 docs, it will just display "more than 500". I don't care about the exact