Hi,
In Lucene individual documents stand alone and can't be related to
other Lucene documents in the index automatically. In other words,
this is something you have to do through clever document/field design.
Otis
--- Urvashi Gadi <[EMAIL PROTECTED]> wrote:
> Hi,
>
> Is there a way to link i
Hi,
Is there a way to link indexed records in Lucene?
I am working with collections and a record might be a part of a
collection and this collection can again be a part of super collection.
So if the search result has the record, is there a way to find out all
the collections under which the reco
You're right, Peter. I did some thorough testing today and I was very
wrong. I'm now using MultiFieldQueryParser as Erik suggested.
~
Daniel Clark, Senior Consultant
Sybase Federal Professional Services
6550 Rock Spring Drive, Suite 800
Bethesda, MD 20817
On 1 Nov 2005, at 11:02, Malcolm wrote:
Hi,
I've been reading my new project bible 'Lucene in Action'
Amen! ;)
about Analysis in Chapter 4 and wondered what others are doing for
indexing XML(if anyone else is, that is!).
Are you folks just writing your own or utilising the current Lucene
: IndexWriter writer = new IndexWriter(directory, new StandardAnalyzer(),
: true);
:
: so if I try to close() it in a finally or something it throws a null
: pointer exception since the exception was throw in the constructor.
:
: I'm simulating the exception by hand-creating the index directory and
Hey-
I currently have an index with user information...one of the fields that I have
in a document is CurrentJob...I was asked to add in past jobs as wellbut
the rules are that I can configure the weighting (mainly boosting) for absolute
jobs (that this is a current job or a not current jo
Well, I'm running on linux and I thought the problem was that the writer
was not being closed but the ioexception is thrown at:
IndexWriter writer = new IndexWriter(directory, new StandardAnalyzer(),
true);
so if I try to close() it in a finally or something it throws a null
pointer exception sin
1) how do you simulate the exception?
2) you didn't say you got a lock timeout error, you said you got a
"couldnt delete the lock file" exception ... is your second test
forcably trying to unlock the index?
3) are you running this test on a windows machine?
4) can you post your unit test?
I have 2 junit tests. The first opens on index writer and then simulates
have an IOException get throw when trying to add a document. The test
that runs after than is just a normal test of the search. After the
first test completes a lock file is left in /tmp. Now, if I run the test
suite normally
I'm currently indexing the INEX collection and then performing queries on
the Format features within the text. I am using a wide range of the XML
features. The reason I asked about the XML Analysis is I am interesting in
opinions and reasons for adding a wide range of discussion to my
dissertat
Not sure I am understanding your question correctly, but I think you
want to pick your Analyzer based on what is in your content (i.e.
language, usage of special symbols, etc.), not based on what the format
of your content is (i.e. XML).
Malcolm wrote:
Hi,
I'm just asking for opinions on Ana
Ok... just got confused because you mentioned XML. Unless you're
actually indexing the raw XML in some of your fields, the fact that
you're indexing XML documents as your source content is irrelevant to
your choice of Analyzer.
Choice of indexer really depends on your specific project requirements
Hi,
I'm just asking for opinions on Analyzer's for the indexing. For example
Otis in his article uses the WhitespaceAnalyzer and the Sandbox program uses
the StandardAnalyzer.I am just gauging opinions on the subject with regard
to XML.
I'm using a mix of the Sandbox XMLDocumentHandlerSAX and a
Not exactly sure what you're asking with regards to Analyzers and
parsing XML...
But for parsing and indexing XML documents with Lucene, you can find a
lot of material out there by searching the list archives and using
google. However, the document I found most helpful was this piece
written by Ot
Hi,
I've been reading my new project bible 'Lucene in Action' about Analysis in
Chapter 4 and wondered what others are doing for indexing XML(if anyone else
is, that is!).
Are you folks just writing your own or utilising the current Lucene analysis
libraries?
thanks,
Malcolm Clark
Is using a QueryParser to parse a query using the same, single instance of
Analyzer thread-safe?
Or should I create a new Analyzer each time?
-
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PR
Such an analyzer already exists, in Lucene's Subversion repository,
under contrib/analyzers/: KeywordAnalyzer.
Robert Watkins wrote:
One approach for matching your queries with Luke would be to write a
custom Analyzer that does absolutely nothing to the terms. Then, if you
put this Analyzer in
One approach for matching your queries with Luke would be to write a
custom Analyzer that does absolutely nothing to the terms. Then, if you
put this Analyzer in your classpath when running Luke you can select it
as the Analyzer you want Luke to use to tokenize your query. This is
not, of course,
One approach for matching your queries with Luke would be to write a
custom Analyzer that does absolutely nothing to the terms. Then, if you
put this Analyzer in your classpath when running Luke you can select it
as the Analyzer you want Luke to use to tokenize your query. This is
not, of course,
One more question...if I do use a tool for replication and I only have the
indexer running on one machine...say it creates 10 file *.cfs...the tool
replicates all the files...then the machine with the indexer compresses those
files and then are now all gone...how will they be removed from the sl
On Thu, 2005-10-27 at 16:35 -0400, Sharma, Siddharth wrote:
> My index has 4 keyword fields and one unindexed field.
> I want to search by the 4 keyword fields and return the one unindexed field.
>
> I can iterate over the documents via Luke.
> But when I search for the same values that I see via
On 1 Nov 2005, at 08:17, Michael D. Curtin wrote:
tcorbet wrote:
I have an index over the titles to .mp3 songs.
It is not unreasonable for the user to want to
see the results from: "Show me Everything".
I understand that title:* is not a valid wildcard query.
I understand that title:[a* TO z
tcorbet wrote:
I have an index over the titles to .mp3 songs.
It is not unreasonable for the user to want to
see the results from: "Show me Everything".
I understand that title:* is not a valid wildcard query.
I understand that title:[a* TO z*] is a valid wildcard query.
What I cannot underst
Oren Shir wrote:
My documents contain a field called SORT_ID, which contains an int that
increases with every document added to the index. I want my results to be
sorted by it.
Which approach will prove the best performance:
1) Zero pad SORT_ID field and sort by it as plain text.
2) Sort using
On 1 Nov 2005, at 06:03, Oren Shir wrote:
Hi,
My documents contain a field called SORT_ID, which contains an int
that
increases with every document added to the index. I want my results
to be
sorted by it.
Which approach will prove the best performance:
1) Zero pad SORT_ID field and sor
On 1 Nov 2005, at 01:43, tcorbet wrote:
I have an index over the titles to .mp3 songs.
It is not unreasonable for the user to want to
see the results from: "Show me Everything".
I understand that title:* is not a valid wildcard query.
I understand that title:[a* TO z*] is a valid wildcard query
Hi,
My documents contain a field called SORT_ID, which contains an int that
increases with every document added to the index. I want my results to be
sorted by it.
Which approach will prove the best performance:
1) Zero pad SORT_ID field and sort by it as plain text.
2) Sort using SortField for
I have an index over the titles to .mp3 songs.
It is not unreasonable for the user to want to
see the results from: "Show me Everything".
I understand that title:* is not a valid wildcard query.
I understand that title:[a* TO z*] is a valid wildcard query.
What I cannot understand is this behavi
28 matches
Mail list logo