Yes, that was it.
Thanks again!
El 14/12/2016 a las 9:44, Jaime escribió:
No, sorry, I think I'm not using SortField.Type.LONG. I'd give it a try.
Thank you very much.
El 13/12/2016 a las 18:53, Chris Hostetter escribió:
How are you constructing your SortField at query time
mally self contained reproducible code
demonstrating your problem? (ie: create an index with 2 docs, then do
a simple serach for both and sort them and show that the order is wrong)
: Date: Tue, 13 Dec 2016 18:30:09 +0100
: From: Jaime
: Reply-To: java-user@lucene.apache.org
: To: java-user@lucene.
Hello,
With Lucene 6.1.0, I'm trying to search sorting the results by a long
column. This is my indexing code:
doc.add(new LongPoint(name, Long.parseLong(value)));
doc.add(new StoredField(name, value));
doc.add(new NumericDocValuesField(SORT_FIELD_PREFIX + name,
Long.parseLong(val
juil. 2016 à 13:46, Jaime a écrit :
I use Lucene 6.1.0 (no solr) with NRT search. At runtime, my index seems
to grow a lot.
I think that for each commit, a new copy of the index is made. This
makes sense as an old searcher could be using the previous version.
However, the old copies don't se
e and reopen
the Index Writer.
I've tried calling IndexWriter.deleteUnusedFiles after each commit, but
didn't help.
Is there any way to free space without closing the IndexWriter?
--
Jaime Pardos
ESTRUCTURE MEDIA SYSTEMS, S.L.
Avda. de Madrid nº 120 nave 10, 28500, Arganda del Rey
nalysis/uk/UkrainianMorfologikAnalyzer.java
Ahmet
On Thursday, June 23, 2016 6:47 PM, Jaime wrote:
Hello,
I want to change the input text before tokenizing. I think I just need
to use some characters as word separators, and maybe remove some others
completely.
I was planning to use MappingCharFilter
Hello,
I want to change the input text before tokenizing. I think I just need
to use some characters as word separators, and maybe remove some others
completely.
I was planning to use MappingCharFilterFactory to replace some chars
with " " and others with "", but I feel like I'm not in the r