Re: Creating composite query in lucene

2016-03-09 Thread sandeep das
th SHOULD. > > -- Jack Krupansky > > On Tue, Mar 8, 2016 at 4:38 AM, sandeep das wrote: > > > Hi, > > > > I'm using lucene-5.2.0 and in query interface I wish to compose a query > > like > > "a=x and (b=y or d=z)" > > > > > > Whic

Creating composite query in lucene

2016-03-08 Thread sandeep das
Hi, I'm using lucene-5.2.0 and in query interface I wish to compose a query like "a=x and (b=y or d=z)" Which can be described as If any document has value "x" for field "a" and field "b" has value "y" or field "d" has value "z" then that document should be chosen. There are three fields in my d

Lucene-5.2.0 on HDFS

2016-02-28 Thread sandeep das
Hi All, I was trying to create indexes on HDFSDirectory. So I tried to use lucene-hdfs-directory-4.7.0 but it seems to be incompatible with lucene-5.2.0. The class HdfsDirectory requires to create an instance of BufferedIndexOutput which has been deprecated in the lucene-5.2.0 hence my application

Compression technique for stored fields

2016-02-23 Thread sandeep das
Hi guys, While running my application I noticed that the lz4 is used as compression technique for stored field. Is there any option by which I can change it to snappy? Regards, Sandeep

Re: Profiling lucene 5.2.0 based tool

2016-02-23 Thread sandeep das
ce > improvements here with the new dimensional points feature. See > https://www.elastic.co/blog/lucene-points-6.0 ... but not 6.0 is not > yet released though it should be soon now. > > Mike McCandless > > http://blog.mikemccandless.com > > > On Tue, Feb 23, 2016 a

Re: Profiling lucene 5.2.0 based tool

2016-02-22 Thread sandeep das
Audenaerde wrote: > Hi Sandeep, > > How many threads do you use to do the indexing? The benchmarks of Lucene > are done on >20 threads IIRC. > > -Rob > > On Tue, Feb 23, 2016 at 8:01 AM, sandeep das wrote: > > > Hi, > > > > I've implemented

Profiling lucene 5.2.0 based tool

2016-02-22 Thread sandeep das
Hi, I've implemented a tool using lucene-5.2.0 to index my CSV files. The tool is reading data from CSV files(residing on disk) and creating indexes on local disk. It is able to process 3.5 MBps data. There are overall 46 fields being added in one document. They are only of three data types 1. Int