Ramirez,

If you partition your data correctly speed will be ~proportional. But there's 
always an upper limit - a slow range query that executes on a single node 
(using cluster key) will always be a slow.




Cheers,

Jens
—
Sent from Mailbox

On Sun, Jul 6, 2014 at 8:04 AM, Rameez Thonnakkal <ssram...@gmail.com>
wrote:

> Won't the performeance improve significantly if you increase the number of
> nodes even in a commodity hardware profile.
> On 5 Jul 2014 01:38, "Jens Rantil" <jens.ran...@tink.se> wrote:
>> Hi Mike,
>>
>> To learn get subsecond performance on your queries using _any_ database
>> you need to use proper indexing. Like Jeremy said, Solr will do this.
>>
>> If you'd like to try to solve this using Cassandra you need to learn the
>> difference between partition and clustering in your primary key and
>> understand you need a clustering to do any kind of range query.
>>
>> Also, COUNTs in Cassandra are generally fairly slow.
>>
>> Cheers,
>> Jens
>> —
>> Sent from Mailbox <https://www.dropbox.com/mailbox>
>>
>>
>> On Tue, Jun 24, 2014 at 10:09 AM, Mike Carter <jaloos...@gmail.com> wrote:
>>
>>> Hello!
>>>
>>>
>>> I'm a beginner in C* and I'm quite struggling with it.
>>>
>>> I’d like to measure the performance of some Cassandra-Range-Queries. The
>>> idea is to execute multidimensional range-queries on Cassandra. E.g. there
>>> is a given table of 1million rows with 10 columns and I like to execute
>>> some queries like “select count(*) from testable where d=1 and v1<10 and v2
>>> >20 and v3 <45 and v4>70 … allow filtering”.  This kind of queries is very
>>> slow in C* and soon the tables are bigger, I get a read-timeout probably
>>> caused by long scan operations.
>>>
>>> In further tests I like to extend the dimensions to more than 200
>>> hundreds and the rows to 100millions, but actually I can’t handle this
>>> small table. Should reorganize the data or is it impossible to perform such
>>> high multi-dimensional queries on Cassandra?
>>>
>>>
>>>
>>>
>>>
>>> The setup:
>>>
>>> Cassandra is installed on a single node with 2 TB disk space and 180GB
>>> Ram.
>>>
>>> Connected to Test Cluster at localhost:9160.
>>>
>>> [cqlsh 4.1.1 | Cassandra 2.0.7 | CQL spec 3.1.1 | Thrift protocol 19.39.0]
>>>
>>>
>>>
>>> Keyspace:
>>>
>>> CREATE KEYSPACE test WITH replication = {
>>>
>>>   'class': 'SimpleStrategy',
>>>
>>>   'replication_factor': '1'
>>>
>>> };
>>>
>>>
>>>
>>>
>>>
>>> Table:
>>>
>>> CREATE TABLE testc21 (
>>>
>>>   key int,
>>>
>>>   d int,
>>>
>>>   v1 int,
>>>
>>>   v10 int,
>>>
>>>   v2 int,
>>>
>>>   v3 int,
>>>
>>>   v4 int,
>>>
>>>   v5 int,
>>>
>>>   v6 int,
>>>
>>>   v7 int,
>>>
>>>   v8 int,
>>>
>>>   v9 int,
>>>
>>>   PRIMARY KEY (key)
>>>
>>> ) WITH
>>>
>>>   bloom_filter_fp_chance=0.010000 AND
>>>
>>>   caching='ROWS_ONLY' AND
>>>
>>>   comment='' AND
>>>
>>>   dclocal_read_repair_chance=0.000000 AND
>>>
>>>   gc_grace_seconds=864000 AND
>>>
>>>   index_interval=128 AND
>>>
>>>   read_repair_chance=0.100000 AND
>>>
>>>   replicate_on_write='true' AND
>>>
>>>   populate_io_cache_on_flush='false' AND
>>>
>>>   default_time_to_live=0 AND
>>>
>>>   speculative_retry='99.0PERCENTILE' AND
>>>
>>>   memtable_flush_period_in_ms=0 AND
>>>
>>>   compaction={'class': 'SizeTieredCompactionStrategy'} AND
>>>
>>>   compression={'sstable_compression': 'LZ4Compressor'};
>>>
>>>
>>>
>>> CREATE INDEX testc21_d_idx ON testc21 (d);
>>>
>>>
>>>
>>>  select * from testc21 limit 10;
>>>
>>> key    | d | v1 | v10 | v2 | v3 | v4  | v5 | v6 | v7 | v8 | v9
>>>
>>> --------+---+----+-----+----+----+-----+----+----+----+----+-----
>>>
>>>  302602 | 1 | 56 |  55 | 26 | 45 |  67 | 75 | 25 | 50 | 26 |  54
>>>
>>>  531141 | 1 | 90 |  77 | 86 | 42 |  76 | 91 | 47 | 31 | 77 |  27
>>>
>>>  693077 | 1 | 67 |  71 | 14 | 59 | 100 | 90 | 11 | 15 |  6 |  19
>>>
>>>    4317 | 1 | 70 |  77 | 44 | 77 |  41 | 68 | 33 |  0 | 99 |  14
>>>
>>>  927961 | 1 | 15 |  97 | 95 | 80 |  35 | 36 | 45 |  8 | 11 | 100
>>>
>>>  313395 | 1 | 68 |  62 | 56 | 85 |  14 | 96 | 43 |  6 | 32 |   7
>>>
>>>  368168 | 1 |  3 |  63 | 55 | 32 |  18 | 95 | 67 | 78 | 83 |  52
>>>
>>>  671830 | 1 | 14 |  29 | 28 | 17 |  42 | 42 |  4 |  6 | 61 |  93
>>>
>>>   62693 | 1 | 26 |  48 | 15 | 22 |  73 | 94 | 86 |  4 | 66 |  63
>>>
>>>  488360 | 1 |  8 |  57 | 86 | 31 |  51 |  9 | 40 | 52 | 91 |  45
>>>
>>> Mike
>>>
>>
>>

Reply via email to