I would preferably spin 2 JVMs inside the same hardware (if you double
everything) than having to deal with what Jeff stated.
Also certain operations are not really found of a large number of vnodes
(eg. repair). There was a lot of improvements in the 3.x release cycle, but
I do still tend to redu
I am doing some on-the-job-learning on this newer feature of the 3.x line,
where the token generation algorithm will compensate for different size nodes
in a cluster. In fact, it is one of the main reasons I upgraded to 3.0.13,
because I have a number of original nodes in a cluster that are abou
Hi
what is the most efficient way to get a distinct key list from a big table
(aprox 20 mil inserts per minute) ?
equivalent to *select distinct key from my_table *for this table
*CREATE TABLE my_table (*
*key text,*
*timestamp bigint,*
*value double,*
*PRIMARY KEY (key, time
hi:
Is this sensor data, hence timestamp? Ho w are you generating this 'key'
field?Can you have only the 'key' field as primary key? Even if not, since
that field is a part of the PK may make such queries fast.
However, are there other attributes thst can be added that define unique
business key
Hi
Is there anything I need to do after dropping a column and adding in a column
to flush cassandra of the changes? We are experiencing issues with our front
end application, and the developers are asking if the issue was caused by the
change in schema, as I've done the drop and add of a column
@Sagar,
A query to get the data looks like this (primary key values included in the
query).
SELECT * FROM table WHERE pk1='2269202-onstreet_high' AND pk2=2017 AND
time='2017-07-18 03:15:00+';
(in actual practice, the queries in our code would use query a range of
time values).
@Cristophe
I
Regrettably, this may be a manifestation of Cassandra-13004, which could
corrupt data being read at the time you issued the ALTER TABLE command
What type of issues are you seeing? Anything in the logs?
--
Jeff Jirsa
> On Aug 18, 2017, at 8:41 AM, Jonathan Baynes
> wrote:
>
> Hi
>
> Is t
Thanks Jeff I have a horrible feeling it may be. I'll get the errors from the
dev guys Monday and email the group, hopefully I can tie this down.
Thanks
Sent from my iPhone
> On 18 Aug 2017, at 17:24, Jeff Jirsa wrote:
>
> Cassandra-13004
___
Hi,
I use cassandra-count (github
https://github.com/brianmhess/cassandra-count) to count records in a
table, but I have wrong results.
When I export data with cqlsh /copy to csv, I have 1M records in my test
table, when I use cassandra-count I have different results for each node :
build/ca
I have a table that uses LeveledCompactionStrategy on Cassandra 2.2. At the
moment, it has two SSTables, both in level 1, one that's repaired and one that
isn't.
$ sstablemetadata lb-135366-big-Data.db | head
SSTable: /home/cassandra/data/my_keyspace/my_table/lb-135366-big
Partitioner: org.apache
There seem to be a lot of SSTables in a repaired state and a lot in an
unrepaired state. For example, for this one table, the logs report
TRACE [main] 2017-08-15 23:50:30,732 LeveledManifest.java:473 - L0 contains 2
SSTables (176997267 bytes) in Manifest@1217144872
TRACE [main] 2017-08-15 23:50:
Cassandra version 3.9, -unit version 3.1.3.2.
In my (first ever) unit test, I've coded:
@BeforeClass public static void initFakeCassandra() throws
InterruptedException, IOException, TTransportException
{
EmbeddedCassandraServerHelper.startEmbeddedCassandra( 2L );
}
Execution crashes d
You need to run an incremental repair for sstables to be marked repaired.
However only if all of the data in that Sstable is repaired during the
repair will you end up with it being marked repaired, otherwise an
anticompaction will occur and split the unrepaired data into its own
sstable.
It's pret
You can SELECT DISTINCT in CQL, however I would recommend against such a
pattern as it is very unlikely to be efficient, and prone to errors. A
distinct query will search every partition for the first live cell, which
could be buried behind a lot of tombstones. It's safe to say at some point
you wi
For the example provided by you , are you saying you are getting two rows
for same pk1,pk2,time?
It may be a problem with your inserts when you are inserting multiple
distinct rows or to validate all nodes are in sync try fetching using
CONSISTENCY ALL in cql.
On 18-Aug-2017 9:37 PM, "Nathan McL
15 matches
Mail list logo