Hello Sean,
here my schema and RF:
-
CREATE KEYSPACE my_keyspace WITH replication = {'class':
'NetworkTopologyStrategy', 'DC1': '1'} AND durable_writes = true;
CREATE TABLE my_keyspace.my_table (
pkey text,
event_da
Hello,
I’d start with describing my use case and how I’d like to use Cassandra to
solve my storage needs.
We're processing a stream of events for various happenings. Every event have a
unique happening_id.
One happening may have many events, usually ~ 20-100 events. I’d like to store
only the l
Your RF is only 1, so the data only exists on one node. This is not typically
how Cassandra is used. If you need the high availability and low latency, you
typically set RF to 3 per DC.
How many event_datetime records can you have per pkey? How many pkeys (roughly)
do you have? In general, you
You say the events are incremental updates. I am interpreting this to mean only
some columns are updated. Others should keep their original values.
You are correct that inserting null creates a tombstone.
Can you only insert the columns that actually have new values? Just skip the
columns with
Depending on the use case, creating separate prepared statements for each
combination of set / unset values in large INSERT/UPDATE statements may be
prohibitive.
Instead, you can look into driver level support for UNSET values. Requires
Cassandra 2.2 or later IIRC.
See:
Java Driver:
https://docs
Hello,
The problem is I can't know the combination of set/unset values. From my
perspective every value should be set. The event from Kafka represents the
complete state of the happening at certain point in time. In my table I
want to store the latest event so the most recent state of the happenin
Hi All,
While the pending node get streaming of token ranges from other nodes,
all coordinator would send new writes to it so that it would not miss
any new data, correct?
I have two (maybe silly) questions here:
Given the CL is ONE,
a) what if the coordinator haven't meet the pending node via go