Re: Cassandra-stress testing

2019-08-21 Thread Ben Slater
Whether 600m rows per hour is good or bad depends on the hardware you are using (do you have 1 node or 1000? 2 cores each or 16?) and the data you are writing (is it 10 bytes per row or 100kb?). In general, I think you will need to supply a lot more context about your use case and set up to get an

Re: Disk space utilization by from some Cassandra

2019-08-21 Thread Jon Haddad
This advice hasn't been valid for a long time now for most use cases. The only time you need to reserve 50% disk space is if you're going to be running major compactions against a table in your cluster that occupies 50% of its total disk space. Nowadays, that's far less common than it was when yo

Re: Cassandra copy command

2019-08-21 Thread Rahul Reddy
Thanks Jean, I have dc1 and dc2 existing. added dc3 from dc1 and dc4 from dc2. If I want to run repair on one node in dc3 from dc1 only is it possible? On Wed, Aug 21, 2019, 8:11 AM Jean Carlo wrote: > Hello Rahul, > > To ensure the consistency among the DCs, it is enough to run a repair > comm

Re: Disk space utilization by from some Cassandra

2019-08-21 Thread Stefan Miklosovic
Hi, for example compaction uses a lot of disk space. It is quite common so it is not safe to have your disk utilised like on 85% because compactions would not have room to comapact and that node would be stuck. This happens in production quite often. Hence, having it on 50% and having big buffer

Disk space utilization by from some Cassandra

2019-08-21 Thread yanga.zukelwa
Good day, I’m running the monitoring script for disk space utilization set the benchmark to 50%. Currently am getting the alerts from some of the nodes About disk space greater than 50%. Is there a way, I can quickly figure out why the space has increased and how I can maintain the disk space

Re: Cassandra copy command

2019-08-21 Thread Jean Carlo
Hello Rahul, To ensure the consistency among the DCs, it is enough to run a repair command. You can do it using http://cassandra-reaper.io/ or runing the commande *nodetool repair* with the respectively options in every node. You do not need to count the rows in every DC to ensure cassandra is s

Re: TWCS target number of time windows

2019-08-21 Thread Oleksandr Shulgin
On Wed, Aug 21, 2019 at 1:47 PM Jeff Jirsa wrote: > Wild ass guess based on what worked for my previous employer (for whom I > wrote twcs), but note that if you model it such that partitions are > contained to a single window/sstable that number is probably irrelevant and > may be able to be made

Re: Cassandra copy command

2019-08-21 Thread Rahul Reddy
Yep I did run rebuild on each new node On Wed, Aug 21, 2019, 7:25 AM Stefan Miklosovic < stefan.mikloso...@instaclustr.com> wrote: > Hi Rahul, > > how did you add that dc3 to cluster? The rule of thumb here is to do > rebuild from each node, for example like here > > https://docs.datastax.com/e

Re: TWCS target number of time windows

2019-08-21 Thread Jeff Jirsa
Wild ass guess based on what worked for my previous employer (for whom I wrote twcs), but note that if you model it such that partitions are contained to a single window/sstable that number is probably irrelevant and may be able to be made arbitrarily high > On Aug 21, 2019, at 4:48 PM, Oleksa

RE: Cassandra-stress testing

2019-08-21 Thread yanga.zukelwa
Sent from Mail for Windows 10 Thanks for feedback. Just to elaborate more, I am currently writing 600m rows per hour and need to understand if this is about on target or if there are better ways to write or perhaps structure the keyspaces and table structures. And I can use the Cassandra St

Re: Cassandra copy command

2019-08-21 Thread Stefan Miklosovic
Hi Rahul, how did you add that dc3 to cluster? The rule of thumb here is to do rebuild from each node, for example like here https://docs.datastax.com/en/archived/cassandra/3.0/cassandra/operations/opsAddDCToCluster.html On Wed, 21 Aug 2019 at 12:57, Rahul Reddy wrote: > > Hi sefan, > > I'm addi

TWCS target number of time windows

2019-08-21 Thread Oleksandr Shulgin
Hi, In the TWCS readme[1] it says: You should target fewer than 50 buckets per table based on your TTL. What is this recommendation based upon? The docs from Datastax[2], for example do not mention anything like that. Is the recommendation still actual today for the version 3.0.$latest? Wha

Re: Cassandra copy command

2019-08-21 Thread Rahul Reddy
Hi sefan, I'm adding new DC3 to exiting cluster and see discripencies couple of millions in Nodetool cfstats in new DC. My table size is 50gb I'm trying to run copy entire table. Copy table to 'full_tablr.csv' with delimiter ','; If I run above command from dc3. Does it get the data only from d

Re: Cassandra copy command

2019-08-21 Thread Stefan Miklosovic
Hi Rahul, what is your motivation behind this? Why do you want to make sure the count is same? What is the purpose of that? All you should care about is that Cassandra will return you right results. It was designed from the very bottom to do that for you, you should not be bothered too much about

Cassandra copy command

2019-08-21 Thread Rahul Reddy
Hello, I have 3 datacenters . Want to make sure record count is same in all dc's . If I run copy command node1 in dc1 does it get the data from only dc1? Nodetool cfstats I'm seeing discrepancies in partitions count is it because we didn't run cleanup after adding few nodes and remove them?. To ru