Whether 600m rows per hour is good or bad depends on the hardware you are
using (do you have 1 node or 1000? 2 cores each or 16?) and the data you
are writing (is it 10 bytes per row or 100kb?).
In general, I think you will need to supply a lot more context about your
use case and set up to get an
This advice hasn't been valid for a long time now for most use cases. The
only time you need to reserve 50% disk space is if you're going to be
running major compactions against a table in your cluster that occupies 50%
of its total disk space. Nowadays, that's far less common than it was when
yo
Thanks Jean,
I have dc1 and dc2 existing. added dc3 from dc1 and dc4 from dc2. If I want
to run repair on one node in dc3 from dc1 only is it possible?
On Wed, Aug 21, 2019, 8:11 AM Jean Carlo wrote:
> Hello Rahul,
>
> To ensure the consistency among the DCs, it is enough to run a repair
> comm
Hi,
for example compaction uses a lot of disk space. It is quite common so
it is not safe to have your disk utilised like on 85% because
compactions would not have room to comapact and that node would be
stuck. This happens in production quite often.
Hence, having it on 50% and having big buffer
Good day,
I’m running the monitoring script for disk space utilization set the benchmark
to 50%. Currently am getting the alerts from some of the nodes
About disk space greater than 50%.
Is there a way, I can quickly figure out why the space has increased and how I
can maintain the disk space
Hello Rahul,
To ensure the consistency among the DCs, it is enough to run a repair
command.
You can do it using http://cassandra-reaper.io/
or runing the commande *nodetool repair* with the respectively options in
every node.
You do not need to count the rows in every DC to ensure cassandra is s
On Wed, Aug 21, 2019 at 1:47 PM Jeff Jirsa wrote:
> Wild ass guess based on what worked for my previous employer (for whom I
> wrote twcs), but note that if you model it such that partitions are
> contained to a single window/sstable that number is probably irrelevant and
> may be able to be made
Yep I did run rebuild on each new node
On Wed, Aug 21, 2019, 7:25 AM Stefan Miklosovic <
stefan.mikloso...@instaclustr.com> wrote:
> Hi Rahul,
>
> how did you add that dc3 to cluster? The rule of thumb here is to do
> rebuild from each node, for example like here
>
> https://docs.datastax.com/e
Wild ass guess based on what worked for my previous employer (for whom I wrote
twcs), but note that if you model it such that partitions are contained to a
single window/sstable that number is probably irrelevant and may be able to be
made arbitrarily high
> On Aug 21, 2019, at 4:48 PM, Oleksa
Sent from Mail for Windows 10
Thanks for feedback.
Just to elaborate more, I am currently writing 600m rows per hour and need to
understand if this is about on target or if there are better ways to write or
perhaps structure the keyspaces and table structures.
And I can use the Cassandra St
Hi Rahul,
how did you add that dc3 to cluster? The rule of thumb here is to do
rebuild from each node, for example like here
https://docs.datastax.com/en/archived/cassandra/3.0/cassandra/operations/opsAddDCToCluster.html
On Wed, 21 Aug 2019 at 12:57, Rahul Reddy wrote:
>
> Hi sefan,
>
> I'm addi
Hi,
In the TWCS readme[1] it says:
You should target fewer than 50 buckets per table based on your TTL.
What is this recommendation based upon? The docs from Datastax[2], for
example do not mention anything like that.
Is the recommendation still actual today for the version 3.0.$latest? Wha
Hi sefan,
I'm adding new DC3 to exiting cluster and see discripencies couple of
millions in Nodetool cfstats in new DC.
My table size is 50gb
I'm trying to run copy entire table.
Copy table to 'full_tablr.csv' with delimiter ',';
If I run above command from dc3. Does it get the data only from d
Hi Rahul,
what is your motivation behind this? Why do you want to make sure the
count is same? What is the purpose of that? All you should care about
is that Cassandra will return you right results. It was designed from
the very bottom to do that for you, you should not be bothered too
much about
Hello,
I have 3 datacenters . Want to make sure record count is same in all dc's .
If I run copy command node1 in dc1 does it get the data from only dc1?
Nodetool cfstats I'm seeing discrepancies in partitions count is it because
we didn't run cleanup after adding few nodes and remove them?. To ru
15 matches
Mail list logo