Okay I found that when building the package, the new helper dh_python2 must be
used to fix that issue.
—
Cyril Scetbon
> On Mar 18, 2018, at 11:08 AM, Cyril Scetbon wrote:
>
> Hey guys,
>
> Having to still use Cassandra 2.1, I have installed it on Ubuntu Xenial and I
> have an issue with cql
Worth noting that if you have racks == RF you only need to repair one rack
to repair all the data in the cluster if you *don't* use -pr. Also note
that full repairs on >=3.0 case anti-compactions and will mark things as
repaired, so once you start repairs you need to keep repairing to ensure
you do
“* 1000 statements in in each batch” sounds like you are doing batching in
both cases. I wouldn't expect things to get better with larger sizes than
that. We’ve generally found more like 100 is the sweet spot but I’m sure it’s
data specific.
On Sun, 18 Mar 2018 at 21:17 onmstester onmstester
wrot
Hey guys,
Having to still use Cassandra 2.1, I have installed it on Ubuntu Xenial and I
have an issue with cqlsh. I was able to fix it by installing python-support and
a fix from 2.1.16. However I’d like to know if there is a way to do it without
installing an old package (python-support) on Xe
I'm using a queue of 100 ExecuteAsyncs * 1000 statements in in each batch =
100K insert queue in non-batch scenario.
Using more than 1000 statememnts per batch throws batch limit exception and
some documents recommend no to change batch_size_limit??!
Sent using Zoho Mail
On Sun, 18
When you say batch was worth than async in terms of throughput are you
comparing throughput with the same number of threads or something? I would
have thought if you have much less CPU usage on the client with batching
and your Cassandra cluster doesn’t sound terribly stressed then there is
room to
Input data does not preserve good locality and I've already tested batch
insert, it was worse than executeAsync in case of throughput but much less CPU
usage at client side.
Sent using Zoho Mail
On Sun, 18 Mar 2018 12:46:02 +0330 Ben Slater
wrote
You will probably find grouping writes into small batches improves overall
performance (if you are not doing it already). See the following
presentation for some more info:
https://www.slideshare.net/Instaclustr/microbatching-highperformance-writes
Cheers
Ben
On Sun, 18 Mar 2018 at 19:23 onmstest
I need to insert some millions records in seconds in Cassandra. Using one
client with asyncExecute with folllowing configs:
maxConnectionsPerHost = 5
maxRequestsPerHost = 32K
maxAsyncQueue at client side = 100K
I could achieve 25% of throughtput i needed, client CPU is more than 80% and
inc