You can increase it if you're sure that it fits your use case. For an
explanation of why batch size vs number of statements, see the discussion
in CASSANDRA-6487. Cheers!

On Mon, Nov 13, 2017 at 6:31 PM, Tim Moore <tim.mo...@lightbend.com> wrote:

> Hi,
>
> I'm trying to understand some of the details of the
> batch_size_warn_threshold_in_kb/batch_size_fail_threshold_in_kb settings.
>
> Specifically, why are the thresholds measured in kb rather than the number
> of partitions affected?
>
> We have run into the limit in a situation where there is a batch with
> writes to two tables (because we wanted to ensure atomicity of the writes
> in the case of a failure). In some situations, the data inserted into one
> of these tables can be large enough to push the total batch size over the
> limit.
>
> In this specific case, we were able to rewrite things so that it could be
> split into separate statement executions with the application handling
> retry on failure so atomicity is not needed. It left us wondering, however,
> whether executing this as a batch was really problematic, or if the warning
> from Cassandra was spurious in this case.
>
> Is it really the total data size that matters, or just the number of
> affected partitions?
>
> Thanks,
> Tim
>
> --
> Tim Moore
> *Senior Engineer, Lagom, Lightbend, Inc.*
> tim.mo...@lightbend.com
> +61 420 981 589 <+61%20420%20981%20589>
> Skype: timothy.m.moore
>
> <https://www.lightbend.com/>
>

Reply via email to