It is ScyllaDB specific. Scylla divides data not only among nodes, but also internally within a node among cores (=shards in our terminology). In the past we had problems with shards being over- and under-utilized (just like your cluster), so this simulator was developed to validate the solution.

On 07/11/2017 10:27 AM, Loic Lambiel wrote:
Thanks for the hint and tool !

By the way, what does the --shards parameter means ?

Thanks

Loic

On 07/10/2017 05:20 PM, Avi Kivity wrote:
32 tokens is too few for 33 nodes. I have a sharding simulator [1] and
it shows


$ ./shardsim --vnodes 32 --nodes 33 --shards 1
33 nodes, 32 vnodes, 1 shards
maximum node overcommit:  1.42642
maximum shard overcommit: 1.426417


So 40% overcommit over the average. Since some nodes can be
undercommitted, this easily explains the 2X difference (40% overcommit +
30% undercommit = 2X).


Newer versions of Cassandra have better token selection and will suffer
less from this.



[1] https://github.com/avikivity/shardsim


On 07/10/2017 04:02 PM, Loic Lambiel wrote:
Hi,

One of our clusters is becoming somehow unbalanced, at least some of the
nodes:

(output edited to remove unnecessary information)
--  Address         Load       Tokens  Owns (effective)   Rack
UN  192.168.1.22   2.99 TB    32      10.6%               RACK1
UN  192.168.1.23   3.35 TB    32      11.7%               RACK1
UN  192.168.1.20   3.22 TB    32      11.3%               RACK1
UN  192.168.1.21   3.21 TB    32      11.2%               RACK1
UN  192.168.1.18   2.87 TB    32      10.3%               RACK1
UN  192.168.1.19   3.49 TB    32      12.0%               RACK1
UN  192.168.1.16   5.32 TB    32      12.9%               RACK1
UN  192.168.1.17   3.77 TB    32      12.0%               RACK1
UN  192.168.1.26   4.46 TB    32      11.2%               RACK1
UN  192.168.1.24   3.24 TB    32      11.4%               RACK1
UN  192.168.1.25   3.31 TB    32      11.2%               RACK1
UN  192.168.1.134  2.75 TB    18      7.2%                RACK1
UN  192.168.1.135  2.52 TB    18      6.0%                RACK1
UN  192.168.1.132  1.85 TB    18      6.8%                RACK1
UN  192.168.1.133  2.41 TB    18      5.7%                RACK1
UN  192.168.1.130  2.95 TB    18      7.1%                RACK1
UN  192.168.1.131  2.82 TB    18      6.7%                RACK1
UN  192.168.1.128  3.04 TB    18      7.1%                RACK1
UN  192.168.1.129  2.47 TB    18      7.2%                RACK1
UN  192.168.1.14   5.63 TB    32      13.4%               RACK1
UN  192.168.1.15   2.95 TB    32      10.4%               RACK1
UN  192.168.1.12   3.83 TB    32      12.4%               RACK1
UN  192.168.1.13   2.71 TB    32      9.5%                RACK1
UN  192.168.1.10   3.51 TB    32      11.9%               RACK1
UN  192.168.1.11   2.96 TB    32      10.3%               RACK1
UN  192.168.1.126  2.48 TB    18      6.7%                RACK1
UN  192.168.1.127  2.23 TB    18      5.5%                RACK1
UN  192.168.1.124  2.05 TB    18      5.5%                RACK1
UN  192.168.1.125  2.33 TB    18      5.8%                RACK1
UN  192.168.1.122  1.99 TB    18      5.1%                RACK1
UN  192.168.1.123  2.44 TB    18      5.7%                RACK1
UN  192.168.1.120  3.58 TB    28      11.4%               RACK1
UN  192.168.1.121  2.33 TB    18      6.8%                RACK1

Notice the node 192.168.1.14 owns 13.4%  / 5.63TB while node
192.168.1.13 owns only 9.5% / 2.71TB, which is almost twice the load.
They both have 32 tokens.

The cluster is running:

* Cassandra 2.1.16 (initially bootstrapped running 2.1.2, with vnodes
enabled)
* RF=3 with single DC and single rack. LCS as the compaction strategy,
JBOD storage
* Partitioner: org.apache.cassandra.dht.Murmur3Partitioner
* Node cleanup performed on all nodes

Almost all of the cluster load comes from a single CF:

CREATE TABLE blobstore.block (
      inode uuid,
      version timeuuid,
      block bigint,
      offset bigint,
      chunksize int,
      payload blob,
      PRIMARY KEY ((inode, version, block), offset)
) WITH CLUSTERING ORDER BY (offset ASC)
      AND bloom_filter_fp_chance = 0.01
      AND caching = '{"keys":"ALL", "rows_per_partition":"NONE"}'
      AND comment = ''
      AND compaction = {'tombstone_threshold': '0.1',
'tombstone_compaction_interval': '60', 'unchecked_tombstone_compaction':
'false', 'class':
'org.apache.cassandra.db.compaction.LeveledCompactionStrategy'}
      AND compression = {'sstable_compression':
'org.apache.cassandra.io.compress.LZ4Compressor'}
      AND dclocal_read_repair_chance = 0.1
      AND default_time_to_live = 0
      AND gc_grace_seconds = 172000
      AND max_index_interval = 2048
      AND memtable_flush_period_in_ms = 0
      AND min_index_interval = 128
      AND read_repair_chance = 0.0
      AND speculative_retry = '99.0PERCENTILE';

The payload column is almost the same size in each record.

I understand that an unbalanced cluster may be the result of a bad
Primary key, which I believe isn't the case here.

Any clue on what could be the cause ? How can I re-balance it without
any decommission ?

My understanding is that nodetool move may only be used when not using
the vnodes feature.

Any help appreciated, thanks !

----
Loic Lambiel

---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org
For additional commands, e-mail: user-h...@cassandra.apache.org



---------------------------------------------------------------------
To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org
For additional commands, e-mail: user-h...@cassandra.apache.org

Reply via email to