Did you make sure auto_bootstrap property is indeed set to [true] when you added the node?
From: Junaid Nasir [mailto:jna...@an10.io] Sent: Monday, June 05, 2017 6:29 AM To: Akhil Mehra <akhilme...@gmail.com> Cc: Vladimir Yudovin <vla...@winguzone.com>; user@cassandra.apache.org Subject: Re: Convert single node C* to cluster (rebalancing problem) not evenly, i have setup a new cluster with subset of data (around 5gb). using the configuration above I am getting these results Datacenter: datacenter1 ======================= Status=Up/Down |/ State=Normal/Leaving/Joining/Moving -- Address Load Tokens Owns (effective) Host ID Rack UN 10.128.2.1 4.86 GiB 256 44.9% e4427611-c247-42ee-9404-371e177f5f17 rack1 UN 10.128.2.10 725.03 MiB 256 55.1% 690d5620-99d3-4ae3-aebe-8f33af54a08b rack1 is there anything else I can tweak/check to make the distribution even? On Sat, Jun 3, 2017 at 3:30 AM, Akhil Mehra <akhilme...@gmail.com<mailto:akhilme...@gmail.com>> wrote: So now the data is evenly balanced in both nodes? Refer to the following documentation to get a better understanding of the roc_address and the broadcast_rpc_address https://www.instaclustr.com/demystifying-cassandras-broadcast_address/<https://urldefense.proofpoint.com/v2/url?u=https-3A__www.instaclustr.com_demystifying-2Dcassandras-2Dbroadcast-5Faddress_&d=DwMFaQ&c=LFYZ-o9_HUMeMTSQicvjIg&r=FsmDztdsVuIKml8IDhdHdg&m=57WqcUduTb1GA2Ij5E1fXgw3Cf21HYBK_4l2HVryPrk&s=MaTA43pugg78xQNfaOQElhyvd8k7CjVqZPr3IWALdWI&e=>. I am surprised that your node started up with rpc_broadcast_address set as this is an unsupported property. I am assuming you are using Cassandra version 3.10. Regards, Akhil On 2/06/2017, at 11:06 PM, Junaid Nasir <jna...@an10.io<mailto:jna...@an10.io>> wrote: I am able to get it working. I added a new node with following changes #rpc_address:0.0.0.0 rpc_address: 10.128.1.11 #rpc_broadcast_address:10.128.1.11 rpc_address was set to 0.0.0.0, (I ran into a problem previously regarding remote connection and made these changes https://stackoverflow.com/questions/12236898/apache-cassandra-remote-access<https://urldefense.proofpoint.com/v2/url?u=https-3A__stackoverflow.com_questions_12236898_apache-2Dcassandra-2Dremote-2Daccess&d=DwMFaQ&c=LFYZ-o9_HUMeMTSQicvjIg&r=FsmDztdsVuIKml8IDhdHdg&m=57WqcUduTb1GA2Ij5E1fXgw3Cf21HYBK_4l2HVryPrk&s=oj8BCLiyBDqqVQNqfGohFeujtqlzAkd-uwS878d4mg4&e=>) should it be happening? On Thu, Jun 1, 2017 at 6:31 PM, Vladimir Yudovin <vla...@winguzone.com<mailto:vla...@winguzone.com>> wrote: Did you run "nodetool cleanup" on first node after second was bootstrapped? It should clean rows not belonging to node after tokens changed. Best regards, Vladimir Yudovin, Winguzone<https://urldefense.proofpoint.com/v2/url?u=https-3A__winguzone.com_-3Ffrom-3Dlist&d=DwMFaQ&c=LFYZ-o9_HUMeMTSQicvjIg&r=FsmDztdsVuIKml8IDhdHdg&m=57WqcUduTb1GA2Ij5E1fXgw3Cf21HYBK_4l2HVryPrk&s=Q1M5YRAsw0iUQKOIulEmO72RhdENQCRhpqZSjgxxHos&e=> - Cloud Cassandra Hosting ---- On Wed, 31 May 2017 03:55:54 -0400 Junaid Nasir <jna...@an10.io<mailto:jna...@an10.io>> wrote ---- Cassandra ensure that adding or removing nodes are very easy and that load is balanced between nodes when a change is made. but it's not working in my case. I have a single node C* deployment (with 270 GB of data) and want to load balance the data on multiple nodes, I followed this guide<https://urldefense.proofpoint.com/v2/url?u=https-3A__docs.datastax.com_en_cassandra_2.1_cassandra_operations_ops-5Fadd-5Fnode-5Fto-5Fcluster-5Ft.html&d=DwMFaQ&c=LFYZ-o9_HUMeMTSQicvjIg&r=FsmDztdsVuIKml8IDhdHdg&m=57WqcUduTb1GA2Ij5E1fXgw3Cf21HYBK_4l2HVryPrk&s=xnLuX4qqSZxVuY6Gz2NXRdc6TG8J7at9kDhkhBnWpnM&e=> `nodetool status` shows 2 nodes but load is not balanced between them Datacenter: dc1 =============== Status=Up/Down |/ State=Normal/Leaving/Joining/Moving -- Address Load Tokens Owns (effective) Host ID Rack UN 10.128.0.7 270.75 GiB 256 48.6% 1a3f6faa-4376-45a8-9c20-11480ae5664c rack1 UN 10.128.0.14 414.36 KiB 256 51.4% 66a89fbf-08ba-4b5d-9f10-55d52a199b41 rack1 I also ran 'nodetool repair' on new node but result is same. any pointers would be appreciated :) conf file of new node cluster_name: 'cluster1' - seeds: "10.128.0.7" num_tokens: 256 endpoint_snitch: GossipingPropertyFileSnitch Thanks, Junaid