*nodetool ring* will give you the tokens for each node on the ring. Each node has the token range between the previous node's token and its token - so the token range for each node is the interval (previous_token, this_token]. The first node in the ring has the range between the last node's token and its token (the "wrapping range").
Patrick Bannister On Sun, Apr 15, 2018 at 11:23 AM, Faraz Mateen <fmat...@an10.io> wrote: > *UPDATE* - I created schema for all the tables in one of the keypsaces, > copied data to new directories and ran nodetool refresh. However, a lot of > data seems to be missing. > > I ran nodetool repair on all three nodes one by one. First two nodes took > around 20 minutes (each) to complete. Third node took a lot of time to > repair and did not complete even in 14 hours. Eventually I had to stop it > manually. > > *nodetool compactionstats *give me the "pending tasks by table name" > traceback which can be viewed here: > https://gist.github.com/farazmateen/10adce4b2477457f0e20fc95176f66a3 > > *nodetool netstats* shows a lot of dropped gossip messages on all the > nodes. Here is the output from one of the nodes: > > Mode: NORMALNot sending any streams.Read Repair Statistics:Attempted: > 0Mismatch (Blocking): 1Mismatch (Background): 2Pool Name > Active Pending Completed DroppedLarge messages n/a > 0 92 1Small messages n/a > 0 355491 0Gossip messages n/a 5 > 3726945 286613 > > Is the problem related to token ranges? How can I find out token range for > each node? > What can I do to further debug and root cause this? > > On Tue, Apr 10, 2018 at 4:28 PM, Faraz Mateen <fmat...@an10.io> wrote: > >> Sorry for the late reply. I was trying to figure out some other approach >> to it. >> >> @Kurt - My previous cluster has 3 nodes but replication factor is 2. I am >> not exactly sure how I would handle the tokens. Can you explain that a bit? >> >> @Michael - Actually, my DC/OS cluster has an older version than my >> previous cluster. However both of them have hash with their data >> directories. Previous cluster is on version 3.9 while new DC/OS cluster is >> on 3.0.16. >> >> >> On Fri, Apr 6, 2018 at 2:35 PM, kurt greaves <k...@instaclustr.com> >> wrote: >> >>> Without looking at the code I'd say maybe the keyspaces are displayed >>> purely because the directories exist (but it seems unlikely). The process >>> you should follow instead is to exclude the system keyspaces for each node >>> and manually apply your schema, then upload your CFs into the correct >>> directory. Note this only works when RF=#nodes, if you have more nodes you >>> need to take tokens into account when restoring. >>> >>> >>> On Fri., 6 Apr. 2018, 17:16 Affan Syed, <as...@an10.io> wrote: >>> >>>> Michael, >>>> >>>> both of the folders are with hash, so I dont think that would be an >>>> issue. >>>> >>>> What is strange is why the tables dont show up if the keyspaces are >>>> visible. Shouldnt that be a meta data that can be edited once and then be >>>> visible? >>>> >>>> Affan >>>> >>>> - Affan >>>> >>>> On Thu, Apr 5, 2018 at 7:55 PM, Michael Shuler <mich...@pbandjelly.org> >>>> wrote: >>>> >>>>> On 04/05/2018 09:04 AM, Faraz Mateen wrote: >>>>> > >>>>> > For example, if the table is *data_main_bim_dn_10*, its data >>>>> directory >>>>> > is named data_main_bim_dn_10-a73202c02bf311e8b5106b13f463f8b9. I >>>>> created >>>>> > a new table with the same name through cqlsh. This resulted in >>>>> creation >>>>> > of another directory with a different hash i.e. >>>>> > data_main_bim_dn_10-c146e8d038c611e8b48cb7bc120612c9. I copied all >>>>> data >>>>> > from the former to the latter. >>>>> > >>>>> > Then I ran *"nodetool refresh ks1 data_main_bim_dn_10"*. After that >>>>> I >>>>> > was able to access all data contents through cqlsh. >>>>> > >>>>> > Now, the problem is, I have around 500 tables and the method I >>>>> mentioned >>>>> > above is quite cumbersome. Bulkloading through sstableloader or >>>>> remote >>>>> > seeding are also a couple of options but they will take a lot of >>>>> time. >>>>> > Does anyone know an easier way to shift all my data to new setup on >>>>> DC/OS? >>>>> >>>>> For upgrade support from older versions of C* that did not have the >>>>> hash >>>>> on the data directory, the table data dir can be just >>>>> `data_main_bim_dn_10` without the appended hash, as in your example. >>>>> >>>>> Give that a quick test to see if that simplifies things for you. >>>>> >>>>> -- >>>>> Kind regards, >>>>> Michael >>>>> >>>>> --------------------------------------------------------------------- >>>>> To unsubscribe, e-mail: user-unsubscr...@cassandra.apache.org >>>>> For additional commands, e-mail: user-h...@cassandra.apache.org >>>>> >>>>> >>>> >> >> >> -- >> Faraz Mateen >> > > > > -- > Faraz Mateen >