It is a Cassandra bug. The workaround is to change system_distributed
keyspce replication strategy to something as below:

      alter keyspace  system_distributed with replication = {'class':
'NetworkTopologyStrategy', 'DC1': '3', 'DC2': '3', 'DC3': '3'};

You may see similar problem for other system keyspaces. Do the same thing.

Cheers,

Yabin

On Thu, Sep 22, 2016 at 1:44 PM, Timo Ahokas <timo.aho...@gmail.com> wrote:

> Hi Alain,
>
> Our normal user keyspaces have RF3 in all DCs, e.g:
>
> create keyspace reporting with replication = {'class':
> 'NetworkTopologyStrategy', 'DC1': '3', 'DC2': '3', 'DC3': '3'};
>
> Any idea would it be safe to change the system_distributed keyspace to
> match this?
>
> -Timo
>
> On 22 September 2016 at 19:23, Timo Ahokas <timo.aho...@gmail.com> wrote:
>
>> Hi Alain,
>>
>> Thanks a lot for a helping out!
>>
>> Some of the basic keyspace / cluster info you requested:
>>
>> # echo "DESCRIBE KEYSPACE system_distributed;" | cqlsh
>>
>> CREATE KEYSPACE system_distributed WITH replication = {'class':
>> 'SimpleStrategy', 'replication_factor': '3'}  AND durable_writes = true;
>>
>> CREATE TABLE system_distributed.repair_history (
>>
>>    keyspace_name text,
>>
>>    columnfamily_name text,
>>
>>    id timeuuid,
>>
>>    coordinator inet,
>>
>>    exception_message text,
>>
>>    exception_stacktrace text,
>>
>>    finished_at timestamp,
>>
>>    parent_id timeuuid,
>>
>>    participants set<inet>,
>>
>>    range_begin text,
>>
>>    range_end text,
>>
>>    started_at timestamp,
>>
>>    status text,
>>
>>    PRIMARY KEY ((keyspace_name, columnfamily_name), id)
>>
>> ) WITH CLUSTERING ORDER BY (id ASC)
>>
>>    AND bloom_filter_fp_chance = 0.01
>>
>>    AND caching = {'keys': 'ALL', 'rows_per_partition': 'NONE'}
>>
>>    AND comment = 'Repair history'
>>
>>    AND compaction = {'class': 'org.apache.cassandra.db.compa
>> ction.SizeTieredCompactionStrategy', 'max_threshold': '32',
>> 'min_threshold': '4'}
>>
>>    AND compression = {'chunk_length_in_kb': '64', 'class': '
>> org.apache.cassandra.io.compress.LZ4Compressor'}
>>
>>    AND crc_check_chance = 1.0
>>
>>    AND dclocal_read_repair_chance = 0.0
>>
>>    AND default_time_to_live = 0
>>
>>    AND gc_grace_seconds = 0
>>
>>    AND max_index_interval = 2048
>>
>>    AND memtable_flush_period_in_ms = 3600000
>>
>>    AND min_index_interval = 128
>>
>>    AND read_repair_chance = 0.0
>>
>>    AND speculative_retry = '99PERCENTILE';
>>
>> CREATE TABLE system_distributed.parent_repair_history (
>>
>>    parent_id timeuuid PRIMARY KEY,
>>
>>    columnfamily_names set<text>,
>>
>>    exception_message text,
>>
>>    exception_stacktrace text,
>>
>>    finished_at timestamp,
>>
>>    keyspace_name text,
>>
>>    requested_ranges set<text>,
>>
>>    started_at timestamp,
>>
>>    successful_ranges set<text>
>>
>> ) WITH bloom_filter_fp_chance = 0.01
>>
>>    AND caching = {'keys': 'ALL', 'rows_per_partition': 'NONE'}
>>
>>    AND comment = 'Repair history'
>>
>>    AND compaction = {'class': 'org.apache.cassandra.db.compa
>> ction.SizeTieredCompactionStrategy', 'max_threshold': '32',
>> 'min_threshold': '4'}
>>
>>    AND compression = {'chunk_length_in_kb': '64', 'class': '
>> org.apache.cassandra.io.compress.LZ4Compressor'}
>>
>>    AND crc_check_chance = 1.0
>>
>>    AND dclocal_read_repair_chance = 0.0
>>
>>    AND default_time_to_live = 0
>>
>>    AND gc_grace_seconds = 0
>>
>>    AND max_index_interval = 2048
>>
>>    AND memtable_flush_period_in_ms = 3600000
>>
>>    AND min_index_interval = 128
>>
>>    AND read_repair_chance = 0.0
>>
>>    AND speculative_retry = '99PERCENTILE';
>>
>>
>> CREATE TABLE system_distributed.repair_history (
>>
>>    keyspace_name text,
>>
>>    columnfamily_name text,
>>
>>    id timeuuid,
>>
>>    coordinator inet,
>>
>>    exception_message text,
>>
>>    exception_stacktrace text,
>>
>>    finished_at timestamp,
>>
>>    parent_id timeuuid,
>>
>>    participants set<inet>,
>>
>>    range_begin text,
>>
>>    range_end text,
>>
>>    started_at timestamp,
>>
>>    status text,
>>
>>    PRIMARY KEY ((keyspace_name, columnfamily_name), id)
>>
>> ) WITH CLUSTERING ORDER BY (id ASC)
>>
>>    AND bloom_filter_fp_chance = 0.01
>>
>>    AND caching = {'keys': 'ALL', 'rows_per_partition': 'NONE'}
>>
>>    AND comment = 'Repair history'
>>
>>    AND compaction = {'class': 'org.apache.cassandra.db.compa
>> ction.SizeTieredCompactionStrategy', 'max_threshold': '32',
>> 'min_threshold': '4'}
>>
>>    AND compression = {'chunk_length_in_kb': '64', 'class': '
>> org.apache.cassandra.io.compress.LZ4Compressor'}
>>
>>    AND crc_check_chance = 1.0
>>
>>    AND dclocal_read_repair_chance = 0.0
>>
>>    AND default_time_to_live = 0
>>
>>    AND gc_grace_seconds = 0
>>
>>    AND max_index_interval = 2048
>>
>>    AND memtable_flush_period_in_ms = 3600000
>>
>>    AND min_index_interval = 128
>>
>>    AND read_repair_chance = 0.0
>>
>>    AND speculative_retry = '99PERCENTILE';
>>
>> CREATE TABLE system_distributed.parent_repair_history (
>>
>>    parent_id timeuuid PRIMARY KEY,
>>
>>    columnfamily_names set<text>,
>>
>>    exception_message text,
>>
>>    exception_stacktrace text,
>>
>>    finished_at timestamp,
>>
>>    keyspace_name text,
>>
>>    requested_ranges set<text>,
>>
>>    started_at timestamp,
>>
>>    successful_ranges set<text>
>>
>> ) WITH bloom_filter_fp_chance = 0.01
>>
>>    AND caching = {'keys': 'ALL', 'rows_per_partition': 'NONE'}
>>
>>    AND comment = 'Repair history'
>>
>>    AND compaction = {'class': 'org.apache.cassandra.db.compa
>> ction.SizeTieredCompactionStrategy', 'max_threshold': '32',
>> 'min_threshold': '4'}
>>
>>    AND compression = {'chunk_length_in_kb': '64', 'class': '
>> org.apache.cassandra.io.compress.LZ4Compressor'}
>>
>>    AND crc_check_chance = 1.0
>>
>>    AND dclocal_read_repair_chance = 0.0
>>
>>    AND default_time_to_live = 0
>>
>>    AND gc_grace_seconds = 0
>>
>>    AND max_index_interval = 2048
>>
>>    AND memtable_flush_period_in_ms = 3600000
>>
>>    AND min_index_interval = 128
>>
>>    AND read_repair_chance = 0.0
>>
>>    AND speculative_retry = '99PERCENTILE';
>>
>>
>>
>> # nodetool status
>>
>> Datacenter: DC1
>>
>> ===============
>>
>> Status=Up/Down
>>
>> |/ State=Normal/Leaving/Joining/Moving
>>
>> --  Address         Load       Tokens       Owns    Host ID
>>                               Rack
>>
>> UN  xxx.xxx.145.5    693,63 GB  256          ?
>>       6f1a0fdd-e3f9-474d-9a49-7bfeeadb3f56  RAC1
>>
>> UN  xxx.xxx.145.225  648,55 GB  256          ?
>>       f900847a-63e4-44c5-b4d7-e439c7cb6a8e  RAC1
>>
>> UN  xxx.xxx.145.160  608,31 GB  256          ?
>>       d257e76d-9e40-4215-94c7-3076c8ff4b7f  RAC1
>>
>> UN  xxx.xxx.145.67   552,93 GB  256          ?
>>       1d47cbdd-cdf1-45b6-aa0e-0c6123899dca  RAC1
>>
>> UN  xxx.xxx.145.227  636,68 GB  256          ?
>>       47e5f207-f9fd-4a86-be8a-66e7630d1baa  RAC1
>>
>> UN  xxx.xxx.146.105  610,9 GB   256          ?
>>       8edf1aaa-49d1-4e4b-9f09-99c4ab6136c2  RAC1
>>
>> UN  xxx.xxx.147.136  666,82 GB  256          ?
>>       bafbf6a2-cff9-489f-a2dd-fc6e8cb08ff6  RAC1
>>
>> UN  xxx.xxx.146.213  609,79 GB  256          ?
>>       6416275c-7570-48a9-957f-2daca71d31aa  RAC1
>>
>> UN  xxx.xxx.146.20   664,44 GB  256          ?
>>       b016df7e-f694-4ef3-928c-8783853e9a07  RAC1
>>
>> UN  xxx.xxx.146.209  615,44 GB  256          ?
>>       898e6d98-1b92-4e86-b52c-f851fd4fda71  RAC1
>>
>> UN  xxx.xxx.146.241  668,91 GB  256          ?
>>       0b5d4c6c-4b7c-4265-92bc-ad74464d85cc  RAC1
>>
>> UN  xxx.xxx.147.211  641,33 GB  256          ?
>>       16cdc4a7-b694-4125-91d6-05b9099cb765  RAC1
>>
>> UN  xxx.xxx.147.125  647,03 GB  256          ?
>>       2e97ed0a-039c-413b-9693-a87fadf40f82  RAC1
>>
>> Datacenter: DC2
>>
>> ===============
>>
>> Status=Up/Down
>>
>> |/ State=Normal/Leaving/Joining/Moving
>>
>> --  Address         Load       Tokens       Owns    Host ID
>>                               Rack
>>
>> UN  xxx.xxx.7.99     18,76 MB   256          ?
>>       d7b907ad-15f5-4c79-962c-c604a5723a7b  RAC1
>>
>> UN  xxx.xxx.6.135    16,04 MB   256          ?
>>       463f480a-baf3-4230-86b7-1106251ebfad  RAC1
>>
>> UN  xxx.xxx.7.229    17,36 MB   256          ?
>>       9487a975-6183-43b8-9208-cd8e09a0ae18  RAC1
>>
>> UN  xxx.xxx.7.5      14,01 MB   256          ?
>>       ae039e49-4d79-4e4e-87bd-921cd6b3291a  RAC1
>>
>> UN  xxx.xxx.7.4      14,93 MB   256          ?
>>       122a47fb-b5ca-46d1-aae9-e6993ab58b66  RAC1
>>
>> UN  xxx.xxx.6.10     16,77 MB   256          ?
>>       bbb66068-bf06-438d-81ee-965e201e8fff  RAC1
>>
>> UN  xxx.xxx.6.15     14,95 MB   256          ?
>>       668a864d-9fd3-41b7-88fb-824e75e71953  RAC1
>>
>> UN  xxx.xxx.7.140    17,38 MB   256          ?
>>       7b016c96-eaa1-4ee1-8657-f4260c70ed37  RAC1
>>
>> UN  xxx.xxx.7.113    19,14 MB   256          ?
>>       46c06c44-ce2f-4ab6-9597-a1314cecf9bc  RAC1
>>
>> UN  xxx.xxx.6.118    16,7 MB    256          ?
>>       9c3c3107-a1d3-4254-ad10-909713a38f8c  RAC1
>>
>> UN  xxx.xxx.6.248    17,29 MB   256          ?
>>       35ff4d3d-d993-468b-9a54-88b40ceec6d4  RAC1
>>
>> UN  xxx.xxx.5.24     16,55 MB   256          ?
>>       5f1f34bd-110f-4d60-9af5-a3abd01b55a5  RAC1
>>
>> UN  xxx.xxx.7.189    16,63 MB   256          ?
>>       be7cbf84-5838-487a-8bd4-b340a1c70fab  RAC1
>>
>> UN  xxx.xxx.5.124    20,37 MB   256          ?
>>       638f2656-fb92-4b70-ba2a-251a749c4c58  RAC1
>>
>> UN  xxx.xxx.6.60     24,57 MB   256          ?
>>       cf16209a-a9a0-4f27-9341-c76d47e50261  RAC1
>>
>> Datacenter: DC3
>>
>> ===============
>>
>> Status=Up/Down
>>
>> |/ State=Normal/Leaving/Joining/Moving
>>
>> --  Address         Load       Tokens       Owns    Host ID
>>                               Rack
>>
>> UN  xxx.xxx.151.102  389,41 GB  256          ?
>>       1740a473-e304-467c-a682-d1b4b0595ffa  RAC1
>>
>> UN  xxx.xxx.149.161  367,82 GB  256          ?
>>       3a5322d4-e49f-45ed-85b5-fd658502859c  RAC1
>>
>> UN  xxx.xxx.149.226  390,88 GB  256          ?
>>       b8ca4576-2632-4198-ac87-10243c0c554e  RAC1
>>
>> UN  xxx.xxx.151.162  408,35 GB  256          ?
>>       54d3dd90-f9ab-47c2-ae31-5f3e87b91e2a  RAC1
>>
>> UN  xxx.xxx.149.109  369,33 GB  256          ?
>>       9172c7d8-0c55-4e8e-a17b-89fdb0dce878  RAC1
>>
>> UN  xxx.xxx.150.172  362,32 GB  256          ?
>>       ba394a29-1a0c-4f50-ab85-4db19011b190  RAC1
>>
>> UN  xxx.xxx.149.238  388,98 GB  256          ?
>>       a3d7228c-ccb4-4787-a4bb-f7720aeedc8e  RAC1
>>
>> UN  xxx.xxx.151.232  435,31 GB  256          ?
>>       500a43ab-ae77-4a07-876c-171cb34c549b  RAC1
>>
>> UN  xxx.xxx.151.43   410,69 GB  256          ?
>>       b8bc80e2-2107-447a-85e4-57a39dc9c595  RAC1
>>
>> UN  xxx.xxx.151.139  407,47 GB  256          ?
>>       ecfa4ba7-7783-47a4-8b17-aadc91a3e776  RAC1
>>
>> UN  xxx.xxx.151.213  375,05 GB  256          ?
>>       9bf53ee1-53d4-4d18-a58e-0b0a17e18a69  RAC1
>>
>> UN  xxx.xxx.149.177  401,91 GB  256          ?
>>       b903faf1-1ae9-45ad-bdce-3c9377458a03  RAC1
>>
>> UN  xxx.xxx.150.145  388,76 GB  256          ?
>>       1c4e4232-db27-4cc1-9985-9eb7f0b984d1  RAC1
>>
>> UN  xxx.xxx.149.48   385,43 GB  256          ?
>>       ad3ea388-203c-4b26-a368-934a6105cc6e  RAC1
>>
>> UN  xxx.xxx.150.189  384,52 GB  256          ?
>>       f361ebad-b0a6-47b7-a55c-245c98f84508  RAC1
>>
>> UN  xxx.xxx.151.220  357,56 GB  256          ?
>>       feb814e6-6d2f-4cef-ae3b-4924c1cbac60  RAC1
>>
>> UN  xxx.xxx.149.121  355,64 GB  256          ?
>>       47fbb104-6a5a-49c0-b086-3f14c853c83b  RAC1
>>
>> UN  xxx.xxx.151.218  416,57 GB  256          ?
>>       bbb21d16-da85-4cfd-87d4-2333c8b02dad  RAC1
>>
>> UN  xxx.xxx.150.26   383,06 GB  256          ?
>>       1ca0085d-93a5-4650-891a-b45f988150a4  RAC1
>>
>> Note: Non-system keyspaces don't have the same replication settings,
>> effective ownership information is meaningless
>>
>>
>>
>> # nodetool status system_distributed
>>
>> Datacenter: DC1
>>
>> ===============
>>
>> Status=Up/Down
>>
>> |/ State=Normal/Leaving/Joining/Moving
>>
>> --  Address         Load       Tokens       Owns (effective)  Host ID
>>                               Rack
>>
>> UN  xxx.xxx.145.5    693,63 GB  256          6,2%
>>              6f1a0fdd-e3f9-474d-9a49-7bfeeadb3f56  RAC1
>>
>> UN  xxx.xxx.145.225  648,55 GB  256          6,8%
>>              f900847a-63e4-44c5-b4d7-e439c7cb6a8e  RAC1
>>
>> UN  xxx.xxx.145.160  608,31 GB  256          6,5%
>>              d257e76d-9e40-4215-94c7-3076c8ff4b7f  RAC1
>>
>> UN  xxx.xxx.145.67   552,93 GB  256          6,1%
>>              1d47cbdd-cdf1-45b6-aa0e-0c6123899dca  RAC1
>>
>> UN  xxx.xxx.145.227  636,68 GB  256          6,0%
>>              47e5f207-f9fd-4a86-be8a-66e7630d1baa  RAC1
>>
>> UN  xxx.xxx.146.105  610,9 GB   256          6,1%
>>              8edf1aaa-49d1-4e4b-9f09-99c4ab6136c2  RAC1
>>
>> UN  xxx.xxx.147.136  666,82 GB  256          6,3%
>>              bafbf6a2-cff9-489f-a2dd-fc6e8cb08ff6  RAC1
>>
>> UN  xxx.xxx.146.213  609,79 GB  256          6,0%
>>              6416275c-7570-48a9-957f-2daca71d31aa  RAC1
>>
>> UN  xxx.xxx.146.20   664,44 GB  256          7,0%
>>              b016df7e-f694-4ef3-928c-8783853e9a07  RAC1
>>
>> UN  xxx.xxx.146.209  615,44 GB  256          6,6%
>>              898e6d98-1b92-4e86-b52c-f851fd4fda71  RAC1
>>
>> UN  xxx.xxx.146.241  668,91 GB  256          6,2%
>>              0b5d4c6c-4b7c-4265-92bc-ad74464d85cc  RAC1
>>
>> UN  xxx.xxx.147.211  641,33 GB  256          6,5%
>>              16cdc4a7-b694-4125-91d6-05b9099cb765  RAC1
>>
>> UN  xxx.xxx.147.125  647,03 GB  256          6,3%
>>              2e97ed0a-039c-413b-9693-a87fadf40f82  RAC1
>>
>> Datacenter: DC2
>>
>> ===============
>>
>> Status=Up/Down
>>
>> |/ State=Normal/Leaving/Joining/Moving
>>
>> --  Address         Load       Tokens       Owns (effective)  Host ID
>>                               Rack
>>
>> UN  xxx.xxx.7.99     18,76 MB   256          6,3%
>>              d7b907ad-15f5-4c79-962c-c604a5723a7b  RAC1
>>
>> UN  xxx.xxx.6.135    16,04 MB   256          6,1%
>>              463f480a-baf3-4230-86b7-1106251ebfad  RAC1
>>
>> UN  xxx.xxx.7.229    17,36 MB   256          5,9%
>>              9487a975-6183-43b8-9208-cd8e09a0ae18  RAC1
>>
>> UN  xxx.xxx.7.5      14,01 MB   256          6,2%
>>              ae039e49-4d79-4e4e-87bd-921cd6b3291a  RAC1
>>
>> UN  xxx.xxx.7.4      14,93 MB   256          6,4%
>>              122a47fb-b5ca-46d1-aae9-e6993ab58b66  RAC1
>>
>> UN  xxx.xxx.6.10     16,77 MB   256          6,4%
>>              bbb66068-bf06-438d-81ee-965e201e8fff  RAC1
>>
>> UN  xxx.xxx.6.15     14,95 MB   256          6,1%
>>              668a864d-9fd3-41b7-88fb-824e75e71953  RAC1
>>
>> UN  xxx.xxx.7.140    17,38 MB   256          6,7%
>>              7b016c96-eaa1-4ee1-8657-f4260c70ed37  RAC1
>>
>> UN  xxx.xxx.7.113    19,14 MB   256          6,8%
>>              46c06c44-ce2f-4ab6-9597-a1314cecf9bc  RAC1
>>
>> UN  xxx.xxx.6.118    16,7 MB    256          6,7%
>>              9c3c3107-a1d3-4254-ad10-909713a38f8c  RAC1
>>
>> UN  xxx.xxx.6.248    17,29 MB   256          6,9%
>>              35ff4d3d-d993-468b-9a54-88b40ceec6d4  RAC1
>>
>> UN  xxx.xxx.5.24     16,55 MB   256          6,8%
>>              5f1f34bd-110f-4d60-9af5-a3abd01b55a5  RAC1
>>
>> UN  xxx.xxx.7.189    16,63 MB   256          6,2%
>>              be7cbf84-5838-487a-8bd4-b340a1c70fab  RAC1
>>
>> UN  xxx.xxx.5.124    20,37 MB   256          6,3%
>>              638f2656-fb92-4b70-ba2a-251a749c4c58  RAC1
>>
>> UN  xxx.xxx.6.60     24,57 MB   256          6,4%
>>              cf16209a-a9a0-4f27-9341-c76d47e50261  RAC1
>>
>> Datacenter: DC3
>>
>> ===============
>>
>> Status=Up/Down
>>
>> |/ State=Normal/Leaving/Joining/Moving
>>
>> --  Address         Load       Tokens       Owns (effective)  Host ID
>>                               Rack
>>
>> UN  xxx.xxx.151.102  389,41 GB  256          6,4%
>>              1740a473-e304-467c-a682-d1b4b0595ffa  RAC1
>>
>> UN  xxx.xxx.149.161  367,82 GB  256          6,3%
>>              3a5322d4-e49f-45ed-85b5-fd658502859c  RAC1
>>
>> UN  xxx.xxx.149.226  390,88 GB  256          6,2%
>>              b8ca4576-2632-4198-ac87-10243c0c554e  RAC1
>>
>> UN  xxx.xxx.151.162  408,35 GB  256          6,4%
>>              54d3dd90-f9ab-47c2-ae31-5f3e87b91e2a  RAC1
>>
>> UN  xxx.xxx.149.109  369,33 GB  256          6,2%
>>              9172c7d8-0c55-4e8e-a17b-89fdb0dce878  RAC1
>>
>> UN  xxx.xxx.150.172  362,32 GB  256          6,0%
>>              ba394a29-1a0c-4f50-ab85-4db19011b190  RAC1
>>
>> UN  xxx.xxx.149.238  388,98 GB  256          6,4%
>>              a3d7228c-ccb4-4787-a4bb-f7720aeedc8e  RAC1
>>
>> UN  xxx.xxx.151.232  435,31 GB  256          6,6%
>>              500a43ab-ae77-4a07-876c-171cb34c549b  RAC1
>>
>> UN  xxx.xxx.151.43   410,69 GB  256          6,2%
>>              b8bc80e2-2107-447a-85e4-57a39dc9c595  RAC1
>>
>> UN  xxx.xxx.151.139  407,47 GB  256          6,2%
>>              ecfa4ba7-7783-47a4-8b17-aadc91a3e776  RAC1
>>
>> UN  xxx.xxx.151.213  375,05 GB  256          6,5%
>>              9bf53ee1-53d4-4d18-a58e-0b0a17e18a69  RAC1
>>
>> UN  xxx.xxx.149.177  401,91 GB  256          6,6%
>>              b903faf1-1ae9-45ad-bdce-3c9377458a03  RAC1
>>
>> UN  xxx.xxx.150.145  388,76 GB  256          7,1%
>>              1c4e4232-db27-4cc1-9985-9eb7f0b984d1  RAC1
>>
>> UN  xxx.xxx.149.48   385,43 GB  256          6,2%
>>              ad3ea388-203c-4b26-a368-934a6105cc6e  RAC1
>>
>> UN  xxx.xxx.150.189  384,52 GB  256          6,4%
>>              f361ebad-b0a6-47b7-a55c-245c98f84508  RAC1
>>
>> UN  xxx.xxx.151.220  357,56 GB  256          6,1%
>>              feb814e6-6d2f-4cef-ae3b-4924c1cbac60  RAC1
>>
>> UN  xxx.xxx.149.121  355,64 GB  256          6,4%
>>              47fbb104-6a5a-49c0-b086-3f14c853c83b  RAC1
>>
>> UN  xxx.xxx.151.218  416,57 GB  256          6,3%
>>              bbb21d16-da85-4cfd-87d4-2333c8b02dad  RAC1
>> UN  xxx.xxx.150.26   383,06 GB  256          6,7%
>>              1ca0085d-93a5-4650-891a-b45f988150a4  RAC1
>>
>> DC1 and DC3 are the old data centers. DC2 is the new one being added (as
>> seen from the data loads).
>>
>> For the snitch we are using GossipingPropertyFileSnitch and a
>> cassandra-rackdc.properties with config such as:
>> dc=DC1
>> rack=RAC1
>>
>> Just noticed that we also have cassandra-topology.properties present on
>> the nodes, but it's up-to-date with all the nodes from the 3 data centers.
>>
>> I was wondering on whether the replication settings for the
>> system_distributed keyspace might need a change, but didn't find any yet
>> documentation pointing to that.
>>
>> Best regards,
>> Timo
>>
>> On 22 September 2016 at 18:00, Alain RODRIGUEZ <arodr...@gmail.com>
>> wrote:
>>
>>> It could be a bug.
>>>
>>> Yet I am not very aware of this system_distributed keyspace, but from
>>> what I see, it is using a simple strategy:
>>>
>>> root@tlp-cassandra-2:~# echo "DESCRIBE KEYSPACE system_distributed;" |
>>> cqlsh $(hostname -I | awk '{print $1}')
>>>
>>> CREATE KEYSPACE system_distributed WITH replication = {'class':
>>> 'SimpleStrategy', 'replication_factor': '3'}  AND durable_writes = true;
>>>
>>> Let's first check some stuff. Could you share the output of:
>>>
>>>
>>>    - echo "DESCRIBE KEYSPACE system_distributed;" | cqlsh
>>>    [ip_address_of_the_server]
>>>    - nodetool status
>>>    - nodetool status system_distributed
>>>    - Let us know about the snitch you are using and the corresponding
>>>    configuration.
>>>
>>>
>>> I am trying to make sure the command you used is expected to work, given
>>> your setup.
>>>
>>> My guess is this you might need to alter this keyspace accordingly to
>>> your cluster setup.
>>>
>>> Just guessing, hope that helps.
>>>
>>> C*heers,
>>> -----------------------
>>> Alain Rodriguez - @arodream - al...@thelastpickle.com
>>> France
>>>
>>> The Last Pickle - Apache Cassandra Consulting
>>> http://www.thelastpickle.com
>>>
>>> 2016-09-22 15:47 GMT+02:00 Timo Ahokas <timo.aho...@gmail.com>:
>>>
>>>> Hi,
>>>>
>>>> We have a Cassandra 3.0.8 cluster (recently upgraded from 2.1.15)
>>>> currently running in two data centers (13 and 19 nodes, RF3 in both). We
>>>> are adding a third data center before decommissioning one of the earlier
>>>> ones. Installing Cassandra (3.0.8) goes fine and all the nodes join the
>>>> cluster (not set to bootstrap, as documented in
>>>> https://docs.datastax.com/en/cassandra/3.0/cassandra/operati
>>>> ons/opsAddDCToCluster.html).
>>>>
>>>> When trying to rebuild nodes in the new DC from a previous DC (nodetool
>>>> rebuild -- DC1), we get the following error:
>>>>
>>>> Unable to find sufficient sources for streaming range
>>>> (597769692463489739,597931451954862346] in keyspace system_distributed
>>>>
>>>> The same error occurs which ever of the 2 existing DCs we try to
>>>> rebuild from.
>>>>
>>>> We run pr repairs (nodetool repair -pr) on all nodes twice a week via
>>>> cron.
>>>>
>>>> Any advice on how to get the rebuild started?
>>>>
>>>> Best regards,
>>>> Timo
>>>>
>>>>
>>>>
>>>>
>>>>
>>>
>>
>

Reply via email to