(Should still be able to complete, unless you’re running out of disk or memory 
or similar, but that’s why it’s streaming more than you expect)


-- 
Jeff Jirsa


> On Oct 15, 2017, at 1:51 PM, Jeff Jirsa <jji...@gmail.com> wrote:
> 
> I
> You’re adding the new node as rac3
> 
> The rack aware policy is going to make sure you get the rack diversity you 
> asked for by making sure one replica of each partition is in rac3, which is 
> going to blow up that instance
> 
> 
> 
> -- 
> Jeff Jirsa
> 
> 
>> On Oct 15, 2017, at 1:42 PM, Stefano Ortolani <ostef...@gmail.com> wrote:
>> 
>> Hi Jeff,
>> 
>> this my third attempt bootstrapping the node so I tried several tricks that 
>> might partially explain the output I am posting.
>> 
>> * To make the bootstrap incremental, I have been throttling the streams on 
>> all nodes to 1Mbits. I have selectively unthrottling one node at a time 
>> hoping that would unlock some routines compacting away redundant data 
>> (you'll see that nodetool netstats reports back fewer nodes than nodetool 
>> status).
>> * Since compactions have had the tendency of getting stuck (hundreds pending 
>> but none executing) in previous bootstraps, I've tried issuing a manual 
>> "nodetool compact" on the boostrapping node.
>> 
>> Having said that, this is the output of the commands,
>> 
>> Thanks a lot,
>> Stefano
>> 
>> nodetool status
>> Datacenter: DC1
>> ===============
>> Status=Up/Down
>> |/ State=Normal/Leaving/Joining/Moving
>> --  Address      Load       Tokens       Owns    Host ID                     
>>           Rack
>> UN  X.Y.33.8   342.4 GB   256          ?       
>> afaae414-30cc-439d-9785-1b7d35f74529  RAC1
>> UN  X.Y.81.4   325.98 GB  256          ?       
>> 00a96a5d-3bfd-497f-91f3-973b75146162  RAC2
>> UN  X.Y.33.4   348.81 GB  256          ?       
>> 1d8e6588-e25b-456a-8f29-0dedc35bda8e  RAC1
>> UN  X.Y.33.5   384.99 GB  256          ?       
>> 13d03fd2-7528-466b-b4b5-1b46508e2465  RAC1
>> UN  X.Y.81.5   336.27 GB  256          ?       
>> aa161400-6c0e-4bde-bcb3-b2e7e7840196  RAC2
>> UN  X.Y.33.6   377.22 GB  256          ?       
>> 43a393ba-6805-4e33-866f-124360174b28  RAC1
>> UN  X.Y.81.6   329.61 GB  256          ?       
>> 4c3c64ae-ef4f-4986-9341-573830416997  RAC2
>> UN  X.Y.33.7   344.25 GB  256          ?       
>> 03d81879-dc0d-4118-92e3-b3013dfde480  RAC1
>> UN  X.Y.81.7   324.93 GB  256          ?       
>> 24bbf4b6-9427-4ed1-a751-a55cc24cc756  RAC2
>> UN  X.Y.81.1   323.8 GB   256          ?       
>> 26244100-0565-4567-ae9c-0fc5346f5558  RAC2
>> UJ  X.Y.177.2  724.5 GB   256          ?       
>> e269a06b-c0c0-43a6-922c-f04c98898e0d  RAC3
>> UN  X.Y.81.2   337.83 GB  256          ?       
>> 09e29429-15ff-44d6-9742-ac95c83c4d9e  RAC2
>> UN  X.Y.81.3   326.4 GB   256          ?       
>> feaa7b27-7ab8-4fc2-b64a-c9df3dd86d97  RAC2
>> UN  X.Y.33.3   350.4 GB   256          ?       
>> cc115991-b7e7-4d06-87b5-8ad5efd45da5  RAC1
>> 
>> 
>> nodetool netstats -H | grep "Already received" -B 1
>>     /X.Y.81.4
>>         Receiving 1992 files, 103.68 GB total. Already received 515 files, 
>> 23.32 GB total
>> --
>>     /X.Y.81.7
>>         Receiving 1936 files, 89.35 GB total. Already received 554 files, 
>> 23.32 GB total
>> --
>>     /X.Y.81.5
>>         Receiving 1926 files, 95.69 GB total. Already received 545 files, 
>> 23.31 GB total
>> --
>>     /X.Y.81.2
>>         Receiving 1992 files, 100.81 GB total. Already received 537 files, 
>> 23.32 GB total
>> --
>>     /X.Y.81.3
>>         Receiving 1958 files, 104.72 GB total. Already received 503 files, 
>> 23.31 GB total
>> --
>>     /X.Y.81.1
>>         Receiving 2034 files, 104.51 GB total. Already received 520 files, 
>> 23.33 GB total
>> --
>>     /X.Y.81.6
>>         Receiving 1962 files, 96.19 GB total. Already received 547 files, 
>> 23.32 GB total
>> --
>>     /X.Y.33.5
>>         Receiving 2121 files, 97.44 GB total. Already received 601 files, 
>> 23.32 GB total
>> 
>> nodetool tpstats
>> Pool Name                    Active   Pending      Completed   Blocked  All 
>> time blocked
>> MutationStage                     0         0      828367015         0       
>>           0
>> ViewMutationStage                 0         0              0         0       
>>           0
>> ReadStage                         0         0              0         0       
>>           0
>> RequestResponseStage              0         0             13         0       
>>           0
>> ReadRepairStage                   0         0              0         0       
>>           0
>> CounterMutationStage              0         0              0         0       
>>           0
>> MiscStage                         0         0              0         0       
>>           0
>> CompactionExecutor                1         1          12150         0       
>>           0
>> MemtableReclaimMemory             0         0           7368         0       
>>           0
>> PendingRangeCalculator            0         0             14         0       
>>           0
>> GossipStage                       0         0         599329         0       
>>           0
>> SecondaryIndexManagement          0         0              0         0       
>>           0
>> HintsDispatcher                   0         0              0         0       
>>           0
>> MigrationStage                    0         0             27         0       
>>           0
>> MemtablePostFlush                 0         0           8112         0       
>>           0
>> ValidationExecutor                0         0              0         0       
>>           0
>> Sampler                           0         0              0         0       
>>           0
>> MemtableFlushWriter               0         0           7368         0       
>>           0
>> InternalResponseStage             0         0             25         0       
>>           0
>> AntiEntropyStage                  0         0              0         0       
>>           0
>> CacheCleanupExecutor              0         0              0         0       
>>           0
>> 
>> Message type           Dropped
>> READ                         0
>> RANGE_SLICE                  0
>> _TRACE                       0
>> HINT                         0
>> MUTATION                     1
>> COUNTER_MUTATION             0
>> BATCH_STORE                  0
>> BATCH_REMOVE                 0
>> REQUEST_RESPONSE             0
>> PAGED_RANGE                  0
>> READ_REPAIR                  0
>> 
>> nodetool compactionstats -H
>> pending tasks: 776
>>                                      id   compaction type         keyspace   
>>                 table   completed     total    unit   progress
>>    24d039f2-b1e6-11e7-ac57-3d25e38b2f5c        Compaction   keyspace_1   
>> table_1     4.85 GB   7.67 GB   bytes     63.25%
>> Active compaction remaining time :        n/a
>> 
>> 
>>> On Sun, Oct 15, 2017 at 9:27 PM, Jeff Jirsa <jji...@gmail.com> wrote:
>>> Can you post (anonymize as needed) nodetool status, nodetool netstats, 
>>> nodetool tpstats, and nodetool compctionstats ?
>>> 
>>> -- 
>>> Jeff Jirsa
>>> 
>>> 
>>>> On Oct 15, 2017, at 1:14 PM, Stefano Ortolani <ostef...@gmail.com> wrote:
>>>> 
>>>> Hi Jeff,
>>>> 
>>>> that would be 3.0.15, single disk, vnodes enabled (num_tokens 256).
>>>> 
>>>> Stefano
>>>> 
>>>>> On Sun, Oct 15, 2017 at 9:11 PM, Jeff Jirsa <jji...@gmail.com> wrote:
>>>>> What version?
>>>>> 
>>>>> Single disk or JBOD?
>>>>> 
>>>>> Vnodes?
>>>>> 
>>>>> -- 
>>>>> Jeff Jirsa
>>>>> 
>>>>> 
>>>>>> On Oct 15, 2017, at 12:49 PM, Stefano Ortolani <ostef...@gmail.com> 
>>>>>> wrote:
>>>>>> 
>>>>>> Hi all,
>>>>>> 
>>>>>> I have been trying "-Dcassandra.disable_stcs_in_l0=true", but no luck so 
>>>>>> far. 
>>>>>> Based on the source code it seems that this option doesn't affect 
>>>>>> compactions while bootstrapping.
>>>>>> 
>>>>>> I am getting quite confused as it seems I am not able to bootstrap a 
>>>>>> node if I don't have at least 6/7 times the disk space used by other 
>>>>>> nodes.
>>>>>> This is weird. The host I am bootstrapping is using a SSD. Also 
>>>>>> compaction throughput is unthrottled (set to 0) and the compacting 
>>>>>> threads are set to 8.
>>>>>> Nevertheless, primary ranges from other nodes are being streamed, but 
>>>>>> data is never compacted away.
>>>>>> 
>>>>>> Does anybody know anything else I could try?
>>>>>> 
>>>>>> Cheers,
>>>>>> Stefano
>>>>>> 
>>>>>>> On Fri, Oct 13, 2017 at 3:58 PM, Stefano Ortolani <ostef...@gmail.com> 
>>>>>>> wrote:
>>>>>>> Other little update: at the same time I see the number of pending tasks 
>>>>>>> stuck (in this case at 1847); restarting the node doesn't help, so I 
>>>>>>> can't really force the node to "digest" all those compactions. In the 
>>>>>>> meanwhile the disk occupied is already twice the average load I have on 
>>>>>>> other nodes.
>>>>>>> 
>>>>>>> Feeling more and more puzzled here :S
>>>>>>> 
>>>>>>>> On Fri, Oct 13, 2017 at 1:28 PM, Stefano Ortolani <ostef...@gmail.com> 
>>>>>>>> wrote:
>>>>>>>> I have been trying to add another node to the cluster (after upgrading 
>>>>>>>> to 3.0.15) and I just noticed through "nodetool netstats" that all 
>>>>>>>> nodes have been streaming to the joining node approx 1/3 of their 
>>>>>>>> SSTables, basically their whole primary range (using RF=3)?
>>>>>>>> 
>>>>>>>> Is this expected/normal? 
>>>>>>>> I was under the impression only the necessary SSTables were going to 
>>>>>>>> be streamed...
>>>>>>>> 
>>>>>>>> Thanks for the help,
>>>>>>>> Stefano
>>>>>>>> 
>>>>>>>> 
>>>>>>>> On Wed, Aug 23, 2017 at 1:37 PM, kurt greaves <k...@instaclustr.com> 
>>>>>>>> wrote:
>>>>>>>>>> But if it also streams, it means I'd still be under-pressure if I am 
>>>>>>>>>> not mistaken. I am under the assumption that the compactions are the 
>>>>>>>>>> by-product of streaming too many SStables at the same time, and not 
>>>>>>>>>> because of my current write load.
>>>>>>>>> 
>>>>>>>>> Ah yeah I wasn't thinking about the capacity problem, more of the 
>>>>>>>>> performance impact from the node being backed up with compactions. If 
>>>>>>>>> you haven't already, you should try disable stcs in l0 on the joining 
>>>>>>>>> node. You will likely still need to do a lot of compactions, but 
>>>>>>>>> generally they should be smaller. The  option is 
>>>>>>>>> -Dcassandra.disable_stcs_in_l0=true
>>>>>>>>>>  I just noticed you were mentioning L1 tables too. Why would that 
>>>>>>>>>> affect the disk footprint?
>>>>>>>>> 
>>>>>>>>> If you've been doing a lot of STCS in L0, you generally end up with 
>>>>>>>>> some large SSTables. These will eventually have to be compacted with 
>>>>>>>>> L1. Could also be suffering the problem of streamed SSTables causing 
>>>>>>>>> large cross-level compactions in the higher levels as well.
>>>>>>>>> ​
>>>>>>>> 
>>>>>>> 
>>>>>> 
>>>> 
>> 

Reply via email to