Hi Jeff, this my third attempt bootstrapping the node so I tried several tricks that might partially explain the output I am posting.
* To make the bootstrap incremental, I have been throttling the streams on all nodes to 1Mbits. I have selectively unthrottling one node at a time hoping that would unlock some routines compacting away redundant data (you'll see that nodetool netstats reports back fewer nodes than nodetool status). * Since compactions have had the tendency of getting stuck (hundreds pending but none executing) in previous bootstraps, I've tried issuing a manual "nodetool compact" on the boostrapping node. Having said that, this is the output of the commands, Thanks a lot, Stefano *nodetool status* Datacenter: DC1 =============== Status=Up/Down |/ State=Normal/Leaving/Joining/Moving -- Address Load Tokens Owns Host ID Rack UN X.Y.33.8 342.4 GB 256 ? afaae414-30cc-439d-9785-1b7d35f74529 RAC1 UN X.Y.81.4 325.98 GB 256 ? 00a96a5d-3bfd-497f-91f3-973b75146162 RAC2 UN X.Y.33.4 348.81 GB 256 ? 1d8e6588-e25b-456a-8f29-0dedc35bda8e RAC1 UN X.Y.33.5 384.99 GB 256 ? 13d03fd2-7528-466b-b4b5-1b46508e2465 RAC1 UN X.Y.81.5 336.27 GB 256 ? aa161400-6c0e-4bde-bcb3-b2e7e7840196 RAC2 UN X.Y.33.6 377.22 GB 256 ? 43a393ba-6805-4e33-866f-124360174b28 RAC1 UN X.Y.81.6 329.61 GB 256 ? 4c3c64ae-ef4f-4986-9341-573830416997 RAC2 UN X.Y.33.7 344.25 GB 256 ? 03d81879-dc0d-4118-92e3-b3013dfde480 RAC1 UN X.Y.81.7 324.93 GB 256 ? 24bbf4b6-9427-4ed1-a751-a55cc24cc756 RAC2 UN X.Y.81.1 323.8 GB 256 ? 26244100-0565-4567-ae9c-0fc5346f5558 RAC2 UJ X.Y.177.2 724.5 GB 256 ? e269a06b-c0c0-43a6-922c-f04c98898e0d RAC3 UN X.Y.81.2 337.83 GB 256 ? 09e29429-15ff-44d6-9742-ac95c83c4d9e RAC2 UN X.Y.81.3 326.4 GB 256 ? feaa7b27-7ab8-4fc2-b64a-c9df3dd86d97 RAC2 UN X.Y.33.3 350.4 GB 256 ? cc115991-b7e7-4d06-87b5-8ad5efd45da5 RAC1 *nodetool netstats -H | grep "Already received" -B 1* /X.Y.81.4 Receiving 1992 files, 103.68 GB total. Already received 515 files, 23.32 GB total -- /X.Y.81.7 Receiving 1936 files, 89.35 GB total. Already received 554 files, 23.32 GB total -- /X.Y.81.5 Receiving 1926 files, 95.69 GB total. Already received 545 files, 23.31 GB total -- /X.Y.81.2 Receiving 1992 files, 100.81 GB total. Already received 537 files, 23.32 GB total -- /X.Y.81.3 Receiving 1958 files, 104.72 GB total. Already received 503 files, 23.31 GB total -- /X.Y.81.1 Receiving 2034 files, 104.51 GB total. Already received 520 files, 23.33 GB total -- /X.Y.81.6 Receiving 1962 files, 96.19 GB total. Already received 547 files, 23.32 GB total -- /X.Y.33.5 Receiving 2121 files, 97.44 GB total. Already received 601 files, 23.32 GB total *nodetool tpstats* Pool Name Active Pending Completed Blocked All time blocked MutationStage 0 0 828367015 0 0 ViewMutationStage 0 0 0 0 0 ReadStage 0 0 0 0 0 RequestResponseStage 0 0 13 0 0 ReadRepairStage 0 0 0 0 0 CounterMutationStage 0 0 0 0 0 MiscStage 0 0 0 0 0 CompactionExecutor 1 1 12150 0 0 MemtableReclaimMemory 0 0 7368 0 0 PendingRangeCalculator 0 0 14 0 0 GossipStage 0 0 599329 0 0 SecondaryIndexManagement 0 0 0 0 0 HintsDispatcher 0 0 0 0 0 MigrationStage 0 0 27 0 0 MemtablePostFlush 0 0 8112 0 0 ValidationExecutor 0 0 0 0 0 Sampler 0 0 0 0 0 MemtableFlushWriter 0 0 7368 0 0 InternalResponseStage 0 0 25 0 0 AntiEntropyStage 0 0 0 0 0 CacheCleanupExecutor 0 0 0 0 0 Message type Dropped READ 0 RANGE_SLICE 0 _TRACE 0 HINT 0 MUTATION 1 COUNTER_MUTATION 0 BATCH_STORE 0 BATCH_REMOVE 0 REQUEST_RESPONSE 0 PAGED_RANGE 0 READ_REPAIR 0 *nodetool compactionstats -H* pending tasks: 776 id compaction type keyspace table completed total unit progress 24d039f2-b1e6-11e7-ac57-3d25e38b2f5c Compaction keyspace_1 table_1 4.85 GB 7.67 GB bytes 63.25% Active compaction remaining time : n/a On Sun, Oct 15, 2017 at 9:27 PM, Jeff Jirsa <jji...@gmail.com> wrote: > Can you post (anonymize as needed) nodetool status, nodetool netstats, > nodetool tpstats, and nodetool compctionstats ? > > -- > Jeff Jirsa > > > On Oct 15, 2017, at 1:14 PM, Stefano Ortolani <ostef...@gmail.com> wrote: > > Hi Jeff, > > that would be 3.0.15, single disk, vnodes enabled (num_tokens 256). > > Stefano > > On Sun, Oct 15, 2017 at 9:11 PM, Jeff Jirsa <jji...@gmail.com> wrote: > >> What version? >> >> Single disk or JBOD? >> >> Vnodes? >> >> -- >> Jeff Jirsa >> >> >> On Oct 15, 2017, at 12:49 PM, Stefano Ortolani <ostef...@gmail.com> >> wrote: >> >> Hi all, >> >> I have been trying "-Dcassandra.disable_stcs_in_l0=true", but no luck so >> far. >> Based on the source code it seems that this option doesn't affect >> compactions while bootstrapping. >> >> I am getting quite confused as it seems I am not able to bootstrap a node >> if I don't have at least 6/7 times the disk space used by other nodes. >> This is weird. The host I am bootstrapping is using a SSD. Also >> compaction throughput is unthrottled (set to 0) and the compacting threads >> are set to 8. >> Nevertheless, primary ranges from other nodes are being streamed, but >> data is never compacted away. >> >> Does anybody know anything else I could try? >> >> Cheers, >> Stefano >> >> On Fri, Oct 13, 2017 at 3:58 PM, Stefano Ortolani <ostef...@gmail.com> >> wrote: >> >>> Other little update: at the same time I see the number of pending tasks >>> stuck (in this case at 1847); restarting the node doesn't help, so I can't >>> really force the node to "digest" all those compactions. In the meanwhile >>> the disk occupied is already twice the average load I have on other nodes. >>> >>> Feeling more and more puzzled here :S >>> >>> On Fri, Oct 13, 2017 at 1:28 PM, Stefano Ortolani <ostef...@gmail.com> >>> wrote: >>> >>>> I have been trying to add another node to the cluster (after upgrading >>>> to 3.0.15) and I just noticed through "nodetool netstats" that all nodes >>>> have been streaming to the joining node approx 1/3 of their SSTables, >>>> basically their whole primary range (using RF=3)? >>>> >>>> Is this expected/normal? >>>> I was under the impression only the necessary SSTables were going to be >>>> streamed... >>>> >>>> Thanks for the help, >>>> Stefano >>>> >>>> >>>> On Wed, Aug 23, 2017 at 1:37 PM, kurt greaves <k...@instaclustr.com> >>>> wrote: >>>> >>>>> But if it also streams, it means I'd still be under-pressure if I am >>>>>> not mistaken. I am under the assumption that the compactions are the >>>>>> by-product of streaming too many SStables at the same time, and not >>>>>> because >>>>>> of my current write load. >>>>>> >>>>> Ah yeah I wasn't thinking about the capacity problem, more of the >>>>> performance impact from the node being backed up with compactions. If you >>>>> haven't already, you should try disable stcs in l0 on the joining node. >>>>> You >>>>> will likely still need to do a lot of compactions, but generally they >>>>> should be smaller. The option is -Dcassandra.disable_stcs_in_l0=true >>>>> >>>>>> I just noticed you were mentioning L1 tables too. Why would that >>>>>> affect the disk footprint? >>>>> >>>>> If you've been doing a lot of STCS in L0, you generally end up with >>>>> some large SSTables. These will eventually have to be compacted with L1. >>>>> Could also be suffering the problem of streamed SSTables causing large >>>>> cross-level compactions in the higher levels as well. >>>>> >>>>> >>>> >>>> >>> >> >