Sometimes, the Pending can change from 128 to 129, 125 etc.
On Wed, Jul 13, 2016 at 10:32 AM, Yuan Fang <y...@kryptoncloud.com> wrote: > $nodetool tpstats > > ... > Pool Name Active Pending Completed > Blocked All time blocked > Native-Transport-Requests 128 128 1420623949 1 > 142821509 > ... > > > > What is this? Is it normal? > > On Tue, Jul 12, 2016 at 3:03 PM, Yuan Fang <y...@kryptoncloud.com> wrote: > >> Hi Jonathan, >> >> Here is the result: >> >> ubuntu@ip-172-31-44-250:~$ iostat -dmx 2 10 >> Linux 3.13.0-74-generic (ip-172-31-44-250) 07/12/2016 _x86_64_ (4 CPU) >> >> Device: rrqm/s wrqm/s r/s w/s rMB/s wMB/s >> avgrq-sz avgqu-sz await r_await w_await svctm %util >> xvda 0.01 2.13 0.74 1.55 0.01 0.02 >> 27.77 0.00 0.74 0.89 0.66 0.43 0.10 >> xvdf 0.01 0.58 237.41 52.50 12.90 6.21 >> 135.02 2.32 8.01 3.65 27.72 0.57 16.63 >> >> Device: rrqm/s wrqm/s r/s w/s rMB/s wMB/s >> avgrq-sz avgqu-sz await r_await w_await svctm %util >> xvda 0.00 7.50 0.00 2.50 0.00 0.04 >> 32.00 0.00 1.60 0.00 1.60 1.60 0.40 >> xvdf 0.00 0.00 353.50 0.00 24.12 0.00 >> 139.75 0.49 1.37 1.37 0.00 0.58 20.60 >> >> Device: rrqm/s wrqm/s r/s w/s rMB/s wMB/s >> avgrq-sz avgqu-sz await r_await w_await svctm %util >> xvda 0.00 0.00 0.00 1.00 0.00 0.00 >> 8.00 0.00 0.00 0.00 0.00 0.00 0.00 >> xvdf 0.00 2.00 463.50 35.00 30.69 2.86 >> 137.84 0.88 1.77 1.29 8.17 0.60 30.00 >> >> Device: rrqm/s wrqm/s r/s w/s rMB/s wMB/s >> avgrq-sz avgqu-sz await r_await w_await svctm %util >> xvda 0.00 0.00 0.00 1.00 0.00 0.00 >> 8.00 0.00 0.00 0.00 0.00 0.00 0.00 >> xvdf 0.00 0.00 99.50 36.00 8.54 4.40 >> 195.62 1.55 3.88 1.45 10.61 1.06 14.40 >> >> Device: rrqm/s wrqm/s r/s w/s rMB/s wMB/s >> avgrq-sz avgqu-sz await r_await w_await svctm %util >> xvda 0.00 5.00 0.00 1.50 0.00 0.03 >> 34.67 0.00 1.33 0.00 1.33 1.33 0.20 >> xvdf 0.00 1.50 703.00 195.00 48.83 23.76 >> 165.57 6.49 8.36 1.66 32.51 0.55 49.80 >> >> Device: rrqm/s wrqm/s r/s w/s rMB/s wMB/s >> avgrq-sz avgqu-sz await r_await w_await svctm %util >> xvda 0.00 0.00 0.00 1.00 0.00 0.04 >> 72.00 0.00 0.00 0.00 0.00 0.00 0.00 >> xvdf 0.00 2.50 149.50 69.50 10.12 6.68 >> 157.14 0.74 3.42 1.18 8.23 0.51 11.20 >> >> Device: rrqm/s wrqm/s r/s w/s rMB/s wMB/s >> avgrq-sz avgqu-sz await r_await w_await svctm %util >> xvda 0.00 5.00 0.00 2.50 0.00 0.03 >> 24.00 0.00 0.00 0.00 0.00 0.00 0.00 >> xvdf 0.00 0.00 61.50 22.50 5.36 2.75 >> 197.64 0.33 3.93 1.50 10.58 0.88 7.40 >> >> Device: rrqm/s wrqm/s r/s w/s rMB/s wMB/s >> avgrq-sz avgqu-sz await r_await w_await svctm %util >> xvda 0.00 0.00 0.00 0.50 0.00 0.00 >> 8.00 0.00 0.00 0.00 0.00 0.00 0.00 >> xvdf 0.00 0.00 375.00 0.00 24.84 0.00 >> 135.64 0.45 1.20 1.20 0.00 0.57 21.20 >> >> Device: rrqm/s wrqm/s r/s w/s rMB/s wMB/s >> avgrq-sz avgqu-sz await r_await w_await svctm %util >> xvda 0.00 1.00 0.00 6.00 0.00 0.03 >> 9.33 0.00 0.00 0.00 0.00 0.00 0.00 >> xvdf 0.00 0.00 542.50 23.50 35.08 2.83 >> 137.16 0.80 1.41 1.15 7.23 0.49 28.00 >> >> Device: rrqm/s wrqm/s r/s w/s rMB/s wMB/s >> avgrq-sz avgqu-sz await r_await w_await svctm %util >> xvda 0.00 3.50 0.50 1.50 0.00 0.02 >> 24.00 0.00 0.00 0.00 0.00 0.00 0.00 >> xvdf 0.00 1.50 272.00 153.50 16.18 18.67 >> 167.73 14.32 33.66 1.39 90.84 0.81 34.60 >> >> >> >> On Tue, Jul 12, 2016 at 12:34 PM, Jonathan Haddad <j...@jonhaddad.com> >> wrote: >> >>> When you have high system load it means your CPU is waiting for >>> *something*, and in my experience it's usually slow disk. A disk connected >>> over network has been a culprit for me many times. >>> >>> On Tue, Jul 12, 2016 at 12:33 PM Jonathan Haddad <j...@jonhaddad.com> >>> wrote: >>> >>>> Can do you do: >>>> >>>> iostat -dmx 2 10 >>>> >>>> >>>> >>>> On Tue, Jul 12, 2016 at 11:20 AM Yuan Fang <y...@kryptoncloud.com> >>>> wrote: >>>> >>>>> Hi Jeff, >>>>> >>>>> The read being low is because we do not have much read operations >>>>> right now. >>>>> >>>>> The heap is only 4GB. >>>>> >>>>> MAX_HEAP_SIZE=4GB >>>>> >>>>> On Thu, Jul 7, 2016 at 7:17 PM, Jeff Jirsa <jeff.ji...@crowdstrike.com >>>>> > wrote: >>>>> >>>>>> EBS iops scale with volume size. >>>>>> >>>>>> >>>>>> >>>>>> A 600G EBS volume only guarantees 1800 iops – if you’re exhausting >>>>>> those on writes, you’re going to suffer on reads. >>>>>> >>>>>> >>>>>> >>>>>> You have a 16G server, and probably a good chunk of that allocated to >>>>>> heap. Consequently, you have almost no page cache, so your reads are >>>>>> going >>>>>> to hit the disk. Your reads being very low is not uncommon if you have no >>>>>> page cache – the default settings for Cassandra (64k compression chunks) >>>>>> are really inefficient for small reads served off of disk. If you drop >>>>>> the >>>>>> compression chunk size (4k, for example), you’ll probably see your read >>>>>> throughput increase significantly, which will give you more iops for >>>>>> commitlog, so write throughput likely goes up, too. >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> *From: *Jonathan Haddad <j...@jonhaddad.com> >>>>>> *Reply-To: *"user@cassandra.apache.org" <user@cassandra.apache.org> >>>>>> *Date: *Thursday, July 7, 2016 at 6:54 PM >>>>>> *To: *"user@cassandra.apache.org" <user@cassandra.apache.org> >>>>>> *Subject: *Re: Is my cluster normal? >>>>>> >>>>>> >>>>>> >>>>>> What's your CPU looking like? If it's low, check your IO with iostat >>>>>> or dstat. I know some people have used Ebs and say it's fine but ive been >>>>>> burned too many times. >>>>>> >>>>>> On Thu, Jul 7, 2016 at 6:12 PM Yuan Fang <y...@kryptoncloud.com> >>>>>> wrote: >>>>>> >>>>>> Hi Riccardo, >>>>>> >>>>>> >>>>>> >>>>>> Very low IO-wait. About 0.3%. >>>>>> >>>>>> No stolen CPU. It is a casssandra only instance. I did not see any >>>>>> dropped messages. >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> ubuntu@cassandra1:/mnt/data$ nodetool tpstats >>>>>> >>>>>> Pool Name Active Pending Completed >>>>>> Blocked All time blocked >>>>>> >>>>>> MutationStage 1 1 929509244 >>>>>> 0 0 >>>>>> >>>>>> ViewMutationStage 0 0 0 >>>>>> 0 0 >>>>>> >>>>>> ReadStage 4 0 4021570 >>>>>> 0 0 >>>>>> >>>>>> RequestResponseStage 0 0 731477999 >>>>>> 0 0 >>>>>> >>>>>> ReadRepairStage 0 0 165603 >>>>>> 0 0 >>>>>> >>>>>> CounterMutationStage 0 0 0 >>>>>> 0 0 >>>>>> >>>>>> MiscStage 0 0 0 >>>>>> 0 0 >>>>>> >>>>>> CompactionExecutor 2 55 92022 >>>>>> 0 0 >>>>>> >>>>>> MemtableReclaimMemory 0 0 1736 >>>>>> 0 0 >>>>>> >>>>>> PendingRangeCalculator 0 0 6 >>>>>> 0 0 >>>>>> >>>>>> GossipStage 0 0 345474 >>>>>> 0 0 >>>>>> >>>>>> SecondaryIndexManagement 0 0 0 >>>>>> 0 0 >>>>>> >>>>>> HintsDispatcher 0 0 4 >>>>>> 0 0 >>>>>> >>>>>> MigrationStage 0 0 35 >>>>>> 0 0 >>>>>> >>>>>> MemtablePostFlush 0 0 1973 >>>>>> 0 0 >>>>>> >>>>>> ValidationExecutor 0 0 0 >>>>>> 0 0 >>>>>> >>>>>> Sampler 0 0 0 >>>>>> 0 0 >>>>>> >>>>>> MemtableFlushWriter 0 0 1736 >>>>>> 0 0 >>>>>> >>>>>> InternalResponseStage 0 0 5311 >>>>>> 0 0 >>>>>> >>>>>> AntiEntropyStage 0 0 0 >>>>>> 0 0 >>>>>> >>>>>> CacheCleanupExecutor 0 0 0 >>>>>> 0 0 >>>>>> >>>>>> Native-Transport-Requests 128 128 347508531 >>>>>> 2 15891862 >>>>>> >>>>>> >>>>>> >>>>>> Message type Dropped >>>>>> >>>>>> READ 0 >>>>>> >>>>>> RANGE_SLICE 0 >>>>>> >>>>>> _TRACE 0 >>>>>> >>>>>> HINT 0 >>>>>> >>>>>> MUTATION 0 >>>>>> >>>>>> COUNTER_MUTATION 0 >>>>>> >>>>>> BATCH_STORE 0 >>>>>> >>>>>> BATCH_REMOVE 0 >>>>>> >>>>>> REQUEST_RESPONSE 0 >>>>>> >>>>>> PAGED_RANGE 0 >>>>>> >>>>>> READ_REPAIR 0 >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> On Thu, Jul 7, 2016 at 5:24 PM, Riccardo Ferrari <ferra...@gmail.com> >>>>>> wrote: >>>>>> >>>>>> Hi Yuan, >>>>>> >>>>>> >>>>>> >>>>>> You machine instance is 4 vcpus that is 4 threads (not cores!!!), >>>>>> aside from any Cassandra specific discussion a system load of 10 on a 4 >>>>>> threads machine is way too much in my opinion. If that is the running >>>>>> average system load I would look deeper into system details. Is that IO >>>>>> wait? Is that CPU Stolen? Is that a Cassandra only instance or are there >>>>>> other processes pushing the load? >>>>>> >>>>>> What does your "nodetool tpstats" say? Hoe many dropped messages do >>>>>> you have? >>>>>> >>>>>> >>>>>> >>>>>> Best, >>>>>> >>>>>> >>>>>> >>>>>> On Fri, Jul 8, 2016 at 12:34 AM, Yuan Fang <y...@kryptoncloud.com> >>>>>> wrote: >>>>>> >>>>>> Thanks Ben! For the post, it seems they got a little better but >>>>>> similar result than i did. Good to know it. >>>>>> >>>>>> I am not sure if a little fine tuning of heap memory will help or >>>>>> not. >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> On Thu, Jul 7, 2016 at 2:58 PM, Ben Slater < >>>>>> ben.sla...@instaclustr.com> wrote: >>>>>> >>>>>> Hi Yuan, >>>>>> >>>>>> >>>>>> >>>>>> You might find this blog post a useful comparison: >>>>>> >>>>>> >>>>>> https://www.instaclustr.com/blog/2016/01/07/multi-data-center-apache-spark-and-apache-cassandra-benchmark/ >>>>>> <https://urldefense.proofpoint.com/v2/url?u=https-3A__www.instaclustr.com_blog_2016_01_07_multi-2Ddata-2Dcenter-2Dapache-2Dspark-2Dand-2Dapache-2Dcassandra-2Dbenchmark_&d=CwMFaQ&c=08AGY6txKsvMOP6lYkHQpPMRA1U6kqhAwGa8-0QCg3M&r=yfYEBHVkX6l0zImlOIBID0gmhluYPD5Jje-3CtaT3ow&m=Ltg5YUTZbI4Ixf7UjzKW636Llz6zXXurTveCLptZwio&s=MU4-NWBjvVO95HnxQtkYk4xkApq4X4IiVy8tPCgj4KU&e=> >>>>>> >>>>>> >>>>>> >>>>>> Although the focus is on Spark and Cassandra and multi-DC there are >>>>>> also some single DC benchmarks of m4.xl >>>>>> <https://urldefense.proofpoint.com/v2/url?u=http-3A__m4.xl&d=CwQFaQ&c=08AGY6txKsvMOP6lYkHQpPMRA1U6kqhAwGa8-0QCg3M&r=yfYEBHVkX6l0zImlOIBID0gmhluYPD5Jje-3CtaT3ow&m=Ltg5YUTZbI4Ixf7UjzKW636Llz6zXXurTveCLptZwio&s=m3DfZk3YOaf0W2OvACsqDWXp-vdlkP-cC0WnEouZwkk&e=> >>>>>> clusters plus some discussion of how we went about benchmarking. >>>>>> >>>>>> >>>>>> >>>>>> Cheers >>>>>> >>>>>> Ben >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> On Fri, 8 Jul 2016 at 07:52 Yuan Fang <y...@kryptoncloud.com> wrote: >>>>>> >>>>>> Yes, here is my stress test result: >>>>>> >>>>>> Results: >>>>>> >>>>>> op rate : 12200 [WRITE:12200] >>>>>> >>>>>> partition rate : 12200 [WRITE:12200] >>>>>> >>>>>> row rate : 12200 [WRITE:12200] >>>>>> >>>>>> latency mean : 16.4 [WRITE:16.4] >>>>>> >>>>>> latency median : 7.1 [WRITE:7.1] >>>>>> >>>>>> latency 95th percentile : 38.1 [WRITE:38.1] >>>>>> >>>>>> latency 99th percentile : 204.3 [WRITE:204.3] >>>>>> >>>>>> latency 99.9th percentile : 465.9 [WRITE:465.9] >>>>>> >>>>>> latency max : 1408.4 [WRITE:1408.4] >>>>>> >>>>>> Total partitions : 1000000 [WRITE:1000000] >>>>>> >>>>>> Total errors : 0 [WRITE:0] >>>>>> >>>>>> total gc count : 0 >>>>>> >>>>>> total gc mb : 0 >>>>>> >>>>>> total gc time (s) : 0 >>>>>> >>>>>> avg gc time(ms) : NaN >>>>>> >>>>>> stdev gc time(ms) : 0 >>>>>> >>>>>> Total operation time : 00:01:21 >>>>>> >>>>>> END >>>>>> >>>>>> >>>>>> >>>>>> On Thu, Jul 7, 2016 at 2:49 PM, Ryan Svihla <r...@foundev.pro> wrote: >>>>>> >>>>>> Lots of variables you're leaving out. >>>>>> >>>>>> >>>>>> >>>>>> Depends on write size, if you're using logged batch or not, what >>>>>> consistency level, what RF, if the writes come in bursts, etc, etc. >>>>>> However, that's all sort of moot for determining "normal" really you >>>>>> need a >>>>>> baseline as all those variables end up mattering a huge amount. >>>>>> >>>>>> >>>>>> >>>>>> I would suggest using Cassandra stress as a baseline and go from >>>>>> there depending on what those numbers say (just pick the defaults). >>>>>> >>>>>> Sent from my iPhone >>>>>> >>>>>> >>>>>> On Jul 7, 2016, at 4:39 PM, Yuan Fang <y...@kryptoncloud.com> wrote: >>>>>> >>>>>> yes, it is about 8k writes per node. >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> On Thu, Jul 7, 2016 at 2:18 PM, daemeon reiydelle <daeme...@gmail.com> >>>>>> wrote: >>>>>> >>>>>> Are you saying 7k writes per node? or 30k writes per node? >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> *.......Daemeon C.M. ReiydelleUSA (+1) 415.501.0198 >>>>>> <%28%2B1%29%20415.501.0198>London (+44) (0) 20 8144 9872 >>>>>> <%28%2B44%29%20%280%29%2020%208144%209872>* >>>>>> >>>>>> >>>>>> >>>>>> On Thu, Jul 7, 2016 at 2:05 PM, Yuan Fang <y...@kryptoncloud.com> >>>>>> wrote: >>>>>> >>>>>> writes 30k/second is the main thing. >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> On Thu, Jul 7, 2016 at 1:51 PM, daemeon reiydelle <daeme...@gmail.com> >>>>>> wrote: >>>>>> >>>>>> Assuming you meant 100k, that likely for something with 16mb of >>>>>> storage (probably way small) where the data is more that 64k hence will >>>>>> not >>>>>> fit into the row cache. >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> *.......Daemeon C.M. ReiydelleUSA (+1) 415.501.0198 >>>>>> <%28%2B1%29%20415.501.0198>London (+44) (0) 20 8144 9872 >>>>>> <%28%2B44%29%20%280%29%2020%208144%209872>* >>>>>> >>>>>> >>>>>> >>>>>> On Thu, Jul 7, 2016 at 1:25 PM, Yuan Fang <y...@kryptoncloud.com> >>>>>> wrote: >>>>>> >>>>>> >>>>>> >>>>>> I have a cluster of 4 m4.xlarge nodes(4 cpus and 16 gb memory and >>>>>> 600GB ssd EBS). >>>>>> >>>>>> I can reach a cluster wide write requests of 30k/second and read >>>>>> request about 100/second. The cluster OS load constantly above 10. Are >>>>>> those normal? >>>>>> >>>>>> >>>>>> >>>>>> Thanks! >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> Best, >>>>>> >>>>>> >>>>>> >>>>>> Yuan >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> -- >>>>>> >>>>>> ———————— >>>>>> >>>>>> Ben Slater >>>>>> >>>>>> Chief Product Officer >>>>>> >>>>>> Instaclustr: Cassandra + Spark - Managed | Consulting | Support >>>>>> >>>>>> +61 437 929 798 >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>> >> >