Yes, there we see some swap thrashing.
On Wed, Feb 20, 2013 at 10:04 PM, Parth Jagirdar <parth.jagir...@citrix.com> wrote: > Marcus, > > > I attempted login into UI while running the log. > > [root@localhost management]# vmstat 1 > procs -----------memory---------- ---swap-- -----io---- --system-- > -----cpu----- > r b swpd free buff cache si so bi bo in cs us sy id > wa st > 0 0 191132 70904 10340 16192 36 28 40 91 9 1 0 0 89 > 10 0 > 0 0 191132 70904 10340 16204 0 0 0 0 46 75 0 0 > 100 0 0 > 0 0 191132 70904 10356 16204 0 0 0 36 72 221 1 0 91 > 8 0 > 0 0 191132 70904 10372 16188 0 0 0 44 73 130 0 1 88 > 11 0 > 0 0 191132 70780 10420 16208 0 0 4 276 83 191 1 0 75 > 24 0 > 0 0 191132 70780 10452 16192 0 0 0 120 106 309 1 0 77 > 22 0 > 0 0 191132 70780 10468 16200 0 0 0 40 91 183 1 1 90 > 8 0 > 0 0 191132 70780 10468 16216 0 0 0 0 47 128 0 0 > 100 0 0 > 0 0 191132 70780 10484 16216 0 0 0 36 70 136 0 0 94 > 6 0 > 0 0 191132 70656 10500 16200 0 0 0 40 66 116 1 0 91 > 8 0 > 0 0 191132 70656 10500 16216 0 0 0 0 47 94 0 0 > 100 0 0 > 0 1 189504 66216 10400 17940 2192 100 3928 172 404 579 9 2 5 > 84 0 > 1 1 188772 60220 10552 21992 1000 0 5220 68 412 741 7 2 21 > 69 0 > 1 2 187352 49316 7832 30344 1660 32 10052 32 833 1015 28 3 0 > 69 0 > 0 4 188816 52420 1392 25716 1488 2872 3168 3240 663 870 19 3 0 > 78 0 > 1 1 187388 51808 1372 25040 2476 1104 3084 1260 675 813 15 3 0 > 82 0 > 0 1 187360 54040 1500 24980 32 0 1048 0 447 379 6 1 0 > 93 0 > 0 1 187360 53916 1516 25004 0 0 924 52 309 283 1 0 0 > 99 0 > 0 1 195476 64076 1272 20624 0 8116 32 8156 312 308 1 2 0 > 97 0 > 0 0 203084 71920 1264 19412 0 7608 0 7608 256 173 0 2 89 > 9 0 > 0 0 203076 71324 1376 20132 64 0 868 40 192 232 2 0 65 > 33 0 > 0 0 203076 71328 1392 20108 0 0 0 68 75 144 1 0 85 > 14 0 > 0 0 203076 71084 1392 20392 0 0 268 0 66 132 0 1 96 > 3 0 > 0 0 203076 71084 1408 20392 0 0 0 36 60 122 0 0 94 > 6 0 > 0 0 203076 71076 1424 20376 0 0 0 36 77 148 1 0 92 > 7 0 > 0 1 203072 70696 1472 20460 96 0 168 280 196 1080 7 1 66 > 26 0 > 0 0 202900 68704 1512 21236 656 0 1432 104 338 760 10 1 10 > 79 0 > 0 0 201804 65728 1540 21984 1184 0 1904 64 547 1117 26 2 40 > 33 0 > 0 2 161904 122500 1540 22640 68 0 776 0 407 477 23 2 57 > 18 0 > 1 0 161384 122132 1556 22748 36 0 92 60 970 200 92 0 0 > 8 0 > 0 1 160840 119512 1836 23228 676 0 1432 76 772 866 58 2 0 > 40 0 > 0 0 160776 119636 1836 23516 196 0 500 0 104 199 1 0 63 > 36 0 > 0 0 160776 119636 1852 23520 0 0 0 44 83 251 2 0 92 > 6 0 > 0 0 160776 119644 1868 23504 0 0 0 40 64 117 0 1 90 > 9 0 > 0 0 160776 119644 1868 23520 0 0 0 0 46 91 0 0 > 100 0 0 > 0 1 160764 119456 1888 23556 28 0 32 164 71 121 0 0 87 > 13 0 > 0 0 160764 119208 1952 23572 0 0 4 288 392 1083 4 1 66 > 29 0 > 0 0 160764 119192 1952 23596 0 0 0 0 42 69 0 0 > 100 0 0 > 0 0 160764 119192 1968 23596 0 0 0 40 60 127 1 0 92 > 7 0 > 0 0 160764 119192 1984 23584 0 0 4 36 71 135 0 1 91 > 8 0 > 0 0 160764 119192 1984 23600 0 0 0 0 46 89 0 0 > 100 0 0 > 0 0 160764 119192 2000 23600 0 0 0 36 59 121 1 0 92 > 7 0 > 0 0 160764 119192 2016 23584 0 0 0 36 82 196 0 0 93 > 7 0 > 0 0 160764 119192 2016 23600 0 0 0 0 38 69 0 0 > 100 0 0 > 0 0 160764 119192 2032 23600 0 0 0 36 63 130 1 0 91 > 8 0 > 0 0 160764 119192 2048 23584 0 0 0 36 67 132 0 0 94 > 6 0 > 0 0 160764 119192 2096 23584 0 0 0 272 89 193 0 0 76 > 24 0 > > > ...Parth > > > On 2/20/13 8:59 PM, "Marcus Sorensen" <shadow...@gmail.com> wrote: > >>Well, it doesn't seem to be actively swapping at this point, but I >>think it's got active memory swapped out and being used as >>occasionally wait% goes up significantly. At any rate this system is >>severely memory limited. >> >>On Wed, Feb 20, 2013 at 9:52 PM, Parth Jagirdar >><parth.jagir...@citrix.com> wrote: >>> Marcus, >>> >>> vmstat 1 output >>> >>> >>> [root@localhost management]# vmstat 1 >>> procs -----------memory---------- ---swap-- -----io---- --system-- >>> -----cpu----- >>> r b swpd free buff cache si so bi bo in cs us sy >>>id >>> wa st >>> 0 1 190820 72380 10904 15852 36 28 40 92 9 1 0 0 >>>89 >>> 10 0 >>> 0 0 190820 72256 10932 15828 0 0 0 56 63 130 0 0 >>>88 >>> 12 0 >>> 1 0 190820 72256 10932 15844 0 0 0 0 53 153 1 0 >>>99 >>> 0 0 >>> 0 0 190820 72256 10948 15844 0 0 0 44 89 253 2 0 >>>88 >>> 10 0 >>> 0 0 190820 72256 10964 15828 0 0 0 72 64 135 0 0 >>>88 >>> 12 0 >>> 0 0 190820 72256 10964 15844 0 0 0 0 43 76 0 0 >>> 100 0 0 >>> 0 0 190820 72256 10980 15844 0 0 0 36 86 244 1 1 >>>91 >>> 7 0 >>> 0 0 190820 72256 10996 15828 0 0 0 44 57 112 0 1 >>>88 >>> 11 0 >>> 0 0 190820 72256 10996 15844 0 0 0 0 45 88 0 0 >>> 100 0 0 >>> 0 0 190820 72256 11012 15844 0 0 0 36 100 264 1 1 >>>91 >>> 7 0 >>> 0 0 190820 72132 11044 15824 0 0 4 96 106 211 1 0 >>>80 >>> 19 0 >>> 0 0 190820 72132 11092 15856 0 0 0 368 81 223 0 1 >>>74 >>> 25 0 >>> 0 0 190820 72132 11108 15856 0 0 0 36 78 145 0 1 >>>93 >>> 6 0 >>> 0 0 190820 72132 11124 15840 0 0 0 40 55 106 1 0 >>>90 >>> 9 0 >>> 0 0 190820 72132 11124 15856 0 0 0 0 47 96 0 0 >>> 100 0 0 >>> 0 0 190820 72132 11140 15856 0 0 0 36 61 113 0 0 >>>85 >>> 15 0 >>> 0 0 190820 72008 11156 15840 0 0 0 36 61 158 0 0 >>>93 >>> 7 0 >>> 0 0 190820 72008 11156 15856 0 0 0 0 41 82 0 0 >>> 100 0 0 >>> 0 0 190820 72008 11172 15856 0 0 0 36 74 149 1 0 >>>94 >>> 5 0 >>> 0 0 190820 72008 11188 15840 0 0 0 36 60 117 0 0 >>>93 >>> 7 0 >>> 0 0 190820 72008 11188 15856 0 0 0 0 43 91 0 0 >>> 100 0 0 >>> 1 0 190820 72008 11252 15860 0 0 4 312 108 243 1 1 >>>68 >>> 30 0 >>> 0 0 190820 72008 11268 15844 0 0 0 36 60 128 0 0 >>>92 >>> 8 0 >>> 0 0 190820 72008 11268 15860 0 0 0 0 36 67 0 0 >>> 100 0 0 >>> 0 0 190820 71884 11284 15860 0 0 0 104 84 139 0 1 >>>83 >>> 16 0 >>> 0 0 190820 71884 11300 15844 0 0 0 60 55 111 0 0 >>>69 >>> 31 0 >>> 0 0 190820 71884 11300 15860 0 0 0 0 53 121 1 0 >>>99 >>> 0 0 >>> 0 0 190820 71884 11316 15860 0 0 0 40 67 130 0 0 >>>87 >>> 13 0 >>> 0 0 190820 71884 11332 15844 0 0 0 40 58 130 0 0 >>>90 >>> 10 0 >>> 0 0 190820 71884 11332 15864 0 0 0 0 59 824 1 1 >>>98 >>> 0 0 >>> 1 0 190820 71884 11348 15864 0 0 0 40 113 185 1 0 >>>67 >>> 32 0 >>> 0 0 190820 71744 11412 15852 0 0 4 540 100 238 0 0 >>>67 >>> 33 0 >>> 0 0 190820 71744 11412 15868 0 0 0 0 55 159 1 0 >>>99 >>> 0 0 >>> 0 0 190820 71744 11428 15868 0 0 0 40 89 246 2 1 >>>90 >>> 7 0 >>> 0 0 190820 71620 11444 15852 0 0 0 72 65 135 0 0 >>>93 >>> 7 0 >>> 0 0 190820 71620 11444 15868 0 0 0 0 40 74 0 0 >>> 100 0 0 >>> 0 0 190820 71620 11460 15868 0 0 0 52 75 216 1 0 >>>92 >>> 7 0 >>> 0 0 190820 71620 11476 15852 0 0 0 44 53 109 0 0 >>>89 >>> 11 0 >>> 0 0 190820 71620 11476 15868 0 0 0 0 43 87 0 0 >>> 100 0 0 >>> 0 0 190820 71620 11496 15868 0 0 4 36 83 143 0 1 >>>90 >>> 9 0 >>> 0 0 190820 71620 11512 15852 0 0 0 40 78 869 1 0 >>>91 >>> 8 0 >>> 0 1 190820 71628 11524 15856 0 0 0 188 94 145 0 0 >>>87 >>> 13 0 >>> 0 0 190820 71496 11576 15872 0 0 4 132 96 214 1 0 >>>80 >>> 19 0 >>> 0 0 190820 71496 11592 15856 0 0 0 36 94 128 1 0 >>>92 >>> 7 0 >>> 0 0 190820 71496 11592 15872 0 0 0 0 115 164 0 0 >>> 100 0 0 >>> 0 0 190820 71496 11608 15876 0 0 0 36 130 200 0 0 >>>87 >>> 13 0 >>> 0 0 190820 71496 11624 15860 0 0 0 36 141 218 1 1 >>>91 >>> 7 0 >>> 0 0 190820 71504 11624 15876 0 0 0 0 105 119 0 0 >>> 100 0 0 >>> 0 0 190820 71504 11640 15876 0 0 0 36 140 218 1 0 >>>90 >>> 9 0 >>> procs -----------memory---------- ---swap-- -----io---- --system-- >>> -----cpu----- >>> r b swpd free buff cache si so bi bo in cs us sy >>>id >>> wa st >>> 0 0 190820 71504 11656 15860 0 0 0 36 131 169 1 0 >>>92 >>> 7 0 >>> 0 0 190820 71504 11656 15876 0 0 0 0 115 146 0 0 >>> 100 0 0 >>> 0 0 190820 71380 11672 15876 0 0 0 36 128 173 0 1 >>>91 >>> 8 0 >>> 0 0 190820 71380 11736 15860 0 0 0 308 146 279 1 0 >>>69 >>> 30 0 >>> 0 0 190820 71380 11736 15876 0 0 0 0 59 82 0 0 >>> 100 0 0 >>> 0 0 190820 71380 11760 15876 0 0 4 64 90 174 1 1 >>>86 >>> 12 0 >>> >>> ...Parth >>> >>> >>> >>> >>> On 2/20/13 8:46 PM, "Parth Jagirdar" <parth.jagir...@citrix.com> wrote: >>> >>>>JAVA_OPTS="-Djava.awt.headless=true >>>>-Dcom.sun.management.jmxremote.port=45219 >>>>-Dcom.sun.management.jmxremote.authenticate=false >>>>-Dcom.sun.management.jmxremote.ssl=false -Xmx512m -Xms512m >>>>-XX:+HeapDumpOnOutOfMemoryError >>>>-XX:HeapDumpPath=/var/log/cloudstack/management/ -XX:PermSize=256M" >>>> >>>>Which did not help. >>>> >>>>-------------- >>>> >>>>[root@localhost management]# cat /proc/meminfo >>>>MemTotal: 1016656 kB >>>>MemFree: 68400 kB >>>>Buffers: 9108 kB >>>>Cached: 20984 kB >>>>SwapCached: 17492 kB >>>>Active: 424152 kB >>>>Inactive: 433152 kB >>>>Active(anon): 409812 kB >>>>Inactive(anon): 417412 kB >>>>Active(file): 14340 kB >>>>Inactive(file): 15740 kB >>>>Unevictable: 0 kB >>>>Mlocked: 0 kB >>>>SwapTotal: 2031608 kB >>>>SwapFree: 1840900 kB >>>>Dirty: 80 kB >>>>Writeback: 0 kB >>>>AnonPages: 815460 kB >>>>Mapped: 11408 kB >>>>Shmem: 4 kB >>>>Slab: 60120 kB >>>>SReclaimable: 10368 kB >>>>SUnreclaim: 49752 kB >>>>KernelStack: 5216 kB >>>>PageTables: 6800 kB >>>>NFS_Unstable: 0 kB >>>>Bounce: 0 kB >>>>WritebackTmp: 0 kB >>>>CommitLimit: 2539936 kB >>>>Committed_AS: 1596896 kB >>>>VmallocTotal: 34359738367 kB >>>>VmallocUsed: 7724 kB >>>>VmallocChunk: 34359718200 kB >>>>HardwareCorrupted: 0 kB >>>>AnonHugePages: 503808 kB >>>>HugePages_Total: 0 >>>>HugePages_Free: 0 >>>>HugePages_Rsvd: 0 >>>>HugePages_Surp: 0 >>>>Hugepagesize: 2048 kB >>>>DirectMap4k: 6144 kB >>>>DirectMap2M: 1038336 kB >>>>[root@localhost management]# >>>>----------------------------- >>>> >>>> PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND >>>> >>>> >>>> 9809 cloud 20 0 2215m 785m 4672 S 0.7 79.1 1:59.40 java >>>> >>>> >>>> 1497 mysql 20 0 700m 15m 3188 S 0.3 1.5 23:04.58 mysqld >>>> >>>> >>>> 1 root 20 0 19348 300 296 S 0.0 0.0 0:00.73 init >>>> >>>> >>>> >>>> >>>> >>>>On 2/20/13 8:26 PM, "Sailaja Mada" <sailaja.m...@citrix.com> wrote: >>>> >>>>>Hi, >>>>> >>>>>Cloudstack Java process statistics are given below when it stops >>>>>responding are given below : >>>>> >>>>>top - 09:52:03 up 4 days, 21:43, 2 users, load average: 0.06, 0.05, >>>>>0.02 >>>>>Tasks: 1 total, 0 running, 1 sleeping, 0 stopped, 0 zombie >>>>>Cpu(s): 1.7%us, 0.7%sy, 0.0%ni, 97.3%id, 0.3%wa, 0.0%hi, 0.0%si, >>>>>0.0%st >>>>>Mem: 1014860k total, 947632k used, 67228k free, 5868k >>>>>buffers >>>>>Swap: 2031608k total, 832320k used, 1199288k free, 26764k cached >>>>> >>>>> PID USER PR NI VIRT RES SHR S %CPU %MEM TIME+ COMMAND >>>>>12559 cloud 20 0 3159m 744m 4440 S 2.3 75.1 6:38.39 java >>>>> >>>>>Thanks, >>>>>Sailaja.M >>>>> >>>>>-----Original Message----- >>>>>From: Marcus Sorensen [mailto:shadow...@gmail.com] >>>>>Sent: Thursday, February 21, 2013 9:35 AM >>>>>To: cloudstack-dev@incubator.apache.org >>>>>Subject: Re: [DISCUSS] Management Server Memory Requirements >>>>> >>>>>Yes, these are great data points, but so far nobody has responded on >>>>>that >>>>>ticket with the information required to know if the slowness is related >>>>>to memory settings or swapping. That was just a hunch on my part from >>>>>being a system admin. >>>>> >>>>>How much memory do these systems have that experience issues? What does >>>>>/proc/meminfo say during the issues? Does adjusting the tomcat6.conf >>>>>memory settings make a difference (see ticket comments)? How much >>>>>memory >>>>>do the java processes list as resident in top? >>>>>On Feb 20, 2013 8:53 PM, "Parth Jagirdar" <parth.jagir...@citrix.com> >>>>>wrote: >>>>> >>>>>> +1 Performance degradation is dramatic and I too have observed this >>>>>>issue. >>>>>> >>>>>> I have logged my comments into 1339. >>>>>> >>>>>> >>>>>> ŠParth >>>>>> >>>>>> On 2/20/13 7:34 PM, "Srikanteswararao Talluri" >>>>>> <srikanteswararao.tall...@citrix.com> wrote: >>>>>> >>>>>> >To add to what Marcus mentioned, >>>>>> >Regarding bug CLOUDSTACK-1339 : I have observed this issue within >>>>>> >5-10 min of starting management server and there has been a lot of >>>>>> >API requests through automated tests. It is observed that Management >>>>>> >server not only slows down but also goes down after a while. >>>>>> > >>>>>> >~Talluri >>>>>> > >>>>>> >-----Original Message----- >>>>>> >From: Marcus Sorensen [mailto:shadow...@gmail.com] >>>>>> >Sent: Thursday, February 21, 2013 7:22 >>>>>> >To: cloudstack-dev@incubator.apache.org >>>>>> >Subject: [DISCUSS] Management Server Memory Requirements >>>>>> > >>>>>> >When Javelin was merged, there was an email sent out stating that >>>>>> >devs should set their MAVEN_OPTS to use 2g of heap, and 512M of >>>>>> >permanent memory. Subsequently, there have also been several >>>>>>e-mails >>>>>> >and issues where devs have echoed this recommendation, and >>>>>>presumably >>>>>> >it fixed issues. I've seen the MS run out of memory myself and >>>>>> >applied those recommendations. >>>>>> > >>>>>> >Is this what we want to provide in the tomcat config for a package >>>>>> >based install as well? It's effectively saying that the minimum >>>>>> >requirements for the management server are something like 3 or 4 GB >>>>>> >(to be safe for other running tasks) of RAM, right? >>>>>> > >>>>>> >There is currently a bug filed that may or may not have to do with >>>>>> >this, CLOUDSTACK-1339. Users report mgmt server slowness, going >>>>>> >unresponsive for minutes at a time, but the logs seem to show >>>>>> >business as usual. User reports that java is taking 75% of RAM, >>>>>> >depending on what else is going on they may be swapping. Settings in >>>>>> >the code for an install are currently at 2g/512M, I've been running >>>>>> >this on a 4GB server for awhile now, java is at 900M, but I haven't >>>>>> >been pounding it with requests or anything. >>>>>> > >>>>>> >This bug might not have anything to do with the memory settings, but >>>>>> >I figured it would be good to nail down what our minimum >>>>>>requirements >>>>>> >are for 4.1 >>>>>> >>>>>> >>>> >>> >