Hi, If after it decreases in size it stays there it might be similar to:
7111576 arc shrinks in the absence of memory pressure Also, see document: ZFS ARC can shrink down without memory pressure result in slow performance [ID 1404581.1] Specifically, check if arc_no_grow is set to 1 after the cache size is decreased, and if it stays that way. The fix is in one of the SRUs and I think it should be in 11.1 I don't know if it was fixed in Illumos or even if Illumos was affected by this at all. -- Robert Milkowski http://milek.blogspot.com > -----Original Message----- > From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss- > boun...@opensolaris.org] On Behalf Of Chris Nagele > Sent: 20 October 2012 18:47 > To: zfs-discuss@opensolaris.org > Subject: [zfs-discuss] ARC de-allocation with large ram > > Hi. We're running OmniOS as a ZFS storage server. For some reason, our > arc cache will grow to a certain point, then suddenly drops. I used > arcstat to catch it in action, but I was not able to capture what else > was going on in the system at the time. I'll do that next. > > read hits miss hit% l2read l2hits l2miss l2hit% arcsz l2size > 166 166 0 100 0 0 0 0 85G 225G > 5.9K 5.9K 0 100 0 0 0 0 85G 225G > 755 715 40 94 40 0 40 0 84G 225G > 17K 17K 0 100 0 0 0 0 67G 225G > 409 395 14 96 14 0 14 0 49G 225G > 388 364 24 93 24 0 24 0 41G 225G > 37K 37K 20 99 20 6 14 30 40G 225G > > For reference, it's a 12TB pool with 512GB SSD L2 ARC and 198GB RAM. > We have nothing else running on the system except NFS. We are also not > using dedupe. Here is the output of memstat at one point: > > # echo ::memstat | mdb -k > Page Summary Pages MB %Tot > ------------ ---------------- ---------------- ---- > Kernel 19061902 74460 38% > ZFS File Data 28237282 110301 56% > Anon 43112 168 0% > Exec and libs 1522 5 0% > Page cache 13509 52 0% > Free (cachelist) 6366 24 0% > Free (freelist) 2958527 11556 6% > > Total 50322220 196571 > Physical 50322219 196571 > > According to "prstat -s rss" nothing else is consuming the memory. > > 592 root 33M 26M sleep 59 0 0:00:33 0.0% fmd/27 > 12 root 13M 11M sleep 59 0 0:00:08 0.0% > svc.configd/21 > 641 root 12M 11M sleep 59 0 0:04:48 0.0% snmpd/1 > 10 root 14M 10M sleep 59 0 0:00:03 0.0% > svc.startd/16 > 342 root 12M 9084K sleep 59 0 0:00:15 0.0% hald/5 > 321 root 14M 8652K sleep 59 0 0:03:00 0.0% nscd/52 > > So far I can't figure out what could be causing this. The only other > thing I can think of is that we have a bunch of zfs send/receive > operations going on as backups across 10 datasets in the pool. I am > not sure how snapshots and send/receive affect the arc. Does anyone > else have any ideas? > > Thanks, > Chris > _______________________________________________ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss