So, vfs.zfs.arc_max="2048M" in /boot/loader.conf was indeed apparently all that was necessary to bring the situation under control. I remember it being a lot more nightmarish, so it's nice to see that it's improved. Thanks for everyone's advice. Per an earlier request, here is the output of "zfs-stats -a" right now (not when the system is running out of memory, but perhaps still interesting):

# zfs-stats -a
------------------------------------------------------------------------
ZFS Subsystem Report                            Mon Apr  4 22:43:43 2011
------------------------------------------------------------------------
System Information:

        Kernel Version:                         802502 (osreldate)
        Hardware Platform:                      amd64
        Processor Architecture:                 amd64

FreeBSD 8.2-STABLE #3: Sat Apr 2 11:48:43 EDT 2011 spawk
10:43PM  up  1:32, 3 users, load averages: 0.13, 0.09, 0.07
------------------------------------------------------------------------
System Memory Statistics:
        Physical Memory:                        8181.32M
        Kernel Memory:                          2134.72M
        DATA:                           99.65%  2127.20M
        TEXT:                           0.35%   7.52M
------------------------------------------------------------------------
ZFS pool information:
        Storage pool Version (spa):             15
        Filesystem Version (zpl):               4
------------------------------------------------------------------------
ARC Misc:
        Deleted:                                583540
        Recycle Misses:                         355
        Mutex Misses:                           11
        Evict Skips:                            11

ARC Size:
        Current Size (arcsize):         100.00% 2048.07M
        Target Size (Adaptive, c):      100.00% 2048.00M
        Min Size (Hard Limit, c_min):   12.50%  256.00M
        Max Size (High Water, c_max):   ~8:1    2048.00M

ARC Size Breakdown:
        Recently Used Cache Size (p):   93.32%  1911.27M
        Frequent;y Used Cache Size (arcsize-p): 6.68%   136.80M

ARC Hash Breakdown:
        Elements Max:                           38168
        Elements Current:               99.18%  37856
        Collisions:                             127822
        Chain Max:                              5
        Chains:                                 4567

ARC Eviction Statistics:
        Evicts Total:                           80383607808
        Evicts Eligible for L2:         9.95%   8001851392
        Evicts Ineligible for L2:       90.05%  72381756416
        Evicts Cached to L2:                    0

ARC Efficiency:
        Cache Access Total:                     1439376
        Cache Hit Ratio:                55.64%  800797
        Cache Miss Ratio:               44.36%  638579
        Actual Hit Ratio:               51.12%  735809

        Data Demand Efficiency:         97.21%
        Data Prefetch Efficiency:       8.78%

        CACHE HITS BY CACHE LIST:
          Anonymously Used:             5.96%   47738
          Most Recently Used (mru):     28.97%  232003
          Most Frequently Used (mfu):   62.91%  503806
          MRU Ghost (mru_ghost):        0.74%   5933
          MFU Ghost (mfu_ghost):        1.41%   11317

        CACHE HITS BY DATA TYPE:
          Demand Data:                  31.17%  249578
          Prefetch Data:                7.47%   59804
          Demand Metadata:              60.33%  483145
          Prefetch Metadata:            1.03%   8270

        CACHE MISSES BY DATA TYPE:
          Demand Data:                  1.12%   7162
          Prefetch Data:                97.31%  621373
          Demand Metadata:              0.91%   5805
          Prefetch Metadata:            0.66%   4239
------------------------------------------------------------------------
VDEV Cache Summary:
        Access Total:                           32382
        Hits Ratio:                     21.62%  7001
        Miss Ratio:                     78.38%  25381
        Delegations:                            11361
------------------------------------------------------------------------
File-Level Prefetch Stats (DMU):

DMU Efficiency:
        Access Total:                           640507
        Hit Ratio:                      92.40%  591801
        Miss Ratio:                     7.60%   48706

        Colinear Access Total:                  48706
        Colinear Hit Ratio:             0.20%   99
        Colinear Miss Ratio:            99.80%  48607

        Stride Access Total:                    533456
        Stride Hit Ratio:               99.97%  533296
        Stride Miss Ratio:              0.03%   160

DMU misc:
        Reclaim successes:                      8857
        Reclaim failures:                       39750
        Stream resets:                          126
        Stream noresets:                        58504
        Bogus streams:                          0
------------------------------------------------------------------------
ZFS Tunable (sysctl):
        kern.maxusers=384
        vfs.zfs.l2c_only_size=0
        vfs.zfs.mfu_ghost_data_lsize=1956972544
        vfs.zfs.mfu_ghost_metadata_lsize=12218880
        vfs.zfs.mfu_ghost_size=1969191424
        vfs.zfs.mfu_data_lsize=103428096
        vfs.zfs.mfu_metadata_lsize=17053184
        vfs.zfs.mfu_size=124921344
        vfs.zfs.mru_ghost_data_lsize=133824512
        vfs.zfs.mru_ghost_metadata_lsize=40669184
        vfs.zfs.mru_ghost_size=174493696
        vfs.zfs.mru_data_lsize=1984430080
        vfs.zfs.mru_metadata_lsize=14490112
        vfs.zfs.mru_size=2005246464
        vfs.zfs.anon_data_lsize=0
        vfs.zfs.anon_metadata_lsize=0
        vfs.zfs.anon_size=0
        vfs.zfs.l2arc_norw=1
        vfs.zfs.l2arc_feed_again=1
        vfs.zfs.l2arc_noprefetch=0
        vfs.zfs.l2arc_feed_min_ms=200
        vfs.zfs.l2arc_feed_secs=1
        vfs.zfs.l2arc_headroom=2
        vfs.zfs.l2arc_write_boost=8388608
        vfs.zfs.l2arc_write_max=8388608
        vfs.zfs.arc_meta_limit=536870912
        vfs.zfs.arc_meta_used=59699288
        vfs.zfs.mdcomp_disable=0
        vfs.zfs.arc_min=268435456
        vfs.zfs.arc_max=2147483648
        vfs.zfs.zfetch.array_rd_sz=1048576
        vfs.zfs.zfetch.block_cap=256
        vfs.zfs.zfetch.min_sec_reap=2
        vfs.zfs.zfetch.max_streams=8
        vfs.zfs.prefetch_disable=0
        vfs.zfs.check_hostid=1
        vfs.zfs.recover=0
        vfs.zfs.txg.write_limit_override=0
        vfs.zfs.txg.synctime=5
        vfs.zfs.txg.timeout=30
        vfs.zfs.scrub_limit=10
        vfs.zfs.vdev.cache.bshift=16
        vfs.zfs.vdev.cache.size=10485760
        vfs.zfs.vdev.cache.max=16384
        vfs.zfs.vdev.aggregation_limit=131072
        vfs.zfs.vdev.ramp_rate=2
        vfs.zfs.vdev.time_shift=6
        vfs.zfs.vdev.min_pending=4
        vfs.zfs.vdev.max_pending=10
        vfs.zfs.cache_flush_disable=0
        vfs.zfs.zil_disable=0
        vfs.zfs.zio.use_uma=0
        vfs.zfs.version.zpl=4
        vfs.zfs.version.spa=15
        vfs.zfs.version.dmu_backup_stream=1
        vfs.zfs.version.dmu_backup_header=2
        vfs.zfs.version.acl=1
        vfs.zfs.debug=0
        vfs.zfs.super_owner=0
        vm.kmem_size=8294764544
        vm.kmem_size_scale=1
        vm.kmem_size_min=0
        vm.kmem_size_max=329853485875
------------------------------------------------------------------------

-Boris
_______________________________________________
freebsd-stable@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-stable
To unsubscribe, send any mail to "freebsd-stable-unsubscr...@freebsd.org"

Reply via email to