Richard - the l2arc is c1t13d0.  What tools can be use to show the l2arc stats?

  raidz1     2.68T   580G    543    453  4.22M  3.70M
    c1t1d0       -      -    258    102   689K   358K
    c1t2d0       -      -    256    103   684K   354K
    c1t3d0       -      -    258    102   690K   359K
    c1t4d0       -      -    260    103   687K   354K
    c1t5d0       -      -    255    101   686K   358K
    c1t6d0       -      -    263    103   685K   354K
    c1t7d0       -      -    259    101   689K   358K
    c1t8d0       -      -    259    103   687K   354K
    c1t9d0       -      -    260    102   689K   358K
    c1t10d0      -      -    263    103   686K   354K
    c1t11d0      -      -    260    102   687K   359K
    c1t12d0      -      -    263    104   684K   354K
  c1t14d0     396K  29.5G      0     65      7  3.61M
cache            -      -      -      -      -      -
  c1t13d0    29.7G  11.1M    157     84  3.93M  6.45M

We've added 16GB to the box bring the overall total to 32GB.
arc_max is set to 8GB:
set zfs:zfs_arc_max = 8589934592

arc_summary output:
ARC Size:
         Current Size:             8192 MB (arcsize)
         Target Size (Adaptive):   8192 MB (c)
         Min Size (Hard Limit):    1024 MB (zfs_arc_min)
         Max Size (Hard Limit):    8192 MB (zfs_arc_max)

ARC Size Breakdown:
         Most Recently Used Cache Size:          39%    3243 MB (p)
         Most Frequently Used Cache Size:        60%    4948 MB (c-p)

ARC Efficency:
         Cache Access Total:             154663786
         Cache Hit Ratio:      41%       64221251       [Defined State for 
buffer]
         Cache Miss Ratio:     58%       90442535       [Undefined State for 
Buffer]
         REAL Hit Ratio:       41%       64221251       [MRU/MFU Hits Only]

         Data Demand   Efficiency:    38%
         Data Prefetch Efficiency:    DISABLED (zfs_prefetch_disable)

        CACHE HITS BY CACHE LIST:
          Anon:                       --%        Counter Rolled.
          Most Recently Used:         17%        11118906 (mru)         [ 
Return Customer ]
          Most Frequently Used:       82%        53102345 (mfu)         [ 
Frequent Customer ]
          Most Recently Used Ghost:   14%        9427708 (mru_ghost)    [ 
Return Customer Evicted, Now Back ]
          Most Frequently Used Ghost:  6%        4344287 (mfu_ghost)    [ 
Frequent Customer Evicted, Now Back ]
        CACHE HITS BY DATA TYPE:
          Demand Data:                84%        54444108
          Prefetch Data:               0%        0
          Demand Metadata:            15%        9777143
          Prefetch Metadata:           0%        0
        CACHE MISSES BY DATA TYPE:
          Demand Data:                96%        87542292
          Prefetch Data:               0%        0
          Demand Metadata:             3%        2900243
          Prefetch Metadata:           0%        0


Also disabled file-level pre-fletch and vdev cache max:
set zfs:zfs_prefetch_disable = 1
set zfs:zfs_vdev_cache_max = 0x1

After reading about some issues with concurrent ios, I tweaked the setting down 
from 35 to 1 and it reduced the response times greatly (2 -> 8ms):
set zfs:zfs_vdev_max_pending=1

It did increased the actv...I'm still unsure about the side-effects here:
    r/s    w/s   Mr/s   Mw/s wait actv wsvc_t asvc_t  %w  %b device
    0.0    0.0    0.0    0.0  0.0  0.0    0.0    0.0   0   0 c0
    0.0    0.0    0.0    0.0  0.0  0.0    0.0    0.0   0   0 c0t0d0
 2295.2  398.7    4.2    7.2  0.0 18.6    0.0    6.9   0 1084 c1
    0.0    0.8    0.0    0.0  0.0  0.0    0.0    0.1   0   0 c1t0d0
  190.3   22.9    0.4    0.0  0.0  1.5    0.0    7.0   0  87 c1t1d0
  180.9   20.6    0.3    0.0  0.0  1.7    0.0    8.5   0  95 c1t2d0
  195.0   43.0    0.3    0.2  0.0  1.6    0.0    6.8   0  93 c1t3d0
  193.2   21.7    0.4    0.0  0.0  1.5    0.0    6.8   0  88 c1t4d0
  195.7   34.8    0.3    0.1  0.0  1.7    0.0    7.5   0  97 c1t5d0
  186.8   20.6    0.3    0.0  0.0  1.5    0.0    7.3   0  88 c1t6d0
  188.4   21.0    0.4    0.0  0.0  1.6    0.0    7.7   0  91 c1t7d0
  189.6   21.2    0.3    0.0  0.0  1.6    0.0    7.4   0  91 c1t8d0
  193.8   22.6    0.4    0.0  0.0  1.5    0.0    7.1   0  91 c1t9d0
  192.6   20.8    0.3    0.0  0.0  1.4    0.0    6.8   0  88 c1t10d0
  195.7   22.2    0.3    0.0  0.0  1.5    0.0    6.7   0  88 c1t11d0
  184.7   20.3    0.3    0.0  0.0  1.4    0.0    6.8   0  84 c1t12d0
    7.3   82.4    0.1    5.5  0.0  0.0    0.0    0.2   0   1 c1t13d0
    1.3   23.9    0.0    1.3  0.0  0.0    0.0    0.2   0   0 c1t14d0

I'm still in talks with the dba in seeing if we can raise the SGA from 4GB to 
6GB to see if it'll help.

The changes that showed a lot of improvement is disabling file/device level 
pre-fletch and reducing concurrent ios from 35 to 1 (tried 10 but it didn't 
help much).  Is there anything else that could be tweaked to increase write 
performance?  Record sizes are set according to 8K and 128K for redo logs.
-- 
This message posted from opensolaris.org
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to