On Sunday, February 7, 2010, Brian McKerr <br...@datamatters.com.au> wrote: > While not strictly a ZFS issue as such I thought I'd post here as this and > the storage forums are my best bet in terms of getting some help. > > I have a machine that I recently set up with b130, b131 and b132. With each > build I have been playing around with ZFS raidz2 and mirroring to do a little > performance testing. This is a 6 SATA port ICH10 Intel motherboard, running > in AHCI mode. OS is on a USB flash. Suffice to say I have noticed that 1 > particular drive out of 6 seems to have very high asvc_t practically all the > time. This is an excerpt from 'iostat -xnM c6t2d0 2'; > > [i] r/s w/s Mr/s Mw/s wait actv wsvc_t asvc_t %w %b device > 70.5 502.0 0.0 4.1 0.0 1.3 0.0 2.2 0 54 c6t2d0 > 50.5 137.5 0.0 3.0 0.0 0.7 0.0 3.9 0 47 c6t2d0 > 71.0 163.5 0.0 4.8 0.0 0.8 0.0 3.4 0 61 c6t2d0 > 13.5 29.5 0.0 1.0 0.0 2.6 0.0 61.4 0 88 c6t2d0 > 1.0 0.5 0.0 0.0 0.0 3.6 0.0 2406.2 0 100 c6t2d0 > 1.0 1.0 0.0 0.0 0.0 4.0 0.0 1993.4 0 100 c6t2d0 > 1.0 1.5 0.0 0.0 0.0 4.0 0.0 1593.8 0 100 c6t2d0 > 2.0 3.0 0.0 0.1 0.0 4.0 0.0 791.6 0 100 c6t2d0 > 1.0 2.0 0.0 0.1 0.0 4.0 0.0 1320.3 0 100 c6t2d0 > 1.0 5.0 0.0 0.3 0.0 3.6 0.0 595.1 0 100 c6t2d0 > [/i] > > and here is the drive shown with the other in the raidz2 pool > > extended device statistics > r/s w/s Mr/s Mw/s wait actv wsvc_t asvc_t %w %b device > 0.0 1.5 0.0 0.0 0.0 0.0 0.0 0.5 0 0 c6t0d0 > 0.0 1.5 0.0 0.0 0.0 0.0 0.0 0.3 0 0 c6t1d0 > 1.0 1.0 0.0 0.0 0.0 4.0 0.0 1994.8 0 100 c6t2d0 > 1.0 1.5 0.0 0.0 0.0 0.0 0.0 5.2 0 1 c6t3d0 > 1.0 1.5 0.0 0.0 0.0 0.0 0.0 6.9 0 1 c6t4d0 > 1.0 1.5 0.0 0.0 0.0 0.0 0.0 10.1 0 2 c6t5d0 > extended device statistics > r/s w/s Mr/s Mw/s wait actv wsvc_t asvc_t %w %b device > 1.0 5.5 0.0 0.2 0.0 0.0 0.0 1.6 0 1 c6t0d0 > 1.0 5.5 0.0 0.2 0.0 0.0 0.0 1.5 0 1 c6t1d0 > 2.0 3.5 0.0 0.1 0.0 4.0 0.0 721.8 0 100 c6t2d0 > 1.0 5.5 0.0 0.2 0.0 0.0 0.0 1.9 0 1 c6t3d0 > 1.0 5.5 0.0 0.2 0.0 0.0 0.0 1.6 0 1 c6t4d0 > 2.0 5.5 0.0 0.2 0.0 0.0 0.0 3.1 0 2 c6t5d0 > extended device statistics > r/s w/s Mr/s Mw/s wait actv wsvc_t asvc_t %w %b device > 0.0 3.5 0.0 0.1 0.0 0.0 0.0 0.4 0 0 c6t0d0 > 0.0 3.5 0.0 0.1 0.0 0.0 0.0 1.8 0 0 c6t1d0 > 1.0 2.0 0.0 0.1 0.0 4.0 0.0 1327.1 0 100 c6t2d0 > 1.0 3.5 0.0 0.1 0.0 0.0 0.0 4.9 0 1 c6t3d0 > 1.0 3.5 0.0 0.1 0.0 0.0 0.0 3.9 0 1 c6t4d0 > 1.0 3.5 0.0 0.1 0.0 0.0 0.0 2.0 0 1 c6t5d0 > > I have seen asvc_t as high as 20000. > > There do not appear to be any errors hardware wise as 'iostat -e' shows > > [i] ---- errors --- > device s/w h/w trn tot > sd0 0 0 0 0 > sd2 0 0 0 0 > sd3 0 0 0 0 > sd4 0 0 0 0 > sd5 0 0 0 0 > sd6 0 0 0 0 > sd7 0 0 0 0 > [/i] > > > > 'zpool iostat -v 2' pauses for anywhere between 3 and 10 seconds before it > prints the stats for that particular drive in the pool; > > capacity operations bandwidth > pool alloc free read write read write > ------------------ ----- ----- ----- ----- ----- ----- > data 185G 5.26T 3 115 8.96K 2.48M > raidz2 185G 5.26T 3 115 8.96K 2.48M > c6t0d0 - - 2 26 2.70K 643K > c6t1d0 - - 2 26 2.49K 643K > > * INSERT ~10 SECOND PAUSE* > > c6t2d0 - - 2 24 2.81K 643K > c6t3d0 - - 2 26 2.75K 643K > c6t4d0 - - 2 26 2.45K 643K > c6t5d0 - - 2 26 2.71K 643K > ------------------ ----- ----- ----- ----- ----- ----- > rpool 3.50G 3.94G 0 0 9.99K 1010 > c5t0d0s0 3.50G 3.94G 0 0 9.99K 1010 > ------------------ ----- ----- ----- ----- ----- ----- > swpool 102K 3.69G 0 0 19 0 > /dev/rdsk/c7t0d0 102K 3.69G 0 0 19 0 > ------------------ ----- ----- ----- ----- ----- ----- > > I have booted up a linux rescue CD that has S.M.A.R.T support (system rescue > CD) and performed the 'long' test on each drive. All drives pass the 'test'. > There also appears to be no system errors with the drives under linux either. > > Can anyone shed any light on this issue, or suggest what I could try next ? I > am sort of discounting hardware problems given that I do not see errors from > the live linux CD. Maybe I should install linux and see if the problem > persists ? > > Cheers. > -- > This message posted from opensolaris.org > _______________________________________________ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
I'd say your easiest two options are swap ports and see if the problem follows the drive. If it does, swap the drive out. --Tim _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss