On Sunday, February 7, 2010, Brian McKerr <br...@datamatters.com.au> wrote:
> While not strictly a ZFS issue as such I thought I'd post here as this and 
> the storage forums are my best bet in terms of getting some help.
>
> I have a machine that I recently set up with b130, b131 and b132. With each 
> build I have been playing around with ZFS raidz2 and mirroring to do a little 
> performance testing. This is a 6 SATA port ICH10 Intel motherboard, running 
> in AHCI mode. OS is on a USB flash. Suffice to say I have noticed that 1 
> particular drive out of 6 seems to have very high asvc_t practically all the 
> time. This is an excerpt from 'iostat -xnM c6t2d0 2';
>
>  [i]   r/s    w/s   Mr/s   Mw/s wait actv wsvc_t asvc_t  %w  %b device
>    70.5  502.0    0.0    4.1  0.0  1.3    0.0    2.2   0  54 c6t2d0
>    50.5  137.5    0.0    3.0  0.0  0.7    0.0    3.9   0  47 c6t2d0
>    71.0  163.5    0.0    4.8  0.0  0.8    0.0    3.4   0  61 c6t2d0
>    13.5   29.5    0.0    1.0  0.0  2.6    0.0   61.4   0  88 c6t2d0
>     1.0    0.5    0.0    0.0  0.0  3.6    0.0 2406.2   0 100 c6t2d0
>     1.0    1.0    0.0    0.0  0.0  4.0    0.0 1993.4   0 100 c6t2d0
>     1.0    1.5    0.0    0.0  0.0  4.0    0.0 1593.8   0 100 c6t2d0
>     2.0    3.0    0.0    0.1  0.0  4.0    0.0  791.6   0 100 c6t2d0
>     1.0    2.0    0.0    0.1  0.0  4.0    0.0 1320.3   0 100 c6t2d0
>     1.0    5.0    0.0    0.3  0.0  3.6    0.0  595.1   0 100 c6t2d0
> [/i]
>
> and here is the drive shown with the other in the raidz2 pool
>
>                     extended device statistics
>     r/s    w/s   Mr/s   Mw/s wait actv wsvc_t asvc_t  %w  %b device
>     0.0    1.5    0.0    0.0  0.0  0.0    0.0    0.5   0   0 c6t0d0
>     0.0    1.5    0.0    0.0  0.0  0.0    0.0    0.3   0   0 c6t1d0
>     1.0    1.0    0.0    0.0  0.0  4.0    0.0 1994.8   0 100 c6t2d0
>     1.0    1.5    0.0    0.0  0.0  0.0    0.0    5.2   0   1 c6t3d0
>     1.0    1.5    0.0    0.0  0.0  0.0    0.0    6.9   0   1 c6t4d0
>     1.0    1.5    0.0    0.0  0.0  0.0    0.0   10.1   0   2 c6t5d0
>                     extended device statistics
>     r/s    w/s   Mr/s   Mw/s wait actv wsvc_t asvc_t  %w  %b device
>     1.0    5.5    0.0    0.2  0.0  0.0    0.0    1.6   0   1 c6t0d0
>     1.0    5.5    0.0    0.2  0.0  0.0    0.0    1.5   0   1 c6t1d0
>     2.0    3.5    0.0    0.1  0.0  4.0    0.0  721.8   0 100 c6t2d0
>     1.0    5.5    0.0    0.2  0.0  0.0    0.0    1.9   0   1 c6t3d0
>     1.0    5.5    0.0    0.2  0.0  0.0    0.0    1.6   0   1 c6t4d0
>     2.0    5.5    0.0    0.2  0.0  0.0    0.0    3.1   0   2 c6t5d0
>                     extended device statistics
>     r/s    w/s   Mr/s   Mw/s wait actv wsvc_t asvc_t  %w  %b device
>     0.0    3.5    0.0    0.1  0.0  0.0    0.0    0.4   0   0 c6t0d0
>     0.0    3.5    0.0    0.1  0.0  0.0    0.0    1.8   0   0 c6t1d0
>     1.0    2.0    0.0    0.1  0.0  4.0    0.0 1327.1   0 100 c6t2d0
>     1.0    3.5    0.0    0.1  0.0  0.0    0.0    4.9   0   1 c6t3d0
>     1.0    3.5    0.0    0.1  0.0  0.0    0.0    3.9   0   1 c6t4d0
>     1.0    3.5    0.0    0.1  0.0  0.0    0.0    2.0   0   1 c6t5d0
>
> I have seen asvc_t as high as 20000.
>
> There do not appear to be any errors hardware wise as 'iostat -e' shows
>
>  [i]          ---- errors ---
> device  s/w h/w trn tot
> sd0       0   0   0   0
> sd2       0   0   0   0
> sd3       0   0   0   0
> sd4       0   0   0   0
> sd5       0   0   0   0
> sd6       0   0   0   0
> sd7       0   0   0   0
> [/i]
>
>
>
> 'zpool iostat -v 2' pauses for anywhere between 3 and 10 seconds before it 
> prints the stats for that particular drive in the pool;
>
>                        capacity     operations    bandwidth
> pool                alloc   free   read  write   read  write
> ------------------  -----  -----  -----  -----  -----  -----
> data                 185G  5.26T      3    115  8.96K  2.48M
>   raidz2             185G  5.26T      3    115  8.96K  2.48M
>     c6t0d0              -      -      2     26  2.70K   643K
>     c6t1d0              -      -      2     26  2.49K   643K
>
> * INSERT ~10 SECOND PAUSE*
>
>     c6t2d0              -      -      2     24  2.81K   643K
>     c6t3d0              -      -      2     26  2.75K   643K
>     c6t4d0              -      -      2     26  2.45K   643K
>     c6t5d0              -      -      2     26  2.71K   643K
> ------------------  -----  -----  -----  -----  -----  -----
> rpool               3.50G  3.94G      0      0  9.99K   1010
>   c5t0d0s0          3.50G  3.94G      0      0  9.99K   1010
> ------------------  -----  -----  -----  -----  -----  -----
> swpool               102K  3.69G      0      0     19      0
>   /dev/rdsk/c7t0d0   102K  3.69G      0      0     19      0
> ------------------  -----  -----  -----  -----  -----  -----
>
> I have booted up a linux rescue CD that has S.M.A.R.T support (system rescue 
> CD) and performed the 'long' test on each drive. All drives pass the 'test'. 
> There also appears to be no system errors with the drives under linux either.
>
> Can anyone shed any light on this issue, or suggest what I could try next ? I 
> am sort of discounting hardware problems given that I do not see errors from 
> the live linux CD. Maybe I should install linux and see if the problem 
> persists ?
>
> Cheers.
> --
> This message posted from opensolaris.org
> _______________________________________________
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss


I'd say your easiest two options are swap ports and see if the problem
follows the drive. If it does, swap the drive out.


--Tim
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to