Carsten Aulbert wrote:

> Put some stress on the system with bonnie and other tools and try to
> find slow disks and see if this could be the main problem but also look
> into more vdevs and then possible move to raidz to somehow compensate
> for lost disk space. Since we have 4 cold spares on the shelf plus a SMS
> warnings on disk failures (that is if fma catches them) the risk
> involved should be tolerable.

First result with bonnie during the "writing intelligently..." phase I
see this in a 2 minute average:

zpool iostats:

               capacity     operations    bandwidth
pool         used  avail   read  write   read  write
----------  -----  -----  -----  -----  -----  -----
atlashome   1.70T  19.2T    225  1.49K   342K   107M
  raidz2     550G  6.28T     74    409   114K  32.6M
    c0t0d0      -      -      0    314  32.3K  2.51M
    c1t0d0      -      -      0    315  31.8K  2.52M
    c4t0d0      -      -      0    313  31.3K  2.52M
    c6t0d0      -      -      0    315  32.3K  2.51M
    c7t0d0      -      -      0    326  32.8K  2.50M
    c0t1d0      -      -      0    309  33.9K  2.52M
    c1t1d0      -      -      0    313  33.4K  2.51M
    c4t1d0      -      -      0    314  33.4K  2.52M
    c5t1d0      -      -      0    308  32.8K  2.52M
    c6t1d0      -      -      0    314  31.3K  2.51M
    c7t1d0      -      -      0    311  31.8K  2.52M
    c0t2d0      -      -      0    309  31.8K  2.52M
    c1t2d0      -      -      0    313  31.8K  2.51M
    c4t2d0      -      -      0    315  31.8K  2.52M
    c5t2d0      -      -      0    307  32.8K  2.52M
  raidz2     567G  6.26T     64    529  96.5K  36.3M
    c6t2d0      -      -      1    368  74.2K  2.79M
    c7t2d0      -      -      1    366  74.2K  2.80M
    c0t3d0      -      -      1    364  75.8K  2.80M
    c1t3d0      -      -      1    365  75.2K  2.80M
    c4t3d0      -      -      1    368  76.8K  2.80M
    c5t3d0      -      -      1    362  76.3K  2.80M
    c6t3d0      -      -      1    366  77.9K  2.80M
    c7t3d0      -      -      1    365  76.8K  2.80M
    c0t4d0      -      -      1    361  76.8K  2.80M
    c1t4d0      -      -      1    363  75.8K  2.80M
    c4t4d0      -      -      1    366  76.3K  2.80M
    c6t4d0      -      -      1    364  78.4K  2.80M
    c7t4d0      -      -      1    370  78.9K  2.79M
    c0t5d0      -      -      1    365  77.3K  2.80M
    c1t5d0      -      -      1    364  74.7K  2.80M
  raidz2     620G  6.64T     86    582   131K  37.9M
    c4t5d0      -      -     18    382  1.16M  2.74M
    c5t5d0      -      -     10    380   674K  2.74M
    c6t5d0      -      -     18    378  1.15M  2.73M
    c7t5d0      -      -      9    384   628K  2.74M
    c0t6d0      -      -     18    377  1.16M  2.74M
    c1t6d0      -      -     10    383   680K  2.75M
    c4t6d0      -      -     19    379  1.21M  2.73M
    c5t6d0      -      -     10    383   691K  2.75M
    c6t6d0      -      -     19    379  1.21M  2.73M
    c7t6d0      -      -     10    383   676K  2.72M
    c0t7d0      -      -     18    374  1.19M  2.75M
    c1t7d0      -      -     10    381   676K  2.74M
    c4t7d0      -      -     19    380  1.22M  2.74M
    c5t7d0      -      -     10    382   696K  2.74M
    c6t7d0      -      -     18    381  1.17M  2.74M
    c7t7d0      -      -      9    386   631K  2.75M
----------  -----  -----  -----  -----  -----  -----

iostat -Mnx 120:
                    extended device statistics
    r/s    w/s   Mr/s   Mw/s wait actv wsvc_t asvc_t  %w  %b device
    0.0    0.0    0.0    0.0  0.0  0.0    0.0    0.0   0   0 c2t0d0
    0.0    0.0    0.0    0.0  0.0  0.0    0.0    0.0   0   0 c3t0d0
    0.0    1.4    0.0    0.0  0.0  0.0    1.5    0.4   0   0 c5t0d0
    0.6  351.5    0.0    2.6  0.4  0.1    1.2    0.2   3   8 c7t0d0
    0.6  336.3    0.0    2.6  0.1  0.1    0.4    0.2   3   7 c0t0d0
    0.6  340.8    0.0    2.6  0.2  0.1    0.6    0.2   3   7 c1t0d0
    0.6  330.6    0.0    2.6  0.1  0.1    0.3    0.2   3   7 c5t1d0
    0.6  336.7    0.0    2.6  0.1  0.1    0.3    0.2   3   7 c4t0d0
    0.6  331.8    0.0    2.6  0.1  0.1    0.3    0.2   3   7 c0t1d0
    0.6  339.0    0.0    2.6  0.4  0.1    1.1    0.2   3   7 c7t1d0
    0.6  335.4    0.0    2.6  0.1  0.1    0.4    0.2   3   7 c1t1d0
    0.6  329.2    0.0    2.6  0.1  0.1    0.3    0.2   3   7 c5t2d0
    0.6  343.7    0.0    2.6  0.3  0.1    0.7    0.2   3   7 c4t1d0
    0.6  331.8    0.0    2.6  0.1  0.1    0.3    0.2   2   7 c0t2d0
    1.2  396.3    0.1    2.9  0.3  0.1    0.7    0.2   4   8 c7t2d0
    0.6  336.7    0.0    2.6  0.1  0.1    0.4    0.2   3   7 c1t2d0
    0.6  341.9    0.0    2.6  0.2  0.1    0.7    0.2   3   7 c4t2d0
    1.3  390.7    0.1    2.9  0.3  0.1    0.8    0.2   4   9 c5t3d0
    1.3  396.7    0.1    2.9  0.3  0.1    0.8    0.2   4   9 c7t3d0
    1.3  393.6    0.1    2.9  0.2  0.1    0.6    0.2   4   9 c0t3d0
    0.0    0.0    0.0    0.0  0.0  0.0    0.0    0.0   0   0 c5t4d0
    1.3  396.2    0.1    2.9  0.2  0.1    0.5    0.2   4   8 c1t3d0
    1.3  399.2    0.1    2.9  0.3  0.1    0.8    0.2   4   9 c4t3d0
    1.3  401.8    0.1    2.9  0.3  0.1    0.8    0.2   4   9 c7t4d0
    1.3  388.5    0.1    2.9  0.2  0.1    0.5    0.2   4   8 c0t4d0
    1.3  391.8    0.1    2.9  0.2  0.1    0.5    0.2   4   9 c1t4d0
    1.3  395.1    0.1    2.9  0.2  0.1    0.6    0.2   4   8 c4t4d0
    9.9  409.7    0.6    2.9  0.8  0.2    1.9    0.4  10  18 c7t5d0
    1.3  395.0    0.1    2.9  0.3  0.1    0.6    0.2   4   9 c0t5d0
   10.6  405.3    0.7    2.9  0.8  0.2    2.0    0.4  11  18 c5t5d0
    1.3  392.8    0.1    2.9  0.2  0.1    0.5    0.2   4   8 c1t5d0
   10.7  407.6    0.7    2.9  0.9  0.2    2.1    0.4  11  19 c7t6d0
   18.6  407.5    1.2    2.9  1.0  0.2    2.4    0.6  15  24 c4t5d0
   10.9  407.8    0.7    2.9  0.8  0.2    2.0    0.4  11  19 c5t6d0
    0.6  337.6    0.0    2.6  0.2  0.1    0.5    0.2   3   7 c6t0d0
   10.7  408.8    0.7    2.9  0.8  0.2    1.9    0.4  11  19 c1t6d0
   10.0  411.6    0.6    2.9  0.8  0.2    1.8    0.4  11  18 c7t7d0
   19.3  403.1    1.2    2.9  1.1  0.3    2.6    0.6  16  26 c4t6d0
    0.6  336.2    0.0    2.6  0.1  0.1    0.4    0.2   3   7 c6t1d0
   11.0  407.7    0.7    2.9  0.8  0.2    1.9    0.4  11  19 c5t7d0
   10.6  406.6    0.7    2.9  0.8  0.2    2.0    0.4  11  19 c1t7d0
   18.5  401.7    1.2    2.9  1.0  0.2    2.5    0.6  15  25 c0t6d0
   19.4  404.8    1.2    2.9  1.0  0.3    2.5    0.6  15  25 c4t7d0
    1.2  397.6    0.1    2.9  0.3  0.1    0.9    0.2   4   9 c6t2d0
   19.0  398.7    1.2    2.9  1.0  0.3    2.5    0.6  15  25 c0t7d0
    1.3  396.1    0.1    2.9  0.2  0.1    0.5    0.2   4   8 c6t3d0
    1.3  392.8    0.1    2.9  0.2  0.1    0.4    0.2   4   8 c6t4d0
   18.4  403.3    1.2    2.9  1.1  0.2    2.5    0.6  15  24 c6t5d0
   19.3  402.7    1.2    2.9  1.1  0.3    2.5    0.6  15  25 c6t6d0
   18.8  406.1    1.2    2.9  1.0  0.2    2.4    0.6  15  25 c6t7d0


Any experts here to say if that's just because bonnie via NFSv3 is a
very special test - if it is I can start something else, suggestions? -
or if some disks are really too busy and slowing down the pool.

Thanks for more insight

Carsten
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to