I can produce the timeout error on multiple, similar servers.
These are storage servers, so no zones or gui running.
Hardware:
Supermicro X7DWN with AOC-USASLP-L8i controller
E1 (single port) backplanes (16 & 24 bay)
(LSILOGICSASX28 A.0 and LSILOGICSASX36 A.1)
up to 36 1Tb WD Sata disks 

This server has 2 x quad core Intel CPU & 16Gb ram.
Disks: WD 1Tb c4t12d0 to c4t47d0 as single raidz pool. (6 disks per set)
Running dev 131.
I see problem on 2009.06 as well.

I note that the latest AOC-USASLP-L8i firmware is LSI Rev 1.26.00.00, which I 
believe does not support MSI. (working on Supermicro to update the firmware)

I have an LSI controller to swap for the  AOC-USASLP-L8i with latest firmware, 
which I can retest with.

After a few hours of light load, no errors appear unless I initiate a scrub.

 iostat -X -e -n
 ---- errors ---
  s/w h/w trn tot device
  0   0   0   0 fd0
  0   9   0   9 c5t1d0
  0   0   0   0 c4t8d0
  0   0   0   0 c4t9d0
  0   0   0   0 c4t12d0
  0   0   0   0 c4t13d0
  0   0   0   0 c4t14d0
  0   0   0   0 c4t15d0
  0   0   0   0 c4t16d0
  0   0   0   0 c4t17d0
  0   0   0   0 c4t18d0
  0   0   0   0 c4t19d0
  0   0   0   0 c4t20d0
  0   0   0   0 c4t21d0
  0   0   0   0 c4t22d0
  0   0   0   0 c4t23d0
  0   0   0   0 c4t30d0
  0   1  10  11 c4t31d0
  0   2  20  22 c4t32d0
  0   0   0   0 c4t33d0
  0   0   0   0 c4t34d0
  0   0   0   0 c4t35d0
  0   0   0   0 c4t36d0
  0   0   0   0 c4t37d0
  0   0   0   0 c4t38d0
  0   0   0   0 c4t39d0
  0   0   0   0 c4t40d0
  0   0   0   0 c4t41d0
  0   0   0   0 c4t42d0
  0   1  10  11 c4t43d0
  0   3  31  34 c4t44d0
  0   1  10  11 c4t45d0
  0   2  20  22 c4t46d0
  0   1  10  11 c4t47d0
  0   0   0   0 c4t48d0
  0   0   0   0 c4t49d0
  0   0   0   0 c4t50d0
  0   0   0   0 c4t51d0
  0   0   0   0 c4t52d0

In this instance, all errors are on the same (24 bay) backplane.
I have also had them on the 16 bay backplane with this 2 chassis configuration.

The problem becomes more of a pain when drives drop off for a short period, 
then reconnect and resilver or occassionally just stop until a reboot or hot 
plug.
The robustness of ZFS certainly helps keep things running.


Mark.
-- 
This message posted from opensolaris.org
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to