On Aug 15, 2010, at 4:59 PM, Ian Collins wrote: > I look after an x4500 for a client and wee keep getting drives marked as > degraded with just over 20 checksum errors. > > Most of these errors appear to be driver or hardware related and thier > frequency increases during a resilver, which can lead to a death spiral. The > increase in errors within a vdev during a resilver (I recently had three > drives in an 8 drive raidz vdev "degraded") points to high read activity > triggering the bug. > > I would like to raise threshold for marking a drive degraded to give me more > time to spot and clear the checksum errors. Is this possible?
There is not a documented system-admin visible interface to this. The settings in question can be set as properties in the zfs-diagnosis.conf file, similar to props set in other FMA modules. The source is also currently available. http://src.opensolaris.org/source/xref/onnv/onnv-gate/usr/src/cmd/fm/modules/common/zfs-diagnosis/zfs_de.c#957 Examples of setting FMA module properties are in /usr/lib/fm/fmd/plugins/cpumem-retire.conf and other .conf files. If you get this to work, please publicly document your changes and why you felt the new settings were better. -- richard -- Richard Elling rich...@nexenta.com +1-760-896-4422 Enterprise class storage for everyone www.nexenta.com _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss