Hi Dan, Dan Pritts wrote: > On Mon, Nov 19, 2007 at 11:10:32AM +0100, Paul Boven wrote: >> Any suggestions on how to further investigate / fix this would be very >> much welcomed. I'm trying to determine whether this is a zfs bug or one >> with the Transtec raidbox, and whether to file a bug with either >> Transtec (Promise) or zfs.
> the way i'd try to do this would be to use the same box under solaris > software RAID, or better yet linux or windows software RAID (to make > sure it's not a solaris device driver problem). > Does pulling the disk then get noticed? If so, it's a zfs bug. Excellent suggestion, and today I had some time to give it a try. I created a 4 disk SVM volume (2x 2-disk stripe, mirrored, with 2 more disks as hot spare). d10 -m /dev/md/rdsk/d11 /dev/md/rdsk/d12 1 d11 1 2 /dev/rdsk/c4t0d0s0 /dev/rdsk/c4t1d0s0 -i 1024b -h hsp001 d12 1 2 /dev/rdsk/c4t2d0s0 /dev/rdsk/c4t3d0s0 -i 1024b -h hsp001 hsp001 c4t4d0s0 c4t5d0s0 I started a write and then pulled a disk. And without any further probing, SVM put a hotspare in place and started resyncing: d10 m 463GB d11 d12 (resync-0%) d11 s 463GB c4t0d0s0 c4t1d0s0 d12 s 463GB c4t2d0s0 (resyncing-c4t4d0s0) c4t3d0s0 hsp001 h - c4t4d0s0 (in-use) c4t5d0s0 This is all on b76. The issue seems to be with zfs indeed. I'm currently downloading b77, and once that is installed I'll have to see whether the fault diagnostics and hot spare handling have indeed improved as several people here have pointed out. Regards, Paul Boven. -- Paul Boven <[EMAIL PROTECTED]> +31 (0)521-596547 Unix/Linux/Networking specialist Joint Institute for VLBI in Europe - www.jive.nl VLBI - It's a fringe science _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss