Hi Dan,

Dan Pritts wrote:
> On Mon, Nov 19, 2007 at 11:10:32AM +0100, Paul Boven wrote:
>> Any suggestions on how to further investigate / fix this would be very
>> much welcomed. I'm trying to determine whether this is a zfs bug or one
>> with the Transtec raidbox, and whether to file a bug with either
>> Transtec (Promise) or zfs.

> the way i'd try to do this would be to use the same box under solaris
> software RAID, or better yet linux or windows software RAID (to make
> sure it's not a solaris device driver problem). 
> Does pulling the disk then get noticed?  If so, it's a zfs bug.  

Excellent suggestion, and today I had some time to give it a try.
I created a 4 disk SVM volume (2x 2-disk stripe, mirrored, with 2 more
disks as hot spare).

d10 -m /dev/md/rdsk/d11 /dev/md/rdsk/d12 1
d11 1 2 /dev/rdsk/c4t0d0s0 /dev/rdsk/c4t1d0s0 -i 1024b -h hsp001
d12 1 2 /dev/rdsk/c4t2d0s0 /dev/rdsk/c4t3d0s0 -i 1024b -h hsp001
hsp001 c4t4d0s0 c4t5d0s0

I started a write and then pulled a disk. And without any further
probing, SVM put a hotspare in place and started resyncing:

d10              m  463GB d11 d12 (resync-0%)
    d11          s  463GB c4t0d0s0 c4t1d0s0
    d12          s  463GB c4t2d0s0 (resyncing-c4t4d0s0) c4t3d0s0
hsp001           h      - c4t4d0s0 (in-use) c4t5d0s0

This is all on b76. The issue seems to be with zfs indeed. I'm currently
downloading b77, and once that is installed I'll have to see whether the
fault diagnostics and hot spare handling have indeed improved as several
people here have pointed out.

Regards, Paul Boven.
-- 
Paul Boven <[EMAIL PROTECTED]> +31 (0)521-596547
Unix/Linux/Networking specialist
Joint Institute for VLBI in Europe - www.jive.nl
VLBI - It's a fringe science
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to