Hello Richard,

Thursday, February 22, 2007, 3:32:07 AM, you wrote:

RE> Nissim Ben Haim wrote:
>> I was asked by a customer considering the x4500 - how much time should 
>> it take to rebuild a failed Disk under RaidZ ?
>> This question keeps popping because customers perceive software RAID as 
>> substantially inferior to HW raids.
>> I could not find someone who has really measured this under several 
>> scenarios.

RE> It is a function of the amount of space used.  As space used -> 0, it
RE> becomes infinitely fast.  As space used -> 100% is approaches the speed
RE> of the I/O subsystem.  In my experience, no hardware RAID array comes
RE> close, they all throttle the resync, though some of them allow you to
RE> tune it a little bit.  The key advantage over a hardware RAID system is
RE> that ZFS knows where the data is and doesn't need to replicate unused
RE> space.  A hardware RAID array doesn't know anything about the data, so
RE> it must reconstruct the entire disk.

RE> Also, the reconstruction is done in time order.  See Jeff Bonwick's blog:
RE> http://blogs.sun.com/bonwick/entry/smokin_mirrors

RE> I've measured resync on some slow IDE disks (*not* an X4500) at an average
RE> of 20 MBytes/s.  So if you have a 500 GByte drive, that would resync a 100%
RE> full file system in about 7 hours versus 11 days for some other systems
RE> (who shall remain nameless :-)

I wish it worked that good.

raid-z2 made of 11 disks on x4500.
When server wasn't doing anything - just resync, with pool almost full
with lots of small files, it took about 2 days to re-sync. I think
with lot of small files and almost full pool classic approach to
resync is actually faster.

With a server under some load (not that big) it took about two weeks!
to re-sync disk.

Then you've got to remember that you can't create new snapshots until
resync finishes or it will start all over again and you never resync.
This is known bug and very, very annoying on x4500 (or any other
server).

You should also keep in mind that current hot-spare support in ZFS is somewhat
lacking - if write fails to a given disk system will panic instead put
hot-spare in. Then after reboot depending on a disk failure type zfs
will note disk is bad and use hot-spare or it will not (if it can
still open that disk).

Also if you have a failing disk (I had) you can manually initiate
re-sync however until it finishes (two weeks???) you can't replace the
old drive online as it needs to export an pool! You also can't detach
a disk from raid-z[12] group until you re-sync new one.

IMHO current hot-spare support in ZFS is very basic only and needs lot
of work to be done.

-- 
Best regards,
 Robert                            mailto:[EMAIL PROTECTED]
                                       http://milek.blogspot.com

_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to