On Sat, 21 Nov 2009, Ted Unangst wrote:

> On Fri, Nov 20, 2009 at 2:08 PM, Jeff Ross <jr...@openvistas.net> wrote:
>> I failed a drive (*NOT* with a nail gun, though, I just removed it ;-) and
>> bioctl correctly showed the drive as failed and the raid running as
>> degraded.  I  re-inserted the same drive but I was unable to find any
magic
>> bioctl -R that would kick off a rebuild.
>>
>> I shutdown that server, removed the failed drive and inserted it into
>> another identical SuperMicro.  System booted, noted an unclean shutdown,
ran
>> fsck and  was at login in short order.
>>
>> So, just for fun I shutdown the second server, took its drive and
>> reinstalled in the second slot of the first again and fired it up.  As the
>> system booted the second drive's light lit solid on so I suspected a
behind
>> the scenes rebuild was going on.  When I got logged in bioctl -v mpi0
shows
>> both drives online, and the raid status is Rebuild.
>
> I have no idea what magic your raid controller has, but unless it
> understands filesystems, this is really asking for trouble.  You have
> two similar but slightly different filesystem images.  You are going
> to "merge" them?

No, I don't think that's the way it is supposed to work.

>
> I assume/hope the the raid card just picked one drive as the winner
> and is going to wipe the other, because that's the only way this could
> work.

Maybe I've always misunderstood hardware raid controllers but I thought
that's the way they are supposed to work, especially with raid1.  One
drive in the set fails, the controller notes that, and if a hot spare is
available it begins rebuilding the array using the hot spare, replacing
the failed drive.

In my scenario, in a 1U server with 2 enclosures I do not have a hot spare
online but I still have a good drive and a failed drive.  Replacing the failed
drive, even with the exact same hard disk, has to result in copying the
good drive to the new drive, right?

Jeff

Reply via email to