On Aug 4, 2007 9:26 PM, Greg Oster <[EMAIL PROTECTED]> wrote:
>
> "L. V. Lammert" writes:
> > On Fri, 3 Aug 2007, Joel Knight wrote:
> >
> > > --- Quoting HDC on 2007/08/02 at 20:26 -0300:
> > >
> > > > Read this...
> > > > http://www.packetmischief.ca/openbsd/doc/raidadmin/<http://www.packetmisc
> > hief
> > > > .ca/openbsd/>
> > > >
> > >
> > > I used to use raidframe and followed the procedures in that doc for
> > > doing so, but now there's no point. If the system requires any type of
> > > raid, go hardware. Long live bio(4).
> > >
> > IF you choose to NOT use a h/w controller, use rsync instead. Permits
> > quick recovery in the case of a drive failure (swap drive cables &
> > reboot), does not require lengthy parity rebuild.
>
> And you only lose the data written since the last rsync...
> and your system probably goes down instead of staying up until you
> can fix it..
>
> RAIDframe, like hardware RAID and rsync, is just another tool.
> Understand the pros and cons of each, but be willing to accept the
> risks associated with whatever you choose... (if you think hardware
> RAID is riskless, then you've never had a 2TB RAID set suddenly
> decide that all components were "offline" and mark them as such!)
>
> For the folks who dislike the "long parity checks"... If you're
> willing to accept a window during which some of your data *might* be
> at risk, change:
>  raidctl -P all
> to something like
>  sleep 3600 ; raidctl -P all &
>

Greg does a minor correction
it should actually be

(sleep 3600 ; raidctl -P all) &

What I do id I comment out

raidctl -P all

in /etc/rc

and put

(sleep 600; raidctl -P all) &

 at the end of  /etc/rc.local.

So the parity rebuild starts only 10 minutes after the system is up an
running from the master disk.

thanks :-)

Siju
> in /etc/rc .  This will, of course, delay the start of the parity
> computation for an hour or so, giving your system a chance to do the
> fscks and get back to multi-user as quickly as possible.
>
> The risk here is as follows (this is for RAID 1.. risks for RAID 5
> are slightly higher):
>   1) even though parity is marked 'dirty', it might actually be in
> sync.  In this case if you have a component failure, your data is
> fine.
>   2) until the parity check is done, only the 'master' component is
> used for reading.  But any writes will be done are mirrored to both
> components.  That means that when the fsck is being done, any
> problems found will be fixed on *both* components, and writes will
> keep the two in sync even before parity is checked.
>   3) Where the risk of data loss comes in is if the master dies
> before the parity check gets done.  In this case, data on the master
> that was not re-written or that was out-of-sync with the slave will
> be lost.  This could result in the loss of pretty much anything.
>
> The important thing here is for you to evaluate your situation and
> decide whether this level of risk is acceptable... For me, I use the
> equivalent to 'sleep 3600' on my home desktop.. and slightly modified
> versions of it on other home servers and other boxen I look after..
> But don't blindly listen to me or anyone else -- learn what the risks
> are for your situation, determine what level of risk you can accept,
> and go from there...
>
> Later...
>
> Greg Oster

Reply via email to