Rich Freeman posted on Fri, 24 Feb 2012 13:47:45 -0500 as excerpted:

> On Fri, Feb 24, 2012 at 1:43 PM, Alexis Ballier <aball...@gentoo.org>
> wrote:
>> moreover the && wont delete the lib if revdep-rebuild failed i think,
>> so it should be even safer to copy/paste :)

FWIW this is the preserved_libs feature/bug I ran into in early testing, 
that convinced me to turn it off.  Running revdep-rebuild manually was 
far safer anyway, since at least then I /knew/ the status of various 
libs, they weren't preserved on first run, then arbitrarily deleted on 
second, even if it still broke remaining depending apps to do so.

So if that was reliably fixed, I'd be FAR happier about enabling 
FEATURES=preserved-libs.  I'm not sure I actually would as I like a bit 
more direct knowledge of stale libs on the system than the automated 
handling gives me, but at least I'd not have to worry about the so-called 
"preserved" libs STILL disappearing and leaving broken packages, if I DID 
enable it!

So definitely ++ on this! =:^)

> Am I the only paranoid person who moves them rather than unlinking them?
>  Oh, if only btrfs were stable...

FWIW, in the rare event it breaks revdep-rebuild or the underlying 
rebuilding itself, I rely on my long set FEATURES=buildpkg and emerge
-K.  In the even rarer event that too is broken, there's always manual 
untarring of the missing lib from the binpkg (I've had to do that once 
when gcc itself was broken due to an unadvised emerge -C that I knew 
might break things given the depclean warning, but also knew I could fix 
with an untar if it came to it, which it did), or if it comes to it, 
booting to backup and using ROOT= to emerge -K back to the working system.


[btrfs status discussion, skip if uninterested.]

I'm not sure if that's a reference to the btrfs snapshots allowing 
rollbacks feature, or a hint that you're running it and worried about its 
stability underneath you...

If it's the latter, you probably already know this, but if it's the 
former, and for others interested...

I recently set the btrfs kernel options and merged btrfs-progs, then read 
up on the wiki and joined the btrfs list, with the plan being to get 
familiar with it and perhaps install it.

>From all the reports about it being an option for various distros, etc, 
now, and all the constant improvement reports, I had /thought/ that the 
biggest issue for stability was the lack of an error-correcting (not just 
detecting) fsck.btrfs, and that the restore tool announced late last 
year, that allows pulling data off of unmountable btrfs volumes was a 
reasonable workaround.

What I found, even allowing for the fact that such lists get the bad 
reports and not the good ones, thus paint a rather worse picture of the 
situation than actually exists for most users, is that...

BTRFS still has a rather longer way to go than I had thought.  It's still 
FAR from stable, even for someone like myself that often runs betas and 
was prepared to keep (and use, if necessary) TESTED backups, etc.  Maybe 
by Q4 this year, but also very possibly not until next year.  I'd 
definitely NOT recommend that anyone run it now, unless you are 
SPECIFICALLY running it for testing and bug reporting purposes with 
"garbage" data (IOW, data that you're NOT depending on, at the btrfs 
level, at all) that you are not only PREPARED to lose, but EXPECT to 
lose, perhaps repeatedly, during your testing.

IOW, there's still known untraced and unfixed active data corruption bugs 
remaining.  Don't put your data on btrfs at this point unless you EXPECT 
to have it corrupted, and want to actively help in tracing and patching 
the problems!

Additionally, for anyone who has been interested in the btrfs RAID 
capacities, striped/raid0 it handles, but its raid1 and raid10 capacities 
are misnamed.   At present, it's strictly two-way-mirror ONLY, there's no 
way to do N-way (N>2) mirroring aside from layering on top of say mdraid, 
at all, and of course layering on top of mdraid loses the data integrity 
guarantees at that level, btrfs still has just the one additional copy it 
can fall back on.  This SERIOUSLY limits btrfs data integrity 
possibilities in a 2+ drive failure scenario.

btrfs raid5/6 isn't available yet, but the current roadmap says kernels 
3.4 or 3.5.  Multi-mirror is supposed to be built on that code, tho the 
mentions of it I've seen are specifically triple-mirror, so it's unclear 
whether arbitrary N-way (N>3) mirroring as in true raid1 will be possible 
even then.  But whether triple-way specifically or N-way (N>=3), given 
it's on top of raid5/6, to be introduced in 3.4/3.5, triple-way mirroring 
thus appears to be 3.5/3.6 at the earliest.

So while I had gotten the picture that btrfs was stabilizing and it was 
mostly over-cautiousness keeping that experimental label around, that's 
definitely NOT the case.  Nobody should really plan on /relying/ on it, 
even with backups, until at least late this year, and very possibly 
looking at 2013 now.

So btrfs is still a ways out. =:^(

Meanwhile, for anyone that's still interested in it at this point, note 
that the homepage wiki currently listed the btrfs-progs package is a 
stale copy on kernel.org, still read-only after the kernel.org breakin.  
The "temporary" but looking more and more permanent location is:

http://btrfs.ipv5.de/index.php?title=Main_Page

Also, regarding the gentoo btrfs-progs package, see my recently filed:

https://bugs.gentoo.org/show_bug.cgi?id=405519

-- 
Duncan - List replies preferred.   No HTML msgs.
"Every nonfree program has a lord, a master --
and if you use the program, he is your master."  Richard Stallman


Reply via email to