On 12/21/12 14:06, Paul Kraus wrote:
On Dec 21, 2012, at 7:49 AM, yudi v wrote:

I am building a new freebsd fileserver to use for backups, will be using 2
disk raid mirroring in a HP microserver n40l.
I have gone through some of the documentation and would like to know what
file systems to choose.

According to the docs, ufs is suggested for the system partitions but
someone on the freebsd irc channel suggested using zfs for the rootfs as
well

Are there any disadvantages of using zfs for the whole system rather than
going with ufs for the system files and zfs for the user data?

        First a disclaimer, I have been working with Solaris since 1995 and 
managed
> lots of data under ZFS, I have only been working with FreeBSD for about the past
> 6 months.

        UFS is clearly very stable and solid, but to get redundancy you need to 
use
> a separate "volume manager".

Slight correction here - you don't need a volume manager (as I understand the term), you'd use the GEOM subsystem, specifically gmirror in this case. See "man gmirror" for details

        ZFS is a completely different way of thinking about managing storage 
(not
> just a filesystem). I prefer ZFS for a number of reasons:

1) End to end data integrity through checksums. With the advent of 1 TB plus
> drives, the uncorrectable error rate (typically 10^-14 or 10^-15) means that > over the life of any drive you *are* now likely to run into uncorrectable errors. > This means that traditional volume managers (which rely on the drive reporting an > bad reads and writes) cannot detect these errors and bad data will be returned to
> the application.

2) Simplicity of management. Since the volume management and filesystem layers
> have been combined, you don't have to manage each separately.

3) Flexibility of storage. Once you build a zpool, the filesystems that reside
> on it share the storage of the entire zpool. This means you don't have to decide > how much space to commit to a given filesystem at creation. It also means that all > the filesystems residing in that one zpool share the performance of all the drives
> in that zpool.

4) Specific to booting off of a ZFS, if you move drives around (as I tend to do 
in
> at least one of my lab systems) the bootloader can still find the root filesystem > under ZFS as it refers to it by zfs device name, not physical drive device name. > Yes, you can tell the bootloader where to find root if you move it, but zfs does
> that automatically.

5) Zero performance penalty snapshots. The only cost to snapshots is the space
> necessary to hold the data. I have managed systems with over 100,000 snapshots.

        I am running two production, one lab, and a bunch of VBox VMs all with 
ZFS.
> The only issue I have seen is one I have also seen under Solaris with ZFS. Certain > kinds of hardware layer faults will cause the zfs management tools (the zpool and > zfs commands) to hang waiting on a blocking I/O that will never return. The data > continuos to be available, you just can't manage the zfs infrastructure until the > device issues are cleared. For example, if you remove a USB drive that hosts a > mounted ZFS, then any attempt to manage that ZFS device will hang (zpool export
> -f <zpool name> hangs until a reboot).

        Previously I had been running (at home) a fileserver under OpenSolaris 
using
> ZFS and it saved my data when I had multiple drive failures. At a certain client > we had a 45 TB configuration built on top of 120 750GB drives. We had multiple > redundancy and could survive a complete failure of 2 of the 5 disk enclosures (yes,
> we tested this in pre-production).

        There are a number of good writeups on how setup a FreeBSD system to 
boot off
> of ZFS, I like this one the best
> http://wiki.freebsd.org/RootOnZFS/GPTZFSBoot/9.0-RELEASE , but I do the zpool/zfs > configuration slightly differently (based on some hard learned lessons on Solaris). > I am writing up my configuration (and why I do it this way), but it is not ready yet.

        Make sure you look at all the information here: 
http://wiki.freebsd.org/ZFS ,
> keeping in mind that lots of it was written before FreeBSD 9. I would NOT use ZFS, > especially for booting, prior to release 9 of FreeBSD. Some of the reason for this
> is the bugs that were fixed in zpool version 28 (included in release 9).

I would agree with all that. My current system uses UFS filesystems for the base install, and ZFS with a raidz zpool for everything else, but that's only because I started using ZFS in REL 8.0 when it was just out of experimental status, and I didn't want to risk having an unbootable system. (That last paragraph suggests I was wise in that decision.) My next machine I'm specing out now will be pure ZFS so I get the boot environment stuff.

_______________________________________________
freebsd-questions@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-questions
To unsubscribe, send any mail to "freebsd-questions-unsubscr...@freebsd.org"

Reply via email to