On 1/2/07, Anders Troberg <[EMAIL PROTECTED]> wrote:
I've ploughed through the documentation, but it's kind of vague on some points and I need to buy some hardware if I'm to test it, so I thought I'd ask first. I'll begin by describing what I want to achieve, and would apreciate if someone could tell me if this is possible or how close I can come. My current situation: I have a lot of data, currently som 10-12 TB spread over about 50 disks in five file servers, adding on the average a new disk every 1-2 months,
Ideally you should add 3-5 disks at a time so you can add raidz(like raid5) groups so the failure of a disk won't cause lost of data. and a new server every year or so. The old paradigm with volumes makes this
a major pain in the rear end, as it becomes very difficult
with ZFS its easier if you keep all disks on one server, just buy the biggest disk box you can and fill with 8x sata controllers. Spreading the disks over multiple servers means you have to use iscsi to share the disks which is just an added headache. Plus you will have to pay for more network infrastructure, plus powering extra CPU's and resources that isn't needed if all disks are attached to one box. to organize data. The large volume also makes backups impractical, as I'm
just an ordinary home user, so fancy tape robots and such are out of my price range. Most of my data is not compressable, and the size varies greatly, from large files (GB-sized) to huge numbers (millions) of tiny files (just a few kB). At the moment, all these are Windows servers, but I plan to switch to some Unix/Linux variant as soon as there is enough benefits (and ZFS sure looks like it could be the juicy bait in that trap). The clients (15 or so) are a mix of Linux, Windows and a bunch of Xboxes (as media players), with the Windows machines gradually being phased out and changed to Linux as fast as I can rewrite my own software (which I can't do without) for Linux. What I want: * The entire storage should be visible as one file system, with one logical file structure where volumes and servers are not even visible, as if it was one huge disk. No paths like /root/server1fs/volume1/dir... in other words.
Yes this is possible, but not advisable, but ZFS allows you to mount your file systems where ever so you won't have to deal with /root/server1fs/dir but you can use ones like mydata/january or mydata/movies or what ever you need. * Software RAID support, even across the network, so I can just add a bunch
of parity disks and survive if a few disks crash. To me, it's well worth it to pony up with the money for 5-10 extra disks if I know that that many disks can fail before I start to lose data. That would be good enough to dispense with the need for proper backups.
yes its possible but not a feature of ZFS, you will need to share disks using iscsi, and then put the shared disks in a zfs pool. But its much better just to put all the disks attached to one server, using external disk boxes and large server cases. * A RAID that allows me to use differently sized disks without losing lots
of disk space. I'm OK if some disk space is lost (ie a file is not striped over all disks, somewhat increasing the stripe size and thereby the size of the parity data), but I don't want to have my 400 GB disks only use the first 160 GB just because I have a shitload of 160 GB disks.
its best to add similar disks in a raidz group, so if you added 5x 160GB disks in a raidz group you would get 4x160GB of data with one drive being used for parity data protecting you from any drive in that group dieing. * Performance does not need to be stellar, but should not be snail-like
either. If it's enough to fill a 100 Mbit network cable, I'm perfectly happy, if it can't fill a 10 Mbit I'm starting to get worried.
if all disks are in one box you shouldn't have any performance problems, because ZFS spreads data amoung all drives in the pool. if you use iscsi to share disks you will have more performance bottle necks but if you have the network bandwidth, you should easily always have at least 100mbit of throughput. * A file system that handles huge numbers of tiny files somewhat
effectively. Many file systems use a full block even for a tiny file, which cause huge overheads when there are many files.
not a problem. ZFS uses variable sized blocks anything from 512bytes to 128k bytes per block, it is even flexible in raidz configurations where a 512 byte file uses just 1k of space, 512 bytes for data, and 512 bytes for parity. * Good interoperability with Linux, Windows and Xbox (actually, this is just
a question of Samba compliance and as such out of scope for this discussion).
Solaris supports, NFSv3 and v4, Samba, FTP, http, and just about any other network protocol you need. Is this doable? If not, how close can I get and what is it that I can't get? yes, its all doable but its best to as many disks in one main server as possible, even if you have to buy new cases and external disk boxes you should see savings simply because its cheaper to power one box, than 5. You will also need to budget differently for disks, adding 3-5 disks at a time instead of a single disk, but your data will be safer. James Dickens uadmin.blogspot.com This message posted from opensolaris.org
_______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
_______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss