Re: [zfs-discuss] Re: Drive Failure w/o Redundancy

2007-06-27 Thread Richard Elling
Erik Trimble wrote: If you had known about the drive sizes beforehand, the you could have done something like this: Partition the drives as follows: A: 1 20GB partition B: 1 20gb & 1 10GB partition C: 1 40GB partition D: 1 40GB partition & 2 10GB paritions then you do: zpool create tank m

Re: [zfs-discuss] Suggestions on 30 drive configuration?

2007-06-27 Thread Boyd Adamson
On 28/06/2007, at 12:29 AM, Victor Latushkin wrote: It is not so easy to predict. ZFS will coalesce writes. A single transaction group may have many different writes in it. Also, raidz[12] is dynamic, and will use what it needs, unlike separate volume managers who do not have any underst

Re: [zfs-discuss] Re: Drive Failure w/o Redundancy

2007-06-27 Thread Erik Trimble
On Wed, 2007-06-27 at 12:03 -0700, Jef Pearlman wrote: > > Jef Pearlman wrote: > > > Absent that, I was considering using zfs and just > > > having a single pool. My main question is this: what > > > is the failure mode of zfs if one of those drives > > > either fails completely or has errors? Do I

Re: [zfs-discuss] Re: Drive Failure w/o Redundancy

2007-06-27 Thread Richard Elling
Jef Pearlman wrote: Perhaps I'm not asking my question clearly. I've already experimented a fair amount with zfs, including creating and destroying a number of pools with and without redundancy, replacing vdevs, etc. Maybe asking by example will clarify what I'm looking for or where I've missed

Re: [zfs-discuss] Re: Drive Failure w/o Redundancy

2007-06-27 Thread Erik Trimble
On Wed, 2007-06-27 at 14:50 -0700, Darren Dunham wrote: > > Darren Dunham wrote: > > >> The problem I've come across with using mirror or raidz for this setup > > >> is that (as far as I know) you can't add disks to mirror/raidz groups, > > >> and if you just add the disk to the pool, you end up in

Re: [zfs-discuss] Re: Drive Failure w/o Redundancy

2007-06-27 Thread Darren Dunham
> Darren Dunham wrote: > >> The problem I've come across with using mirror or raidz for this setup > >> is that (as far as I know) you can't add disks to mirror/raidz groups, > >> and if you just add the disk to the pool, you end up in the same > >> situation as above (with more space but no redund

[zfs-discuss] ReiserFS4 like metadata/search

2007-06-27 Thread Oliver Schinagl
The only thing I haven't found in zfs yet, is metadata etc info. The previous 'next best thing' in FS was of course ReiserFS (4). Reiser3 was quite a nice thing, fast, journaled and all that, but Reiser4 promised to bring all those things that we see emerging now, like cross FS search, any documen

Re: [zfs-discuss] Re: Drive Failure w/o Redundancy

2007-06-27 Thread Neil Perrin
Darren Dunham wrote: The problem I've come across with using mirror or raidz for this setup is that (as far as I know) you can't add disks to mirror/raidz groups, and if you just add the disk to the pool, you end up in the same situation as above (with more space but no redundancy). You can't

[zfs-discuss] Re: [zfs-code] Space allocation failure

2007-06-27 Thread Manoj Joseph
Hi, In brief, what I am trying to do is to use libzpool to access a zpool - like ztest does. Matthew Ahrens wrote: Manoj Joseph wrote: Hi, Replying to myself again. :) I see this problem only if I attempt to use a zpool that already exists. If I create one (using files instead of devices,

Re: [zfs-discuss] Re: Drive Failure w/o Redundancy

2007-06-27 Thread Darren Dunham
> Perhaps I'm not asking my question clearly. I've already experimented > a fair amount with zfs, including creating and destroying a number of > pools with and without redundancy, replacing vdevs, etc. Maybe asking > by example will clarify what I'm looking for or where I've missed the > boat. The

[zfs-discuss] Re: Drive Failure w/o Redundancy

2007-06-27 Thread Jef Pearlman
> Jef Pearlman wrote: > > Absent that, I was considering using zfs and just > > having a single pool. My main question is this: what > > is the failure mode of zfs if one of those drives > > either fails completely or has errors? Do I > > permanently lose access to the entire pool? Can I > > attemp

Re: [zfs-discuss] Re: ZFS usb keys

2007-06-27 Thread Matthew Ahrens
William D. Hathaway wrote: It would be really handy if whoever was responsible for the message at: http://www.sun.com/msg/ZFS-8000-A5 could add data about which zpool versions are supported at specific OS/patch releases. Did you look at http://www.opensolaris.org/os/community/zfs/version/N W

Re: [zfs-discuss] Re: zfs and 2530 jbod

2007-06-27 Thread Frank Cusack
On June 26, 2007 2:13:54 PM -0700 Joel Miller <[EMAIL PROTECTED]> wrote: The 2500 series engineering team is talking with the ZFS folks to understand the various aspects of delivering a complete solution. (There is a lot more to it than "it seems to work"...). Great news, you made my day! Any

Re: [zfs-discuss] Re: Suggestions on 30 drive configuration?

2007-06-27 Thread Dan Saul
I have 8 SATA on the motherboard, 4 PCI cards with 4 SATA each, one PCIe 4x sata card with two, and one PCIe 1x with two. The operating system itself will be on a hard drive attached to one ATA 100 connector. Kind of like a "poor man's" data centre, except not that cheap... It still is estimated

Re: [zfs-discuss] New german white paper on ZFS

2007-06-27 Thread Cindy . Swearingen
Jens, Someone already added it to the ZFS links page, here: http://opensolaris.org/os/community/zfs/links/ I just added a link to the links page from the zfs docs page so it is easier to find. Thanks, Cindy Jens Elkner wrote: On Tue, Jun 19, 2007 at 05:19:05PM +0200, Constantin Gonzalez wro

Re: [zfs-discuss] Re: ZFS usb keys

2007-06-27 Thread Mike Lee
I had a similar situation between x86 and SPARC, version number. When I created the pool on the LOWER rev machine, it was seen by the HIGHER rev machine. This was a USB HDD, not a stick. I can now move the drive between boxes. HTH, Mike Dick Davies wrote: Thanks to everyone for the sanity ch

Re: [zfs-discuss] Re: ZFS usb keys

2007-06-27 Thread Dick Davies
Thanks to everyone for the sanity check - I think it's a platform issue, but not an endian one. The stick was originally DOS-formatted, and the zpool was built on the first fdisk partition. So Sparcs aren't seeing it, but the x86/x64 boxes are. -- Rasputin :: Jack of All Trades - Master of Nuns

Re: [zfs-discuss] Suggestions on 30 drive configuration?

2007-06-27 Thread Victor Latushkin
Richard Elling wrote: Rob Logan wrote: > an array of 30 drives in a RaidZ2 configuration with two hot spares > I don't want to mirror 15 drives to 15 drives ok, so space over speed... and are willing to toss somewhere between 4 and 15 drives for protection. raidz splits the (up to 128k) writ

Re: [zfs-discuss] New german white paper on ZFS

2007-06-27 Thread Jens Elkner
On Tue, Jun 19, 2007 at 05:19:05PM +0200, Constantin Gonzalez wrote: Hi, > http://blogs.sun.com/constantin/entry/new_zfs_white_paper_in Excellent!!! I think it is a pretty good idea, to put the links for the paper and slides on the ZFS Documentation page aka http://www.opensolaris.org/os/commu

Re: [zfs-discuss] Re: ZFS usb keys

2007-06-27 Thread Mark J Musante
On Wed, 27 Jun 2007, [UTF-8] Jürgen Keil wrote: > Yep, I just tried it, and it refuses to "zpool import" the newer pool, > telling me about the incompatible version. So I guess the pool format > isn't the correct explanation for the Dick Davies' (number9) problem. Have you tried creating the poo

Re: [zfs-discuss] Wishlist items

2007-06-27 Thread Mark J Musante
On Tue, 26 Jun 2007 [EMAIL PROTECTED] wrote: > In particular zfs sometimes takes a while to return from zfs snapshot -r > tank/[EMAIL PROTECTED] in the case where there are a great many iscsi shared > volumes underneath. A little progress feedback would go a long way. This would certainly be nice

Re: [zfs-discuss] Re: Re[2]: Re: Re: Re: Snapshots impact on performance

2007-06-27 Thread Victor Latushkin
Gino wrote: Same problem here (snv_60). Robert, did you find any solutions? Couple of week ago I put together an implementation of space maps which completely eliminates loops and recursion from space map alloc operation, and allows to implement different allocation strategies quite easily (

[zfs-discuss] Re: ZFS usb keys

2007-06-27 Thread William D. Hathaway
It would be really handy if whoever was responsible for the message at: http://www.sun.com/msg/ZFS-8000-A5 could add data about which zpool versions are supported at specific OS/patch releases. The current message doesn't help the user figure out how to accomplish their implied task, which is t

[zfs-discuss] Re: ZFS usb keys

2007-06-27 Thread Jürgen Keil
> Shouldn't S10u3 just see the newer on-disk format and > report that fact, rather than complain it is corrupt? Yep, I just tried it, and it refuses to "zpool import" the newer pool, telling me about the incompatible version. So I guess the pool format isn't the correct explanation for the Dick D

[zfs-discuss] Re: Re[2]: Re: Re: Re: Snapshots impact on performance

2007-06-27 Thread Gino
Same problem here (snv_60). Robert, did you find any solutions? gino This message posted from opensolaris.org ___ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

[zfs-discuss] Re: Re: ZFS - SAN and Raid

2007-06-27 Thread Richard L. Hamilton
> Victor Engle wrote: > > Roshan, > > > > As far as I know, there is no problem at all with > using SAN storage > > with ZFS and it does look like you were having an > underlying problem > > with either powerpath or the array. > > Correct. A write failed. > > > The best practices guide on opens