Re: [zfs-discuss] Another user looses his pool (10TB) in this case and 40 days work

2009-07-25 Thread David Magda
On Jul 25, 2009, at 15:32, Frank Middleton wrote: Can you comment on if/how mirroring or raidz mitigates this, or tree corruption in general? I have yet to lose a pool even on a machine with fairly pathological problems, but it is mirrored (and copies=2). Presumably at least on of the drives i

Re: [zfs-discuss] Another user looses his pool (10TB) in this case and 40 days work

2009-07-25 Thread David Magda
On Jul 25, 2009, at 16:30, Carson Gaspar wrote: Frank Middleton wrote: Doesn't this mean /any/ hardware might have this problem, albeit with much lower probability? No. You'll lose unwritten data, but won't corrupt the pool, because the on-disk state will be sane, as long as your iSCSI s

[zfs-discuss] Fishworks iSCSI cache enabled...

2009-07-25 Thread Marcelo Leal
Hello all, Somebody using iSCSI cache enable on 7000 series? I'm talking about OpenSolaris (ZFS) as an iSCSI initiator, because i don't know another filesystem that handles disk caches. So, that option was created for ZFS ;-)? Any suggestions on this? Thanks Leal [ http://www.eall.com.br/

[zfs-discuss] zpool is lain to burnination (bwahahahah!)

2009-07-25 Thread Graeme Clark
Hi Folks, I have a small problem. I've disappeared about 5.9TB of data. My host system was (well, still is) connected to this storage via iSCSI and MPXIO, doing round robin of a pair of GigE ports. I'd like to make a quick excuse before we begin here. I was originally doing raidz2 (there are

Re: [zfs-discuss] Another user looses his pool (10TB) in this case and 40 days work

2009-07-25 Thread Toby Thain
On 25-Jul-09, at 3:32 PM, Frank Middleton wrote: On 07/25/09 02:50 PM, David Magda wrote: Yes, it can be affected. If the snapshot's data structure / record is underneath the corrupted data in the tree then it won't be able to be reached. Can you comment on if/how mirroring or raidz mitigat

Re: [zfs-discuss] The importance of ECC RAM for ZFS

2009-07-25 Thread Marc Bevand
dick hoogendijk nagual.nl> writes: > > I live in Holland and it is not easy to find motherboards that (a) > truly support ECC ram and (b) are (Open)Solaris compatible. Virtually all motherboards for AMD processors support ECC RAM because the memory controller is in the CPU and all AMD CPUs supp

Re: [zfs-discuss] Beginner's question on output of zfs list

2009-07-25 Thread Ian Collins
Axelle Apvrille wrote: Hi, I'm pretty sure this is a beginner's question. The output of my zfs list is shown below. I don't understand how two ZFS file systems (opensolaris and opensolaris-1) can be mounted on the same mountpoint. [Hint: they correspond to different BEs]. They may have the

Re: [zfs-discuss] The importance of ECC RAM for ZFS

2009-07-25 Thread Ian Collins
Michael McCandless wrote: Thanks for the numerous responses everyone! Responding to some of the answers...: ZFS has to trust the storage to have committed the data it claims to have committed in the same way it has to trust the integrity of the RAM it uses for checksummed data. I ho

Re: [zfs-discuss] Another user looses his pool (10TB) in this case and 40 days work

2009-07-25 Thread Carson Gaspar
Frank Middleton wrote: Finally, a number of posters blamed VB for ignoring a flush, but according to the evil tuning guide, without any application syncs, ZFS may wait up to 5 seconds before issuing a synch, and there must be all kinds of failure modes even on bare hardware where it never gets a

[zfs-discuss] Beginner's question on output of zfs list

2009-07-25 Thread Axelle Apvrille
Hi, I'm pretty sure this is a beginner's question. The output of my zfs list is shown below. I don't understand how two ZFS file systems (opensolaris and opensolaris-1) can be mounted on the same mountpoint. [Hint: they correspond to different BEs]. And as I probably should erase one of them (th

Re: [zfs-discuss] Another user looses his pool (10TB) in this case and 40 days work

2009-07-25 Thread Frank Middleton
On 07/25/09 02:50 PM, David Magda wrote: Yes, it can be affected. If the snapshot's data structure / record is underneath the corrupted data in the tree then it won't be able to be reached. Can you comment on if/how mirroring or raidz mitigates this, or tree corruption in general? I have yet t

Re: [zfs-discuss] Another user looses his pool (10TB) in this case and 40 days work

2009-07-25 Thread David Magda
On Jul 25, 2009, at 14:17, roland wrote: thanks for the explanation ! one more question: there are situations where the disks doing strange things (like lying) have caused the ZFS data structures to become wonky. The 'broken' data structure will cause all branches underneath it to be lost--a

Re: [zfs-discuss] Another user looses his pool (10TB) in this case and 40 days work

2009-07-25 Thread roland
thanks for the explanation ! one more question: > there are situations where the disks doing strange things >(like lying) have caused the ZFS data structures to become wonky. The >'broken' data structure will cause all branches underneath it to be >lost--and if it's near the top of the tree, it c

Re: [zfs-discuss] Another user looses his pool (10TB) in this case and 40 days work

2009-07-25 Thread David Magda
On Jul 25, 2009, at 12:24, roland wrote: why can i loose a whole 10TB pool including all the snapshots with the logging/transactional nature of zfs? Because ZFS does not (yet) have an (easy) way to go back a previous state. That's what this bug is about: need a way to rollback to an uber

[zfs-discuss] sharenfs question

2009-07-25 Thread dick hoogendijk
# zfs create store/snaps # zfs set sharenfs='rw=arwen,root=arwen' store/snaps # share -...@store/snaps /store/snaps sec=sys,rw=arwen,root=arwen "" arwen# zfs send -Rv rp...@0906 > /net/westmark/store/snaps/rpool.0906 zsh: permission denied: /net/westmark/store/snaps/rpool.0906 *** BOTH sy

Re: [zfs-discuss] Another user looses his pool (10TB) in this case and 40 days work

2009-07-25 Thread roland
>As soon as you have more then one disk in the equation, then it is >vital that the disks commit their data when requested since otherwise >the data on disk will not be in a consistent state. ok, but doesn`t that refer only to the most recent data? why can i loose a whole 10TB pool including all t

Re: [zfs-discuss] Another user looses his pool (10TB) in this case and 40 days work

2009-07-25 Thread Bob Friesenhahn
On Sat, 25 Jul 2009, roland wrote: When that happens, ZFS believes the data is safely written, but a power cut or >crash can cause severe problems with the pool. didn`t i read a million times that zfs ensures an "always consistent state" and is self healing, too? so, if new blocks are alwa

Re: [zfs-discuss] The importance of ECC RAM for ZFS

2009-07-25 Thread Michael McCandless
Thanks for the numerous responses everyone! Responding to some of the answers...: > ZFS has to trust the storage to have committed the data it > claims to have committed in the same way it has to trust the integrity > of the RAM it uses for checksummed data. I hope that's not true. Ie, I can

Re: [zfs-discuss] Another user looses his pool (10TB) in this case and 40 days work

2009-07-25 Thread roland
>Running this kind of setup absolutely can give you NO garanties at all. >Virtualisation, OSOL/zfs on WinXP. It's nice to play with and see it >"working" but would I TRUST precious data to it? No way! why not? if i write some data trough virtualization layer which goes straight trough to raw disk

Re: [zfs-discuss] SSD's and ZFS...

2009-07-25 Thread Tristan Ball
Bob Friesenhahn wrote: On Fri, 24 Jul 2009, Tristan Ball wrote: I've used 8K IO sizes for all the stage one tests - I know I might get it to go faster with a larger size, but I like to know how well systems will do when I treat them badly! The Stage_1_Ops_thru_run is interesting. 2000+ ops/s

Re: [zfs-discuss] SSD's and ZFS...

2009-07-25 Thread Tristan Ball
The 128G Supertalent Ultradrive ME. The larger version of the drive mentioned in the original post. Sorry, should have made that a little clearer. :-) T Kyle McDonald wrote: Tristan Ball wrote: It just so happens I have one of the 128G and two of the 32G versions in my drawer, waiting to go

Re: [zfs-discuss] ZFS Mirror cloning

2009-07-25 Thread Andrew Gabriel
Richard Elling wrote: That is because you had only one other choice: filesystem level copy. With ZFS I believe you will find that snapshots will allow you to have better control over this. The send/receive process is very, very similar to a mirror resilver, so you are only carrying your previous