On Sep 12, 2010, at 7:49 PM, Michael Eskowitz wrote:

> I recently lost all of the data on my single parity raid z array.  Each of 
> the drives was encrypted with the zfs array built within the encrypted 
> volumes.
> 
> I am not exactly sure what happened.  

Murphy strikes again!

> The files were there and accessible and then they were all gone.  The server 
> apparently crashed and rebooted and everything was lost.  After the crash I 
> remounted the encrypted drives and the zpool was still reporting that roughly 
> 3TB of the 7TB array were used, but I could not see any of the files through 
> the array's mount point.  I unmounted the zpool and then remounted it and 
> suddenly zpool was reporting 0TB were used.  

Were you using zfs send/receive?  If so, then this is the behaviour expected 
when a
session is interrupted. Since the snapshot did not completely arrive at the 
receiver, the
changes are rolled back.  It can take a few minutes for terabytes to be freed.

> I did not remap the virtual device.  The only thing of note that I saw was 
> that the name of storage pool had changed.  Originally it was "Movies" and 
> then it became "Movita".  I am guessing that the file system became corrupted 
> some how.  (zpool status did not report any errors)
> 
> So, my questions are these... 
> 
> Is there anyway to undelete data from a lost raidz array?

It depends entirely on the nature of the loss.  In the case I describe above, 
there is nothing
lost because nothing was there (!)

>  If I build a new virtual device on top of the old one and the drive topology 
> remains the same, can we scan the drives for files from old arrays?

The short answer is no.

> Also, is there any way to repair a corrupted storage pool?

Yes, but it depends entirely on the nature of the corruption.

>  Is it possible to backup the file table or whatever partition index zfs 
> maintains?

The ZFS configuration data is stored redundantly in the pool and checksummed.

> I imagine that you all are going to suggest that I scrub the array, but that 
> is not an option at this point.  I had a backup of all of the data lost as I 
> am moving between file servers so at a certain point I gave up and decided to 
> start fresh.  This doesn't give me a warm fuzzy feeling about zfs, though.

AFAICT, ZFS appears to be working as designed.  Are you trying to kill the 
canary? :-)
 -- richard

-- 
OpenStorage Summit, October 25-27, Palo Alto, CA
http://nexenta-summit2010.eventbrite.com

Richard Elling
rich...@nexenta.com   +1-760-896-4422
Enterprise class storage for everyone
www.nexenta.com





_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to