On Sat, Jun 16, 2012 at 09:58:40AM -0500, Gregg Wonderly wrote:
> 
> On Jun 16, 2012, at 9:49 AM, Scott Aitken wrote:
> 
> > On Sat, Jun 16, 2012 at 09:09:53AM -0500, Gregg Wonderly wrote:
> >> Use 'dd' to replicate as much of lofi/2 as you can onto another device, 
> >> and then 
> >> cable that into place?
> >> 
> >> It looks like you just need to put a functioning, working, but not correct 
> >> device, in that slot so that it will import and then you can 'zpool 
> >> replace' the 
> >> new disk into the pool perhaps?
> >> 
> >> Gregg Wonderly
> >> 
> >> On 6/16/2012 2:02 AM, Scott Aitken wrote:
> >>> On Sat, Jun 16, 2012 at 08:54:05AM +0200, Stefan Ring wrote:
> >>>>> when you say remove the device, I assume you mean simply make it 
> >>>>> unavailable
> >>>>> for import (I can't remove it from the vdev).
> >>>> Yes, that's what I meant.
> >>>> 
> >>>>> root@openindiana-01:/mnt# zpool import -d /dev/lofi
> >>>>> ??pool: ZP-8T-RZ1-01
> >>>>> ?? ??id: 9952605666247778346
> >>>>> ??state: FAULTED
> >>>>> status: One or more devices are missing from the system.
> >>>>> action: The pool cannot be imported. Attach the missing
> >>>>> ?? ?? ?? ??devices and try again.
> >>>>> ?? see: http://www.sun.com/msg/ZFS-8000-3C
> >>>>> config:
> >>>>> 
> >>>>> ?? ?? ?? ??ZP-8T-RZ1-01 ?? ?? ?? ?? ?? ?? ??FAULTED ??corrupted data
> >>>>> ?? ?? ?? ?? ??raidz1-0 ?? ?? ?? ?? ?? ?? ?? ??DEGRADED
> >>>>> ?? ?? ?? ?? ?? ??12339070507640025002 ??UNAVAIL ??cannot open
> >>>>> ?? ?? ?? ?? ?? ??/dev/lofi/5 ?? ?? ?? ?? ?? ONLINE
> >>>>> ?? ?? ?? ?? ?? ??/dev/lofi/4 ?? ?? ?? ?? ?? ONLINE
> >>>>> ?? ?? ?? ?? ?? ??/dev/lofi/3 ?? ?? ?? ?? ?? ONLINE
> >>>>> ?? ?? ?? ?? ?? ??/dev/lofi/1 ?? ?? ?? ?? ?? ONLINE
> >>>>> 
> >>>>> It's interesting that even though 4 of the 5 disks are available, it 
> >>>>> still
> >>>>> can import it as DEGRADED.
> >>>> I agree that it's "interesting". Now someone really knowledgable will
> >>>> need to have a look at this. I can only imagine that somehow the
> >>>> devices contain data from different points in time, and that it's too
> >>>> far apart for the aggressive txg rollback that was added in PSARC
> >>>> 2009/479. Btw, did you try that? Try: zpool import -d /dev/lofi -FVX
> >>>> ZP-8T-RZ1-01.
> >>>> 
> >>> Hi again,
> >>> 
> >>> that got slightly further, but still no dice:
> >>> 
> >>> root@openindiana-01:/mnt#  zpool import -d /dev/lofi -FVX ZP-8T-RZ1-01
> >>> root@openindiana-01:/mnt# zpool list
> >>> NAME           SIZE  ALLOC   FREE    CAP  DEDUP  HEALTH  ALTROOT
> >>> ZP-8T-RZ1-01      -      -      -      -      -  FAULTED  -
> >>> rpool         15.9G  2.17G  13.7G    13%  1.00x  ONLINE  -
> >>> root@openindiana-01:/mnt# zpool status
> >>>   pool: ZP-8T-RZ1-01
> >>>  state: FAULTED
> >>> status: One or more devices could not be used because the label is missing
> >>>         or invalid.  There are insufficient replicas for the pool to 
> >>> continue
> >>>         functioning.
> >>> action: Destroy and re-create the pool from
> >>>         a backup source.
> >>>    see: http://www.sun.com/msg/ZFS-8000-5E
> >>>   scan: none requested
> >>> config:
> >>> 
> >>>         NAME                      STATE     READ WRITE CKSUM
> >>>         ZP-8T-RZ1-01              FAULTED      0     0     1  corrupted 
> >>> data
> >>>           raidz1-0                ONLINE       0     0     6
> >>>             12339070507640025002  UNAVAIL      0     0     0  was 
> >>> /dev/lofi/2
> >>>             /dev/lofi/5           ONLINE       0     0     0
> >>>             /dev/lofi/4           ONLINE       0     0     0
> >>>             /dev/lofi/3           ONLINE       0     0     0
> >>>             /dev/lofi/1           ONLINE       0     0     0
> >>> 
> >>> root@openindiana-01:/mnt# zpool scrub ZP-8T-RZ1-01
> >>> cannot scrub 'ZP-8T-RZ1-01': pool is currently unavailable
> >>> 
> >>> Thanks for your tenacity Stefan.
> >>> Scott
> >>> _______________________________________________
> >>> zfs-discuss mailing list
> >>> zfs-discuss@opensolaris.org
> >>> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
> >>> 
> >> 
> >> 
> > 
> > Hi Greg,
> > 
> > lofi/2 is a dd of a real disk.  I am using disk images because I can roll
> > back, clone etc without using the original drives (which are long gone
> > anyway).
> > 
> > I have tried making /2 unavailable for import, and zfs just moans that it
> > can't be opened.  It fails to import even though I have only one disk 
> > missing
> > of a RAIDZ array.
> 
> My experience is that ZFS will not import a pool with a missing disk.  There 
> has to be something in that slot before the import will occur.  Even if the 
> disk is "corrupt", it needs to be there.  I think this is a failsafe 
> mechanism that tries to keep a pool from going live when you have mistakenly 
> not connected all the drives.  That keeps the disks from becoming 
> chronologically/txn misaligned which can result in data loss, in the right 
> combinations I believe.
> 
> Gregg Wonderly
> 

Hi again Gregg,

not sure if I should be top posting this...

Given I am working with images, it's hard to put just anything "in place" of
lofi/2.  ZFS scans all of the files in the directory for ZFS labels, so just
replacing lofi/2 with an empty file (for example) just means ZFS skips it,
which is the same result as deleting lofi/2 altogether.  I did this, but to
no avail.  ZFS complains about having insufficient replicas.

There is something more going on here which is beyond my capabilities.  Given
the history I outlined earlier, there is almost certainly one or two corrupt
labels on those disks (images).

Scott
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to