On Jun 16, 2012, at 9:49 AM, Scott Aitken wrote:

> On Sat, Jun 16, 2012 at 09:09:53AM -0500, Gregg Wonderly wrote:
>> Use 'dd' to replicate as much of lofi/2 as you can onto another device, and 
>> then 
>> cable that into place?
>> 
>> It looks like you just need to put a functioning, working, but not correct 
>> device, in that slot so that it will import and then you can 'zpool replace' 
>> the 
>> new disk into the pool perhaps?
>> 
>> Gregg Wonderly
>> 
>> On 6/16/2012 2:02 AM, Scott Aitken wrote:
>>> On Sat, Jun 16, 2012 at 08:54:05AM +0200, Stefan Ring wrote:
>>>>> when you say remove the device, I assume you mean simply make it 
>>>>> unavailable
>>>>> for import (I can't remove it from the vdev).
>>>> Yes, that's what I meant.
>>>> 
>>>>> root@openindiana-01:/mnt# zpool import -d /dev/lofi
>>>>> ??pool: ZP-8T-RZ1-01
>>>>> ?? ??id: 9952605666247778346
>>>>> ??state: FAULTED
>>>>> status: One or more devices are missing from the system.
>>>>> action: The pool cannot be imported. Attach the missing
>>>>> ?? ?? ?? ??devices and try again.
>>>>> ?? see: http://www.sun.com/msg/ZFS-8000-3C
>>>>> config:
>>>>> 
>>>>> ?? ?? ?? ??ZP-8T-RZ1-01 ?? ?? ?? ?? ?? ?? ??FAULTED ??corrupted data
>>>>> ?? ?? ?? ?? ??raidz1-0 ?? ?? ?? ?? ?? ?? ?? ??DEGRADED
>>>>> ?? ?? ?? ?? ?? ??12339070507640025002 ??UNAVAIL ??cannot open
>>>>> ?? ?? ?? ?? ?? ??/dev/lofi/5 ?? ?? ?? ?? ?? ONLINE
>>>>> ?? ?? ?? ?? ?? ??/dev/lofi/4 ?? ?? ?? ?? ?? ONLINE
>>>>> ?? ?? ?? ?? ?? ??/dev/lofi/3 ?? ?? ?? ?? ?? ONLINE
>>>>> ?? ?? ?? ?? ?? ??/dev/lofi/1 ?? ?? ?? ?? ?? ONLINE
>>>>> 
>>>>> It's interesting that even though 4 of the 5 disks are available, it still
>>>>> can import it as DEGRADED.
>>>> I agree that it's "interesting". Now someone really knowledgable will
>>>> need to have a look at this. I can only imagine that somehow the
>>>> devices contain data from different points in time, and that it's too
>>>> far apart for the aggressive txg rollback that was added in PSARC
>>>> 2009/479. Btw, did you try that? Try: zpool import -d /dev/lofi -FVX
>>>> ZP-8T-RZ1-01.
>>>> 
>>> Hi again,
>>> 
>>> that got slightly further, but still no dice:
>>> 
>>> root@openindiana-01:/mnt#  zpool import -d /dev/lofi -FVX ZP-8T-RZ1-01
>>> root@openindiana-01:/mnt# zpool list
>>> NAME           SIZE  ALLOC   FREE    CAP  DEDUP  HEALTH  ALTROOT
>>> ZP-8T-RZ1-01      -      -      -      -      -  FAULTED  -
>>> rpool         15.9G  2.17G  13.7G    13%  1.00x  ONLINE  -
>>> root@openindiana-01:/mnt# zpool status
>>>   pool: ZP-8T-RZ1-01
>>>  state: FAULTED
>>> status: One or more devices could not be used because the label is missing
>>>         or invalid.  There are insufficient replicas for the pool to 
>>> continue
>>>         functioning.
>>> action: Destroy and re-create the pool from
>>>         a backup source.
>>>    see: http://www.sun.com/msg/ZFS-8000-5E
>>>   scan: none requested
>>> config:
>>> 
>>>         NAME                      STATE     READ WRITE CKSUM
>>>         ZP-8T-RZ1-01              FAULTED      0     0     1  corrupted data
>>>           raidz1-0                ONLINE       0     0     6
>>>             12339070507640025002  UNAVAIL      0     0     0  was 
>>> /dev/lofi/2
>>>             /dev/lofi/5           ONLINE       0     0     0
>>>             /dev/lofi/4           ONLINE       0     0     0
>>>             /dev/lofi/3           ONLINE       0     0     0
>>>             /dev/lofi/1           ONLINE       0     0     0
>>> 
>>> root@openindiana-01:/mnt# zpool scrub ZP-8T-RZ1-01
>>> cannot scrub 'ZP-8T-RZ1-01': pool is currently unavailable
>>> 
>>> Thanks for your tenacity Stefan.
>>> Scott
>>> _______________________________________________
>>> zfs-discuss mailing list
>>> zfs-discuss@opensolaris.org
>>> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>>> 
>> 
>> 
> 
> Hi Greg,
> 
> lofi/2 is a dd of a real disk.  I am using disk images because I can roll
> back, clone etc without using the original drives (which are long gone
> anyway).
> 
> I have tried making /2 unavailable for import, and zfs just moans that it
> can't be opened.  It fails to import even though I have only one disk missing
> of a RAIDZ array.

My experience is that ZFS will not import a pool with a missing disk.  There 
has to be something in that slot before the import will occur.  Even if the 
disk is "corrupt", it needs to be there.  I think this is a failsafe mechanism 
that tries to keep a pool from going live when you have mistakenly not 
connected all the drives.  That keeps the disks from becoming 
chronologically/txn misaligned which can result in data loss, in the right 
combinations I believe.

Gregg Wonderly
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to