Karthik Krishnamoorthy wrote:
> We did try with this
>
> zpool set failmode=continue <pool> option
>
> and the wait option before pulling running the cp command and pulling 
> out the mirrors and in both cases there was a hang and I have a core 
> dump of the hang as well.
>   

You have to wait for the I/O drivers to declare that the device is
dead.  This can be up to several minutes, depending on the driver.

> Any pointers to the bug opening process ?
>   

http://bugs.opensolaris.org, or bugster if you have an account.
Be sure to indicate which drivers you are using, as this is not likely
a ZFS bug, per se.  Output from prtconf -D should be a minimum.
 -- richard

> Thanks
> Karthik
>
> On 10/15/08 22:27, Neil Perrin wrote:
>   
>> On 10/15/08 23:12, Karthik Krishnamoorthy wrote:
>>     
>>> Neil,
>>>
>>> Thanks for the quick suggestion, the hang seems to happen even with 
>>> the zpool set failmode=continue <pool> option.
>>>
>>> Any other way to recover from the hang ?
>>>       
>> You should set the property before you remove the devices.
>> This should prevent the hang. It isn't used to recover from it.
>>
>> If you did do that then it seems like a bug somewhere in ZFS or the IO 
>> stack
>> below it. In which case you should file a bug.
>>
>> Neil.
>>     
>>> thanks and regards,
>>> Karthik
>>>
>>> On 10/15/08 22:03, Neil Perrin wrote:
>>>       
>>>> Karthik,
>>>>
>>>> The pool failmode property as implemented governs the behaviour when 
>>>> all
>>>> the devices needed are unavailable. The default behaviour is to wait
>>>> (block) until the IO can continue - perhaps by re-enabling the 
>>>> device(s).
>>>> The behaviour you expected can be achieved by "zpool set 
>>>> failmode=continue <pool>",
>>>> as shown in the link you indicated below.
>>>>
>>>> Neil.
>>>>
>>>> On 10/15/08 22:38, Karthik Krishnamoorthy wrote:
>>>>         
>>>>> Hello All,
>>>>>
>>>>>   Summary:
>>>>>   ~~~~~~~~
>>>>>   cp command for mirrored zfs hung when all the disks in the mirrored
>>>>>   pool were unavailable.
>>>>>     Detailed description:
>>>>>   ~~~~~~~~~~~~~~~~~~~~~
>>>>>     The cp command (copy a 1GB file from nfs to zfs) hung when all 
>>>>> the disks
>>>>>   in the mirrored pool (both c1t0d9 and c2t0d9) were removed 
>>>>> physically.
>>>>>            NAME        STATE     READ WRITE CKSUM
>>>>>          test        ONLINE      0     0     0
>>>>>            mirror    ONLINE      0     0     0
>>>>>              c1t0d9  ONLINE      0     0     0
>>>>>              c2t0d9  ONLINE      0     0     0
>>>>>     We think if all the disks in the pool are unavailable, cp 
>>>>> command should
>>>>>   fail with error (not cause hang).
>>>>>     Our request:
>>>>>   ~~~~~~~~~~~~
>>>>>   Please investigate the root cause of this issue.
>>>>>  
>>>>>   How to reproduce:
>>>>>   ~~~~~~~~~~~~~~~~~
>>>>>   1. create a zfs mirrored pool
>>>>>   2. execute cp command from somewhere to the zfs mirrored pool.
>>>>>   3. remove the both of disks physically during cp command working
>>>>>     =  hang happen (cp command never return and we can't kill cp 
>>>>> command)
>>>>>
>>>>> One engineer pointed me to this page  
>>>>> http://opensolaris.org/os/community/arc/caselog/2007/567/onepager/ 
>>>>> and indicated that if all the mirrors are removed zfs enters a hang 
>>>>> like state to prevent the kernel from going into a panic mode and 
>>>>> this type of feature would be an RFE.
>>>>>
>>>>> My questions are
>>>>>
>>>>> Are there any documentation of the "mirror" configuration of zfs 
>>>>> that explains what happens when the underlying
>>>>> drivers detect problems in one of the mirror devices?
>>>>>
>>>>> It seems that the traditional views of "mirror" or "raid-2" would 
>>>>> expect that the
>>>>> mirror would be able to proceed without interruption and that does 
>>>>> not seem to be this case in ZFS.
>>>>> What is the purpose of the mirror, in zfs?  Is it more like an instant
>>>>> backup?  If so, what can the user do to recover, when there is an
>>>>> IO error on one of the devices?
>>>>>
>>>>>
>>>>> Appreciate any pointers and help,
>>>>>
>>>>> Thanks and regards,
>>>>> Karthik
>>>>> _______________________________________________
>>>>> zfs-discuss mailing list
>>>>> zfs-discuss@opensolaris.org
>>>>> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>>>>>           
>
> _______________________________________________
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>   

_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to