Thanks Cindy. I'm running 111b at the moment. I ran a scrub last
night, and it still reports the same status.

r...@weyl:~# uname -a
SunOS weyl 5.11 snv_111b i86pc i386 i86pc Solaris
r...@weyl:~# zpool status -x
  pool: tank
 state: DEGRADED
status: One or more devices could not be opened.  Sufficient replicas exist for
        the pool to continue functioning in a degraded state.
action: Attach the missing device and online it using 'zpool online'.
   see: http://www.sun.com/msg/ZFS-8000-2Q
 scrub: scrub completed after 2h40m with 0 errors on Mon Jun 28 01:23:12 2010
config:

        NAME                       STATE     READ WRITE CKSUM
        tank                       DEGRADED     0     0     0
          mirror                   DEGRADED     0     0     0
            spare                  DEGRADED 1.37M     0     0
              9828443264686839751  UNAVAIL      0     0     0  was
/dev/dsk/c6t1d0s0
              c7t1d0               DEGRADED     0     0 1.37M  too many errors
            c9t0d0                 ONLINE       0     0     0
          mirror                   ONLINE       0     0     0
            c7t0d0                 ONLINE       0     0     0
            c5t1d0                 ONLINE       0     0     0
        spares
          c7t1d0                   INUSE     currently in use

errors: No known data errors
r...@weyl:~#



On Mon, Jun 28, 2010 at 14:55, Cindy Swearingen
<cindy.swearin...@oracle.com> wrote:
> Hi Donald,
>
> I think this is just a reporting error in the zpool status output,
> depending on what Solaris release is.
>
> Thanks,
>
> Cindy
>
> On 06/27/10 15:13, Donald Murray, P.Eng. wrote:
>>
>> Hi,
>>
>> I awoke this morning to a panic'd opensolaris zfs box. I rebooted it
>> and confirmed it would panic each time it tried to import the 'tank'
>> pool. Once I disconnected half of one of the mirrored disks, the box
>> booted cleanly and the pool imported without a panic.
>>
>> Because this box has a hot spare, it began resilvering automatically.
>> This is the first time I've resilvered to a hot spare, so I'm not sure
>> whether the output below [1]  is normal.
>>
>> In particular, I think it's odd that the spare has an equal number of
>> read and cksum errors. Is this normal? Is my spare a piece of junk,
>> just like the disk it replaced?
>>
>>
>> [1]
>> r...@weyl:~# zpool status tank
>>  pool: tank
>>  state: DEGRADED
>> status: One or more devices could not be opened.  Sufficient replicas
>> exist for
>>        the pool to continue functioning in a degraded state.
>> action: Attach the missing device and online it using 'zpool online'.
>>   see: http://www.sun.com/msg/ZFS-8000-2Q
>>  scrub: resilver in progress for 3h42m, 97.34% done, 0h6m to go
>> config:
>>
>>        NAME                       STATE     READ WRITE CKSUM
>>        tank                       DEGRADED     0     0     0
>>          mirror                   DEGRADED     0     0     0
>>            spare                  DEGRADED 1.36M     0     0
>>              9828443264686839751  UNAVAIL      0     0     0  was
>> /dev/dsk/c6t1d0s0
>>              c7t1d0               DEGRADED     0     0 1.36M  too many
>> errors
>>            c9t0d0                 ONLINE       0     0     0
>>          mirror                   ONLINE       0     0     0
>>            c7t0d0                 ONLINE       0     0     0
>>            c5t1d0                 ONLINE       0     0     0
>>        spares
>>          c7t1d0                   INUSE     currently in use
>>
>> errors: No known data errors
>> r...@weyl:~#
>> _______________________________________________
>> zfs-discuss mailing list
>> zfs-discuss@opensolaris.org
>> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to