Bug ID is 6793967.

This problem just happened again.
% zpool status pool1
  pool: pool1
 state: DEGRADED
 scrub: resilver completed after 0h48m with 0 errors on Mon Jan  5 12:30:52 2009
config:

        NAME           STATE     READ WRITE CKSUM
        pool1          DEGRADED     0     0     0
          raidz2       DEGRADED     0     0     0
            c4t8d0s0   ONLINE       0     0     0
            c4t9d0s0   ONLINE       0     0     0
            c4t10d0s0  ONLINE       0     0     0
            c4t11d0s0  ONLINE       0     0     0
            c4t12d0s0  REMOVED      0     0     0
            c4t13d0s0  ONLINE       0     0     0

errors: No known data errors

% zpool status -x
all pools are healthy
%
# zpool online pool1 c4t12d0s0
% zpool status -x
  pool: pool1
 state: ONLINE
status: One or more devices is currently being resilvered.  The pool will
        continue to function, possibly in a degraded state.
action: Wait for the resilver to complete.
 scrub: resilver in progress for 0h0m, 0.12% done, 2h38m to go
config:

        NAME           STATE     READ WRITE CKSUM
        pool1          ONLINE       0     0     0
          raidz2       ONLINE       0     0     0
            c4t8d0s0   ONLINE       0     0     0
            c4t9d0s0   ONLINE       0     0     0
            c4t10d0s0  ONLINE       0     0     0
            c4t11d0s0  ONLINE       0     0     0
            c4t12d0s0  ONLINE       0     0     0
            c4t13d0s0  ONLINE       0     0     0

errors: No known data errors
%

Ben

> I just put in a (low priority) bug report on this.
> 
> Ben
> 
> > This post from close to a year ago never received
> a
> > response.  We just had this same thing happen to
> > another server that is running Solaris 10 U6.  One
> of
> > the disks was marked as removed and the pool
> > degraded, but 'zpool status -x' says all pools are
> > healthy.  After doing an 'zpool online' on the
> disk
> > it resilvered in fine.  Any ideas why 'zpool
> status
> > -x' reports all healthy while 'zpool status' shows
> a
> > pool in degraded mode?
> > 
> > thanks,
> > Ben
> > 
> > > We run a cron job that does a 'zpool status -x'
> to
> > > check for any degraded pools.  We just happened
> to
> > > find a pool degraded this morning by running
> > 'zpool
> > > status' by hand and were surprised that it was
> > > degraded as we didn't get a notice from the cron
> > > job.
> > > 
> > > # uname -srvp
> > > SunOS 5.11 snv_78 i386
> > > 
> > > # zpool status -x
> > > all pools are healthy
> > > 
> > > # zpool status pool1
> > >   pool: pool1
> > > tate: DEGRADED
> > >  scrub: none requested
> > > onfig:
> > > 
> > >         NAME         STATE     READ WRITE CKSUM
> > > pool1        DEGRADED     0     0     0
> > >           raidz1     DEGRADED     0     0     0
> > >   c1t8d0   REMOVED      0     0     0
> > >           c1t9d0   ONLINE       0     0     0
> > >   c1t10d0  ONLINE       0     0     0
> > >           c1t11d0  ONLINE       0     0     0
> > > No known data errors
> > > 
> > > I'm going to look into it now why the disk is
> > listed
> > > as removed.
> > > 
> > > Does this look like a bug with 'zpool status
> -x'?
> > > 
> > > Ben
-- 
This message posted from opensolaris.org
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to