I woke up this morning and so you're messages, unfortunately I had to
reboot, the server completely froze.
Now I have that :
pool: nas
state: DEGRADED
status: One or more devices is currently being resilvered. The pool will
continue to function, possibly in a degraded state.
action: Wait for the resilver to complete.
scan: resilver in progress since Wed Oct 23 08:19:42 2013
5.81G scanned out of 22.2T at 49.2M/s, 131h43m to go
15.6M resilvered, 0.03% done
config:
NAME STATE READ WRITE CKSUM
nas DEGRADED 0 0 0
raidz1-0 DEGRADED 0 0 0
c8t50024E9004993E6Ed0p0 ONLINE 0 0 0
c8t50024E92062E7524d0 ONLINE 0 0 0
c8t50024E900495BE84d0p0 ONLINE 0 0 0
c8t50014EE25A5EEC23d0p0 ONLINE 0 0 0
c8t50024E9003F03980d0p0 ONLINE 0 0 0
c8t50014EE2B0D3EFC8d0 ONLINE 0 0 0
c8t50014EE6561DDB4Cd0p0 DEGRADED 0 0 0 too many errors
c8t50024E9003F03A09d0p0 ONLINE 0 0 0
raidz1-1 ONLINE 0 0 0
c50t8d0 ONLINE 0 0 0 (resilvering)
c2d0 ONLINE 0 0 0 (resilvering)
c1d0 ONLINE 0 0 0 (resilvering)
c50t11d0 ONLINE 0 0 0
c50t10d0 ONLINE 0 0 0 (resilvering)
Le 23/10/2013 08:00, Jason Matthews a écrit :
first, dont reboot. if you do you might not be able remount the pool. the data
you see is from the disks that are functioning. listing the files and copying
complete files are two different things. if you dont have a backup you may need
to copy whatever partial data you can from the broken pool.
now let's start by getting the disks back in good shape.
clear the degraded disk
zpool clear c8t50014EE6561DDB4Cd0p0
reseat the missing disks in the hopes they come back then clear them
check cfgadm -al and make sure they are connected and configured
when you reseat them check the messages (or dmesg) to see if the system notices
the re-insertion. if it does see the disk installed clear the disks in the pool
in effort to bring the pool back to an operational state.
Sent from Jasons' hand held
On Oct 22, 2013, at 5:04 PM, Clement BRIZARD <[email protected]> wrote:
Hello everybody,
I have a problem with my pool, I had some slowdowns lately on my nfs share of
my zfs pool. A weekly scrub began and is still running but it worries me, it
currently returne that
pool: nas
state: UNAVAIL
status: One or more devices are faulted in response to IO failures.
action: Make sure the affected devices are connected, then run 'zpool clear'.
see: http://illumos.org/msg/ZFS-8000-HC
scan: scrub in progress since Sun Oct 20 19:29:23 2013
15.2T scanned out of 22.2T at 84.0M/s, 24h5m to go
1.29G repaired, 68.67% done
config:
NAME STATE READ WRITE CKSUM
nas UNAVAIL 63 2 0 insufficient
replicas
raidz1-0 DEGRADED 0 0 0
c8t50024E9004993E6Ed0p0 ONLINE 0 0 0
c8t50024E92062E7524d0 ONLINE 0 0 0
c8t50024E900495BE84d0p0 ONLINE 0 0 0
c8t50014EE25A5EEC23d0p0 ONLINE 0 0 0
c8t50024E9003F03980d0p0 ONLINE 0 0 1 (repairing)
c8t50014EE2B0D3EFC8d0 ONLINE 0 0 0
c8t50014EE6561DDB4Cd0p0 DEGRADED 0 0 211 too many errors
(repairing)
c8t50024E9003F03A09d0p0 ONLINE 0 0 18 (repairing)
raidz1-1 UNAVAIL 131 9 0 insufficient
replicas
c50t8d0 REMOVED 0 0 0 (repairing)
c2d0 ONLINE 0 0 0 (repairing)
c1d0 ONLINE 0 0 0 (repairing)
c50t11d0 ONLINE 0 0 0 (repairing)
c50t10d0 REMOVED 0 0 0
errors: 10972861 data errors, use '-v' for a list
really weird, I haven't disconnected any disk. For several hours even if it
said that the pool was unavailable I was browsing on it via nfs. I can't
anymore.
What do you think I should do ?
_______________________________________________
OpenIndiana-discuss mailing list
[email protected]
http://openindiana.org/mailman/listinfo/openindiana-discuss
_______________________________________________
OpenIndiana-discuss mailing list
[email protected]
http://openindiana.org/mailman/listinfo/openindiana-discuss
_______________________________________________
OpenIndiana-discuss mailing list
[email protected]
http://openindiana.org/mailman/listinfo/openindiana-discuss