I'm still a student so when I'll get a job I will change all that.
apparently mixing vdevs with different number of disk is not the best
idea either
Le 23/10/2013 09:21, Mark a écrit :
One suggestion for the future for better reliability.
Avoid using WDC WD20EARS, they aren't great, and hopeless on SAS
controllers.
Avoid mixing disk brands or models in a zpool.
The different behaviour can trigger issues.
On 23/10/2013 7:46 p.m., Clement BRIZARD wrote:
I cleared the "degraded" disk. we will see what happens in 131hours
pool: nas
state: ONLINE
status: One or more devices is currently being resilvered. The pool
will
continue to function, possibly in a degraded state.
action: Wait for the resilver to complete.
scan: resilver in progress since Wed Oct 23 08:25:56 2013
2.23G scanned out of 22.2T at 48.6M/s, 133h22m to go
6.10M resilvered, 0.01% done
config:
NAME STATE READ WRITE CKSUM
CAP Product
nas ONLINE 0 0 0
raidz1-0 ONLINE 0 0 0
c8t50024E9004993E6Ed0p0 ONLINE 0 0 0 2
TB SAMSUNG HD204UI
c8t50024E92062E7524d0 ONLINE 0 0 0 2
TB SAMSUNG HD204UI
c8t50024E900495BE84d0p0 ONLINE 0 0 0 2
TB SAMSUNG HD204UI
c8t50014EE25A5EEC23d0p0 ONLINE 0 0 0 2
TB WDC WD20EARS-00M
c8t50024E9003F03980d0p0 ONLINE 0 0 0 2
TB SAMSUNG HD204UI
c8t50014EE2B0D3EFC8d0 ONLINE 0 0 0 2
TB WDC WD20EARX-00P
c8t50014EE6561DDB4Cd0p0 ONLINE 0 0 0 2
TB WDC WD20EARS-00M
c8t50024E9003F03A09d0p0 ONLINE 0 0 0 2
TB SAMSUNG HD204UI
raidz1-1 ONLINE 0 0 0
c50t8d0 ONLINE 0 0 0
(resilvering) 2 TB ST2000DL004 HD20
c2d0 ONLINE 0 0 0
(resilvering) 2 TB
c1d0 ONLINE 0 0 0
(resilvering) 2 TB
c50t11d0 ONLINE 0 0 0 2
TB SAMSUNG HD204UI
c50t10d0 ONLINE 0 0 0
(resilvering) 2 TB SAMSUNG HD204UI
Le 23/10/2013 08:43, Clement BRIZARD a écrit :
I woke up this morning and so you're messages, unfortunately I had to
reboot, the server completely froze.
Now I have that :
pool: nas
state: DEGRADED
status: One or more devices is currently being resilvered. The pool
will
continue to function, possibly in a degraded state.
action: Wait for the resilver to complete.
scan: resilver in progress since Wed Oct 23 08:19:42 2013
5.81G scanned out of 22.2T at 49.2M/s, 131h43m to go
15.6M resilvered, 0.03% done
config:
NAME STATE READ WRITE CKSUM
nas DEGRADED 0 0 0
raidz1-0 DEGRADED 0 0 0
c8t50024E9004993E6Ed0p0 ONLINE 0 0 0
c8t50024E92062E7524d0 ONLINE 0 0 0
c8t50024E900495BE84d0p0 ONLINE 0 0 0
c8t50014EE25A5EEC23d0p0 ONLINE 0 0 0
c8t50024E9003F03980d0p0 ONLINE 0 0 0
c8t50014EE2B0D3EFC8d0 ONLINE 0 0 0
c8t50014EE6561DDB4Cd0p0 DEGRADED 0 0 0 too many
errors
c8t50024E9003F03A09d0p0 ONLINE 0 0 0
raidz1-1 ONLINE 0 0 0
c50t8d0 ONLINE 0 0 0
(resilvering)
c2d0 ONLINE 0 0 0
(resilvering)
c1d0 ONLINE 0 0 0
(resilvering)
c50t11d0 ONLINE 0 0 0
c50t10d0 ONLINE 0 0 0
(resilvering)
Le 23/10/2013 08:00, Jason Matthews a écrit :
first, dont reboot. if you do you might not be able remount the pool.
the data you see is from the disks that are functioning. listing the
files and copying complete files are two different things. if you
dont have a backup you may need to copy whatever partial data you can
from the broken pool.
now let's start by getting the disks back in good shape.
clear the degraded disk
zpool clear c8t50014EE6561DDB4Cd0p0
reseat the missing disks in the hopes they come back then clear them
check cfgadm -al and make sure they are connected and configured
when you reseat them check the messages (or dmesg) to see if the
system notices the re-insertion. if it does see the disk installed
clear the disks in the pool in effort to bring the pool back to an
operational state.
Sent from Jasons' hand held
On Oct 22, 2013, at 5:04 PM, Clement BRIZARD <clem...@brizou.fr>
wrote:
Hello everybody,
I have a problem with my pool, I had some slowdowns lately on my nfs
share of my zfs pool. A weekly scrub began and is still running but
it worries me, it currently returne that
pool: nas
state: UNAVAIL
status: One or more devices are faulted in response to IO failures.
action: Make sure the affected devices are connected, then run
'zpool clear'.
see: http://illumos.org/msg/ZFS-8000-HC
scan: scrub in progress since Sun Oct 20 19:29:23 2013
15.2T scanned out of 22.2T at 84.0M/s, 24h5m to go
1.29G repaired, 68.67% done
config:
NAME STATE READ WRITE CKSUM
nas UNAVAIL 63 2 0
insufficient replicas
raidz1-0 DEGRADED 0 0 0
c8t50024E9004993E6Ed0p0 ONLINE 0 0 0
c8t50024E92062E7524d0 ONLINE 0 0 0
c8t50024E900495BE84d0p0 ONLINE 0 0 0
c8t50014EE25A5EEC23d0p0 ONLINE 0 0 0
c8t50024E9003F03980d0p0 ONLINE 0 0 1 (repairing)
c8t50014EE2B0D3EFC8d0 ONLINE 0 0 0
c8t50014EE6561DDB4Cd0p0 DEGRADED 0 0 211 too many
errors (repairing)
c8t50024E9003F03A09d0p0 ONLINE 0 0 18 (repairing)
raidz1-1 UNAVAIL 131 9 0
insufficient replicas
c50t8d0 REMOVED 0 0 0 (repairing)
c2d0 ONLINE 0 0 0 (repairing)
c1d0 ONLINE 0 0 0 (repairing)
c50t11d0 ONLINE 0 0 0 (repairing)
c50t10d0 REMOVED 0 0 0
errors: 10972861 data errors, use '-v' for a list
really weird, I haven't disconnected any disk. For several hours
even if it said that the pool was unavailable I was browsing on it
via nfs. I can't anymore.
What do you think I should do ?
_______________________________________________
OpenIndiana-discuss mailing list
OpenIndiana-discuss@openindiana.org
http://openindiana.org/mailman/listinfo/openindiana-discuss
_______________________________________________
OpenIndiana-discuss mailing list
OpenIndiana-discuss@openindiana.org
http://openindiana.org/mailman/listinfo/openindiana-discuss
_______________________________________________
OpenIndiana-discuss mailing list
OpenIndiana-discuss@openindiana.org
http://openindiana.org/mailman/listinfo/openindiana-discuss
_______________________________________________
OpenIndiana-discuss mailing list
OpenIndiana-discuss@openindiana.org
http://openindiana.org/mailman/listinfo/openindiana-discuss
_______________________________________________
OpenIndiana-discuss mailing list
OpenIndiana-discuss@openindiana.org
http://openindiana.org/mailman/listinfo/openindiana-discuss
_______________________________________________
OpenIndiana-discuss mailing list
OpenIndiana-discuss@openindiana.org
http://openindiana.org/mailman/listinfo/openindiana-discuss