Pare ok acum, Consistency Policy : resync Nu uita sa faci mdadm.conf si mkinitrd astfel incat la reboot sa pastreze acelasi device
mdadm --detail --scan > /etc/mdadm/mdadm.conf sau /etc/mdadm.conf Sent from mobile > On Dec 30, 2017, at 13:05, Paul Lacatus (Personal) <[email protected]> > wrote: > > Am reasamblat aria ca /dev/md0 . La inceput nu a vrut sa adauge si /dev/sda1 > , cu --readd nu a vrut cu --add l-a adaugat si acum ii face rebuilding > > [root@datavault ~]# mdadm /dev/md0 --add /dev/sda1 > mdadm: added /dev/sda1 > [root@datavault ~]# mdadm -D /dev/md0 > /dev/md0: > Version : 0.90 > Creation Time : Sat Dec 18 20:07:41 2010 > Raid Level : raid5 > Array Size : 2930279808 (2794.53 GiB 3000.61 GB) > Used Dev Size : 976759936 (931.51 GiB 1000.20 GB) > Raid Devices : 4 > Total Devices : 4 > Preferred Minor : 0 > Persistence : Superblock is persistent > > Update Time : Sat Dec 30 13:00:12 2017 > State : clean, degraded, recovering > Active Devices : 3 > Working Devices : 4 > Failed Devices : 0 > Spare Devices : 1 > > Layout : left-symmetric > Chunk Size : 64K > > Consistency Policy : resync > > Rebuild Status : 0% complete > > UUID : 188f7506:c03bd2ac:cdc3d78b:f6534f77 > Events : 0.14849 > > Number Major Minor RaidDevice State > 0 8 17 0 active sync /dev/sdb1 > 1 8 33 1 active sync /dev/sdc1 > 2 8 49 2 active sync /dev/sdd1 > 4 8 1 3 spare rebuilding /dev/sda1 > > > >> On 30-Dec-17 12:47, Paul Lacatus (Personal) wrote: >> La serverul meu de fisiere de acasa mi-a picat sursa . Pana sa ma prind ca >> e sursa de vina , facea probleme cand erau toate hdd alimentate , s-a dus >> si centos 6.9 care dadea kernel panic not tainted. Cu ocazia asta am >> reinstalat cu un centos 7.4 >> >> treaba e ca acum raid-ul fost /dev/md127 a devenit inactiv : >> >>> [root@datavault ~]# mdadm --detail /dev/md127 >>> /dev/md127: >>> Version : 0.90 >>> Raid Level : raid0 >>> Total Devices : 4 >>> Preferred Minor : 0 >>> Persistence : Superblock is persistent >>> >>> State : inactive >>> >>> UUID : 188f7506:c03bd2ac:cdc3d78b:f6534f77 >>> Events : 0.14838 >>> >>> Number Major Minor RaidDevice >>> >>> - 8 1 - /dev/sda1 >>> - 8 17 - /dev/sdb1 >>> - 8 33 - /dev/sdc1 >>> - 8 49 - /dev/sdd1 >> >> la examinare pe fiecare disc in parte datele par a fi ok dar aria nu pleaca >> din cauza testelor facute . Trebuie sa recreez aria sau se poate reporni ? >> >>> cat raid.status >>> /dev/sda1: >>> Magic : a92b4efc >>> Version : 0.90.00 >>> UUID : 188f7506:c03bd2ac:cdc3d78b:f6534f77 >>> Creation Time : Sat Dec 18 20:07:41 2010 >>> Raid Level : raid5 >>> Used Dev Size : 976759936 (931.51 GiB 1000.20 GB) >>> Array Size : 2930279808 (2794.53 GiB 3000.61 GB) >>> Raid Devices : 4 >>> Total Devices : 4 >>> Preferred Minor : 127 >>> >>> Update Time : Thu Dec 28 11:38:37 2017 >>> State : clean >>> Active Devices : 4 >>> Working Devices : 4 >>> Failed Devices : 0 >>> Spare Devices : 0 >>> Checksum : 279913d8 - correct >>> Events : 14838 >>> >>> Layout : left-symmetric >>> Chunk Size : 64K >>> >>> Number Major Minor RaidDevice State >>> this 3 8 1 3 active sync /dev/sda1 >>> >>> 0 0 8 17 0 active sync /dev/sdb1 >>> 1 1 8 33 1 active sync /dev/sdc1 >>> 2 2 8 49 2 active sync /dev/sdd1 >>> 3 3 8 1 3 active sync /dev/sda1 >>> /dev/sdb1: >>> Magic : a92b4efc >>> Version : 0.90.00 >>> UUID : 188f7506:c03bd2ac:cdc3d78b:f6534f77 >>> Creation Time : Sat Dec 18 20:07:41 2010 >>> Raid Level : raid5 >>> Used Dev Size : 976759936 (931.51 GiB 1000.20 GB) >>> Array Size : 2930279808 (2794.53 GiB 3000.61 GB) >>> Raid Devices : 4 >>> Total Devices : 3 >>> Preferred Minor : 127 >>> >>> Update Time : Sat Dec 30 10:19:54 2017 >>> State : clean >>> Active Devices : 3 >>> Working Devices : 3 >>> Failed Devices : 1 >>> Spare Devices : 0 >>> Checksum : 279ba477 - correct >>> Events : 14846 >>> >>> Layout : left-symmetric >>> Chunk Size : 64K >>> >>> Number Major Minor RaidDevice State >>> this 0 8 17 0 active sync /dev/sdb1 >>> >>> 0 0 8 17 0 active sync /dev/sdb1 >>> 1 1 8 33 1 active sync /dev/sdc1 >>> 2 2 8 49 2 active sync /dev/sdd1 >>> 3 3 0 0 3 faulty removed >>> /dev/sdc1: >>> Magic : a92b4efc >>> Version : 0.90.00 >>> UUID : 188f7506:c03bd2ac:cdc3d78b:f6534f77 >>> Creation Time : Sat Dec 18 20:07:41 2010 >>> Raid Level : raid5 >>> Used Dev Size : 976759936 (931.51 GiB 1000.20 GB) >>> Array Size : 2930279808 (2794.53 GiB 3000.61 GB) >>> Raid Devices : 4 >>> Total Devices : 3 >>> Preferred Minor : 127 >>> >>> Update Time : Sat Dec 30 10:20:32 2017 >>> State : active >>> Active Devices : 3 >>> Working Devices : 3 >>> Failed Devices : 1 >>> Spare Devices : 0 >>> Checksum : 279b6ab1 - correct >>> Events : 14847 >>> >>> Layout : left-symmetric >>> Chunk Size : 64K >>> >>> Number Major Minor RaidDevice State >>> this 1 8 33 1 active sync /dev/sdc1 >>> >>> 0 0 8 17 0 active sync /dev/sdb1 >>> 1 1 8 33 1 active sync /dev/sdc1 >>> 2 2 8 49 2 active sync /dev/sdd1 >>> 3 3 0 0 3 faulty removed >>> /dev/sdd1: >>> Magic : a92b4efc >>> Version : 0.90.00 >>> UUID : 188f7506:c03bd2ac:cdc3d78b:f6534f77 >>> Creation Time : Sat Dec 18 20:07:41 2010 >>> Raid Level : raid5 >>> Used Dev Size : 976759936 (931.51 GiB 1000.20 GB) >>> Array Size : 2930279808 (2794.53 GiB 3000.61 GB) >>> Raid Devices : 4 >>> Total Devices : 3 >>> Preferred Minor : 127 >>> >>> Update Time : Sat Dec 30 10:20:32 2017 >>> State : active >>> Active Devices : 3 >>> Working Devices : 3 >>> Failed Devices : 1 >>> Spare Devices : 0 >>> Checksum : 279b6ac3 - correct >>> Events : 14847 >>> >>> Layout : left-symmetric >>> Chunk Size : 64K >>> >>> Number Major Minor RaidDevice State >>> this 2 8 49 2 active sync /dev/sdd1 >>> >>> 0 0 8 17 0 active sync /dev/sdb1 >>> 1 1 8 33 1 active sync /dev/sdc1 >>> 2 2 8 49 2 active sync /dev/sdd1 >>> 3 3 0 0 3 faulty removed >>> [root@datavault ~]# >> >> >> >> _______________________________________________ >> RLUG mailing list >> [email protected] >> http://lists.lug.ro/mailman/listinfo/rlug_lists.lug.ro > > > _______________________________________________ > RLUG mailing list > [email protected] > http://lists.lug.ro/mailman/listinfo/rlug_lists.lug.ro _______________________________________________ RLUG mailing list [email protected] http://lists.lug.ro/mailman/listinfo/rlug_lists.lug.ro
