Am reasamblat aria ca /dev/md0 . La inceput nu a vrut sa adauge si /dev/sda1 , cu --readd nu a vrut cu --add l-a adaugat si acum ii face rebuilding

[root@datavault ~]# mdadm /dev/md0 --add /dev/sda1
mdadm: added /dev/sda1
[root@datavault ~]# mdadm -D /dev/md0
/dev/md0:
           Version : 0.90
     Creation Time : Sat Dec 18 20:07:41 2010
        Raid Level : raid5
        Array Size : 2930279808 (2794.53 GiB 3000.61 GB)
     Used Dev Size : 976759936 (931.51 GiB 1000.20 GB)
      Raid Devices : 4
     Total Devices : 4
   Preferred Minor : 0
       Persistence : Superblock is persistent

       Update Time : Sat Dec 30 13:00:12 2017
             State : clean, degraded, recovering
    Active Devices : 3
   Working Devices : 4
    Failed Devices : 0
     Spare Devices : 1

            Layout : left-symmetric
        Chunk Size : 64K

Consistency Policy : resync

    Rebuild Status : 0% complete

              UUID : 188f7506:c03bd2ac:cdc3d78b:f6534f77
            Events : 0.14849

    Number   Major   Minor   RaidDevice State
       0       8       17        0      active sync /dev/sdb1
       1       8       33        1      active sync /dev/sdc1
       2       8       49        2      active sync /dev/sdd1
       4       8        1        3      spare rebuilding   /dev/sda1



On 30-Dec-17 12:47, Paul Lacatus (Personal) wrote:
La serverul meu de fisiere de acasa mi-a picat sursa .  Pana sa ma prind ca e sursa de vina , facea probleme cand erau toate hdd alimentate ,  s-a dus si centos 6.9 care dadea kernel panic not tainted. Cu ocazia asta am reinstalat cu un centos 7.4

treaba e ca acum raid-ul fost /dev/md127 a devenit inactiv :

[root@datavault ~]# mdadm --detail /dev/md127
/dev/md127:
           Version : 0.90
        Raid Level : raid0
     Total Devices : 4
   Preferred Minor : 0
       Persistence : Superblock is persistent

             State : inactive

              UUID : 188f7506:c03bd2ac:cdc3d78b:f6534f77
            Events : 0.14838

    Number   Major   Minor   RaidDevice

       -       8        1        -        /dev/sda1
       -       8       17        -        /dev/sdb1
       -       8       33        -        /dev/sdc1
       -       8       49        -        /dev/sdd1

la examinare pe fiecare disc in parte datele par a fi ok dar aria nu pleaca din cauza testelor facute  . Trebuie sa recreez aria sau se poate reporni ?

 cat raid.status
/dev/sda1:
          Magic : a92b4efc
        Version : 0.90.00
           UUID : 188f7506:c03bd2ac:cdc3d78b:f6534f77
  Creation Time : Sat Dec 18 20:07:41 2010
     Raid Level : raid5
  Used Dev Size : 976759936 (931.51 GiB 1000.20 GB)
     Array Size : 2930279808 (2794.53 GiB 3000.61 GB)
   Raid Devices : 4
  Total Devices : 4
Preferred Minor : 127

    Update Time : Thu Dec 28 11:38:37 2017
          State : clean
 Active Devices : 4
Working Devices : 4
 Failed Devices : 0
  Spare Devices : 0
       Checksum : 279913d8 - correct
         Events : 14838

         Layout : left-symmetric
     Chunk Size : 64K

      Number   Major   Minor   RaidDevice State
this     3       8        1        3      active sync /dev/sda1

   0     0       8       17        0      active sync /dev/sdb1
   1     1       8       33        1      active sync /dev/sdc1
   2     2       8       49        2      active sync /dev/sdd1
   3     3       8        1        3      active sync /dev/sda1
/dev/sdb1:
          Magic : a92b4efc
        Version : 0.90.00
           UUID : 188f7506:c03bd2ac:cdc3d78b:f6534f77
  Creation Time : Sat Dec 18 20:07:41 2010
     Raid Level : raid5
  Used Dev Size : 976759936 (931.51 GiB 1000.20 GB)
     Array Size : 2930279808 (2794.53 GiB 3000.61 GB)
   Raid Devices : 4
  Total Devices : 3
Preferred Minor : 127

    Update Time : Sat Dec 30 10:19:54 2017
          State : clean
 Active Devices : 3
Working Devices : 3
 Failed Devices : 1
  Spare Devices : 0
       Checksum : 279ba477 - correct
         Events : 14846

         Layout : left-symmetric
     Chunk Size : 64K

      Number   Major   Minor   RaidDevice State
this     0       8       17        0      active sync /dev/sdb1

   0     0       8       17        0      active sync /dev/sdb1
   1     1       8       33        1      active sync /dev/sdc1
   2     2       8       49        2      active sync /dev/sdd1
   3     3       0        0        3      faulty removed
/dev/sdc1:
          Magic : a92b4efc
        Version : 0.90.00
           UUID : 188f7506:c03bd2ac:cdc3d78b:f6534f77
  Creation Time : Sat Dec 18 20:07:41 2010
     Raid Level : raid5
  Used Dev Size : 976759936 (931.51 GiB 1000.20 GB)
     Array Size : 2930279808 (2794.53 GiB 3000.61 GB)
   Raid Devices : 4
  Total Devices : 3
Preferred Minor : 127

    Update Time : Sat Dec 30 10:20:32 2017
          State : active
 Active Devices : 3
Working Devices : 3
 Failed Devices : 1
  Spare Devices : 0
       Checksum : 279b6ab1 - correct
         Events : 14847

         Layout : left-symmetric
     Chunk Size : 64K

      Number   Major   Minor   RaidDevice State
this     1       8       33        1      active sync /dev/sdc1

   0     0       8       17        0      active sync /dev/sdb1
   1     1       8       33        1      active sync /dev/sdc1
   2     2       8       49        2      active sync /dev/sdd1
   3     3       0        0        3      faulty removed
/dev/sdd1:
          Magic : a92b4efc
        Version : 0.90.00
           UUID : 188f7506:c03bd2ac:cdc3d78b:f6534f77
  Creation Time : Sat Dec 18 20:07:41 2010
     Raid Level : raid5
  Used Dev Size : 976759936 (931.51 GiB 1000.20 GB)
     Array Size : 2930279808 (2794.53 GiB 3000.61 GB)
   Raid Devices : 4
  Total Devices : 3
Preferred Minor : 127

    Update Time : Sat Dec 30 10:20:32 2017
          State : active
 Active Devices : 3
Working Devices : 3
 Failed Devices : 1
  Spare Devices : 0
       Checksum : 279b6ac3 - correct
         Events : 14847

         Layout : left-symmetric
     Chunk Size : 64K

      Number   Major   Minor   RaidDevice State
this     2       8       49        2      active sync /dev/sdd1

   0     0       8       17        0      active sync /dev/sdb1
   1     1       8       33        1      active sync /dev/sdc1
   2     2       8       49        2      active sync /dev/sdd1
   3     3       0        0        3      faulty removed
[root@datavault ~]#



_______________________________________________
RLUG mailing list
[email protected]
http://lists.lug.ro/mailman/listinfo/rlug_lists.lug.ro


_______________________________________________
RLUG mailing list
[email protected]
http://lists.lug.ro/mailman/listinfo/rlug_lists.lug.ro

Raspunde prin e-mail lui