On Sun, Aug 07, 2005 at 11:19:12PM -0500, Colin Ingram wrote:
> calling all lvm/raid gurus....

No guru, but have just suffered through something similar. I just built
an lvm on top of raid10 and certainly went through all of your issues
below


> 
> I'm running debian sarge and I have/had a 200GB raid 1 devices with LVM2 
> on top.  The raid one was made from hdb1 and hdc1.  Recently when I was 
> trying to change the size of a logical volume.  I noticed that my vg was 
> using /dev/hdc1 as a physical volume instead of /dev/md0. 

before we go further, I had to do two tweaks to the /etc/lvm/lvm.conf.
First, you need to make sure you've got:
    md_component_detection = 1

Then, I needed to add a filter to remove all of my /dev/hd* drives so
that lvm would only be allowed to see the /dev/md* "drives". There was
already a line there to remove /dev/cdrom, so I hijacked that:
    filter = [ "r|/dev/cdrom|", "r|/dev/hd.*|" ]

In the end, I'm left with the following in my /etc/lvm/.cache
        # This file is automatically maintained by lvm.

        persistent_filter_cache {
                valid_devices=[
                        "/dev/md5",
                        "/dev/md0",
                        "/dev/md1"
                ]
        }

md0 is raid1 /, md1 is raid1 swap, and md5 is raid10 over md2/3/4

That resolved my found duplicate PV issues

> 
> $ pvscan
>      Found duplicate PV YDMoRSv4EKZXHNPuchid5hPIwmXDCBCm: using 
> /dev/hdb1 not /dev/hdc1
>      PV /dev/hda4   VG LVM    lvm2 [68.96 GB / 0    free]
>      PV /dev/hdc1   VG data   lvm2 [186.30 GB / 4.00 MB free]
>      Total: 2 [255.26 GB] / in use: 2 [255.26 GB] / in no VG: 0 [0   ]
> 
> some additional digging and I found that my raid devices is inactive
> 
> $ mdadm -D /dev/md0
>    mdadm: md device /dev/md0 does not appear to be active.
> 
> but the mdsuperblock still seems present on both devices
> 
> $mdadm -E /dev/hdb1
> /dev/hdb1:
>          Magic : a92b4efc
>        Version : 00.90.00
>           UUID : 9dc1df9b:771959ca:53ed751d:7b3b40ba
>  Creation Time : Sun Apr 17 13:24:05 2005
>     Raid Level : raid1
>   Raid Devices : 2
>  Total Devices : 2
> Preferred Minor : 0
> 
>    Update Time : Fri May 13 16:52:37 2005
>          State : clean
> Active Devices : 2
> Working Devices : 2
> Failed Devices : 0
>  Spare Devices : 0
>       Checksum : 1dcab098 - correct
>         Events : 0.482221
> 
> 
>      Number   Major   Minor   RaidDevice State
> this     1       3       65        1      active sync   
> /dev/.static/dev/hdb1
> 
>   0     0      22        1        0      active sync   
> /dev/.static/dev/hdc1
>   1     1       3       65        1      active sync   
> /dev/.static/dev/hdb1
> 
> $ mdadm /dev/hdc1
> /dev/hdc1:
>          Magic : a92b4efc
>        Version : 00.90.00
>           UUID : 9dc1df9b:771959ca:53ed751d:7b3b40ba
>  Creation Time : Sun Apr 17 13:24:05 2005
>     Raid Level : raid1
>   Raid Devices : 2
>  Total Devices : 2
> Preferred Minor : 0
> 
>    Update Time : Fri May 13 16:52:37 2005
>          State : clean
> Active Devices : 2
> Working Devices : 2
> Failed Devices : 0
>  Spare Devices : 0
>       Checksum : 1dcab069 - correct
>         Events : 0.482221
> 
> 
>      Number   Major   Minor   RaidDevice State
> this     0      22        1        0      active sync   
> /dev/.static/dev/hdc1
> 
>   0     0      22        1        0      active sync   
> /dev/.static/dev/hdc1
>   1     1       3       65        1      active sync   
> /dev/.static/dev/hdb1
> 
> 
> I'm not sure how this happen but I've come up with a couple of 
> possibilities that happen several months ago. 
> 
> 1. I migrated these disk from a previous test box for this server.  I 
> may have messed on the initial install into the new box. 
> 2. I replaced hdb1 as hdc1 when doing some  maintenance. 
> 
> In any case LVM recognized the UUID of one of disks and continued 
> normally. 
> 
> Now I'm try to fix the situation.  All the data on the disks are backup 
> so I could just start fresh but I would like to repair the situation if 
> I can.  Here is the steps I've come up with.  Please check them out for 
> pitfalls etc or missing commands.
> 
> Outline:
> To fix the LVM, I will treat the situation as if I am migrating the 
> disks to a new system.  Instead of moving drives I will be repairing raid. 
> To fix Raid I need to determine which disk LVM has been using then 
> restart the RAID with that disk then add the second disk as a spare.
> 
> Detail:
> 1. Umount file systems
> 
> # unmount /backup
> # unmount /share/ftp
> # unmount /share
> 
> 2. Mark Volume Group as inactive
> 
> # vgchange -an data
> 
> 3. Export volume Group
> 
> # vgexport data
> 
> 4. Mount Disk and examine contents for recent data
> 
> # mount /dev/hdb1 /mnt/hdb
> # mount /dev/hdc1 /mnt/hdc
> 
> 5.  Restart array with devices with most recent data
> 
> #  mdadm -A /dev/md0 /dev/hdb1 or /dev/hdc1
> (will this start the array degraded or do I need to create a new array 
> with mdadm -C /dev/md0 --level=1 --raid-devices=/dev/hdb1 missing)
> 
> 6. Add second devices as spare
> 
> # mdadm /dev/md0 --add /dev/hdc1 or /dev/hdb1
> 
> 7.  Fix LVM (this is assuming that pvscan now recognizes PV /dev/md0 as 
> being apart of VG data
> 
> #  vgimport data
> #  vgchange -ay data
> #  mount /share
> #  mount /share/ftp
> #  mount /backup
> 
> 8.  Set up mdadm in monitor mode and have a beer!
> 
> 
> Thanks for the help in advance
> 
> Colin
> 
> 
> -- 
> To UNSUBSCRIBE, email to [EMAIL PROTECTED] 
> with a subject of "unsubscribe". Trouble? Contact 
> [EMAIL PROTECTED]
> 
> 


-- 
To UNSUBSCRIBE, email to [EMAIL PROTECTED] 
with a subject of "unsubscribe". Trouble? Contact [EMAIL PROTECTED]

Reply via email to