Alvin Oga wrote:
what exaactly did you type BEFORE you removed the bad disk ??
raidhotadd, raidhotremove, etc, etc.. is required ( aka good idea )
I use mdadm to manage my array. The command was
mdadm --set-faulty /dev/md1 /dev/hda1 --remove /dev/md1 /dev/hda1
(repeated for the other partitions substituting md2,hda2, md3,hda3 etc;
only some of the partitions had been marked as failed in the old drive I
tried re-adding them but they failed again after a few days so I swapped
the drive)
some bios cannot boot off of additional ide cards ( pci cards )
and mixing with onboard ide might be confusing the system
My system wouldn't boot at all with the new drive in place until I went
into the bios and explicitly told it to boot off the drive on the ide
card. The drives attached to the card were listed in the bios setup
screen as boot options so I'm pretty sure my bios is compatible with
booting off raid. Since Windows will only boot off the drive with bios
device code 0x80 I believe that the bios may be switching the codes
around and this is what is confusing lilo. I don't know why this is a
problem though since I thought that since version 22.5 lilo didn't care
about device codes.
LILO version 22.6.1
too old ...
you should be using lilo-22.7.1 for raid stuff if you want a working
bootable raid system
It's the version that comes with sarge. I set up the array with the LILO
that came with woody, surely upgrading isn't necessary? I'll look for a
backport.
lba32
boot=/dev/md1
raid-extra-boot=/dev/hda,/dev/hde
why do you use "extra-boot" ??
- try commenting that line out
AFAIK that line is essential. It tells lilo to write the mbr to both
members of the raid array, so that if one drive dies the system remains
bootable. If I'm wrong about this please let me know. I set up the array
using the instructions at http://alioth.debian.org/projects/rootraiddoc/
this implies /dev/md1 is your /boot ??
and that your rootfs ( /bin, /etc, /sbin .. ) is in /dev/md3 ??
That is correct, however this was my first Linux system and I went a bit
overboard on the partitioning: tmp, usr, var, and home each have their
own partition.
for more debugging ...
what is the contents of /etc/fstab
# /etc/fstab: static file system information.
#
# <file system> <mount point> <type> <options> <dump>
<pass>
/dev/md3 / ext3 errors=remount-ro 0 1
/dev/md2 none swap sw 0 0
proc /proc proc defaults 0 0
/dev/fd0 /floppy auto user,noauto 0 0
/dev/cdrom /cdrom iso9660 ro,user,noauto 0 0
/dev/md5 /tmp ext3 defaults 0 2
/dev/md6 /usr ext3 defaults 0 2
/dev/md7 /var ext3 defaults 0 2
/dev/md8 /home ext3 defaults 0 2
/dev/md1 /boot ext3 defaults 0 2
what is the contents of /etc/mdadm.conf
:~# cat /etc/mdadm/mdadm.conf
DEVICE /dev/hda* /dev/hde*
ARRAY /dev/md1 level=raid1 disks=2 UUID=f126c8af:1fd56337:a653c52b:3e50c388
devices=/dev/hda1,/dev/hde1
ARRAY /dev/md2 level=raid1 disks=2 UUID=6a91045f:babc1717:3a48db49:e10249c1
devices=/dev/hda2,/dev/hde2
ARRAY /dev/md3 level=raid1 disks=2 UUID=b32dcb66:a802ec78:795cdb4e:3ed450ae
devices=/dev/hda3,/dev/hde3
ARRAY /dev/md5 level=raid1 disks=2 UUID=173c680c:200676a5:29113ce8:5e0c54e9
devices=/dev/hda5,/dev/hde5
ARRAY /dev/md6 level=raid1 disks=2 UUID=73a930b9:674865c7:1c5030f1:55ae62a9
devices=/dev/hda6,/dev/hde6
ARRAY /dev/md7 level=raid1 disks=2 UUID=641c416c:2ffd8890:3cce2a6f:811f694b
devices=/dev/hda7,/dev/hde7
ARRAY /dev/md8 level=raid1 disks=2 UUID=850a410c:1fa77dd9:5234f468:43814e67
devices=/dev/hda8,/dev/hde8
what is the output of "cat /proc/mdstat"
:~# cat /proc/mdstat
Personalities : [raid1] [raid5]
read_ahead 1024 sectors
md1 : active raid1 hde1[1] hda1[0]
16000 blocks [2/2] [UU]
md2 : active raid1 hde2[1] hda2[0]
490304 blocks [2/2] [UU]
md3 : active raid1 hde3[1] hda3[0]
144576 blocks [2/2] [UU]
md5 : active raid1 hde5[1] hda5[0]
96640 blocks [2/2] [UU]
md6 : active raid1 hde6[1] hda6[0]
4000128 blocks [2/2] [UU]
md7 : active raid1 hde7[1] hda7[0]
4000128 blocks [2/2] [UU]
md8 : active raid1 hde8[1] hda8[0]
31273088 blocks [2/2] [UU]
Thanks for any help/advice you can give.
--
To UNSUBSCRIBE, email to [EMAIL PROTECTED]
with a subject of "unsubscribe". Trouble? Contact [EMAIL PROTECTED]