I should unsubscribe; still too new to understand you folks! JAWs
Date: Wed, 6 Mar 2013 19:22:24 +0000 From: debian-user-digest-requ...@lists.debian.org Subject: debian-user-digest Digest V2013 #252 To: debian-user-dig...@lists.debian.org --Forwarded Message Attachment-- debian-user-digest Digest Volume 2013 : Issue 252 Today's Topics: Re: Iceweasel and chromium stable an [ Sven Joachim <svenj...@gmx.de> ] Re: RAID1 all bootable [ Shane Johnson <sdj@rasmussenequipme ] Raid 5 [ Dick Thomas <xpd...@gmail.com> ] Re: Raid 5 [ Gary Dale <garyd...@rogers.com> ] Re: Raid 5 [ Adam Wolfe <kadamwo...@gmail.com> ] Re: Raid 5 [ Dick Thomas <xpd...@gmail.com> ] Re: Raid 5 [ Gary Dale <garyd...@rogers.com> ] Re: RAID1 all bootable [ Francesco Pietra <chiendarret@gmail ] Re: Raid 5 [ Gary Dale <garyd...@rogers.com> ] --Forwarded Message Attachment-- Date: Wed, 6 Mar 2013 19:00:15 +0100 From: svenj...@gmx.de Subject: Re: Iceweasel and chromium stable and secure versions To: debian-user@lists.debian.org On 2013-03-06 04:39 +0100, Steven Rosenberg wrote: > On Thu, Feb 28, 2013 at 9:49 AM, Sven Joachim <svenj...@gmx.de> wrote: >> On 2013-02-28 18:18 +0100, Henry Jensen wrote: >> >>> I noticed that chromium got updated to the newest stable and secure >>> version 25.0.1364.97. Can we expect that Chromium will receive further >>> updates in the future during the lifetime of Debian 7.0, perhaps as >>> part of the Update-Repository (formerly known as volatile)? >> >> It is planned to regularly push new upstream versions of Chromium into >> Debian 7.0, AFAIK as security updates. >> >>> And what about Iceweasel. Upstream Firefox 10.x is EOL now. >> >> Iceweasel will get security support, but no new major versions. If you >> want those, look on http://mozilla.debian.net/. > > > That's a nice compromise. I remember Chromium getting mighty old in > the Lenny days. Oh, you misremember. ;-) Chromium was never included in Lenny, it only hit Debian in 2010. The "mighty old" version is in Squeeze, it has been abandoned for ~1.5 years (last security update is from September 2011). Looking at the latest build log[1], I doubt that Chromium will be supportable on i386 in the Wheezy time frame. The memory requirements for linking it are likely to increase further in the future. Cheers, Sven 1. https://buildd.debian.org/status/package.php?p=chromium-browser --Forwarded Message Attachment-- Date: Wed, 6 Mar 2013 11:02:34 -0700 From: s...@rasmussenequipment.com Subject: Re: RAID1 all bootable To: chiendar...@gmail.com CC: debian-user@lists.debian.org <snip> As far as I can remember, I already posted for this system root@.....:/home/francesco# cat /proc/mdstat Personalities : [raid1] md1 : active raid1 sda2[0] sdb2[1] á á á 487759680 blocks super 1.2 [2/2] [UU] md0 : active raid1 sda1[0] sdb1[1] á á á 191296 blocks super 1.2 [2/2] [UU] unused devices: <none> root@.....:/home/francesco# francesco@.....:~$ df -h Filesystem á á á á á áSize áUsed Avail Use% Mounted on rootfs á á á á á á á á938M á185M á705M á21% / udev á á á á á á á á á 10M á á 0 á 10M á 0% /dev tmpfs á á á á á á á á 807M á628K á807M á 1% /run /dev/mapper/vg1-root á938M á185M á705M á21% / tmpfs á á á á á á á á 5.0M á á 0 á5.0M á 0% /run/lock tmpfs á á á á á á á á 1.6G á 84K á1.6G á 1% /run/shm /dev/md0 á á á á á á á176M á 19M á148M á11% /boot /dev/mapper/vg1-home á395G á284G á 91G á76% /home /dev/mapper/vg1-opt á 9.2G á1.5G á7.3G á17% /opt /dev/mapper/vg1-tmp á 2.8G á 69M á2.6G á 3% /tmp /dev/mapper/vg1-usr á á28G á4.3G á 22G á17% /usr /dev/mapper/vg1-var á 9.2G á840M á7.9G á10% /var francesco@.....:~$ the "deadly command' "grub-install /dev/sdb" áwas run with the system started as above. Thanks francesco pietra <snip>Francesco,The df -h shows us what is mounted but not if the drives are partitioned or not. áCan you do fdisk -l áand send us the output of that?Also, you replied directly to me without the mailing list. áI have included it in the CC so that everyone can share in the knowledge. áPlease make sure áyou always reply to the list.(Reply-all works real good for this.) Thanksá -- Shane D. JohnsonIT AdministratorRasmussen Equipment --Forwarded Message Attachment-- Date: Wed, 6 Mar 2013 18:37:26 +0000 From: xpd...@gmail.com Subject: Raid 5 To: debian-user@lists.debian.org What is the best way to setup a raid 5 array (4* 2TB drives) should I make raid 5 for my system and /home then raid 0 or 1 for the boot, or should I buy a 5th drive for system/boot and install in the standard way? as this is my 1st time on debian and not sure what would be best Dick Thomas About.me http://about.me/dick.thomas Blog: www.xpd259.co.uk G+: www.google.com/profiles/xpd259 gpg key: C791809B --Forwarded Message Attachment-- Date: Wed, 6 Mar 2013 13:43:06 -0500 From: garyd...@rogers.com Subject: Re: Raid 5 To: debian-user@lists.debian.org CC: debian-user@lists.debian.org On 06/03/13 01:37 PM, Dick Thomas wrote: > What is the best way to setup a raid 5 array (4* 2TB drives) > should I make raid 5 for my system and /home > then raid 0 or 1 for the boot, or should I buy a 5th drive for > system/boot and install in the standard way? > as this is my 1st time on debian and not sure what would be best > Make one large RAID5 array then partition the RAID array as you like (/, /home, swap, etc..). This is bootable using grub so there is no need for a separate /boot. --Forwarded Message Attachment-- Date: Wed, 6 Mar 2013 13:47:26 -0500 From: kadamwo...@gmail.com Subject: Re: Raid 5 To: debian-user@lists.debian.org I had one h**l of a time doing this over the weekend. What finally worked for me was creating LOGICAL partitions on each drive and setting them as used for RAID volume devices. This gave me /dev/sda5, /dev/sdb5 etc etc. When grub did it's install, it added all the /dev/sda1 etc partitions and rebooted fine. When I tried primary partitions, grub would just fail and I'd have to restart the whole install process over. NOTE: when booting from the install cd i had to [tab] the 'install' menu entry and add "dmraid=true". Then after the initial install, it would still fail to boot. Back to the install cd, choose 'rescue', [tab] and add 'dmraid=true' again. Then get thee to a shell and 'grub-install --recheck /dev/sdaX' for each partition. On 03/06/2013 01:37 PM, Dick Thomas wrote: > What is the best way to setup a raid 5 array (4* 2TB drives) > should I make raid 5 for my system and /home > then raid 0 or 1 for the boot, or should I buy a 5th drive for > system/boot and install in the standard way? > as this is my 1st time on debian and not sure what would be best > > > Dick Thomas > About.me http://about.me/dick.thomas > Blog: www.xpd259.co.uk > G+: www.google.com/profiles/xpd259 > gpg key: C791809B > > --Forwarded Message Attachment-- Date: Wed, 6 Mar 2013 18:55:20 +0000 From: xpd...@gmail.com Subject: Re: Raid 5 To: kadamwo...@gmail.com CC: debian-user@lists.debian.org About.me http://about.me/dick.thomas Blog: www.xpd259.co.uk G+: www.google.com/profiles/xpd259 gpg key: C791809B On 6 March 2013 18:47, Adam Wolfe <kadamwo...@gmail.com> wrote: > I had one h**l of a time doing this over the weekend. > > What finally worked for me was creating LOGICAL partitions on each drive and > setting them as used for RAID volume devices. > This gave me /dev/sda5, /dev/sdb5 etc etc. > When grub did it's install, it added all the /dev/sda1 etc partitions and > rebooted fine. > > When I tried primary partitions, grub would just fail and I'd have to > restart the whole install process over. > > NOTE: when booting from the install cd i had to [tab] the 'install' menu > entry and add "dmraid=true". > Then after the initial install, it would still fail to boot. > Back to the install cd, choose 'rescue', [tab] and add 'dmraid=true' again. > Then get thee to a shell and 'grub-install --recheck /dev/sdaX' for each > partition. > > > > > On 03/06/2013 01:37 PM, Dick Thomas wrote: >> >> What is the best way to setup a raid 5 array (4* 2TB drives) >> should I make raid 5 for my system and /home >> then raid 0 or 1 for the boot, or should I buy a 5th drive for >> system/boot and install in the standard way? >> as this is my 1st time on debian and not sure what would be best >> >> >> Dick Thomas >> About.me http://about.me/dick.thomas >> Blog: www.xpd259.co.uk >> G+: www.google.com/profiles/xpd259 >> gpg key: C791809B >> >> > > > -- > To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org with a subject > of "unsubscribe". Trouble? Contact listmas...@lists.debian.org > Archive: http://lists.debian.org/51378f3e.20...@gmail.com > I've tried installing from my "hardware" motherboard raid and that just fails even with dmraid=true atm i've got a raid 0 /boot raid 5 encrypted then /lvm /lvm/swap /lvm/root //var/log but wasn't sure if that was the way to do it or am I just confusing matters more :) --Forwarded Message Attachment-- Date: Wed, 6 Mar 2013 13:59:49 -0500 From: garyd...@rogers.com Subject: Re: Raid 5 To: debian-user@lists.debian.org CC: debian-user@lists.debian.org On 06/03/13 01:47 PM, Adam Wolfe wrote: > I had one h**l of a time doing this over the weekend. > > What finally worked for me was creating LOGICAL partitions on each > drive and setting them as used for RAID volume devices. > This gave me /dev/sda5, /dev/sdb5 etc etc. > When grub did it's install, it added all the /dev/sda1 etc partitions > and rebooted fine. > > When I tried primary partitions, grub would just fail and I'd have to > restart the whole install process over. > > NOTE: when booting from the install cd i had to [tab] the 'install' > menu entry and add "dmraid=true". > Then after the initial install, it would still fail to boot. > Back to the install cd, choose 'rescue', [tab] and add 'dmraid=true' > again. Then get thee to a shell and 'grub-install --recheck > /dev/sdaX' for each partition. > > > > On 03/06/2013 01:37 PM, Dick Thomas wrote: >> What is the best way to setup a raid 5 array (4* 2TB drives) >> should I make raid 5 for my system and /home >> then raid 0 or 1 for the boot, or should I buy a 5th drive for >> system/boot and install in the standard way? >> as this is my 1st time on debian and not sure what would be best >> Sorry but this isn't difficult (although it may affect top-posters more than bottom posters :) ). The Debian installer allows you to create a whole-disk RAID array then partition it. You have a single RAID 5 array with some number of primary partitions (up to 4 - I use 2, / and /home, with swap files rather than swap partitions but traditionalists may prefer a swap partition). Grub treats the array like a disk drive and has no problem booting from it. One issue you may have with Squeeze (I recommend Wheezy instead) is that the UUID for / in grub.cfg may be wrong. Simply replace it with the correct one (probably for /dev/md0p1) and everything will work. You will have to repeat this anytime update-grub is run. This is not an issue with Wheezy. --Forwarded Message Attachment-- Date: Wed, 6 Mar 2013 20:03:12 +0100 From: chiendar...@gmail.com Subject: Re: RAID1 all bootable To: s...@rasmussenequipment.com; lsore...@csclub.uwaterloo.ca; debian-am...@lists.debian.org; debian-user@lists.debian.org Sorry, I forgot both the list and the appropriate output; root@....:/home/francesco# fdisk -l Disk /dev/sda: 500.1 GB, 500107862016 bytes 255 heads, 63 sectors/track, 60801 cylinders, total 976773168 sectors Units = sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 4096 bytes I/O size (minimum/optimal): 4096 bytes / 4096 bytes Disk identifier: 0x000f1911 Device Boot Start End Blocks Id System /dev/sda1 * 2048 385023 191488 fd Linux raid autodetect /dev/sda2 385024 976166911 487890944 fd Linux raid autodetect Disk /dev/sdb: 500.1 GB, 500107862016 bytes 255 heads, 63 sectors/track, 60801 cylinders, total 976773168 sectors Units = sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 4096 bytes I/O size (minimum/optimal): 4096 bytes / 4096 bytes Disk identifier: 0x0000cca6 Device Boot Start End Blocks Id System /dev/sdb1 * 2048 385023 191488 fd Linux raid autodetect /dev/sdb2 385024 976166911 487890944 fd Linux raid autodetect Disk /dev/md0: 195 MB, 195887104 bytes 2 heads, 4 sectors/track, 47824 cylinders, total 382592 sectors Units = sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 4096 bytes I/O size (minimum/optimal): 4096 bytes / 4096 bytes Disk identifier: 0x00000000 Disk /dev/md0 doesn't contain a valid partition table Disk /dev/md1: 499.5 GB, 499465912320 bytes 2 heads, 4 sectors/track, 121939920 cylinders, total 975519360 sectors Units = sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 4096 bytes I/O size (minimum/optimal): 4096 bytes / 4096 bytes Disk identifier: 0x00000000 Disk /dev/md1 doesn't contain a valid partition table Disk /dev/mapper/vg1-root: 998 MB, 998244352 bytes 255 heads, 63 sectors/track, 121 cylinders, total 1949696 sectors Units = sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 4096 bytes I/O size (minimum/optimal): 4096 bytes / 4096 bytes Disk identifier: 0x00000000 Disk /dev/mapper/vg1-root doesn't contain a valid partition table Disk /dev/mapper/vg1-swap: 15.0 GB, 14998831104 bytes 255 heads, 63 sectors/track, 1823 cylinders, total 29294592 sectors Units = sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 4096 bytes I/O size (minimum/optimal): 4096 bytes / 4096 bytes Disk identifier: 0x00000000 Disk /dev/mapper/vg1-swap doesn't contain a valid partition table Disk /dev/mapper/vg1-usr: 30.0 GB, 29997662208 bytes 255 heads, 63 sectors/track, 3647 cylinders, total 58589184 sectors Units = sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 4096 bytes I/O size (minimum/optimal): 4096 bytes / 4096 bytes Disk identifier: 0x00000000 Disk /dev/mapper/vg1-usr doesn't contain a valid partition table Disk /dev/mapper/vg1-opt: 9999 MB, 9999220736 bytes 255 heads, 63 sectors/track, 1215 cylinders, total 19529728 sectors Units = sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 4096 bytes I/O size (minimum/optimal): 4096 bytes / 4096 bytes Disk identifier: 0x00000000 Disk /dev/mapper/vg1-opt doesn't contain a valid partition table Disk /dev/mapper/vg1-var: 9999 MB, 9999220736 bytes 255 heads, 63 sectors/track, 1215 cylinders, total 19529728 sectors Units = sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 4096 bytes I/O size (minimum/optimal): 4096 bytes / 4096 bytes Disk identifier: 0x00000000 Disk /dev/mapper/vg1-var doesn't contain a valid partition table Disk /dev/mapper/vg1-tmp: 2998 MB, 2998927360 bytes 255 heads, 63 sectors/track, 364 cylinders, total 5857280 sectors Units = sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 4096 bytes I/O size (minimum/optimal): 4096 bytes / 4096 bytes Disk identifier: 0x00000000 Disk /dev/mapper/vg1-tmp doesn't contain a valid partition table Disk /dev/mapper/vg1-home: 430.0 GB, 430000046080 bytes 255 heads, 63 sectors/track, 52277 cylinders, total 839843840 sectors Units = sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes / 4096 bytes I/O size (minimum/optimal): 4096 bytes / 4096 bytes Disk identifier: 0x00000000 Disk /dev/mapper/vg1-home doesn't contain a valid partition table root@....:/home/francesco# root@.....:/home/francesco# cat /proc/mdstat Personalities : [raid1] md1 : active raid1 sda2[0] sdb2[1] 487759680 blocks super 1.2 [2/2] [UU] md0 : active raid1 sda1[0] sdb1[1] 191296 blocks super 1.2 [2/2] [UU] unused devices: <none> root@.....:/home/francesco# francesco@.....:~$ df -h Filesystem Size Used Avail Use% Mounted on rootfs 938M 185M 705M 21% / udev 10M 0 10M 0% /dev tmpfs 807M 628K 807M 1% /run /dev/mapper/vg1-root 938M 185M 705M 21% / tmpfs 5.0M 0 5.0M 0% /run/lock tmpfs 1.6G 84K 1.6G 1% /run/shm /dev/md0 176M 19M 148M 11% /boot /dev/mapper/vg1-home 395G 284G 91G 76% /home /dev/mapper/vg1-opt 9.2G 1.5G 7.3G 17% /opt /dev/mapper/vg1-tmp 2.8G 69M 2.6G 3% /tmp /dev/mapper/vg1-usr 28G 4.3G 22G 17% /usr /dev/mapper/vg1-var 9.2G 840M 7.9G 10% /var francesco@.....:~$ This is the currest status. "grub-install /dev/sdb" was run with that situation (deriving from install with amd64 wheezy B$ installer downloaded on Feb 1, 2013. That installation ended with grub-install /dev/sda update grub Thanks francesco pietra On Wed, Mar 6, 2013 at 7:02 PM, Shane Johnson <s...@rasmussenequipment.com> wrote: > > > <snip> > >> As far as I can remember, I already posted for this system >> >> root@.....:/home/francesco# cat /proc/mdstat >> Personalities : [raid1] >> md1 : active raid1 sda2[0] sdb2[1] >> 487759680 blocks super 1.2 [2/2] [UU] >> >> md0 : active raid1 sda1[0] sdb1[1] >> 191296 blocks super 1.2 [2/2] [UU] >> >> unused devices: <none> >> root@.....:/home/francesco# >> >> francesco@.....:~$ df -h >> Filesystem Size Used Avail Use% Mounted on >> rootfs 938M 185M 705M 21% / >> udev 10M 0 10M 0% /dev >> tmpfs 807M 628K 807M 1% /run >> /dev/mapper/vg1-root 938M 185M 705M 21% / >> tmpfs 5.0M 0 5.0M 0% /run/lock >> tmpfs 1.6G 84K 1.6G 1% /run/shm >> /dev/md0 176M 19M 148M 11% /boot >> /dev/mapper/vg1-home 395G 284G 91G 76% /home >> /dev/mapper/vg1-opt 9.2G 1.5G 7.3G 17% /opt >> /dev/mapper/vg1-tmp 2.8G 69M 2.6G 3% /tmp >> /dev/mapper/vg1-usr 28G 4.3G 22G 17% /usr >> /dev/mapper/vg1-var 9.2G 840M 7.9G 10% /var >> francesco@.....:~$ >> >> >> the "deadly command' "grub-install /dev/sdb" was run with the system >> started as above. >> >> Thanks >> francesco pietra >> <snip> > > Francesco, > The df -h shows us what is mounted but not if the drives are partitioned or > not. Can you do fdisk -l and send us the output of that? > Also, you replied directly to me without the mailing list. I have included > it in the CC so that everyone can share in the knowledge. Please make sure > you always reply to the list.(Reply-all works real good for this.) > > Thanks > > > -- > Shane D. Johnson > IT Administrator > Rasmussen Equipment > > --Forwarded Message Attachment-- Date: Wed, 6 Mar 2013 14:06:23 -0500 From: garyd...@rogers.com Subject: Re: Raid 5 To: debian-user@lists.debian.org CC: debian-user@lists.debian.org On 06/03/13 01:55 PM, Dick Thomas wrote: > About.me http://about.me/dick.thomas > Blog: www.xpd259.co.uk > G+: www.google.com/profiles/xpd259 > gpg key: C791809B > > > On 6 March 2013 18:47, Adam Wolfe<kadamwo...@gmail.com> wrote: >> I had one h**l of a time doing this over the weekend. >> >> What finally worked for me was creating LOGICAL partitions on each drive and >> setting them as used for RAID volume devices. >> This gave me /dev/sda5, /dev/sdb5 etc etc. >> When grub did it's install, it added all the /dev/sda1 etc partitions and >> rebooted fine. >> >> When I tried primary partitions, grub would just fail and I'd have to >> restart the whole install process over. >> >> NOTE: when booting from the install cd i had to [tab] the 'install' menu >> entry and add "dmraid=true". >> Then after the initial install, it would still fail to boot. >> Back to the install cd, choose 'rescue', [tab] and add 'dmraid=true' again. >> Then get thee to a shell and 'grub-install --recheck /dev/sdaX' for each >> partition. >> >> >> >> >> On 03/06/2013 01:37 PM, Dick Thomas wrote: >>> What is the best way to setup a raid 5 array (4* 2TB drives) >>> should I make raid 5 for my system and /home >>> then raid 0 or 1 for the boot, or should I buy a 5th drive for >>> system/boot and install in the standard way? >>> as this is my 1st time on debian and not sure what would be best >>> >>> >>> Dick Thomas >>> About.me http://about.me/dick.thomas >>> Blog: www.xpd259.co.uk >>> G+: www.google.com/profiles/xpd259 >>> gpg key: C791809B >>> >>> >> >> -- >> To UNSUBSCRIBE, email to debian-user-requ...@lists.debian.org with a subject >> of "unsubscribe". Trouble? Contact listmas...@lists.debian.org >> Archive: http://lists.debian.org/51378f3e.20...@gmail.com >> > > I've tried installing from my "hardware" motherboard raid and that > just fails even with dmraid=true > atm i've got a > raid 0 /boot > > raid 5 encrypted then /lvm > /lvm/swap > /lvm/root > //var/log > > but wasn't sure if that was the way to do it > or am I just confusing matters more :) > Do NOT use the hardware RAID. It's just a crippled form of software RAID. Ignore the advice from Adam Wolfe - it's nonsense. Use the Debian installer (advanced mode) to create the RAID 5 array on drives with just one partition (whole disk) as /dev/md0. Then partition the RAID 5 array into / and /home. Install and reboot. If you are using Wheezy this will work directly. If you are using Squeeze then you may need to fix the UUID in /boot/grub.cfg. I've done this successfully several times. It just works.