Re: New FAQ entry? (was IBM xSeries stop responding during RAID1 reconstruction)

2006-06-21 Thread David Greaves
OK :) David Niccolo Rigacci wrote: > Thanks to the several guys in this list, I have solved my problem > and elaborated this, can be a new FAQ entry? > > > > Q: Sometimes when a RAID volume is resyncing, the system seems to > locks-up: every disk activity is blocked until resync is done. > > A:

Re: Large single raid and XFS or two small ones and EXT3?

2006-06-24 Thread David Greaves
Adam Talbot wrote: > OK, this topic I relay need to get in on. > I have spent the last few week bench marking my new 1.2TB, 6 disk, RAID6 > array. Very interesting. Thanks. Did you get around to any 'tuning'. Things like raid chunk size, external logs for xfs, blockdev readahead on the underlying

Re: raid issues after power failure

2006-07-02 Thread David Greaves
Francois Barre wrote: > 2006/7/1, Ákos Maróy <[EMAIL PROTECTED]>: >> Neil Brown wrote: >> > Try adding '--force' to the -A line. >> > That tells mdadm to try really hard to assemble the array. >> >> thanks, this seems to have solved the issue... >> >> >> Akos >> >> > > Well, Neil, I'm wondering, >

Re: [PATCH] enable auto=yes by default when using udev

2006-07-03 Thread David Greaves
Neil Brown wrote: > I guess I could test for both, but then udev might change > again I'd really like a more robust check. > > Maybe I could test if /dev was a mount point? IIRC you can have diskless machines with a shared root and nfs mounted static /dev/ David -- - To unsubscribe from th

Re: SWRaid Wiki

2006-07-11 Thread David Greaves
Francois Barre wrote: > Hello David, all, > > You pointed the http://linux-raid.osdl.org as a future ressource for > SwRAID and MD knowledge base. Yes. it's not ready for public use yet so I've not announced it formally - I just mention it to people when things pop up. > > In fact, the TODO page

md reports: unknown partition table

2006-07-18 Thread David Greaves
Hi After a powercut I'm trying to mount an array and failing :( teak:~# mdadm --assemble /dev/media --auto=p /dev/sd[bcdef]1 mdadm: /dev/media has been started with 5 drives. Good However: teak:~# mount /media mount: /dev/media1 is not a valid block device teak:~# dd if=/dev/media1 of=/dev/nul

Re: md reports: unknown partition table - fixed.

2006-07-18 Thread David Greaves
David Greaves wrote: > Hi > > After a powercut I'm trying to mount an array and failing :( A reboot after tidying up /dev/ fixed it. The first time through I'd forgotten to update the boot scripts and they were assembling the wrong UUID. That was fine; I realised this

Serious XFS bug in 2.6.17 kernels - FYI

2006-07-20 Thread David Greaves
Just an FYI for my friends here who may be running 2.6.17.x kernels and using XFS and who may not be monitoring lkml :) There is a fairly serious corruption problem that has recently been discussed on lkml and affects all 2.6.17 before -stable .7 (not yet released) Essentially the fs can be corru

Re: host based mirror distance in a fc-based SAN environment

2006-07-26 Thread David Greaves
Stefan Majer wrote: > Hi, > > im curious if there are some numbers out up to which distance its possible > to mirror (raid1) 2 FC-LUNs. We have 2 datacenters with a effective > distance of 11km. The fabrics in one datacenter are connected to the > fabrics in the other datacenter with 5 dark fibre

Re: let md auto-detect 128+ raid members, fix potential race condition

2006-07-31 Thread David Greaves
Alexandre Oliva wrote: > On Jul 30, 2006, Neil Brown <[EMAIL PROTECTED]> wrote: > >> 1/ >> It just isn't "right". We don't mount filesystems from partitions >> just because they have type 'Linux'. We don't enable swap on >> partitions just because they have type 'Linux swap'. So wh

Re: [PATCH] md: new bitmap sysfs interface

2006-08-03 Thread David Greaves
Neil Brown wrote: > write-bits-here-to-dirty-them-in-the-bitmap > > is probably (no, definitely) too verbose. > Any better suggestions? It's not actually a bitmap is it? It takes a number or range and *operates* on a bitmap. so: dirty-chunk-in-bitmap or maybe: dirty-bitmap-chunk David -

Re: raid5/lvm setup questions

2006-08-05 Thread David Greaves
Shane wrote: > Hello all, > > I'm building a new server which will use a number of disks > and am not sure of the best way to go about the setup. > There will be 4 320gb SATA drives installed at first. I'm > just wondering how to set the system up for upgradability. > I'll be using raid5 but no

Re: raid5/lvm setup questions

2006-08-07 Thread David Greaves
Shane wrote: > On Mon, Aug 07, 2006 at 08:57:13PM +0100, Nix wrote: >> On 5 Aug 2006, David Greaves prattled cheerily: >>> As an example of the cons: I've just set up lvm2 over my raid5 and whilst >>> testing snapshots, the first thing that happened was a kernel

Re: raid5/lvm setup questions

2006-08-07 Thread David Greaves
Nix wrote: > On 5 Aug 2006, David Greaves prattled cheerily: that's me :) >> As an example of the cons: I've just set up lvm2 over my raid5 and whilst >> testing snapshots, the first thing that happened was a kernel BUG and an >> oops... > > I've been b

Re: Resize on dirty array?

2006-08-09 Thread David Greaves
No, it wasn't *less* reliable than a single drive; you benefited as soon as a James Peverill wrote: > > In this case the raid WAS the backup... however it seems it turned out > to be less reliable than the single disks it was supporting. In the > future I think I'll make sure my disks have varyin

Re: Kernel RAID support

2006-09-03 Thread David Greaves
Richard Scobie wrote: > Josh Litherland wrote: >> On Sun, 2006-09-03 at 15:56 +1200, Richard Scobie wrote: >> >>> I am building 2.6.18rc5-mm1 and I cannot find the entry under "make >>> config", to enable the various RAID options. >> >> >> Under "Device Drivers", switch on "Multi-device support".

Re: Simulating Drive Failure on Mirrored OS drive

2006-10-02 Thread David Greaves
andy liebman wrote: > I tried simply unplugging one drive from its power and from its SATA > connector. The OS didn't like that at all. My KDE session kept running, > but I could no longer open any new terminals. I couldn't become root in > an existing terminal that was already running. And I could

Re: Recipe for Mirrored OS Drives

2006-10-02 Thread David Greaves
andy liebman wrote: > A few weeks ago, I promised that I would put my "recipe" here for > creating "mirrored OS drives from an existing OS Drive". This "recipe" > combines what I learned from MANY OTHER sometimes conflicting documents > on the same subject -- documents that were probably developed

Re: [PATCH 003 of 6] md: Remove 'experimental' classification from raid5 reshape.

2006-10-02 Thread David Greaves
Typo in first line of this patch :) > I have had enough success reports not^H^H^H to believe that this > is safe for 2.6.19. - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majord

Re: mdadm and raidtools - noob

2006-10-02 Thread David Greaves
Mark Ryden wrote: > Hello linux-raid list, > > I want to create a Linux Software RAID1 on linux FC5 (x86_64), > from SATA II disks. I am a noob in this. No problems. > I looked for it and saw that as far as I understand, > raidtools is quite old - from 2003. > for exanple, http://people.redhat.c

Re: Recipe for Mirrored OS Drives

2006-10-02 Thread David Greaves
andy liebman wrote: > >> >> Feel free to add it here: >> http://linux-raid.osdl.org/index.php/Main_Page >> >> I haven't been able to do much for a few weeks (typical - I find some >> time and >> use it all up just getting the basic setup done - still it's started!) >> >> David >> > > Any hints on

Re: Recipe for Mirrored OS Drives

2006-10-03 Thread David Greaves
Nix wrote: > On 2 Oct 2006, David Greaves spake: >> I suggest you link from http://linux-raid.osdl.org/index.php/RAID_Boot > > The pages don't really have the same purpose. RAID_Boot is `how to boot > your RAID system using initramfs'; this is `how to set up a RAID

Re: Need help recovering a raid5 array

2006-10-24 Thread David Greaves
[EMAIL PROTECTED] wrote: > Hello all, Hi First off, don't do anything else without reading up or talking on here :) The list archive has got a lot of good material - 'help' is usually a good search term!!! > > I had a disk fail in a raid 5 array (4 disk array, no spares), and am > having troub

Re: Raid5 or 6 here... ?

2006-10-24 Thread David Greaves
Gordon Henderson wrote: >1747 ?S< 724:25 [md9_raid5] > > It's kernel 2.6.18 and Wasn't the module merged to raid456 in 2.6.18? Are your mdx_raid6's earlier kernels. My raid 6 is on 2.7.17 and says _raid6 Could it be that the combined kernel thread is called mdX_raid5 David - To

Re: Raid5 or 6 here... ?

2006-10-24 Thread David Greaves
David Greaves wrote: > Gordon Henderson wrote: >>1747 ?S< 724:25 [md9_raid5] >> >> It's kernel 2.6.18 and > > Wasn't the module merged to raid456 in 2.6.18? > > Are your mdx_raid6's earlier kernels. My raid 6 is on 2.7.17 and sa

Re: Relabeling UUID

2006-12-13 Thread David Greaves
Neil Brown wrote: > Patches to the man page to add useful examples are always welcome. And if people would like to be more verbose, the wiki is available at http://linux-raid.osdl.org/ It's now kinda useful but definitely not fully migrated from the old RAID FAQ. David - To unsubscribe from thi

Re: Frequent SATA errors / port timeouts in 2.6.18.3?

2006-12-14 Thread David Greaves
Patrik Jonsson wrote: > Hi all, > this may not be the best list for this question, but I figure that the > number of disks connected to users here should be pretty big... > > I upgraded from 2.6.17-rc4 to 2.6.18.3 about a week ago, and I've since > had 3 drives kicked out of my 10-drive RAID5 arra

Re: mdadm: what if - crashed OS

2007-01-05 Thread David Greaves
Assuming you can allow some downtime, get yourself a rescue CD such as 'RIP' This will let you boot into the machine and run mdadm commands. You don't mention kernel/mdadm versions so you may want to check they're close on the rescue CD. Then try looking at the manpage around --assemble. In part

Re: mdadm --grow failed

2007-02-18 Thread David Greaves
Marc Marais wrote: [snip] > Unfortunately one of the drives timed out during the operation (not a read > error - just a timeout - which I would've thought would be retried but > anyway...): > Help appreciated. (I do have a full backup of course but that's a last > resort with my luck I'd get a r

Re: Manually hacking superblocks

2007-04-13 Thread David Greaves
Lasse Kärkkäinen wrote: > I managed to mess up a RAID-5 array by mdadm -adding a few failed disks > back, trying to get the array running again. Unfortunately, -add didn't > do what I expected, but instead made spares out of the failed disks. The > disks failed due to loose SATA cabling and the dat

Re: Partitioned arrays initially missing from /proc/partitions

2007-04-23 Thread David Greaves
Hi Neil I think this is a bug. Essentially if I create an auto=part md device then I get md_d0p? partitions. If I stop the array and just re-assemble, I don't. It looks like the same (?) problem as Mike (see below - Mike do you have a patch?) but I'm on 2.6.20.7 with mdadm v2.5.6 FWIW I upgrade

Re: Multiple disk failure, but slot numbers are corrupt and preventing assembly.

2007-04-23 Thread David Greaves
There is some odd stuff in there: /dev/sda1: Active Devices : 4 Working Devices : 4 Failed Devices : 0 Events : 0.115909229 /dev/sdb1: Active Devices : 5 Working Devices : 4 Failed Devices : 1 Events : 0.115909230 /dev/sdc1: Active Devices : 8 Working Devices : 8 Failed Devices : 1 Events : 0.11

Re: Multiple disk failure, but slot numbers are corrupt and preventing assembly.

2007-04-24 Thread David Greaves
Leon Woestenberg wrote: > On 4/24/07, Leon Woestenberg <[EMAIL PROTECTED]> wrote: >> Hello, >> >> On 4/23/07, David Greaves <[EMAIL PROTECTED]> wrote: >> > There is some odd stuff in there: >> > >> [EMAIL PROTECTED] ~]# mdadm

Re: Partitioned arrays initially missing from /proc/partitions

2007-04-24 Thread David Greaves
Neil Brown wrote: > This problem is very hard to solve inside the kernel. > The partitions will not be visible until the array is opened *after* > it has been created. Making the partitions visible before that would > be possible, but would be very easy. > > I think the best solution is Mike's so

Re: Partitioned arrays initially missing from /proc/partitions

2007-04-24 Thread David Greaves
Mike Accetta wrote: > David Greaves writes: > > ... >> It looks like the same (?) problem as Mike (see below - Mike do you have a >> patch?) but I'm on 2.6.20.7 with mdadm v2.5.6 > ... > > We have since started assembling the array from the initrd using >

Re: Partitioned arrays initially missing from /proc/partitions

2007-04-24 Thread David Greaves
David Greaves wrote: > currently recompiling the kernel to allow autorun... Which of course won't work because I'm on 1.2 superblocks: md: Autodetecting RAID arrays. md: invalid raid superblock magic on sdb1 md: sdb1 has invalid sb, not importing! md: invalid raid superblock mag

Re: Partitioned arrays initially missing from /proc/partitions

2007-04-24 Thread David Greaves
Neil Brown wrote: > This problem is very hard to solve inside the kernel. > The partitions will not be visible until the array is opened *after* > it has been created. Making the partitions visible before that would > be possible, but would be very easy. > > I think the best solution is Mike's so

Re: Partitioned arrays initially missing from /proc/partitions

2007-04-24 Thread David Greaves
Neil Brown wrote: > On Tuesday April 24, [EMAIL PROTECTED] wrote: >> Neil Brown wrote: >>> This problem is very hard to solve inside the kernel. >>> The partitions will not be visible until the array is opened *after* >>> it has been created. Making the partitions visible before that would >>> be

Re: Partitioned arrays initially missing from /proc/partitions

2007-04-24 Thread David Greaves
Neil Brown wrote: > On Tuesday April 24, [EMAIL PROTECTED] wrote: >> Neil, isn't it easy to just do this after an assemble? > > Yes, but it should not be needed, and I'd like to understand why it > is. > One of the last things do_md_run does is >mddev->changed = 1; > > When you next open /dev

Re: Multiple disk failure, but slot numbers are corrupt and preventing assembly.

2007-04-24 Thread David Greaves
Leon Woestenberg wrote: > David, > > thanks for all the advice so far. No problem :) > In first instance we were searching for ways to tell mdadm what we > know about the array (through mdadm.conf) but from all advice we got > we have to take the 'usual' non-syncing-recreate approach. > > We wi

Re: Multiple disk failure, but slot numbers are corrupt and preventing assembly.

2007-04-26 Thread David Greaves
Bill Davidsen wrote: > Leon Woestenberg wrote: >> We will try to make disk clones first. Will dd suffice or do I need >> something more fancy that maybe copes with source drive read errors in >> a better fashion? > > Yes to both. dd will be fine in most cases, and I suggest using noerror > to con

Re: RAID rebuild on Create

2007-04-30 Thread David Greaves
Jan Engelhardt wrote: > Hi list, > > > when a user does `mdadm -C /dev/md0 -l -n > `, the array gets rebuilt for at least RAID1 and RAID5, even if > the disk contents are most likely not of importance (otherwise we would > not be creating a raid array right now). Could not this needless resy

Re: raid10 on centos 5

2007-05-04 Thread David Greaves
Ruslan Sivak wrote: > So a custom kernel is needed? Is there a way to do a kickstart install > with the new kernel? Or better yet, put it on the install cd? have you tried: modprobe raid10 ? David - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message

Re: Partitioned arrays initially missing from /proc/partitions

2007-05-07 Thread David Greaves
eeded. How do other block devices initialise their partitions on 'discovery'? David David Greaves wrote: > Neil Brown wrote: >> On Tuesday April 24, [EMAIL PROTECTED] wrote: >>> Neil, isn't it easy to just do this after an assemble? >> Yes, but it should not

Re: Swapping out for larger disks

2007-05-08 Thread David Greaves
Brad Campbell wrote: > G'day all, > > I've got 3 arrays here. A 3 drive raid-5, a 10 drive raid-5 and a 15 > drive raid-6. They are all currently 250GB SATA drives. > > I'm contemplating an upgrade to 500GB drives on one or more of the > arrays and wondering the best way to do the physical swap.

Re: removed disk && md-device

2007-05-10 Thread David Greaves
Neil Brown wrote: > On Wednesday May 9, [EMAIL PROTECTED] wrote: >> Neil Brown <[EMAIL PROTECTED]> [2007.04.02.0953 +0200]: >>> Hmmm... this is somewhat awkward. You could argue that udev should be >>> taught to remove the device from the array before removing the device >> >from /dev. But I'm no

Re: removed disk && md-device

2007-05-11 Thread David Greaves
[Repost - didn't seem to make it to the lists, sorry cc's] Sorry, rushed email - it wasn't clear. I think there is something important here though. Oh, it may be worth distinguishing between a drive identifier (/dev/sdb) and a drive slot (md0, slot2). Neil Brown wrote: > On Thursday May 10, [EMAI

Re: removed disk && md-device

2007-05-11 Thread David Greaves
Sorry, rushed email - it wasn't clear. I think there is something important here though. Oh, it may be worth distinguishing between a drive identifier (/dev/sdb) and a drive slot (md0, slot2). Neil Brown wrote: > On Thursday May 10, [EMAIL PROTECTED] wrote: >> Neil Brown wrote: >>> On Wednesday M

Re: removed disk && md-device

2007-05-11 Thread David Greaves
[Another repost - DNS/ISP/MX problems didn't seem to make it to any of the lists or debian bugtrack, sorry cc's] Sorry, rushed email - it wasn't clear. I think there is something important here though. Oh, it may be worth distinguishing between a drive identifier (/dev/sdb) and a drive slot (md0,

Re: how to synchronize two devices (RAID-1, but not really?)

2007-05-15 Thread David Greaves
Tomasz Chmielewski wrote: > Peter Rabbitson schrieb: >> Tomasz Chmielewski wrote: >>> I have a RAID-10 setup of four 400 GB HDDs. As the data grows by several >>> GBs a day, I want to migrate it somehow to RAID-5 on separate disks in a >>> separate machine. >>> >>> Which would be easy, if I didn't

Re: Raid-5 long write wait while reading

2007-05-30 Thread David Greaves
Holger Kiehl wrote: > If you are using ext2/3 check if when creating the filesystem the stride > option helps you, see: > http://tldp.org/HOWTO/Software-RAID-HOWTO-5.html#ss5.11 > > There is a newer howto or wiki but I forgot its location. http://linux-raid.osdl.org/ David - To unsubscribe from

Re: limits on raid

2007-06-16 Thread David Greaves
Neil Brown wrote: On Friday June 15, [EMAIL PROTECTED] wrote: As I understand the way raid works, when you write a block to the array, it will have to read all the other blocks in the stripe and recalculate the parity and write it out. Your u

Re: limits on raid

2007-06-16 Thread David Greaves
[EMAIL PROTECTED] wrote: On Sat, 16 Jun 2007, Neil Brown wrote: I want to test several configurations, from a 45 disk raid6 to a 45 disk raid0. at 2-3 days per test (or longer, depending on the tests) this becomes a very slow process. Are you suggesting the code that is written to enhance data

Re: [linux-lvm] 2.6.22-rc4 XFS fails after hibernate/resume

2007-06-17 Thread David Greaves
David Robinson wrote: David Greaves wrote: This isn't a regression. I was seeing these problems on 2.6.21 (but 22 was in -rc so I waited to try it). I tried 2.6.22-rc4 (with Tejun's patches) to see if it had improved - no. Note this is a different (desktop) machine to that i

Re: [linux-lvm] 2.6.22-rc4 XFS fails after hibernate/resume

2007-06-18 Thread David Greaves
David Greaves wrote: David Robinson wrote: David Greaves wrote: This isn't a regression. I was seeing these problems on 2.6.21 (but 22 was in -rc so I waited to try it). I tried 2.6.22-rc4 (with Tejun's patches) to see if it had improved - no. Note this is a different (desktop)

Re: XFS Tunables for High Speed Linux SW RAID5 Systems?

2007-06-18 Thread David Greaves
David Chinner wrote: On Fri, Jun 15, 2007 at 04:36:07PM -0400, Justin Piszcz wrote: Hi, I was wondering if the XFS folks can recommend any optimizations for high speed disk arrays using RAID5? [sysctls snipped] None of those options will make much difference to performance. mkfs parameters

Re: resync to last 27h - usually 3. what's this?

2007-06-18 Thread David Greaves
Dexter Filmore wrote: 1661 minutes is *way* too long. it's a 4x250GiB sATA array and usually takes 3 hours to resync or check, for that matter. So, what's this? kernel, mdadm verisons? I seem to recall a long fixed ETA calculation bug some time back... David - To unsubscribe from this list:

Re: [linux-lvm] 2.6.22-rc4 XFS fails after hibernate/resume

2007-06-18 Thread David Greaves
at 08:49:34AM +0100, David Greaves wrote: David Greaves wrote: So doing: xfs_freeze -f /scratch sync echo platform > /sys/power/disk echo disk > /sys/power/state # resume xfs_freeze -u /scratch Works (for now - more usage testing tonight) Verrry interesting. Good :) What you were seeing

Re: Software based SATA RAID-5 expandable arrays?

2007-06-19 Thread David Greaves
Dexter Filmore wrote: Why dontcha just cut all the "look how big my ePenis is" chatter and tell us what you wanna do? Nobody gives a rat if your ultra1337 sound cards needs a 10 megawatt power supply. Chill Dexter. How many faults have you seen on this list attributed to poor PSUs? How many

Re: raid5 recover after a 2 disk failure

2007-06-19 Thread David Greaves
Frank Jenkins wrote: So here's the /proc/mdstat prior to the array failure: I'll take a look through this and see if I can see any problems Frank. Bit busy now - give me a few minutes. David - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to

Re: [linux-lvm] 2.6.22-rc5 XFS fails after hibernate/resume

2007-06-19 Thread David Greaves
David Greaves wrote: I'm going to have to do some more testing... done David Chinner wrote: On Mon, Jun 18, 2007 at 08:49:34AM +0100, David Greaves wrote: David Greaves wrote: So doing: xfs_freeze -f /scratch sync echo platform > /sys/power/disk echo disk > /sys/power/sta

Re: raid5 recover after a 2 disk failure

2007-06-19 Thread David Greaves
All looked OK, a few comments... Frank Jenkins wrote: Logically this comes first... This should force the array back into a useable state, yes? (assuming that I'm correct and sde isn't really busted). correct. Since you have a spare you may want to use ddrescue to transfer data from sde to t

Re: [linux-lvm] 2.6.22-rc5 XFS fails after hibernate/resume

2007-06-19 Thread David Greaves
Tejun Heo wrote: Hello, again... David Greaves wrote: Good :) Now, not so good :) Oh, crap. :-) So I hibernated last night and resumed this morning. Before hibernating I froze and sync'ed. After resume I thawed it. (Sorry Dave) Here are some photos of the screen during resume.

Re: [linux-lvm] 2.6.22-rc5 XFS fails after hibernate/resume

2007-06-19 Thread David Greaves
Rafael J. Wysocki wrote: This is on 2.6.22-rc5 Is the Tejun's patch http://www.sisk.pl/kernel/hibernation_and_suspend/2.6.22-rc5/patches/30-block-always-requeue-nonfs-requests-at-the-front.patch applied on top of that? 2.6.22-rc5 includes it. (but, when I was testing rc4, I did apply this

Re: limits on raid

2007-06-21 Thread David Greaves
Neil Brown wrote: This isn't quite right. Thanks :) Firstly, it is mdadm which decided to make one drive a 'spare' for raid5, not the kernel. Secondly, it only applies to raid5, not raid6 or raid1 or raid10. For raid6, the initial resync (just like the resync after an unclean shutdown) reads

Re: limits on raid

2007-06-21 Thread David Greaves
[EMAIL PROTECTED] wrote: On Thu, 21 Jun 2007, David Chinner wrote: one of the 'killer features' of zfs is that it does checksums of every file on disk. so many people don't consider the disk infallable. several other filesystems also do checksums both bitkeeper and git do checksums of files t

Re: [linux-lvm] 2.6.22-rc5 XFS fails after hibernate/resume

2007-06-21 Thread David Greaves
been away, back now... Tejun Heo wrote: David Greaves wrote: Tejun Heo wrote: How reproducible is the problem? Does the problem go away or occur more often if you change the drive you write the memory image to? I don't think there should be activity on the sda drive during resume i

Re: limits on raid

2007-06-22 Thread David Greaves
Neil Brown wrote: On Thursday June 21, [EMAIL PROTECTED] wrote: I didn't get a comment on my suggestion for a quick and dirty fix for -assume-clean issues... Bill Davidsen wrote: How about a simple solution which would get an array on line and still be safe? All it would take is a flag which

Re: limits on raid

2007-06-22 Thread David Greaves
[EMAIL PROTECTED] wrote: On Fri, 22 Jun 2007, David Greaves wrote: That's not a bad thing - until you look at the complexity it brings - and then consider the impact and exceptions when you do, eg hardware acceleration? md information fed up to the fs layer for xfs? simple long

Re: Degraded array on every reboot

2007-06-22 Thread David Greaves
John Hendrikx wrote: I'm not sure why this keeps going wrong, but I do know I made a mistake when initially reconstructing the array. What I did was the following: # mdadm /dev/md1 --add /dev/hde Releazing that I didn't want to add the complete drive (/dev/hde) but only one of its partitions (

Re: limits on raid

2007-06-22 Thread David Greaves
Bill Davidsen wrote: David Greaves wrote: [EMAIL PROTECTED] wrote: On Fri, 22 Jun 2007, David Greaves wrote: If you end up 'fiddling' in md because someone specified --assume-clean on a raid5 [in this case just to save a few minutes *testing time* on system with a heavily choked

Re: mdadm usage: creating arrays with helpful names?

2007-06-27 Thread David Greaves
Richard Michael wrote: How do I create an array with a helpful name? i.e. "/dev/md/storage"? The mdadm man page hints at this in the discussion of the --auto option in the ASSEMBLE MODE section, but doesn't clearly indicate how it's done. Must I create the device nodes by hand first using MAKED

Re: Fastest Chunk Size w/XFS For MD Software RAID = 1024k

2007-06-28 Thread David Greaves
David Chinner wrote: On Wed, Jun 27, 2007 at 07:20:42PM -0400, Justin Piszcz wrote: For drives with 16MB of cache (in this case, raptors). That's four (4) drives, right? I'm pretty sure he's using 10 - email a few days back... Justin Piszcz wrote: Running test with 10 RAPTOR 150 hard drives

Re: mdadm usage: creating arrays with helpful names?

2007-06-28 Thread David Greaves
(back on list for google's benefit ;) and because there are some good questions and I don't know all the answers... ) Oh, and Neil 'cos there may be a bug ... Richard Michael wrote: On Wed, Jun 27, 2007 at 08:49:22AM +0100, David Greaves wrote: http://linux-raid.osd

Re: [linux-pm] Re: [linux-lvm] 2.6.22-rc4 XFS fails after hibernate/resume

2007-06-29 Thread David Greaves
David Chinner wrote: On Fri, Jun 29, 2007 at 12:16:44AM +0200, Rafael J. Wysocki wrote: There are two solutions possible, IMO. One would be to make these workqueues freezable, which is possible, but hacky and Oleg didn't like that very much. The second would be to freeze XFS from within the hib

Re: [linux-pm] Re: [linux-lvm] 2.6.22-rc4 XFS fails after hibernate/resume

2007-06-29 Thread David Greaves
David Chinner wrote: On Fri, Jun 29, 2007 at 08:40:00AM +0100, David Greaves wrote: What happens if a filesystem is frozen and I hibernate? Will it be thawed when I resume? If you froze it yourself, then you'll have to thaw it yourself. So hibernate will not attempt to re-freeze a froz

Re: [linux-lvm] 2.6.22-rc5 XFS fails after hibernate/resume

2007-06-29 Thread David Greaves
David Greaves wrote: been away, back now... again... David Greaves wrote: When I move the swap/resume partition to a different controller (ie when I broke the / mirror and used the freed space) the problem seems to go away. No, it's not gone away - but it's taking longer to show

Re: [linux-pm] Re: [linux-lvm] 2.6.22-rc4 XFS fails after hibernate/resume

2007-06-29 Thread David Greaves
Rafael J. Wysocki wrote: On Friday, 29 June 2007 09:54, David Greaves wrote: David Chinner wrote: On Fri, Jun 29, 2007 at 08:40:00AM +0100, David Greaves wrote: What happens if a filesystem is frozen and I hibernate? Will it be thawed when I resume? If you froze it yourself, then you'll

Re: [linux-lvm] 2.6.22-rc5 XFS fails after hibernate/resume

2007-07-02 Thread David Greaves
Rafael J. Wysocki wrote: On Monday, 2 July 2007 12:56, Tejun Heo wrote: David Greaves wrote: Tejun Heo wrote: It's really weird tho. The PHY RDY status changed events are coming from the device which is NOT used while resuming There is an obvious problem there though Tejun (the errors

Re: [linux-lvm] 2.6.22-rc5 XFS fails after hibernate/resume

2007-07-02 Thread David Greaves
Rafael J. Wysocki wrote: On Monday, 2 July 2007 16:32, David Greaves wrote: Rafael J. Wysocki wrote: On Monday, 2 July 2007 12:56, Tejun Heo wrote: David Greaves wrote: Tejun Heo wrote: It's really weird tho. The PHY RDY status changed events are coming from the device which is NOT

Re: Proposed enhancement to mdadm: Allow "--write-behind=" to be done in grow mode.

2007-07-03 Thread David Greaves
Ian Dall wrote: There doesn't seem to be any designated place to send bug reports and feature requests to mdadm, so I hope I am doing the right thing by sending it here. I have a small patch to mdamd which allows the write-behind amount to be set a array grow time (instead of currently only at g

Re: RAID5 not being reassembled correctly after device swap

2007-07-03 Thread David Greaves
Michael Frotscher wrote: Hello all, I guess you can say that I'm at my wit's end. I really don't get it. An RAID array is suppose to recognize its members purely by its uuid, isn't it? So technically, I can remove a drive from one bus, reconnect it to another giving it a new device name and t

Re: RAID5 not being reassembled correctly after device swap

2007-07-04 Thread David Greaves
Michael Frotscher wrote: I forgot, in case it's of any help. Also do mdadm --examine /dev/hd[abc]3 David - To unsubscribe from this list: send the line "unsubscribe linux-raid" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html

Re: nested raid possible?

2007-07-06 Thread David Greaves
Simon Valiquette wrote: [EMAIL PROTECTED] un jour écrivit: I have a raid-5 array of six 250gb hard drives. I just bought four 500gb hard drives. Simon raised some sensible points. Also consider a raid5 of the 4x500 drives and using LVM to join them to a raid 5 of the 6x250s The benefit usi

Re: mdadm create to existing raid5

2007-07-13 Thread David Greaves
Guy Watkins wrote: } [EMAIL PROTECTED] On Behalf Of Jon Collette } I wasn't thinking and did a mdadm --create to my existing raid5 instead } of --assemble. The syncing process ran and now its not mountable. Is } there anyway to recover from this? Maybe. Not really sure. But don't do anything

Re: mdadm create to existing raid5

2007-07-13 Thread David Greaves
David Greaves wrote: For a simple 4 device array I there are 24 permutations - doable by hand, if you have 5 devices then it's 120, 6 is 720 - getting tricky ;) Oh, wait, for 4 devices there are 24 permutations - and you need to do it 4 times, substituting 'missing' for each

Re: Raid array is not automatically detected.

2007-07-16 Thread David Greaves
Bryan Christ wrote: I do have the type set to 0xfd. Others have said that auto-assemble only works on RAID 0 and 1, but just as Justin mentioned, I too have another box with RAID5 that gets auto assembled by the kernel (also no initrd). I expected the same behavior when I built this array--ag

Re: Raid array is not automatically detected.

2007-07-18 Thread David Greaves
dean gaudet wrote: On Mon, 16 Jul 2007, David Greaves wrote: Bryan Christ wrote: I do have the type set to 0xfd. Others have said that auto-assemble only works on RAID 0 and 1, but just as Justin mentioned, I too have another box with RAID5 that gets auto assembled by the kernel (also no

Re: Raid array is not automatically detected.

2007-07-18 Thread David Greaves
Bryan Christ wrote: I'm now very confused... It's all that top-posting... When I run mdadm --examine /dev/md0 I get the error message: No superblock detected on /dev/md0 However, when I run mdadm -D /dev/md0 the report clearly states "Superblock is persistent"

Re: Software based SATA RAID-5 expandable arrays?

2007-08-01 Thread David Greaves
Michael wrote: I have removed the drives from my machine, the problem Im having is that I dont know the order (ports) they go back into the machine. Does anyone know how to determine the order, or how to fix the drive array if the order is not correct? If you are not attempting a (complex) re

Re: Problem creating software raid - Device or resource busy.

2007-08-03 Thread David Greaves
Basheer wrote: Hi, Hi I'm not familiar with your kernel or mdadm versions - please say. Drive 1 -- /dev/sda1 = swap /dev/sda2 = linux raid (os & data is here) /dev/sda3 = linux raid Are you describing partition types here (ie 0xfd = raid autodetect in fdisk)? Drive 2 -

Re: [RFD] Layering: Use-Case Composers (was: DRBD - what is it, anyways? [compare with e.g. NBD + MD raid])

2007-08-13 Thread David Greaves
Paul Clements wrote: Well, if people would like to see a timeout option, I actually coded up a patch a couple of years ago to do just that, but I never got it into mainline because you can do almost as well by doing a check at user-level (I basically ping the nbd connection periodically and if

Re: [RFD] Layering: Use-Case Composers (was: DRBD - what is it, anyways? [compare with e.g. NBD + MD raid])

2007-08-13 Thread David Greaves
[EMAIL PROTECTED] wrote: per the message below MD (or DM) would need to be modified to work reasonably well with one of the disk components being over an unreliable link (like a network link) are the MD/DM maintainers interested in extending their code in this direction? or would they prefer

Re: [RFD] Layering: Use-Case Composers (was: DRBD - what is it, anyways? [compare with e.g. NBD + MD raid])

2007-08-13 Thread David Greaves
[EMAIL PROTECTED] wrote: Would this just be relevant to network devices or would it improve support for jostled usb and sata hot-plugging I wonder? good question, I suspect that some of the error handling would be similar (for devices that are unreachable not haning the system for example), b

Re: SWAP file on a RAID-10 array possible?

2007-08-15 Thread David Greaves
Tomas France wrote: Hi everyone, I apologize for asking such a fundamental question on the Linux-RAID list but the answers I found elsewhere have been contradicting one another. So, is it possible to have a swap file on a RAID-10 array? yes. mkswap /dev/md swapon /dev/md Should you use RAI

Re: SWAP file on a RAID-10 array possible?

2007-08-15 Thread David Greaves
Tomas France wrote: Thanks for the answer, David! you're welome By the way, does anyone know if there is a comprehensive how-to on software RAID with mdadm available somewhere? I mean a website where I could get answers to questions like "How to convert your system from no RAID to RAID-1, fr

Re: Moving RAID distro

2007-08-15 Thread David Greaves
Richard Grundy wrote: Hello, I was just wonder if it's possible to move my RAID5 array to another distro, same machine just a different flavor of Linux. Yes. The only problem will be if it is the root filesystem (unlikely). Would it just be a case of running: sudo mdadm --create --verbose

Re: 4 Port eSATA RAID5/JBOD PCI-E 8x Controller

2007-08-21 Thread David Greaves
Richard Scobie wrote: This looks like a potentially good, cheap candidate for md use. Although Linux support is not explicitly mentioned, SiI 3124 is used. Thanks Richard. FWIW I find this kind of info useful. David - To uns

Re: Patch for boot-time assembly of v1.x-metadata-based soft (MD) arrays: reasoning and future plans

2007-08-27 Thread David Greaves
Dan Williams wrote: On 8/26/07, Abe Skolnik <[EMAIL PROTECTED]> wrote: Because you can rely on the configuration file to be certain about which disks to pull in and which to ignore. Without the config file the auto-detect routine may not always do the right thing because it will need to make as

Re: Spare disk could not sleep / standby

2005-03-08 Thread David Greaves
Neil Brown wrote: As the event count needs to be updated every time the superblock is modified, the event count will be updated forever active->clean or clean->active transition. All the drives in an array must have the same value for the event count, so the spares need to be updated even though t

  1   2   >