Matthew Hagerty wrote:
Can anyone shed some light on this, give me some options to try? What
happened to kernel panics and such when there were serious errors
going on? The only glimmer of information I have is that *one* time
there was an error on the console about there not being any RAID
controller available. I did purchase a spare controller and I'm about
to swap it out and see if it helps, but for some reason I doubt it.
If a controller like that was failing, I would certainly hope to see
some serious error messages or panics going on.
I have been running FreeBSD since version 1.01 and have never had a
box so unstable in the last 12 or so years, especially one that is
supposed to be "server" quality instead of the make-shift ones I put
together with desktop hardware. And last, I'm getting sick of my
Linux admin friends telling me "told you so! should have run
Linux...", please give me something to stick in their pie holes!
Several times now I have had Linux servers (and production quality ones,
not built by me ones :-)) die in a somewhat similar fashion. In every
case the cause has been either a flaky disk or a flaky disk controller,
or some combination.
What seems to happen is that the disk is entirely "lost" by the OS. At
that point any process which never accesses the disk (i.e. is already in
memory) is able to run but the moment any process tries to access the
disk it locks up. So you can't ssh in to the server, but if you happen
to be logged in, you shell is probably cached and keeps working. If you
typed ls recently, you can run ls (but see nothing or get a cryptic
error message like I/O Error), for example.
Clearly nothing is logged as the disk has gone AWOL. Often the machines
behaved fine after a reboot and then did the same some time later. In
one case, the supposedly transparent "RAID-1" array was completely
broken, but Linux logged precisely nothing to tell you :-( You can
stick that where you like in your Linux friends :-O
This somewhat fits with your symptoms. If the disk vanished, then all
those postgres processes would probably fail unless everything they
needed happened to be cached in RAM. The Web server and PHP scripts
probably are cached in RAM if they are called frequently so you might
well see lots of postgres processes stacked up.
LSI MegaRAID has a CLI of sorts in sysutils/megarc. You might start
with that (and check the RAID BIOS next time the machine reboots).
I'd say that if you have an alternative RAID controller that would be a
good place to start. If LSI do any stndalone diagnostics, you could try
those.
--Alex
PS Kernel's usually panic when some internal state is just too wrong to
continue. A disk or even a controller disappearing isn't going to make
the internal state wrong - it's just a device gone missing - so I would
not be surprised if the machine just locked up.
_______________________________________________
freebsd-hackers@freebsd.org mailing list
http://lists.freebsd.org/mailman/listinfo/freebsd-hackers
To unsubscribe, send any mail to "[EMAIL PROTECTED]"