Chris,

Thank you for the heads up. I was going to add another RAID 5 array
with these drives ... shudderz ... I held off because of reliabilty
issues with the 1T Seagates (I already have one of two over heating).
I have had really good luck with the 750 Gig. Hitachis.

All the best, Dan

On Wed, Nov 5, 2008 at 5:45 PM, Chris q <[EMAIL PROTECTED]> wrote:
>
>
> ---------- Forwarded message ----------
> From: Chris <[EMAIL PROTECTED]>
> Date: Wed, Nov 5, 2008 at 5:45 PM
> Subject: Re: [clug-talk] Fwd: Fwd: asus P5Q -Carefule of seagate 1.5 TB
> drives, they are a bit broken
> To: CLUG General <[email protected]>
>
>
> http://forums.seagate.com/stx/board/message?board.id=ata_drives&thread.id=2390&view=by_date_ascending&page=6
> Problem tentatively solved. Any of you thinking about buying the seagate 1.5
> TB drives, wait until they fix this problem.
>
>
> On Wed, Nov 5, 2008 at 4:15 PM, Chris <[EMAIL PROTECTED]> wrote:
>>
>> ok, here's the part that I don't get perhaps you can explain what's going
>> on. it looks like a bunch of cups stuff is happening then my raid array sata
>> links reset,although that might be a coincidence. Is there any reason why
>> the link is resetting? doesn't make sense to me.
>>
>> Nov  5 12:49:12 serverv2 -- MARK --
>> Nov  5 13:09:12 serverv2 -- MARK --
>> Nov  5 13:29:12 serverv2 -- MARK --
>> Nov  5 13:49:12 serverv2 -- MARK --
>> Nov  5 14:09:12 serverv2 -- MARK --
>> Nov  5 14:29:12 serverv2 -- MARK --
>> Nov  5 14:49:12 serverv2 -- MARK --
>> Nov  5 15:09:12 serverv2 -- MARK --
>> Nov  5 15:29:12 serverv2 -- MARK --
>> Nov  5 15:49:12 serverv2 -- MARK --
>> Nov  5 15:50:55 serverv2 python: hp-systray(init)[6671]: warning: No hp:
>> or hpfax: devices found in any installed CUPS queue. Exiting.
>> Nov  5 15:52:07 serverv2 kernel: [12192.326735] type=1503
>> audit(1225925527.559:4): operation="inode_permission" requested_mask="r::"
>> denied_mask="r::" fsuid=7 name="/proc/6778/net/" pid=6778
>> profile="/usr/sbin/cupsd"
>> Nov  5 15:52:08 serverv2 kernel: [12193.222565] type=1503
>> audit(1225925528.454:5): operation="inode_permission" requested_mask="r::"
>> denied_mask="r::" fsuid=7 name="/proc/6782/net/" pid=6782
>> profile="/usr/sbin/cupsd"
>> Nov  5 15:52:08 serverv2 kernel: [12193.222606] type=1503
>> audit(1225925528.454:6): operation="socket_create" family="ax25"
>> sock_type="dgram" protocol=0 pid=6782 profile="/usr/sbin/cupsd"
>> Nov  5 15:52:08 serverv2 kernel: [12193.222614] type=1503
>> audit(1225925528.454:7): operation="socket_create" family="netrom"
>> sock_type="seqpacket" protocol=0 pid=6782 profile="/usr/sbin/cupsd"
>> Nov  5 15:52:08 serverv2 kernel: [12193.222621] type=1503
>> audit(1225925528.454:8): operation="socket_create" family="rose"
>> sock_type="dgram" protocol=0 pid=6782 profile="/usr/sbin/cupsd"
>> Nov  5 15:52:08 serverv2 kernel: [12193.222628] type=1503
>> audit(1225925528.454:9): operation="socket_create" family="ipx"
>> sock_type="dgram" protocol=0 pid=6782 profile="/usr/sbin/cupsd"
>> Nov  5 15:52:08 serverv2 kernel: [12193.222634] type=1503
>> audit(1225925528.454:10): operation="socket_create" family="appletalk"
>> sock_type="dgram" protocol=0 pid=6782 profile="/usr/sbin/cupsd"
>> Nov  5 15:52:08 serverv2 kernel: [12193.222641] type=1503
>> audit(1225925528.454:11): operation="socket_create" family="econet"
>> sock_type="dgram" protocol=0 pid=6782 profile="/usr/sbin/cupsd"
>> Nov  5 15:52:08 serverv2 kernel: [12193.222648] type=1503
>> audit(1225925528.454:12): operation="socket_create" family="ash"
>> sock_type="dgram" protocol=0 pid=6782 profile="/usr/sbin/cupsd"
>> Nov  5 15:52:08 serverv2 kernel: [12193.222654] type=1503
>> audit(1225925528.454:13): operation="socket_create" family="x25"
>> sock_type="seqpacket" protocol=0 pid=6782 profile="/usr/sbin/cupsd"
>> Nov  5 16:05:21 serverv2 kernel: [12986.184075] ata3: hard resetting link
>> Nov  5 16:05:21 serverv2 kernel: [12986.184077] ata4: hard resetting link
>> Nov  5 16:05:21 serverv2 kernel: [12986.668023] ata4: SATA link up 3.0
>> Gbps (SStatus 123 SControl 300)
>> Nov  5 16:05:21 serverv2 kernel: [12986.668709] ata3: SATA link up 3.0
>> Gbps (SStatus 123 SControl 300)
>> Nov  5 16:05:21 serverv2 kernel: [12986.670396] ata4.00: configured for
>> UDMA/133
>> Nov  5 16:05:21 serverv2 kernel: [12986.670419] ata4: EH complete
>> Nov  5 16:05:21 serverv2 kernel: [12986.670494] sd 3:0:0:0: [sdd]
>> 2930277168 512-byte hardware sectors (1500302 MB)
>> Nov  5 16:05:21 serverv2 kernel: [12986.670517] sd 3:0:0:0: [sdd] Write
>> Protect is off
>> Nov  5 16:05:21 serverv2 kernel: [12986.670556] sd 3:0:0:0: [sdd] Write
>> cache: enabled, read cache: enabled, doesn't support DPO or FUA
>> Nov  5 16:05:21 serverv2 kernel: [12986.670941] ata3.00: configured for
>> UDMA/133
>> Nov  5 16:05:21 serverv2 kernel: [12986.670952] ata3: EH complete
>> Nov  5 16:05:21 serverv2 kernel: [12986.670992] sd 2:0:0:0: [sdc]
>> 2930277168 512-byte hardware sectors (1500302 MB)
>> Nov  5 16:05:21 serverv2 kernel: [12986.671012] sd 2:0:0:0: [sdc] Write
>> Protect is off
>> Nov  5 16:05:21 serverv2 kernel: [12986.671050] sd 2:0:0:0: [sdc] Write
>> cache: enabled, read cache: enabled, doesn't support DPO or FUA
>> Nov  5 16:05:21 serverv2 kernel: [12986.682437] md: super_written gets
>> error=-5, uptodate=0
>> Nov  5 16:05:21 serverv2 kernel: [12986.704202] md: super_written gets
>> error=-5, uptodate=0
>> Nov  5 16:05:21 serverv2 kernel: [12986.757591] RAID5 conf printout:
>> Nov  5 16:05:21 serverv2 kernel: [12986.757598]  --- rd:6 wd:4
>> Nov  5 16:05:21 serverv2 kernel: [12986.757601]  disk 0, o:1, dev:sda1
>> Nov  5 16:05:21 serverv2 kernel: [12986.757604]  disk 1, o:1, dev:sdb1
>> Nov  5 16:05:21 serverv2 kernel: [12986.757606]  disk 2, o:0, dev:sdc1
>> Nov  5 16:05:21 serverv2 kernel: [12986.757608]  disk 3, o:0, dev:sdd1
>> Nov  5 16:05:21 serverv2 kernel: [12986.757610]  disk 4, o:1, dev:sde1
>> Nov  5 16:05:21 serverv2 kernel: [12986.757612]  disk 5, o:1, dev:sdf1
>> Nov  5 16:05:22 serverv2 kernel: [12986.769512] RAID5 conf printout:
>> Nov  5 16:05:22 serverv2 kernel: [12986.769520]  --- rd:6 wd:4
>> Nov  5 16:05:22 serverv2 kernel: [12986.769523]  disk 0, o:1, dev:sda1
>> Nov  5 16:05:22 serverv2 kernel: [12986.769525]  disk 1, o:1, dev:sdb1
>> Nov  5 16:05:22 serverv2 kernel: [12986.769527]  disk 2, o:0, dev:sdc1
>> Nov  5 16:05:22 serverv2 kernel: [12986.769529]  disk 4, o:1, dev:sde1
>> Nov  5 16:05:22 serverv2 kernel: [12986.769531]  disk 5, o:1, dev:sdf1
>> Nov  5 16:05:22 serverv2 kernel: [12986.769549] RAID5 conf printout:
>> Nov  5 16:05:22 serverv2 kernel: [12986.769551]  --- rd:6 wd:4
>> Nov  5 16:05:22 serverv2 kernel: [12986.769552]  disk 0, o:1, dev:sda1
>> Nov  5 16:05:22 serverv2 kernel: [12986.769554]  disk 1, o:1, dev:sdb1
>> Nov  5 16:05:22 serverv2 kernel: [12986.769556]  disk 2, o:0, dev:sdc1
>> Nov  5 16:05:22 serverv2 kernel: [12986.769558]  disk 4, o:1, dev:sde1
>> Nov  5 16:05:22 serverv2 kernel: [12986.769560]  disk 5, o:1, dev:sdf1
>> Nov  5 16:05:22 serverv2 kernel: [12986.789508] RAID5 conf printout:
>> Nov  5 16:05:22 serverv2 kernel: [12986.789513]  --- rd:6 wd:4
>> Nov  5 16:05:22 serverv2 kernel: [12986.789516]  disk 0, o:1, dev:sda1
>> Nov  5 16:05:22 serverv2 kernel: [12986.789518]  disk 1, o:1, dev:sdb1
>> Nov  5 16:05:22 serverv2 kernel: [12986.789520]  disk 4, o:1, dev:sde1
>> Nov  5 16:05:22 serverv2 kernel: [12986.789522]  disk 5, o:1, dev:sdf1
>>
>>
>>
>> On Tue, Nov 4, 2008 at 4:57 PM, Mark Carlson <[EMAIL PROTECTED]>
>> wrote:
>>>
>>> I'm sure someone can help you if you simply state why you think your
>>> RAID array is failing.  What makes you think the array is failing?
>>> Can you not access the file system on it?  Do you get an error message
>>> that says that the array is bad?
>>>
>>> As far as I know, you may not even have created a file system on the
>>> array, let alone mounted it.  This is why you need to provide the
>>> steps you took to create the array.  I cannot stress this enough.
>>> Tell us what you did and we can help you.  "I created a software raid
>>> array" does not cut it.  It's not like we need screen shots or
>>> anything.  If you did it using a GUI, tell us what GUI you used and
>>> what buttons you pressed and anything you typed in.  If you used the
>>> command line, tell us what you typed in.  That's it, that's all...
>>> it's a pretty standard thing to do when you need help with a problem,
>>> even non-computer problems.
>>>
>>> What you've done so far is analogous to going to the doctor and
>>> saying: "I'm sick, give me pills that make me feel better."
>>>
>>> -Mark C.
>>>
>>> On 11/4/08, Chris q <[EMAIL PROTECTED]> wrote:
>>> >
>>> >
>>> >
>>> >
>>> > The stuff down at the bottom about the ata fail seemed important. I'm
>>> > hoping
>>> > someone can tell me why the raid array is failing.
>>> >
>>> >
>>> >
>>> >
>>> >  On Tue, Nov 4, 2008 at 2:11 PM, Mark Carlson <[EMAIL PROTECTED]>
>>> > wrote:
>>> >
>>> > >
>>> > >
>>> > > On 11/2/08, Chris q <[EMAIL PROTECTED]> wrote:
>>> > > > Well, I attached an IDE cdrom and it immediately happened again..
>>> > > > here's the /var/log/messages output, I'm hoping you guys can help
>>> > > > me
>>> > figure
>>> > > > this out.
>>> > >
>>> > > Posting the entire contents of /var/log/messages is helpful
>>> > > sometimes... but not right now.  What in /var/log/messages do you
>>> > > think relates to your problems, and why do you think it has something
>>> > > to do with your cdrom drive?
>>> > >
>>> > > I would like to help you solve your problem, but you are missing some
>>> > > key things here.
>>> > >
>>> > > 1. CLEARLY state your problem.  Include any error messages related to
>>> > > the problem, no more and no less.  I still don't understand what your
>>> > > problem and you seem to be withholding information.
>>> > > 2. How are you setting up the software raid?  Describe your steps.
>>> > > This is often the root cause of Linux problems, misconfiguration,
>>> > > rather than hardware or software errors.
>>> > > 3. Have you tried setting up a RAID 0 array instead of RAID 5?  Use
>>> > > two drives instead of all 6.  Maybe one of your drives is bad.
>>> > >
>>> > >
>>> > > -Mark C.
>>>
>>> _______________________________________________
>>> clug-talk mailing list
>>> [email protected]
>>> http://clug.ca/mailman/listinfo/clug-talk_clug.ca
>>> Mailing List Guidelines (http://clug.ca/ml_guidelines.php)
>>> **Please remove these lines when replying
>>
>
>
>
> _______________________________________________
> clug-talk mailing list
> [email protected]
> http://clug.ca/mailman/listinfo/clug-talk_clug.ca
> Mailing List Guidelines (http://clug.ca/ml_guidelines.php)
> **Please remove these lines when replying
>

_______________________________________________
clug-talk mailing list
[email protected]
http://clug.ca/mailman/listinfo/clug-talk_clug.ca
Mailing List Guidelines (http://clug.ca/ml_guidelines.php)
**Please remove these lines when replying

Reply via email to