Am 20.01.13 16:51, schrieb Edward Ned Harvey (opensolarisisdeadlongliveopensolaris):
From: zfs-discuss-boun...@opensolaris.org [mailto:zfs-discuss-
boun...@opensolaris.org] On Behalf Of Stephan Budach

I am always experiencing chksum errors while scrubbing my zpool(s), but
I never experienced chksum errors while resilvering. Does anybody know
why that would be?
When you resilver, you're not reading all the data on all the drives.  Only 
just enough to resilver, which doesn't include all the data that was previously 
in-sync (maybe a little of it, but mostly not).  Even if you have a completely 
failed drive, replaced with a completely new empty drive, if you have a 3-way 
mirror, you only need to read one good copy of the data in order to write the 
resilver'd data onto the new drive.  So you could still be failing to detect 
cksum errors on the *other* side of the mirror, which wasn't read during the 
resilver.

What's more, when you resilver, the system is just going to write the target 
disk.  Not go back and verify every written block of the target disk.

So, think of a scrub as a "complete, thorough, resilver" whereas "resilver" is 
just a lightweight version, doing only the parts that are known to be out-of sync, and without 
subsequent read verification.
Well, I always used to issue a scrub after resilver, but since we completely "re-designed" our server room, things started to act up and each scrub would at least come up with chksum errors. On the Fire 4170 I only noticed these chksum errors, while on the Dell sometimes the whole thing broke down and ZFS would mark numerous disks as faulted.

This happens on all of my servers, Sun Fire 4170M2,
Dell PE 650 and on any FC storage that I have.
While you apparently have been able to keep the system in production for a 
while, consider yourself lucky.  You have a real problem, and solving it 
probably won't be easy.  Your problem is either hardware, firmware, or drivers. 
 If you have a support contract on the Sun, I would recommend starting there.  
Because the Dell is definitely a configuration that you won't find official 
support for - just a lot of community contributors, who will likely not provide 
a super awesome answer for you super soon.

I know, I dedicated quite some of my time to keep this setup up and running. I do have support coverage for my two Sun Solaris servers, but as you may have experienced as well, you're sometimes better off asking here first… ;)

I have gone over our SAN setup/topology and maybe I have found at leats one issue worth looking at: we do have five QLogic 5600 SanBoxes and one of then basically operates as a core switch, were all other ISLs are hooked up, That is, this switch has 4 ISLs and 12 storage array connects, while the Dell sits on another Sanbox and thus all traffic is routed through that switch.

I don't know, but maybe this a bit too much for this setup and the Dell hosts around 240 drives, which are mostly located on a neighbour switch. I will try and tweak this setup such as that the Dell gets a connection on that Sanbox directly which will vastly reduce the inter-switch-traffic.

I am also seeing these warnings in /var/adm/messages on either the Dell and the my new Sun Server X2:

Jan 20 18:22:10 solaris11b scsi: [ID 243001 kern.warning] WARNING: /pci@0,0/pci8086,3c08@3/pci1077,171@0,1/fp@0,0 (fcp0): Jan 20 18:22:10 solaris11b SCSI command to d_id=0x10601 lun=0x0 failed, Bad FCP response values: rsvd1=0, rsvd2=0, sts-rsvd1=0, sts-rsvd2=0, rsplen=0, senselen=0 Jan 20 18:22:10 solaris11b scsi: [ID 243001 kern.warning] WARNING: /pci@0,0/pci8086,3c08@3/pci1077,171@0,1/fp@0,0 (fcp0): Jan 20 18:22:10 solaris11b SCSI command to d_id=0x30e01 lun=0x1 failed, Bad FCP response values: rsvd1=0, rsvd2=0, sts-rsvd1=0, sts-rsvd2=0, rsplen=0, senselen=0

These are always targeted at LUNs on a remote Sanboxes…



_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to