Sorry, running snv_123, indiana

On Fri, Oct 23, 2009 at 11:16 AM, Jeremy f <rysh...@gmail.com> wrote:

> What bug# is this under? I'm having what I believe is the same problem. Is
> it possible to just take the mpt driver from a prior build in the time
> being?
> The below is from the load the zpool scrub creates. This is on a dell t7400
> workstation with a 1068E oemed lsi. I updated the firmware to the newest
> available from dell. The errors follow whichever of the 4 drives has the
> highest load.
>
> Streaming doesn't seem to trigger it as I can push 60 MiB a second to a
> mirrored rpool all day, it's only when there are a lot of metadata
> operations.
>
>
> Oct 23 06:25:44 systurbo5 scsi: [ID 107833 kern.warning] WARNING: /p...@0
> ,0/pci8086,4...@9/pci8086,3...@0/pci8086,3...@0/pci1028,2...@0 (mpt0):
> Oct 23 06:25:44 systurbo5       Disconnected command timeout for Target 1
> Oct 23 06:27:15 systurbo5 scsi: [ID 107833 kern.warning] WARNING: /p...@0
> ,0/pci8086,4...@9/pci8086,3...@0/pci8086,3...@0/pci1028,2...@0 (mpt0):
> Oct 23 06:27:15 systurbo5       Disconnected command timeout for Target 1
> Oct 23 06:28:26 systurbo5 scsi: [ID 107833 kern.warning] WARNING: /p...@0
> ,0/pci8086,4...@9/pci8086,3...@0/pci8086,3...@0/pci1028,2...@0 (mpt0):
> Oct 23 06:28:26 systurbo5       Disconnected command timeout for Target 1
> Oct 23 06:29:47 systurbo5 scsi: [ID 107833 kern.warning] WARNING: /p...@0
> ,0/pci8086,4...@9/pci8086,3...@0/pci8086,3...@0/pci1028,2...@0 (mpt0):
> Oct 23 06:29:47 systurbo5       Disconnected command timeout for Target 1
> Oct 23 06:30:58 systurbo5 scsi: [ID 107833 kern.warning] WARNING: /p...@0
> ,0/pci8086,4...@9/pci8086,3...@0/pci8086,3...@0/pci1028,2...@0 (mpt0):
> Oct 23 06:30:58 systurbo5       Disconnected command timeout for Target 1
> Oct 23 06:31:28 systurbo5 scsi: [ID 243001 kern.warning] WARNING: /p...@0
> ,0/pci8086,4...@9/pci8086,3...@0/pci8086,3...@0/pci1028,2...@0 (mpt0):
> Oct 23 06:31:28 systurbo5       mpt_handle_event_sync: IOCStatus=0x8000,
> IOCLogInfo=0x31123000
> Oct 23 06:31:28 systurbo5 scsi: [ID 243001 kern.warning] WARNING: /p...@0
> ,0/pci8086,4...@9/pci8086,3...@0/pci8086,3...@0/pci1028,2...@0 (mpt0):
> Oct 23 06:31:28 systurbo5       mpt_handle_event: IOCStatus=0x8000,
> IOCLogInfo=0x31123000
> Oct 23 06:31:29 systurbo5 scsi: [ID 365881 kern.info] /p...@0
> ,0/pci8086,4...@9/pci8086,3...@0/pci8086,3...@0/pci1028,2...@0 (mpt0):
> Oct 23 06:31:29 systurbo5       Log info 0x31123000 received for target 1.
> Oct 23 06:31:29 systurbo5       scsi_status=0x0, ioc_status=0x804b,
> scsi_state=0xc
> Oct 23 06:31:29 systurbo5 scsi: [ID 365881 kern.info] /p...@0
> ,0/pci8086,4...@9/pci8086,3...@0/pci8086,3...@0/pci1028,2...@0 (mpt0):
> Oct 23 06:31:29 systurbo5       Log info 0x31123000 received for target 1.
> Oct 23 06:31:29 systurbo5       scsi_status=0x0, ioc_status=0x804b,
> scsi_state=0xc
> Oct 23 06:31:29 systurbo5 scsi: [ID 365881 kern.info] /p...@0
> ,0/pci8086,4...@9/pci8086,3...@0/pci8086,3...@0/pci1028,2...@0 (mpt0):
> Oct 23 06:31:29 systurbo5       Log info 0x31123000 received for target 1.
> Oct 23 06:31:29 systurbo5       scsi_status=0x0, ioc_status=0x804b,
> scsi_state=0xc
> Oct 23 06:31:29 systurbo5 scsi: [ID 365881 kern.info] /p...@0
> ,0/pci8086,4...@9/pci8086,3...@0/pci8086,3...@0/pci1028,2...@0 (mpt0):
> Oct 23 06:31:29 systurbo5       Log info 0x31123000 received for target 1.
> Oct 23 06:31:29 systurbo5       scsi_status=0x0, ioc_status=0x804b,
> scsi_state=0xc
>
>
> On Fri, Oct 23, 2009 at 7:13 AM, Adam Cheal <ach...@pnimedia.com> wrote:
>
>> Our config is:
>> OpenSolaris snv_118 x64
>> 1 x LSISAS3801E controller
>> 2 x 23-disk JBOD (fully populated, 1TB 7.2k SATA drives)
>> Each of the two external ports on the LSI connects to a 23-disk JBOD.
>> ZFS-wise we use 1 zpool with 2 x 22-disk raidz2 vdevs (1 vdev per JBOD).
>> Each zpool has one ZFS filesystem containing millions of files/directories.
>> This data is served up via CIFS (kernel), which is why we went with snv_118
>> (first release post-2009.06 that had stable CIFS server). Like I mentioned
>> to James, we know that the server won't be a star performance-wise
>> especially because of the wide vdevs but it shouldn't hiccup under load
>> either. A guaranteed way for us to cause these IO errors is to load up the
>> zpool with about 30 TB of data (90% full) then scrub it. Within 30 minutes
>> we start to see the errors, which usually evolves into "failing" disks
>> (because of excessive retry errors) which just makes things worse.
>> --
>> This message posted from opensolaris.org
>> _______________________________________________
>> zfs-discuss mailing list
>> zfs-discuss@opensolaris.org
>> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>>
>
>
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to