comment below... Janåke Rönnblom wrote: > Hi! > > I have a problem with ZFS and most likely the SATA PCI-X controllers. > I run > opensolaris 2008.11 snv_98 and my hardware is Sun Netra x4200 M2 with > 3 SIL3124 PCI-X with 4 eSATA ports each connected to 3 1U diskchassis > which each hold 4 SATA disks manufactured by Seagate model ES.2 > (500 and 750) for a total of 12 disks. Every disk has its own eSATA > cable > connected to the ports on the PCI-X cards. > > The problem I have is that disk access seems to stop for a few seconds > and > then continue. This happens every few seconds and the end result is > that the > performance is terrible and unusable. > > The idea was to use this box for serving iSCSI to a Windows 2003 > Server. However > with IOmeter on the Windows box and looking at Task manager i noticed > that the > speed pulses from 90% to 0% all the time. Investigating further I > noticed that > I get the same behavior during a simple cp on the localhost. > > /usr/X11/bin/scanpci gives me this information > > pci bus 0x0006 cardnum 0x01 function 0x00: vendor 0x1095 device 0x3124 > Silicon Image, Inc. SiI 3124 PCI-X Serial ATA Controller > > pci bus 0x0084 cardnum 0x01 function 0x00: vendor 0x1095 device 0x3124 > Silicon Image, Inc. SiI 3124 PCI-X Serial ATA Controller > > pci bus 0x0088 cardnum 0x01 function 0x00: vendor 0x1095 device 0x3124 > Silicon Image, Inc. SiI 3124 PCI-X Serial ATA Controller > > c5*, c6* and c7* are the eSATA disks. > > zpool create -f zfsatan mirror c5t0d0 c5t1d0 mirror c5t2d0 c5t3d0 > mirror c6t0d0 c6t1d0 mirror c6t2d0 c6t3d0 mirror c7t0d0 c7t1d0 mirror > c7t2d0 c7t3d0 > zfs create zfsatan/fs01 > > -bash-3.2# time dd if=/dev/zero bs=1024x1024x1024 count=8 of=/zfsatan/ > fs01/storfil > 8+0 records in > 8+0 records out > > real 2m58.863s > user 0m0.001s > sys 0m10.636s > > gives me for this run a 8192/178 gives me around 46MBytes / second... > That is really sucky speed > for 12 drives. However this speed varies since the hangups seems to > occur on random and for a > random time. > > If you look at the output from iostat -cxn 1 below you find that the > first one is okay but the > second on the disks are in 100 %w... and it stays at 100 %w for a few > seconds. > > us sy wt id > 0 34 0 66 > extended device statistics > r/s w/s kr/s kw/s wait actv wsvc_t asvc_t %w %b device > 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c3t0d0 > 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c4t0d0 > 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c0t0d0 > 0.0 400.9 0.0 49560.1 14.1 0.5 35.2 1.2 47 48 c5t0d0 > 0.0 156.0 0.0 18327.1 4.6 0.2 29.5 1.1 17 18 c5t1d0 > 0.0 7.0 0.0 132.0 2.7 0.0 386.0 4.9 56 2 c5t2d0 > 0.0 293.0 0.0 36735.2 13.4 0.3 45.6 1.1 89 34 c5t3d0 > 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c8t0d0 > 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c8t1d0 > 0.0 142.0 0.0 17409.8 4.9 0.2 34.9 1.4 20 20 c6t0d0 > 0.0 350.0 0.0 44030.5 12.6 0.4 36.0 1.3 44 44 c6t1d0 > 0.0 291.0 0.0 34599.7 9.6 0.3 33.1 1.2 34 35 c6t2d0 > 0.0 334.0 0.0 40231.0 11.3 0.4 34.0 1.2 39 40 c6t3d0 > 0.0 241.0 0.0 28210.0 18.1 0.3 75.0 1.1 77 27 c7t0d0 > 0.0 317.0 0.0 38064.8 10.6 0.4 33.4 1.2 38 38 c7t1d0 > 0.0 162.0 0.0 18455.7 4.5 0.2 27.6 1.1 18 18 c7t2d0 > 0.0 162.0 0.0 18455.7 4.5 0.2 27.7 1.1 18 18 c7t3d0 > cpu > us sy wt id > 0 22 0 78 > extended device statistics > r/s w/s kr/s kw/s wait actv wsvc_t asvc_t %w %b device > 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c3t0d0 > 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c4t0d0 > 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c0t0d0 > 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c5t0d0 > 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c5t1d0 > 0.0 0.0 0.0 0.0 5.0 0.0 0.0 0.0 100 0 c5t2d0 > 0.0 0.0 0.0 0.0 5.0 0.0 0.0 0.0 100 0 c5t3d0 > 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c8t0d0 > 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c8t1d0 > 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c6t0d0 > 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c6t1d0 > 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c6t2d0 > 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c6t3d0 > 0.0 0.0 0.0 0.0 21.0 0.0 0.0 0.0 100 0 c7t0d0 > 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c7t1d0 > 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c7t2d0 > 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0 c7t3d0 >
In iostat, the wait, wsvc_t, and %w are for I/Os that are queuing to the HBA. Similarly, the actv, asvc_t, and %b are for I/Os that are queuing to a device. > > Perhaps related bugs: > > Disk access stops for minutes with 100% blocking > http://bugs.opensolaris.org/view_bug.do?bug_id=6544624 > > si3124 driver loses interrupts. > http://bugs.opensolaris.org/view_bug.do?bug_id=6566207 > Based on the above iostat data, I would suspect the HBA. CR 6544624 was marked as a dup of CR 6429205 which was fixed in snv_87. CR 6566207 was fixed in snv_71. There may be a new bug lurking here. -- richard > Any ideas? Should I ditch the SIL3224 cards as they seem to have a bad > rep on this > maillist? I have ordered a Sun SG-XPCI8SAS-E-Z which is an SAS PCI-X > card but it will > cost me a lot more money just without adding any extra benefit... > Except that it > might actually work ;) > > -J > > > ----------------------------------------------------- > Janåke Rönnblom > Phone : +46-910-699 180 > Mobile : 070-397 07 43 > URL : http://www.ronnblom.se > ----------------------------------------------------- > "Those who do not understand Unix are condemned to reinvent it, > poorly." -- Henry Spencer > > > > _______________________________________________ > zfs-discuss mailing list > zfs-discuss@opensolaris.org > http://mail.opensolaris.org/mailman/listinfo/zfs-discuss > _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss