I am in the proof-of-concept phase of building a large ZFS/Solaris based SAN box, and am experiencing absolutely poor / unusable performance.
Where to begin... The Hardware setup: Supermicro 4U 24 Drive Bay Chassis Supermicro X8DT3 Server Motherboard 2x Xeon E5520 Nehalem 2.26 Quad Core CPUs 4GB Memory Intel EXPI9404PT 4 port 1000GB Server Network Card (used for ISCSI traffic only) Adaptec 52445 28 Port SATA/SAS Raid Controller connected to 24x Western Digital WD1002FBYS 1TB Enterprise drives. I have configured the 24 drives as single simple volumes in the Adeptec RAID BIOS , and are presenting them to the OS as such. I then, Create a zpool, using raidz2, using all 24 drives, 1 as a hotspare: zpool create tank raidz2 c1t0d0 c1t1d0 [....] c1t22d0 spare c1t23d00 Then create a volume store: zfs create -o canmount=off tank/volumes Then create a 10 TB volume to be presented to our file server: zfs create -V 10TB -o shareiscsi=on tank/volumes/fsrv1data >From here, I discover the iscsi target on our Windows server 2008 R2 File >server, and see the disk is attached in Disk Management. I initialize the >10TB disk fine, and begin to quick format it. Here is where I begin to see >the poor performance issue. The Quick Format took about 45 minutes. And once >the disk is fully mounted, I get maybe 2-5 MB/s average to this disk. I have no clue what I could be doing wrong. To my knowledge, I followed the documentation for setting this up correctly, though I have not looked at any tuning guides beyond the first line saying you shouldn't need to do any of this as the people who picked these defaults know more about it then you. Jumbo Frames are enabled on both sides of the iscsi path, as well as on the switch, and rx/tx buffers increased to 2048 on both sides as well. I know this is not a hardware / iscsi network issue. As another test, I installed Openfiler in a similar configuration (using hardware raid) on this box, and was getting 350-450 MB/S from our fileserver, An "iostat -xndz 1" readout of the "%b% coloum during a file copy to the LUN shows maybe 10-15 seconds of %b at 0 for all disks, then 1-2 seconds of 100, and repeats. Is there anything I need to do to get this usable? Or any additional information I can provide to help solve this problem? As nice as Openfiler is, it doesn't have ZFS, which is necessary to achieve our final goal. -- This message posted from opensolaris.org _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss