On Tue, Aug 4, 2009 at 9:57 AM, Charles Baker<no-re...@opensolaris.org> wrote: >> My testing has shown some serious problems with the >> iSCSI implementation for OpenSolaris. >> >> I setup a VMware vSphere 4 box with RAID 10 >> direct-attached storage and 3 virtual machines: >> - OpenSolaris 2009.06 (snv_111b) running 64-bit >> - CentOS 5.3 x64 (ran yum update) >> - Ubuntu Server 9.04 x64 (ran apt-get upgrade) >> >> I gave each virtual 2 GB of RAM, a 32 GB drive and >> setup a 16 GB iSCSI target on each (the two Linux vms >> used iSCSI Enterprise Target 0.4.16 with blockio). >> VMware Tools was installed on each. No tuning was >> done on any of the operating systems. >> >> I ran two tests for write performance - one one the >> server itself and one from my Mac connected via >> Gigabit (mtu of 1500) iSCSI connection using >> globalSAN’s latest initiator. >> >> Here’s what I used on the servers: >> time dd if=/dev/zero of=/root/testfile bs=1048576k >> count=4 >> and the Mac OS with the iSCSI connected drive >> (formatted with GPT / Mac OS Extended journaled): >> time dd if=/dev/zero of=/Volumes/test/testfile >> bs=1048576k count=4 >> >> The results were very interesting (all calculations >> using 1 MB = 1,084,756 bytes) >> >> For OpenSolaris, the local write performance averaged >> 86 MB/s. I turned on lzjb compression for rpool (zfs >> set compression=lzjb rpool) and it went up to 414 >> MB/s since I’m writing zeros). The average >> performance via iSCSI was an abysmal 16 MB/s (even >> with compression turned on - with it off, 13 MB/s). >> >> For CentOS (ext3), local write performance averaged >> 141 MB/s. iSCSI performance was 78 MB/s (almost as >> fast as local ZFS performance on the OpenSolaris >> server when compression was turned off). >> >> Ubuntu Server (ext4) had 150 MB/s for the local >> write. iSCSI performance averaged 80 MB/s. >> >> One of the main differences between the three virtual >> machines was that the iSCSI target on the Linux >> machines used partitions with no file system. On >> OpenSolaris, the iSCSI target created sits on top of >> ZFS. That creates a lot of overhead (although you do >> get some great features). >> >> Since all the virtual machines were connected to the >> same switch (with the same MTU), had the same amount >> of RAM, used default configurations for the operating >> systems, and sat on the same RAID 10 storage, I’d say >> it was a pretty level playing field. >> >> While jumbo frames will help iSCSI performance, it >> won’t overcome inherit limitations of the iSCSI >> target’s implementation.
If you want to host your VMs from Solaris (Open or not) use NFS right now as the iSCSI implementation is still quite a bit immature and won't perform nearly as good as the Linux implementation. Until comstar stabilizes and replaces iscsitgt I would hold off on iSCSI on Solaris. -Ross _______________________________________________ zfs-discuss mailing list zfs-discuss@opensolaris.org http://mail.opensolaris.org/mailman/listinfo/zfs-discuss