On 2/14/08, Bob Friesenhahn <[EMAIL PROTECTED]> wrote:
>
> Under Solaris 10 on a 4 core Sun Ultra 40 with 20GB RAM, I am setting
> up a Sun StorageTek 2540 with 12 300GB 15K RPM SAS drives and
> connected via load-shared 4Gbit FC links.  This week I have tried many
> different configurations, using firmware managed RAID, ZFS managed
> RAID, and with the controller cache enabled or disabled.
>
> My objective is to obtain the best single-file write performance.
> Unfortunately, I am hitting some sort of write bottleneck and I am not
> sure how to solve it.  I was hoping for a write speed of 300MB/second.
> With ZFS on top of a firmware managed RAID 0 across all 12 drives, I
> hit a peak of 200MB/second.  With each drive exported as a LUN and a
> ZFS pool of 6 pairs, I see a write rate of 154MB/second.  The number
> of drives used has not had much effect on write rate.
>
> Information on my pool is shown at the end of this email.
>
> I am driving the writes using 'iozone' since 'filebench' does not seem
> to want to install/work on Solaris 10.
>
> I am suspecting that the problem is that I am running out of IOPS
> since the drive array indicates a an average IOPS of 214 for one drive
> even though the peak write speed is only 26MB/second (peak read is
> 42MB/second).
>
> Can someone share with me what they think the write bottleneck might
> be and how I can surmount it?
>
> Thanks,
>
> Bob
> ======================================
> Bob Friesenhahn
> [EMAIL PROTECTED], http://www.simplesystems.org/users/bfriesen/
> GraphicsMagick Maintainer,    http://www.GraphicsMagick.org/
>
> % zpool status
>    pool: Sun_2540
>   state: ONLINE
>   scrub: none requested
> config:
>
>          NAME                                       STATE     READ WRITE
> CKSUM
>          Sun_2540                                   ONLINE       0
> 0     0
>            mirror                                   ONLINE       0
> 0     0
>              c4t600A0B80003A8A0B0000096A47B4559Ed0  ONLINE       0
> 0     0
>              c4t600A0B80003A8A0B0000096E47B456DAd0  ONLINE       0
> 0     0
>            mirror                                   ONLINE       0
> 0     0
>              c4t600A0B80003A8A0B0000096147B451BEd0  ONLINE       0
> 0     0
>              c4t600A0B80003A8A0B0000096647B453CEd0  ONLINE       0
> 0     0
>            mirror                                   ONLINE       0
> 0     0
>              c4t600A0B80003A8A0B0000097347B457D4d0  ONLINE       0
> 0     0
>              c4t600A0B800039C9B500000A9C47B4522Dd0  ONLINE       0
> 0     0
>            mirror                                   ONLINE       0
> 0     0
>              c4t600A0B800039C9B500000AA047B4529Bd0  ONLINE       0
> 0     0
>              c4t600A0B800039C9B500000AA447B4544Fd0  ONLINE       0
> 0     0
>            mirror                                   ONLINE       0
> 0     0
>              c4t600A0B800039C9B500000AA847B45605d0  ONLINE       0
> 0     0
>              c4t600A0B800039C9B500000AAC47B45739d0  ONLINE       0
> 0     0
>            mirror                                   ONLINE       0
> 0     0
>              c4t600A0B800039C9B500000AB047B457ADd0  ONLINE       0
> 0     0
>              c4t600A0B800039C9B500000AB447B4595Fd0  ONLINE       0
> 0     0
>
> errors: No known data errors
> freddy:~% zpool iostat
>                 capacity     operations    bandwidth
> pool         used  avail   read  write   read  write
> ----------  -----  -----  -----  -----  -----  -----
> Sun_2540    64.0G  1.57T    808    861  99.8M   105M
> freddy:~% zpool iostat -v
>                                             capacity
> operations    bandwidth
> pool                                     used  avail   read  write
> read  write
>
> --------------------------------------  -----  -----  -----  -----  -----  
> -----
> Sun_2540                                64.0G  1.57T    809    860
> 100M   105M
>    mirror                                10.7G   267G    135    143  16.7M
>   17.6M
>      c4t600A0B80003A8A0B0000096A47B4559Ed0      -      -     66    141
> 8.37M  17.6M
>      c4t600A0B80003A8A0B0000096E47B456DAd0      -      -     67    141
> 8.37M  17.6M
>    mirror                                10.7G   267G    135    143  16.7M
>   17.6M
>      c4t600A0B80003A8A0B0000096147B451BEd0      -      -     66    141
> 8.37M  17.6M
>      c4t600A0B80003A8A0B0000096647B453CEd0      -      -     66    141
> 8.37M  17.6M
>    mirror                                10.7G   267G    134    143  16.7M
>   17.6M
>      c4t600A0B80003A8A0B0000097347B457D4d0      -      -     66    141
> 8.34M  17.6M
>      c4t600A0B800039C9B500000A9C47B4522Dd0      -      -     66    141
> 8.32M  17.6M
>    mirror                                10.7G   267G    134    143  16.6M
>   17.6M
>      c4t600A0B800039C9B500000AA047B4529Bd0      -      -     66    141
> 8.32M  17.6M
>      c4t600A0B800039C9B500000AA447B4544Fd0      -      -     66    141
> 8.30M  17.6M
>    mirror                                10.7G   267G    134    143  16.6M
>   17.6M
>      c4t600A0B800039C9B500000AA847B45605d0      -      -     66    141
> 8.31M  17.6M
>      c4t600A0B800039C9B500000AAC47B45739d0      -      -     66    141
> 8.30M  17.6M
>    mirror                                10.7G   267G    134    143  16.6M
>   17.6M
>      c4t600A0B800039C9B500000AB047B457ADd0      -      -     66    141
> 8.30M  17.6M
>      c4t600A0B800039C9B500000AB447B4595Fd0      -      -     66    141
> 8.29M  17.6M
>
> --------------------------------------  -----  -----  -----  -----  -----  
> -----
>
>
> _______________________________________________
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>


If you're going for best single file write performance, why are you doing
mirrors of the LUNs?  Perhaps I'm misunderstanding why you went from one
giant raid-0 to what is essentially a raid-10.

--Tim
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to