Ah ok, I wont be using dedup anyway just wanted to try. Ill be adding more ram 
though, I guess you can't have too much. Thanks

Erik Trimble <erik.trim...@oracle.com> wrote:

>You can't really do that.
>
>Adding an SSD for L2ARC will help a bit, but L2ARC storage also consumes
>RAM to maintain a cache table of what's in the L2ARC.  Using 2GB of RAM
>with an SSD-based L2ARC (even without Dedup) likely won't help you too
>much vs not having the SSD. 
>
>If you're going to turn on Dedup, you need at least 8GB of RAM to go
>with the SSD.
>
>-Erik
>
>
>On Tue, 2011-01-18 at 18:35 +0000, Michael Armstrong wrote:
>> Thanks everyone, I think overtime I'm gonna update the system to include an 
>> ssd for sure. Memory may come later though. Thanks for everyone's responses
>> 
>> Erik Trimble <erik.trim...@oracle.com> wrote:
>> 
>> >On Tue, 2011-01-18 at 15:11 +0000, Michael Armstrong wrote:
>> >> I've since turned off dedup, added another 3 drives and results have 
>> >> improved to around 148388K/sec on average, would turning on compression 
>> >> make things more CPU bound and improve performance further?
>> >> 
>> >> On 18 Jan 2011, at 15:07, Richard Elling wrote:
>> >> 
>> >> > On Jan 15, 2011, at 4:21 PM, Michael Armstrong wrote:
>> >> > 
>> >> >> Hi guys, sorry in advance if this is somewhat a lowly question, I've 
>> >> >> recently built a zfs test box based on nexentastor with 4x samsung 2tb 
>> >> >> drives connected via SATA-II in a raidz1 configuration with dedup 
>> >> >> enabled compression off and pool version 23. From running bonnie++ I 
>> >> >> get the following results:
>> >> >> 
>> >> >> Version 1.03b       ------Sequential Output------ --Sequential Input- 
>> >> >> --Random-
>> >> >>                   -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- 
>> >> >> --Seeks--
>> >> >> Machine        Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP  
>> >> >> /sec %CP
>> >> >> nexentastor      4G 60582  54 20502   4 12385   3 53901  57 105290  10 
>> >> >> 429.8   1
>> >> >>                   ------Sequential Create------ --------Random 
>> >> >> Create--------
>> >> >>                   -Create-- --Read--- -Delete-- -Create-- --Read--- 
>> >> >> -Delete--
>> >> >>             files  /sec %CP  /sec %CP  /sec %CP  /sec %CP  /sec %CP  
>> >> >> /sec %CP
>> >> >>                16  7181  29 +++++ +++ +++++ +++ 21477  97 +++++ +++ 
>> >> >> +++++ +++
>> >> >> nexentastor,4G,60582,54,20502,4,12385,3,53901,57,105290,10,429.8,1,16,7181,29,+++++,+++,+++++,+++,21477,97,+++++,+++,+++++,+++
>> >> >> 
>> >> >> 
>> >> >> I'd expect more than 105290K/s on a sequential read as a peak for a 
>> >> >> single drive, let alone a striped set. The system has a relatively 
>> >> >> decent CPU, however only 2GB memory, do you think increasing this to 
>> >> >> 4GB would noticeably affect performance of my zpool? The memory is 
>> >> >> only DDR1.
>> >> > 
>> >> > 2GB or 4GB of RAM + dedup is a recipe for pain. Do yourself a favor, 
>> >> > turn off dedup
>> >> > and enable compression.
>> >> > -- richard
>> >> > 
>> >> 
>> >> _______________________________________________
>> >> zfs-discuss mailing list
>> >> zfs-discuss@opensolaris.org
>> >> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>> >
>> >
>> >Compression will help speed things up (I/O, that is), presuming that
>> >you're not already CPU-bound, which it doesn't seem you are.
>> >
>> >If you want Dedup, you pretty much are required to buy an SSD for L2ARC,
>> >*and* get more RAM.
>> >
>> >
>> >These days, I really don't recommend running ZFS as a fileserver without
>> >a bare minimum of 4GB of RAM (8GB for anything other than light use),
>> >even with Dedup turned off. 
>> >
>> >
>> >-- 
>> >Erik Trimble
>> >Java System Support
>> >Mailstop:  usca22-317
>> >Phone:  x67195
>> >Santa Clara, CA
>> >Timezone: US/Pacific (GMT-0800)
>> >
>-- 
>Erik Trimble
>Java System Support
>Mailstop:  usca22-317
>Phone:  x67195
>Santa Clara, CA
>Timezone: US/Pacific (GMT-0800)
>
_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to