You may want to ask your SAN vendor if they have a setting you can make 
to no-op the cache flush. That way you don't have to worry about the 
flush behavior if you change/add different arrays.

Adam N. Copeland wrote:
> Thanks for the replies.
>
> It appears the problem is that we are I/O bound. We have our SAN guy
> looking into possibly moving us to faster spindles. In the meantime, I
> wanted to implement whatever was possible to give us breathing room.
> Turning off atime certainly helped, but we are definitely not completely
> out of the drink yet.
>
> I also found that disabling the ZFS cache flush as per the Evil Tuning
> Guide was a huge boon, considering we're on a battery-backed (non-Sun) SAN.
>
> Thanks,
> Adam
>
> Richard Elling wrote:
>   
>> As it happens, I'm currently involved with a project doing some
>> performance
>> analysis for this... but it is currently a WIP.  Comments below.
>>
>> Robert Milkowski wrote:
>>     
>>> Hello Adam,
>>>
>>> Tuesday, October 21, 2008, 2:00:46 PM, you wrote:
>>>
>>> ANC> We're using a rather large (3.8TB) ZFS volume for our mailstores
>>> on a
>>> ANC> JMS setup. Does anybody have any tips for tuning ZFS for JMS? I'm
>>> ANC> looking for even the most obvious tips, as I am a bit of a
>>> novice. Thanks,
>>>
>>> Well, it's kind of broad topic and it depends on a specific
>>> environment. Then do not tune for the sake of tuning - try to
>>> understand your problem first. Nevertheless you should consider
>>> things like (random order):
>>>
>>> 1. RAID level - you probably will end-up with relatively small random
>>>    IOs - generally avoid RAID-Z
>>>    Of course it could be that RAID-Z in your environment is perfectly
>>>    fine.
>>>   
>>>       
>> There are some write latency-sensitive areas that will begin
>> to cause consternation for large loads.  Storage tuning is very
>> important in this space.  In our case, we're using a ST6540
>> array which has a decent write cache and fast back-end.
>>
>>     
>>> 2. Depending on your workload and disk subsystem ZFS's slog on SSD
>>> could help to improve performance
>>>   
>>>       
>> My experiments show that this is not the main performance
>> issue for large message volumes.
>>
>>     
>>> 3. Disable atime updates on zfs file system
>>>   
>>>       
>> Agree.  JMS doesn't use it, so it just means extra work.
>>
>>     
>>> 4. Enabling compression like lzjb in theory could help - depends on
>>> how weel you data would compress and how much CPU you have left and if
>>> you are mostly IO bond
>>>   
>>>       
>> We have not experimented with this yet, but know that some
>> of the latency-sensitive writes are files with a small number of
>> bytes, which will not compress to be less than one disk block.
>> [opportunities for cleverness are here :-)]
>>
>> There may be a benefit for the message body, but in my tests
>> we are not concentrating on that at this time.
>>
>>     
>>> 5. ZFS recordsize - probably not as in most cases when you read
>>> anything from email you will probably read entire mail anyway.
>>> Nevertheless could be easily checked with dtrace.
>>>   
>>>       
>> This does not seem to be an issue.
>>
>>     
>>> 6. IIRC JMS keeps an index/db file per mailbox - so just maybe L2ARC
>>> on large SSD would help assuming it would nicely cache these files -
>>> would need to be simulated/tested
>>>   
>>>       
>> This does not seem to be an issue, but in our testing the message
>> stores have plenty of memory, and hence, ARC size is on the order
>> of tens of GBytes.
>>
>>     
>>> 7. Disabling vdev pre-fetching in ZFS could help - see ZFS Evile tuning
>>> guide
>>>   
>>>       
>> My experiments showed no benefit by disabling pre-fetch.  However,
>> there are multiple layers of pre-fetching at play when you are using an
>> array, and we haven't done a complete analysis on this yet.  It is clear
>> that we are not bandwidth limited, so prefetching may not hurt.
>>
>>     
>>> Except for #3 and maybe #7 first identify what is your problem and
>>> what are you trying to fix.
>>>
>>>   
>>>       
>> Yep.
>> -- richard
>>
>>     
> _______________________________________________
> zfs-discuss mailing list
> zfs-discuss@opensolaris.org
> http://mail.opensolaris.org/mailman/listinfo/zfs-discuss
>
>   

_______________________________________________
zfs-discuss mailing list
zfs-discuss@opensolaris.org
http://mail.opensolaris.org/mailman/listinfo/zfs-discuss

Reply via email to