On Sat, Jul 18, 2015 at 10:25 PM, Nick Fisk <n...@fisk.me.uk> wrote:
> Hi All,
>
> I’m doing some testing on the new High/Low speed cache tiering flushing and 
> I’m trying to get my head round the effect that changing these 2 settings 
> have on the flushing speed.  When setting the osd_agent_max_ops to 1, I can 
> get up to 20% improvement before the osd_agent_max_high_ops value kicks in 
> for high speed flushing. Which is great for bursty workloads.
>
> As I understand it, these settings loosely effect the number of concurrent 
> operations the cache pool OSD’s will flush down to the base pool.
>
> I may have got completely the wrong idea in my head but I can’t understand 
> how a static default setting will work with different cache/base ratios. For 
> example if I had a relatively small number of very fast cache tier OSD’s 
> (PCI-E SSD perhaps) and a much larger number of base tier OSD’s, would the 
> value need to be increased to ensure sufficient utilisation of the base tier 
> and make sure that the cache tier doesn’t fill up too fast?
>
> Alternatively where the cache tier is based on spinning disks or where the 
> base tier is not as comparatively large, this value may need to be reduced to 
> stop it saturating the disks.
>
> Any Thoughts?

I'm not terribly familiar with these exact values, but I think you've
got it right. We can't make decisions at the level of the entire cache
pool (because sharing that information isn't feasible), so we let you
specify it on a per-OSD basis according to what setup you have.

I've no idea if anybody has gathered up a matrix of baseline good
settings or not.
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to