The hit set count/period is supposed to control whether the object will be
in the cache pool or in the cold stage pool. By setting to 0, the object is
always promoted. This is good for writings but on my use case, for example,
I wouldn't want every read operation to make an object get promoted and
that is what happens. You can't adjust the warm up. I don't know if I'm
being clear.


Em Ter, 6 de jun de 2017 05:26, TYLin <wooer...@gmail.com> escreveu:

>
> On Jun 6, 2017, at 11:18 AM, jiajia zhong <zhong2p...@gmail.com> wrote:
>
>>
>> it's very similar to ours.  but  is there any need to seperate the osds
> for different pools ? why ?
> below's our crushmap.
>
> -98   6.29997 root tier_cache
> -94   1.39999     host cephn1-ssd
>  95   0.70000         osd.95           up  1.00000          1.00000
> 101   0.34999         osd.101          up  1.00000          1.00000
> 102   0.34999         osd.102          up  1.00000          1.00000
> -95   1.39999     host cephn2-ssd
>  94   0.70000         osd.94           up  1.00000          1.00000
> 103   0.34999         osd.103          up  1.00000          1.00000
> 104   0.34999         osd.104          up  1.00000          1.00000
> -96   1.39999     host cephn3-ssd
> 105   0.34999         osd.105          up  1.00000          1.00000
> 106   0.34999         osd.106          up  1.00000          1.00000
>  93   0.70000         osd.93           up  1.00000          1.00000
> -93   0.70000     host cephn4-ssd
>  97   0.34999         osd.97           up  1.00000          1.00000
>  98   0.34999         osd.98           up  1.00000          1.00000
> -97   1.39999     host cephn5-ssd
>  96   0.70000         osd.96           up  1.00000          1.00000
>  99   0.34999         osd.99           up  1.00000          1.00000
> 100   0.34999         osd.100          up  1.00000          1.00000
>
>
> Because ceph cannot distinguish metadata request and data request. If we
> use same osd sets for both metadata cache and data cache, the bandwidth of
> metadata request may be occupied by data request and lead to long response
> time.
>
> Thanks,
> Ting Yi Lin
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to