I'm still using the default values, mostly because I haven't had time to
test.

On Thu, Nov 27, 2014 at 2:44 AM, Andrei Mikhailovsky <and...@arhont.com>
wrote:

> Hi Craig,
>
> Are you keeping the filestore, disk and op threads at their default
> values? or did you also change them?
>
> Cheers
>
>
> Tuning these values depends on a lot more than just the SSDs and HDDs.
> Which kernel and IO scheduler are you using?  Does your HBA do write
> caching?  It also depends on what your goals are.  Tuning for a RadosGW
> cluster is different that for a RDB cluster.  The short answer is that you
> are the only person that can can tell you what your optimal values are.  As
> always, the best benchmark is production load.
>
>
> In my small cluster (5 nodes, 44 osds), I'm optimizing to minimize latency
> during recovery.  When the cluster is healthy, bandwidth and latency are
> more than adequate for my needs.  Even with journals on SSDs, I've found
> that reducing the number of operations and threads has reduced my average
> latency.
>
> I use injectargs to try out new values while I monitor cluster latency.  I
> monitor latency while the cluster is healthy and recovering.  If a change
> is deemed better, only then will I persist the change to ceph.conf.  This
> gives me a fallback that any changes that causes massive problems can be
> undone with a restart or reboot.
>
>
> So far, the configs that I've written to ceph.conf are
> [global]
>   mon osd down out interval = 900
>   mon osd min down reporters = 9
>   mon osd min down reports = 12
>   osd pool default flag hashpspool = true
>
> [osd]
>   osd max backfills = 1
>   osd recovery max active = 1
>   osd recovery op priority = 1
>
>
> I have it on my list to investigate filestore max sync interval.  And now
> that I've pasted that, I need to revisit the min down reports/reporters.  I
> have some nodes with 10 OSDs, and I don't want any one node able to mark
> the rest of the cluster as down (it happened once).
>
>
>
>
> On Sat, Nov 22, 2014 at 6:24 AM, Andrei Mikhailovsky <and...@arhont.com>
> wrote:
>
>> Hello guys,
>>
>> Could some one comment on the optimal or recommended values of various
>> threads values in ceph.conf?
>>
>> At the moment I have the following settings:
>>
>> filestore_op_threads = 8
>> osd_disk_threads = 8
>> osd_op_threads = 8
>> filestore_merge_threshold = 40
>> filestore_split_multiple = 8
>>
>> Are these reasonable for a small cluster made of 7.2K SAS disks with ssd
>> journals with a ratio of 4:1?
>>
>> What are the settings that other people are using?
>>
>> Thanks
>>
>> Andrei
>>
>>
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to