On a Luminous 12.2.7 cluster these are the defaults: ceph daemon osd.x config show
"osd_scrub_max_interval": "604800.000000", "osd_scrub_min_interval": "86400.000000", "osd_scrub_interval_randomize_ratio": "0.500000", "osd_scrub_chunk_max": "25", "osd_scrub_chunk_min": "5", "osd_scrub_priority": "5", "osd_scrub_sleep": "0.000000", "osd_deep_scrub_interval": "604800.000000", "osd_deep_scrub_stride": "524288", "osd_disk_thread_ioprio_class": "", "osd_disk_thread_ioprio_priority": "-1", You can check your differences with the defaults using: ceph daemon osd.x config diff Kind regards, Caspar Op di 11 dec. 2018 om 12:36 schreef Janne Johansson <icepic...@gmail.com>: > Den tis 11 dec. 2018 kl 12:26 skrev Caspar Smit <caspars...@supernas.eu>: > > > > Furthermore, presuming you are running Jewel or Luminous you can change > some settings in ceph.conf to mitigate the deep-scrub impact: > > > > osd scrub max interval = 4838400 > > osd scrub min interval = 2419200 > > osd scrub interval randomize ratio = 1.0 > > osd scrub chunk max = 1 > > osd scrub chunk min = 1 > > osd scrub priority = 1 > > osd scrub sleep = 0.1 > > osd deep scrub interval = 2419200 > > osd deep scrub stride = 1048576 > > osd disk thread ioprio class = idle > > osd disk thread ioprio priority = 7 > > > > It would be interesting to see what the defaults for those were, so > one can see which go up and which go down. > > -- > May the most significant bit of your life be positive. >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com