Thanks huang for the reply. Its is the disk compaction taking more time the disk i/o is completely utilized upto 100% looks like both osd_compact_leveldb_on_mount = false & leveldb_compact_on_mount = false isnt working as expected on ceph v10.2.9 is there a way to turn off compaction?
Also, the reason why we are restarting osd's is due to splitting and we increased split multiple and merge_threshold. Is there a way we would inject it? Is osd restarts the only solution? Thanks In Advance in.linkedin.com/in/nikhilravindra On Thu, Mar 28, 2019 at 3:58 PM huang jun <hjwsm1...@gmail.com> wrote: > Did the time really cost on db compact operation? > or you can turn on debug_osd=20 to see what happens, > what about the disk util during start? > > Nikhil R <nikh.ravin...@gmail.com> 于2019年3月28日周四 下午4:36写道: > > > > CEPH osd restarts are taking too long a time > > below is my ceph.conf > > [osd] > > osd_compact_leveldb_on_mount = false > > leveldb_compact_on_mount = false > > leveldb_cache_size=1073741824 > > leveldb_compression = false > > osd_mount_options_xfs = "rw,noatime,inode64,logbsize=256k" > > osd_max_backfills = 1 > > osd_recovery_max_active = 1 > > osd_recovery_op_priority = 1 > > filestore_split_multiple = 72 > > filestore_merge_threshold = 480 > > osd_max_scrubs = 1 > > osd_scrub_begin_hour = 22 > > osd_scrub_end_hour = 3 > > osd_deep_scrub_interval = 2419200 > > osd_scrub_sleep = 0.1 > > > > looks like both osd_compact_leveldb_on_mount = false & > leveldb_compact_on_mount = false isnt working as expected on ceph v10.2.9 > > > > Any ideas on a fix would be appreciated asap > > in.linkedin.com/in/nikhilravindra > > > > _______________________________________________ > > ceph-users mailing list > > ceph-users@lists.ceph.com > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > > -- > Thank you! > HuangJun >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com