Hello Chris, We don't use SSD as journal. each host has one intel E5-2620 CPU which is 6 cores. the networking (both cluster and data networks) is 10Gbps.
My further questions include, (1) osd_mkfs_type = xfs osd_mkfs_options_xfs = -f filestore_xattr_use_omap = true for XFS filesystem, we should not enable filestore_xattr_use_omap = true, is it? (2) filestore_queue_max_ops = 25000 filestore_queue_max_bytes = 10485760 filestore_queue_committing_max_ops = 5000 filestore_queue_committing_max_bytes = 10485760000 journal_max_write_bytes = 1073714824 journal_max_write_entries = 10000 journal_queue_max_ops = 50000 journal_queue_max_bytes = 10485760000 Since we don't have SSD as journals, all these setup are too large? what are the better values? (3) osd_mount_options_xfs = "rw,noexec,nodev,noatime,nodiratime,nobarrier" What's your suggested options here? Thanks a lot. 2016-05-10 15:31 GMT+08:00 Christian Balzer <ch...@gol.com>: > On Tue, 10 May 2016 11:48:07 +0800 Geocast wrote: > > Hello, > > > We have 21 hosts for ceph OSD servers, each host has 12 SATA disks (4TB > > each), 64GB memory. > No journal SSDs? > What CPU(s) and network? > > > ceph version 10.2.0, Ubuntu 16.04 LTS > > The whole cluster is new installed. > > > > Can you help check what the arguments we put in ceph.conf is reasonable > > or not? > > thanks. > > > > [osd] > > osd_data = /var/lib/ceph/osd/ceph-$id > > osd_journal_size = 20000 > Overkill most likely, but not an issue. > > > osd_mkfs_type = xfs > > osd_mkfs_options_xfs = -f > > filestore_xattr_use_omap = true > > filestore_min_sync_interval = 10 > Are you aware what this does and have you actually tested this (IOPS AND > throughput) with various other setting on your hardware to arrive at this > number? > > > filestore_max_sync_interval = 15 > That's fine in and by itself, unlikely to ever be reached anyway. > > > filestore_queue_max_ops = 25000 > > filestore_queue_max_bytes = 10485760 > > filestore_queue_committing_max_ops = 5000 > > filestore_queue_committing_max_bytes = 10485760000 > > journal_max_write_bytes = 1073714824 > > journal_max_write_entries = 10000 > > journal_queue_max_ops = 50000 > > journal_queue_max_bytes = 10485760000 > Same as above, have you tested these setting (from filestore_queue_max_ops > onward) compared to the defaults? > > With HDDs only I'd expect any benefits to be small and/or things to become > very uneven once the HDDs are saturated. > > > osd_max_write_size = 512 > > osd_client_message_size_cap = 2147483648 > > osd_deep_scrub_stride = 131072 > > osd_op_threads = 8 > > osd_disk_threads = 4 > > osd_map_cache_size = 1024 > > osd_map_cache_bl_size = 128 > > osd_mount_options_xfs = "rw,noexec,nodev,noatime,nodiratime,nobarrier" > The nobarrier part is a a potential recipe for disaster unless you have all > on-disk caches disabled and every other cache battery backed. > > The only devices I trust to mount nobarrier are SSDs with powercaps that > have been proven to do the right thing (Intel DC S amongst them). > > > osd_recovery_op_priority = 4 > > osd_recovery_max_active = 10 > > osd_max_backfills = 4 > > > That's sane enough. > > > [client] > > rbd_cache = true > AFAIK that's the case with recent Ceph versions anyway. > > > rbd_cache_size = 268435456 > > Are you sure that you have 256MB per client to waste on RBD cache? > If so, bully for you, but you might find that depending on your use case a > smaller RBD cache but more VM memory (for pagecache, SLAB, etc) could be > more beneficial. > > > rbd_cache_max_dirty = 134217728 > > rbd_cache_max_dirty_age = 5 > > Christian > -- > Christian Balzer Network/Systems Engineer > ch...@gol.com Global OnLine Japan/Rakuten Communications > http://www.gol.com/ >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com