Also examine your network layout.  Any saturation in the private cluster
network or client facing network will be felt in clients / libvirt /
virtual machines

As OSD count increases...

   - Ensure client network private cluster network seperation - different
   nics, different wires, different switches
   - Add more nics both client side and private cluster network side and
   lag them.
   - If/When your dept's budget suddenly swells...implement 10 gig-e.


Monitor, capacity plan, execute  :)

/Chris C

On Tue, Aug 22, 2017 at 3:02 PM, Maged Mokhtar <mmokh...@petasan.org> wrote:

>
>
> It is likely your 2 spinning disks cannot keep up with the load. Things
> are likely to improve if you double your OSDs hooking them up to your
> existing SSD journal. Technically it would be nice to run a
> load/performance tool (either atop/collectl/sysstat) and measure how busy
> your resources are, but it is most likely your 2 spinning disks will show
> near 100% busy utilization.
>
> filestore_max_sync_interval: i do not recommend decreasing this to 0.1, i
> would keep it at 5 sec
>
> osd_op_threads do not increase this unless you have enough cores.
>
> but adding disks is the way to go
>
> Maged
>
>
>
> On 2017-08-22 20:08, fcid wrote:
>
> Hello everyone,
>
> I've been using ceph to provide storage using RBD for 60 KVM virtual
> machines running on proxmox.
>
> The ceph cluster we have is very small (2 OSDs + 1 mon per node, and a
> total of 3 nodes) and we are having some performace issues, like big
> latency times (apply lat:~0.5 s; commit lat: 0.001 s), which get worse by
> the weekly deep-scrubs.
>
> I wonder if doubling the numbers of OSDs would improve latency times, or
> if there is any other configuration tweak recommended for such small
> cluster. Also, I'm looking forward to read any experience of other users
> using a similiar configuration.
>
> Some technical info:
>
>   - Ceph version: 10.2.5
>
>   - OSDs have SSD journal (one SSD disk per 2 OSDs) and have a spindle for
> backend disk.
>
>   - Using CFQ disk queue scheduler
>
>   - OSD configuration excerpt:
>
> osd_recovery_max_active = 1
> osd_recovery_op_priority = 63
> osd_client_op_priority = 1
> osd_mkfs_options = -f -i size=2048 -n size=64k
> osd_mount_options_xfs = inode64,noatime,logbsize=256k
> osd_journal_size = 20480
> osd_op_threads = 12
> osd_disk_threads = 1
> osd_disk_thread_ioprio_class = idle
> osd_disk_thread_ioprio_priority = 7
> osd_scrub_begin_hour = 3
> osd_scrub_end_hour = 8
> osd_scrub_during_recovery = false
> filestore_merge_threshold = 40
> filestore_split_multiple = 8
> filestore_xattr_use_omap = true
> filestore_queue_max_ops = 2500
> filestore_min_sync_interval = 0.01
> filestore_max_sync_interval = 0.1
> filestore_journal_writeahead = true
>
> Best regards,
>
>
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to