Are your journals on separate disks? What is your ratio of journal disks to
data disks? Are you doing replication size 3 ?

On Mon, Apr 20, 2015 at 9:30 AM, J-P Methot <jpmet...@gtcomm.net> wrote:

> Hi,
>
> This is similar to another thread running right now, but since our current
> setup is completely different from the one described in the other thread, I
> thought it may be better to start a new one.
>
> We are running Ceph Firefly 0.80.8 (soon to be upgraded to 0.80.9). We
> have 6 OSD hosts with 16 OSD each (so a total of 96 OSDs). Each OSD is a
> Samsung SSD 840 EVO on which I can reach write speeds of roughly 400
> MB/sec, plugged in jbod on a controller that can theoretically transfer at
> 6gb/sec. All of that is linked to openstack compute nodes on two bonded
> 10gbps links (so a max transfer rate of 20 gbps).
>
> When I run rados bench from the compute nodes, I reach the network cap in
> read speed. However, write speeds are vastly inferior, reaching about 920
> MB/sec. If I have 4 compute nodes running the write benchmark at the same
> time, I can see the number plummet to 350 MB/sec . For our planned usage,
> we find it to be rather slow, considering we will run a high number of
> virtual machines in there.
>
> Of course, the first thing to do would be to transfer the journal on
> faster drives. However, these are SSDs we're talking about. We don't really
> have access to faster drives. I must find a way to get better write speeds.
> Thus, I am looking for suggestions as to how to make it faster.
>
> I have also thought of options myself like:
> -Upgrading to the latest stable hammer version (would that really give me
> a big performance increase?)
> -Crush map modifications? (this is a long shot, but I'm still using the
> default crush map, maybe there's a change there I could make to improve
> performances)
>
> Any suggestions as to anything else I can tweak would be strongly
> appreciated.
>
> For reference, here's part of my ceph.conf:
>
> [global]
> auth_service_required = cephx
> filestore_xattr_use_omap = true
> auth_client_required = cephx
> auth_cluster_required = cephx
> osd pool default size = 3
>
>
> osd pg bits = 12
> osd pgp bits = 12
> osd pool default pg num = 800
> osd pool default pgp num = 800
>
> [client]
> rbd cache = true
> rbd cache writethrough until flush = true
>
> [osd]
> filestore_fd_cache_size = 1000000
> filestore_omap_header_cache_size = 1000000
> filestore_fd_cache_random = true
> filestore_queue_max_ops = 5000
> journal_queue_max_ops = 1000000
> max_open_files = 1000000
> osd journal size = 10000
>
> --
> ======================
> Jean-Philippe Méthot
> Administrateur système / System administrator
> GloboTech Communications
> Phone: 1-514-907-0050
> Toll Free: 1-(888)-GTCOMM1
> Fax: 1-(514)-907-0750
> jpmet...@gtcomm.net
> http://www.gtcomm.net
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to