FYI -- Jarek ---------- Forwarded message ---------- From: Jaroslaw Owsiewski <jaroslaw.owsiew...@allegrogroup.com> Date: 2016-02-09 12:00 GMT+01:00 Subject: Re: [ceph-users] Increasing time to save RGW objects To: Wade Holler <wade.hol...@gmail.com>
Hi, For example: # ceph --admin-daemon=ceph-osd.98.asok perf dump generaly: ceph --admin-daemon=/path/to/osd.asok help Best Regards -- Jarek 2016-02-09 11:21 GMT+01:00 Wade Holler <wade.hol...@gmail.com>: > Hi there, > > What is the best way to "look at the rgw admin socket " to see what > operations are taking a long time ? > > Best Regards > Wade > > On Mon, Feb 8, 2016 at 12:16 PM Gregory Farnum <gfar...@redhat.com> wrote: > >> On Mon, Feb 8, 2016 at 8:49 AM, Kris Jurka <ju...@ejurka.com> wrote: >> > >> > I've been testing the performance of ceph by storing objects through >> RGW. >> > This is on Debian with Hammer using 40 magnetic OSDs, 5 mons, and 4 RGW >> > instances. Initially the storage time was holding reasonably steady, >> but it >> > has started to rise recently as shown in the attached chart. >> > >> > The test repeatedly saves 100k objects of 55 kB size using multiple >> threads >> > (50) against multiple RGW gateways (4). It uses a sequential >> identifier as >> > the object key and shards the bucket name using id % 100. The buckets >> have >> > index sharding enabled with 64 index shards per bucket. >> > >> > ceph status doesn't appear to show any issues. Is there something I >> should >> > be looking at here? >> > >> > >> > # ceph status >> > cluster 3fc86d01-cf9c-4bed-b130-7a53d7997964 >> > health HEALTH_OK >> > monmap e2: 5 mons at >> > {condor= >> 192.168.188.90:6789/0,duck=192.168.188.140:6789/0,eagle=192.168.188.100:6789/0,falcon=192.168.188.110:6789/0,shark=192.168.188.118:6789/0 >> } >> > election epoch 18, quorum 0,1,2,3,4 >> > condor,eagle,falcon,shark,duck >> > osdmap e674: 40 osds: 40 up, 40 in >> > pgmap v258756: 3128 pgs, 10 pools, 1392 GB data, 27282 kobjects >> > 4784 GB used, 69499 GB / 74284 GB avail >> > 3128 active+clean >> > client io 268 kB/s rd, 1100 kB/s wr, 493 op/s >> >> It's probably a combination of your bucket indices getting larger and >> your PGs getting split into subfolders on the OSDs. If you keep >> running tests and things get slower it's the first; if they speed >> partway back up again it's the latter. >> Other things to check: >> * you can look at your OSD stores and how the object files are divvied up. >> * you can look at the rgw admin socket and/or logs to see what >> operations are the ones taking time >> * you can check the dump_historic_ops on the OSDs to see if there are >> any notably slow ops >> -Greg >> >> > >> > >> > Kris Jurka >> > >> > _______________________________________________ >> > ceph-users mailing list >> > ceph-users@lists.ceph.com >> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> > >> _______________________________________________ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com