On 05/19/2015 11:31 AM, Srikanth Madugundi wrote:
Hi,
I am seeing write performance hit with small files (60K) using radosgw.
The radosgw is configured to run with 600 threads. Here is the write
speed I get with file sizes of 60K
# sudo ceph -s
cluster e445e46e-4d84-4606-9923-16fff64446dc
health HEALTH_OK
monmap e1: 1 mons at {osd187=13.24.0.7:6789/0
<http://13.24.0.7:6789/0>}, election epoch 1, quorum 0 osd187
osdmap e205: 28 osds: 22 up, 22 in
pgmap v17007: 1078 pgs, 9 pools, 154 GB data, 653 kobjects
292 GB used, 8709 GB / 9002 GB avail
1078 active+clean
client io 1117 kB/s rd, *2878 kB/s wr*, 2513 op/s
It appears that you have 22 OSDs and between reads and writes, there are
~114 ops/s per OSD. How many ops/s per disk are you trying to achieve?
#
If I run the same script with larger file sizes(1MB-3MB), I get a better
write speed.
Generally larger files will do better for a variety of reasons, but the
primary one is that the data will consistently be more sequentially laid
out. Assuming your OSDs are on spinning disks, this is a big advantage.
# sudo ceph -s
cluster e445e46e-4d84-4606-9923-16fff64446dc
health HEALTH_OK
monmap e1: 1 mons at {osd187=13.24.0.79:6789/0
<http://13.24.0.79:6789/0>}, election epoch 1, quorum 0 osd187
osdmap e205: 28 osds: 22 up, 22 in
pgmap v16883: 1078 pgs, 9 pools, 125 GB data, 140 kobjects
192 GB used, 8809 GB / 9002 GB avail
1078 active+clean
client io *105 MB/s wr*, 1839 op/s
#
My cluster has 2 OSD hosts running total of 20 osd daemons, 1 mon and 1
radosgw hosts. Is the bottleneck coming from the single radosgw process?
If so, is it possible to run radosgw in multi process mode?
I think before anyone can answer your question, it might help to detail
what your hardware setup is, how you are running the tests, and what
kind of performance you'd like to achieve.
Regards
Srikanth
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com