My hardware setup One OSD host - EL6 - 10 spinning disks with configuration - sda (hpsa0): 450GB (0%) RAID-0 == 1 x 450GB 15K SAS/6 - 31GB Memory - 1Gb/s ethernet line
Monitor and gateway hosts have the same configuration with just one disk. I am benchmarking newstore performance with small files using radosgw. I am hitting a bottleneck when writing data though radosgw. I get very good write throughput when using librados to write small files (60K). Regards Srikanth On Wed, May 20, 2015 at 8:03 AM, Mark Nelson <mnel...@redhat.com> wrote: > On 05/19/2015 11:31 AM, Srikanth Madugundi wrote: > >> Hi, >> >> I am seeing write performance hit with small files (60K) using radosgw. >> The radosgw is configured to run with 600 threads. Here is the write >> speed I get with file sizes of 60K >> >> >> # sudo ceph -s >> cluster e445e46e-4d84-4606-9923-16fff64446dc >> health HEALTH_OK >> monmap e1: 1 mons at {osd187=13.24.0.7:6789/0 >> <http://13.24.0.7:6789/0>}, election epoch 1, quorum 0 osd187 >> osdmap e205: 28 osds: 22 up, 22 in >> pgmap v17007: 1078 pgs, 9 pools, 154 GB data, 653 kobjects >> 292 GB used, 8709 GB / 9002 GB avail >> 1078 active+clean >> client io 1117 kB/s rd, *2878 kB/s wr*, 2513 op/s >> > > It appears that you have 22 OSDs and between reads and writes, there are > ~114 ops/s per OSD. How many ops/s per disk are you trying to achieve? > > # >> >> >> If I run the same script with larger file sizes(1MB-3MB), I get a better >> write speed. >> > > Generally larger files will do better for a variety of reasons, but the > primary one is that the data will consistently be more sequentially laid > out. Assuming your OSDs are on spinning disks, this is a big advantage. > > >> >> # sudo ceph -s >> cluster e445e46e-4d84-4606-9923-16fff64446dc >> health HEALTH_OK >> monmap e1: 1 mons at {osd187=13.24.0.79:6789/0 >> <http://13.24.0.79:6789/0>}, election epoch 1, quorum 0 osd187 >> osdmap e205: 28 osds: 22 up, 22 in >> pgmap v16883: 1078 pgs, 9 pools, 125 GB data, 140 kobjects >> 192 GB used, 8809 GB / 9002 GB avail >> 1078 active+clean >> client io *105 MB/s wr*, 1839 op/s >> # >> >> My cluster has 2 OSD hosts running total of 20 osd daemons, 1 mon and 1 >> radosgw hosts. Is the bottleneck coming from the single radosgw process? >> If so, is it possible to run radosgw in multi process mode? >> > > I think before anyone can answer your question, it might help to detail > what your hardware setup is, how you are running the tests, and what kind > of performance you'd like to achieve. > > >> Regards >> Srikanth >> >> >> >> _______________________________________________ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >> _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com