On Sun, Dec 28, 2014 at 1:25 AM, Lindsay Mathieson
<[email protected]> wrote:
> On Sat, 27 Dec 2014 06:02:32 PM you wrote:
>> Are you able to separate log with data in your setup and check the
>> difference?
>
> Do you mean putting the OSD journal on a separate disk? I have the journals on
> SSD partitions, which has helped a lot, previously I was getting 13 MB/s
>

No, I meant XFS journal, as we are speaking about filestore fs performance.

> Its not a good SSD - Samsung 840 EVO :( one of my plans for the new year is to
> get SSD's with better seq write speed and IOPS
>
> I've been trying to figure out if adding more OSD's will improve my
> performance, I only have 2 OSD's (one per node)

Erm, yes. Two OSDs cannot be considered even for a performance
measurement testbed setup, neither should three or any other small
number. This explains numbers you are getting and impact from
nobarrier option.

>
>>  So, depending on type of your benchmark
>> (sync/async/IOPS-/bandwidth-hungry) you may win something just for
>> crossing journal and data between disks (and increase failure domain
>> for a single disk as well  ).
>
> One does tend to foxus on raw seq read/writes for becnhmarking, but my actual
> usage is solely for hosting KVM images, so really random R/W is probably more
> important.

Ok, then my suggestion may not help as much as it can.

>
> --
> Lindsay
_______________________________________________
ceph-users mailing list
[email protected]
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to