Hello I used real world type data, which is to say rsync, dd , scp, etc.. In addition to simple performance of the VMs themselves (I am using this Ceph cluster as backend to KVM) .
Regardless of which method I used I averaged between 90 to 100MB. On Tue, Nov 24, 2015 at 8:47 AM, Mart van Santen <m...@greenhost.nl> wrote: > Dear Marek, > > I would expect a higher performance, but did you measure this? with rados > bench? Please note, ceph is build for parallel access, so the combined > speed increases with more threads, if this is a single thread measurement, > I wonder how well it reflects the performance of the platform. With rados > bench you can specify how many threads you want to use. > > Regards, > > Mart > > > > > On 11/24/2015 04:37 PM, Marek Dohojda wrote: > > Yeah they are, that is one thing I was planning on changing, What I am > really interested at the moment, is vague expected performance. I mean is > 100MB around normal, very low, or "could be better"? > > On Tue, Nov 24, 2015 at 8:02 AM, Alan Johnson <al...@supermicro.com> > wrote: > >> Are the journals on the same device – it might be better to use the SSDs >> for journaling since you are not getting better performance with SSDs? >> >> >> >> *From:* ceph-users [mailto:ceph-users-boun...@lists.ceph.com] *On Behalf >> Of *Marek Dohojda >> *Sent:* Monday, November 23, 2015 10:24 PM >> *To:* Haomai Wang >> *Cc:* ceph-users@lists.ceph.com >> *Subject:* Re: [ceph-users] Performance question >> >> >> >> Sorry I should have specified SAS is the 100 MB :) , but to be honest >> SSD isn't much faster. >> >> >> >> On Mon, Nov 23, 2015 at 7:38 PM, Haomai Wang < <haomaiw...@gmail.com> >> haomaiw...@gmail.com> wrote: >> >> On Tue, Nov 24, 2015 at 10:35 AM, Marek Dohojda >> <mdoho...@altitudedigital.com> wrote: >> > No SSD and SAS are in two separate pools. >> > >> > On Mon, Nov 23, 2015 at 7:30 PM, Haomai Wang <haomaiw...@gmail.com> >> wrote: >> >> >> >> On Tue, Nov 24, 2015 at 10:23 AM, Marek Dohojda >> >> <mdoho...@altitudedigital.com> wrote: >> >> > I have a Hammer Ceph cluster on 7 nodes with total 14 OSDs. 7 of >> which >> >> > are >> >> > SSD and 7 of which are SAS 10K drives. I get typically about 100MB >> IO >> >> > rates >> >> > on this cluster. >> >> So which pool you get with 100 MB? >> >> >> >> >> >> You mixed up sas and ssd in one pool? >> >> >> >> > >> >> > I have a simple question. Is 100MB within my configuration what I >> >> > should >> >> > expect, or should it be higher? I am not sure if I should be looking >> for >> >> > issues, or just accept what I have. >> >> > >> >> > _______________________________________________ >> >> > ceph-users mailing list >> >> > ceph-users@lists.ceph.com >> >> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >> > >> >> >> >> >> >> >> >> -- >> >> Best Regards, >> >> >> >> Wheat >> > >> > >> >> >> -- >> Best Regards, >> >> Wheat >> >> >> > > > > _______________________________________________ > ceph-users mailing > listceph-us...@lists.ceph.comhttp://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > -- > Mart van Santen > Greenhost > E: m...@greenhost.nl > T: +31 20 4890444 > W: https://greenhost.nl > > A PGP signature can be attached to this e-mail, > you need PGP software to verify it. > My public key is available in keyserver(s) > see: http://tinyurl.com/openpgp-manual > > PGP Fingerprint: CA85 EB11 2B70 042D AF66 B29A 6437 01A1 10A3 D3A5 > > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com