On Fri, 18 Jan 2019, 14:46 Marc Roos <m.r...@f1-outsourcing.eu wrote: > > > [@test]# time cat 50b.img > /dev/null > > real 0m0.004s > user 0m0.000s > sys 0m0.002s > [@test]# time cat 50b.img > /dev/null > > real 0m0.002s > user 0m0.000s > sys 0m0.002s > [@test]# time cat 50b.img > /dev/null > > real 0m0.002s > user 0m0.000s > sys 0m0.001s > [@test]# time cat 50b.img > /dev/null > > real 0m0.002s > user 0m0.001s > sys 0m0.001s > [@test]# > > Luminous, centos7.6 kernel cephfs mount, 10Gbit, ssd meta, hdd data, mds > 2,2Ghz >
Did you drop the caches on your client before reading the file? > > > > -----Original Message----- > From: Alexandre DERUMIER [mailto:aderum...@odiso.com] > Sent: 18 January 2019 15:37 > To: Burkhard Linke > Cc: ceph-users > Subject: Re: [ceph-users] CephFS - Small file - single thread - read > performance. > > Hi, > I don't have so big latencies: > > # time cat 50bytesfile > /dev/null > > real 0m0,002s > user 0m0,001s > sys 0m0,000s > > > (It's on an ceph ssd cluster (mimic), kernel cephfs client (4.18), 10GB > network with small latency too, client/server have 3ghz cpus) > > > > ----- Mail original ----- > De: "Burkhard Linke" <burkhard.li...@computational.bio.uni-giessen.de> > À: "ceph-users" <ceph-users@lists.ceph.com> > Envoyé: Vendredi 18 Janvier 2019 15:29:45 > Objet: Re: [ceph-users] CephFS - Small file - single thread - read > performance. > > Hi, > > On 1/18/19 3:11 PM, jes...@krogh.cc wrote: > > Hi. > > > > We have the intention of using CephFS for some of our shares, which > > we'd like to spool to tape as a part normal backup schedule. CephFS > > works nice for large files but for "small" .. < 0.1MB .. there seem to > > > be a "overhead" on 20-40ms per file. I tested like this: > > > > root@abe:/nfs/home/jk# time cat /ceph/cluster/rsyncbackups/13kbfile > > > /dev/null > > > > real 0m0.034s > > user 0m0.001s > > sys 0m0.000s > > > > And from local page-cache right after. > > root@abe:/nfs/home/jk# time cat /ceph/cluster/rsyncbackups/13kbfile > > > /dev/null > > > > real 0m0.002s > > user 0m0.002s > > sys 0m0.000s > > > > Giving a ~20ms overhead in a single file. > > > > This is about x3 higher than on our local filesystems (xfs) based on > > same spindles. > > > > CephFS metadata is on SSD - everything else on big-slow HDD's (in both > > > cases). > > > > Is this what everyone else see? > > > Each file access on client side requires the acquisition of a > corresponding locking entity ('file capability') from the MDS. This adds > an extra network round trip to the MDS. In the worst case the MDS needs > to request a capability release from another client which still holds > the cap (e.g. file is still in page cache), adding another extra network > round trip. > > > CephFS is not NFS, and has a strong consistency model. This comes at a > price. > > > Regards, > > Burkhard > > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com