Nah, I would use one Filesystem unless you can’t. The backtrace does create another object but IIRC it’s a maximum one IO per create/rename (on the file). On Wed, Jun 13, 2018 at 1:12 PM Webert de Souza Lima <webert.b...@gmail.com> wrote:
> Thanks for clarifying that, Gregory. > > As said before, we use the file layout to resolve the difference of > workloads in those 2 different directories in cephfs. > Would you recommend using 2 filesystems instead? By doing so, each fs > would have it's default data pool accordingly. > > > Regards, > > Webert Lima > DevOps Engineer at MAV Tecnologia > *Belo Horizonte - Brasil* > *IRC NICK - WebertRLZ* > > > On Wed, Jun 13, 2018 at 11:33 AM Gregory Farnum <gfar...@redhat.com> > wrote: > >> The backtrace object Zheng referred to is used only for resolving hard >> links or in disaster recovery scenarios. If the default data pool isn’t >> available you would stack up pending RADOS writes inside of your mds but >> the rest of the system would continue unless you manage to run the mds out >> of memory. >> -Greg >> On Wed, Jun 13, 2018 at 9:25 AM Webert de Souza Lima < >> webert.b...@gmail.com> wrote: >> >>> Thank you Zheng. >>> >>> Does that mean that, when using such feature, our data integrity relies >>> now on both data pools' integrity/availability? >>> >>> We currently use such feature in production for dovecot's index files, >>> so we could store this directory on a pool of SSDs only. The main data pool >>> is made of HDDs and stores the email files themselves. >>> >>> There ain't too many files created, it's just a few files per email >>> user, and basically one directory per user's mailbox. >>> Each mailbox has a index file that is updated upon every new email >>> received or moved, deleted, read, etc. >>> >>> I think in this scenario the overhead may be acceptable for us. >>> >>> >>> Regards, >>> >>> Webert Lima >>> DevOps Engineer at MAV Tecnologia >>> *Belo Horizonte - Brasil* >>> *IRC NICK - WebertRLZ* >>> >>> >>> On Wed, Jun 13, 2018 at 9:51 AM Yan, Zheng <uker...@gmail.com> wrote: >>> >>>> On Wed, Jun 13, 2018 at 3:34 AM Webert de Souza Lima >>>> <webert.b...@gmail.com> wrote: >>>> > >>>> > hello, >>>> > >>>> > is there any performance impact on cephfs for using file layouts to >>>> bind a specific directory in cephfs to a given pool? Of course, such pool >>>> is not the default data pool for this cephfs. >>>> > >>>> >>>> For each file, no matter which pool file data are stored, mds alway >>>> create an object in the default data pool. The object in default data >>>> pool is used for storing backtrace. So files stored in non-default >>>> pool have extra overhead on file creation. For large file, the >>>> overhead can be neglect. But for lots of small files, the overhead may >>>> affect performance. >>>> >>>> >>>> > Regards, >>>> > >>>> > Webert Lima >>>> > DevOps Engineer at MAV Tecnologia >>>> > Belo Horizonte - Brasil >>>> > IRC NICK - WebertRLZ >>>> > _______________________________________________ >>>> > ceph-users mailing list >>>> > ceph-users@lists.ceph.com >>>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>>> >>> _______________________________________________ >>> ceph-users mailing list >>> ceph-users@lists.ceph.com >>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >>> >>
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com