Caching is handled in BlueStore itself, erasure coding happens on a higher layer.
Paul -- Paul Emmerich Looking for help with your Ceph cluster? Contact us at https://croit.io croit GmbH Freseniusstr. 31h 81247 München www.croit.io Tel: +49 89 1896585 90 On Sun, Jun 9, 2019 at 8:43 AM <jes...@krogh.cc> wrote: > Hi. > > I just changed some of my data on CephFS to go to the EC pool instead > of the 3x replicated pool. The data is "write rare / read heavy" data > being served to an HPC cluster. > > To my surprise it looks like the OSD memory caching is done at the > "split object level" not at the "assembled object level", as a > consequence - even though the dataset is fully memory cached it > actually deliveres a very "heavy" cross OSD network traffic to > assemble the objects back. > > Since (as far as I understand) no changes can go the the underlying > object without going though the primary pg - then caching could be > more effectively done at that level. > > The caching on the 3x replica does not retrieve all 3 copies to compare > and verify on a read request (or I at least cannot see any network > traffic supporting that it should be the case). > > Is above configurable? Or would that be a feature/performance request? > > Jesper > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com