Hello,

This is for a cluster currently running at 14.2.7.  Since our cluster is still relatively small we feel a strong need to run our CephFS on an EC Pool (8 + 2) and Crush Failure Domain = OSD to maximize capacity.
I have read and re-read 
https://docs.ceph.com/docs/nautilus/cephfs/createfs/#creating-pools and 
https://docs.ceph.com/docs/nautilus/cephfs/file-layouts/#file-layouts, 
but it still isn't quite clear to me.  Since this topic is mentioned in 
the release notes for 14.2.8 I thought I should probably ask so I can 
configure this correctly.
If using a small replicated pool as the default data pool, how does one 
use a file layout induce the bulk of the data to be stored in the 
secondary EC data pool?  From the links referenced I infer that a file 
layout is required.  Is it possible to have a file layout based solely 
on size?
BTW, we want to do this in a way that we don't have to think about which 
directory goes with which file size or anything like that.  This needs 
to be an internal detail that is completely hidden from the client.
Also, is it possible to insert a replicated data pool as the default on 
an already deployed CephFS, or will I need to create a new FS and copy 
the data over?
Thanks.

-Dave

--
Dave Hall
Binghamton University
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to