Den fre 27 juli 2018 kl 12:24 skrev Anton Aleksandrov <an...@aleksandrov.eu
>:

> Hello,
>
> Might sounds strange, but I could not find answer in google or docs, might
> be called somehow else.
>
> I dont understand pool capacity policy and how to set/define it. I have
> created simple cluster for CephFS on 4 servers, each has 30gb disk - so in
> total 120gb. On top I build replicated metapool with size of 3 and erasure
> pool for data k=2, m=1. Made CephFS, things look good. "ceph df" shows,
> that not all space is used.
>
> *ceph df*
> GLOBAL:
>     SIZE     AVAIL      RAW USED     %RAW USED
>     119G     68941M       53922M         43.89
> POOLS:
>     NAME                ID     USED       %USED     MAX AVAIL     OBJECTS
>     cephfs_data         1      28645M     42.22        39210M       61674
>     cephfs_metadata     2        171M      0.87        19605M        1089
>
> pg_num for medata is 8
> pg_num for data is 40
>
> Am I doing anything wrong? I want as much space for data as possible.
>
>
>
Looks like it says that cephfs_data can write at most 39G, which at 3x
replication makes it consume close to 120G in total. How did this differ
from your expectations?


-- 
May the most significant bit of your life be positive.
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to