On Tue, 16 Jul 2013, Ta Ba Tuan wrote:
> Hi everyone.
> 
> I have 83 osds, and every osds have same 2TB, (Capacity sumary is 166TB)
> I'm using replicate 3 for pools ('data','metadata').
> 
> But when mounting Ceph filesystem from somewhere (using: mount -t ceph
> Monitor_IP:/ /ceph -o name=admin,secret=xxxxxxxxxx")
> then capacity sumary is showed "160TB"?, I used replicate 3 and I think that
> it must return 160TB/3=50TB?
> 
> Filesystem                Size  Used Avail Use% Mounted on
> 192.168.32.90:/    160T  500G  156T   1%  /tmp/ceph_mount
> 
> Please, explain this  help me?

statfs/df show the raw capacity of the cluster, not the usable capacity.  
How much data you can store is a (potentially) complex function of your 
CRUSH rules and replication layout.  If you store 1TB, you'll notice the 
available space will go down by about 2TB (if you're using the default 
2x).

sage
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to