On Wed, 16 Jan 2019, 02:20 David Young <funkypeng...@protonmail.com wrote:
> Hi folks, > > My ceph cluster is used exclusively for cephfs, as follows: > > --- > root@node1:~# grep ceph /etc/fstab > node2:6789:/ /ceph ceph > auto,_netdev,name=admin,secretfile=/root/ceph.admin.secret > root@node1:~# > --- > > "rados df" shows me the following: > > --- > root@node1:~# rados df > POOL_NAME USED OBJECTS CLONES COPIES MISSING_ON_PRIMARY > UNFOUND DEGRADED RD_OPS RD WR_OPS WR > cephfs_metadata 197 MiB 49066 0 98132 0 > 0 0 9934744 55 GiB 57244243 232 GiB > media 196 TiB 51768595 0 258842975 0 > 1 203534 477915206 509 TiB 165167618 292 TiB > > total_objects 51817661 > total_used 266 TiB > total_avail 135 TiB > total_space 400 TiB > root@node1:~# > --- > > But "df" on the mounted cephfs volume shows me: > > --- > root@node1:~# df -h /ceph > Filesystem Size Used Avail Use% Mounted on > 10.20.30.22:6789:/ 207T 196T 11T 95% /ceph > root@node1:~# > --- > > And ceph -s shows me: > > --- > data: > pools: 2 pools, 1028 pgs > objects: 51.82 M objects, 196 TiB > usage: 266 TiB used, 135 TiB / 400 TiB avail > --- > > "media" is an EC pool with size of 5 (4+1), so I can expect 1TB of data to > consume 1.25TB raw space. > > My question is, why does "df" show me I have 11TB free, when "rados df" > shows me I have 135TB (raw) available? > Probabaly because your OSDs are quite unbalanced. What does your 'ceph osd df' look like? > > Thanks! > D > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com