On 3/10/15 11:06, Mateusz Skała wrote:

Hi,

In my cluster is something wrong with free space. In cluster with 10OSD (5*1TB + 5*2TB) ‘ceph –s’ shows:

11425 GB used, 2485 GB / 13910 GB avail

But I have only 2 rbd disks in one pool (‘rbd’):

>>rados df

pool name category KB objects clones degraded unfound rd rd KB wr wr KB

rbd - 3976154023 971434 0 6474 0 11542224 1391869743 742847 385900453

  total used     11988041672       971434

  total avail     2598378648

  total space    14586420320

>>rbd ls

vm-100-disk-1

vm-100-disk-2

>>rbd info vm-100-disk-1

rbd image 'vm-100-disk-1':

        size 16384 MB in 4096 objects

        order 22 (4096 kB objects)

        block_name_prefix: rbd_data.14ef2ae8944a

        format: 2

        features: layering

>>rbd info vm-100-disk-2

rbd image 'vm-100-disk-2':

        size 4096 GB in 1048576 objects

        order 22 (4096 kB objects)

        block_name_prefix: rbd_data.15682ae8944a

        format: 2

        features: layering

So my rbd disks use only 4112GB. Default size of cluster is 2 so used space should be ca 8224GB, why ceph –s shows 11425 GB ?

Can someone explain this situation?

Thanks, Mateusz

Hey,

what does "ceph df" show?

ceph -s shows raw disk usage so there will be some overhead from file system, also maybe you left some files there?




_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to