Hi,
take a look with:

rados df
rbd -p <pool> ls

and with -l option for long output like
rbd -p rbd ls -l
NAME             SIZE PARENT FMT PROT LOCK
vm-127-disk-1  35000M          2          
vm-131-disk-1   8192M          2          
vm-135-disk-1   8192M          2
...

Udo


On 19.06.2014 09:42, wsnote wrote:
> Lewis, thanks very much!
> With your suggestion, I have installed ceph successfully and started
> it normally.
> Now I find another question.When I start ceph cluster, it has used
> 106TB space.
> What's the thing that take so much space?
>
> command: ceph -s
> [root@yun4 ~]# ceph -s
>     cluster cbf88224-1c9f-4d90-9e78-ef72d63ddce6
>      health HEALTH_OK
>      monmap e1: 10 mons at
> {1=222.186.55.4:6789/0,10=222.186.55.13:6789/0,2=222.186.55.5:6789/0,3=222.186.55.6:6789/0,4=222.186.55.7:6789/0,5=222.186.55.8:6789/0,6=222.186.55.9:6789/0,7=222.186.55.10:6789/0,8=222.186.55.11:6789/0,9=222.186.55.12:6789/0},
> election epoch 8, quorum 0,1,2,3,4,5,6,7,8,9 1,2,3,4,5,6,7,8,9,10
>      osdmap e63: 109 osds: 109 up, 109 in
>       pgmap v462: 21312 pgs, 3 pools, 0 bytes data, 0 objects
> *            106 TB used, 289 TB / 396 TB avail*
>                21312 active+clean
>
>
> command: ceph df
> in one server of ceph cluster, it shows:
> [root@yun4 ~]# df -h
> Filesystem      Size  Used Avail Use% Mounted on
> /dev/sda3       1.8T   25G  1.8T   2% /
> tmpfs            16G     0   16G   0% /dev/shm
> /dev/sda1       194M   60M  125M  33% /boot
> /dev/sdb1       3.7T 1001G  2.7T  27% /data/osd1
> /dev/sdk1       3.7T 1001G  2.7T  27% /data/osd10
> /dev/sdl1       3.7T 1001G  2.7T  27% /data/osd11
> /dev/sdc1       3.7T 1001G  2.7T  27% /data/osd2
> /dev/sdd1       3.7T 1001G  2.7T  27% /data/osd3
> /dev/sde1       3.7T 1001G  2.7T  27% /data/osd4
> /dev/sdf1       3.7T 1001G  2.7T  27% /data/osd5
> /dev/sdg1       3.7T 1001G  2.7T  27% /data/osd6
> /dev/sdh1       3.7T 1001G  2.7T  27% /data/osd7
> /dev/sdi1       3.7T 1001G  2.7T  27% /data/osd8
> /dev/sdj1       3.7T 1001G  2.7T  27% /data/osd9
>
> There is 27% space used in every disk.
> What's the possible reason?
>
>

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to