Actually, I have to revise this, Ceph _is_ freeing capacity, but very slowly, roughly 150G every 5 minutes. Is that normal? I feel like capacity is generally freed almost immediately when I've previously deleted pools.
Thanks! -Joe From: Gruher, Joseph R Sent: Thursday, April 03, 2014 10:32 AM To: ceph-users@lists.ceph.com Cc: Gruher, Joseph R Subject: Ceph Not Freeing Capacity After Pools Deleted Hi all- I am testing on Ceph 0.78 running on Ubuntu 13.04 with 3.13 kernel. I had two replication pools and five erasure code pools. Cluster was getting full so I deleted all the EC pools. However, Ceph is not freeing the capacity. Note below there is only 1636G in the two pools but the global stats still report 13652G as used (90.5% when it should be down to 10.8%). Any ideas why Ceph is not freeing the capacity and how to fix it? Thanks! ceph@joceph-admin01:~$ ceph df GLOBAL: SIZE AVAIL RAW USED %RAW USED 15083G 1430G 13652G 90.51 POOLS: NAME ID USED %USED OBJECTS data 0 0 0 0 metadata 1 0 0 0 rbd 2 0 0 0 mycontainers_1 20 839G 5.56 224105 mycontainers_2 21 797G 5.29 213515 ceph@joceph-admin01:~$ ceph status cluster b12ebb71-e4a6-41fa-8246-71cbfa09fb6e health HEALTH_WARN 18 near full osd(s) monmap e1: 2 mons at {mohonpeak01=10.0.0.101:6789/0,mohonpeak02=10.0.0.102:6789/0}, election epoch 10, quorum 0,1 mohonpeak01,mohonpeak02 osdmap e214: 18 osds: 18 up, 18 in pgmap v198720: 2784 pgs, 10 pools, 1637 GB data, 427 kobjects 13652 GB used, 1430 GB / 15083 GB avail 2784 active+clean ceph@joceph-admin01:~$ ceph osd dump epoch 214 fsid b12ebb71-e4a6-41fa-8246-71cbfa09fb6e created 2014-03-24 12:06:28.290970 modified 2014-04-03 10:18:07.714158 flags pool 0 'data' replicated size 1 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 128 pgp_num 128 last_change 84 owner 0 flags hashpspool crash_replay_interval 45 stripe_width 0 pool 1 'metadata' replicated size 1 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 128 pgp_num 128 last_change 86 owner 0 flags hashpspool stripe_width 0 pool 2 'rbd' replicated size 1 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 128 pgp_num 128 last_change 88 owner 0 flags hashpspool stripe_width 0 pool 20 'mycontainers_1' replicated size 1 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 1600 pgp_num 1600 last_change 167 owner 0 flags hashpspool stripe_width 0 pool 21 'mycontainers_2' replicated size 2 min_size 1 crush_ruleset 0 object_hash rjenkins pg_num 800 pgp_num 800 last_change 171 owner 0 flags hashpspool stripe_width 0 max_osd 18 osd.0 up in weight 1 up_from 195 up_thru 201 down_at 194 last_clean_interval [116,185) 10.0.0.101:6815/5203 10.0.1.101:6810/5203 10.0.1.101:6811/5203 10.0.0.101:6816/5203 exists,up 56431fdc-88a2-4c55-a4a5-64596f080962 osd.1 up in weight 1 up_from 196 up_thru 203 down_at 195 last_clean_interval [117,185) 10.0.0.101:6800/5125 10.0.1.101:6800/5125 10.0.1.101:6801/5125 10.0.0.101:6801/5125 exists,up 266f2705-6286-4a1d-82ba-cb5e1fb56e46 osd.2 up in weight 1 up_from 193 up_thru 201 down_at 192 last_clean_interval [117,185) 10.0.0.101:6821/5245 10.0.1.101:6814/5245 10.0.1.101:6815/5245 10.0.0.101:6822/5245 exists,up e363143f-805e-4e4c-8732-fd9f07d7cf31 osd.3 up in weight 1 up_from 195 up_thru 201 down_at 194 last_clean_interval [116,185) 10.0.0.101:6803/5138 10.0.1.101:6802/5138 10.0.1.101:6803/5138 10.0.0.101:6804/5138 exists,up 444df45b-de9f-42a4-92d6-82b479be0a01 osd.4 up in weight 1 up_from 200 up_thru 201 down_at 199 last_clean_interval [116,185) 10.0.0.101:6806/5143 10.0.1.101:6804/5143 10.0.1.101:6805/5143 10.0.0.101:6807/5143 exists,up 47e19cd1-0ca2-4205-ba4d-7e726683097a osd.5 up in weight 1 up_from 197 up_thru 203 down_at 196 last_clean_interval [117,185) 10.0.0.101:6824/5370 10.0.1.101:6816/5370 10.0.1.101:6817/5370 10.0.0.101:6825/5370 exists,up a3aed3a0-da30-4483-ae40-eebf8d4b0fc9 osd.6 up in weight 1 up_from 194 up_thru 201 down_at 193 last_clean_interval [116,185) 10.0.0.101:6818/5226 10.0.1.101:6812/5226 10.0.1.101:6813/5226 10.0.0.101:6819/5226 exists,up 5ba235c2-c304-4a82-80aa-9a9934367347 osd.7 up in weight 1 up_from 196 up_thru 202 down_at 195 last_clean_interval [117,185) 10.0.0.101:6809/5160 10.0.1.101:6806/5160 10.0.1.101:6807/5160 10.0.0.101:6810/5160 exists,up db75e990-7a6e-4fff-9d13-ab50a7139821 osd.8 up in weight 1 up_from 200 up_thru 201 down_at 199 last_clean_interval [116,185) 10.0.0.101:6812/5184 10.0.1.101:6808/5184 10.0.1.101:6809/5184 10.0.0.101:6813/5184 exists,up b134b786-6ee1-4b5b-bead-2885b3bc75c4 osd.9 up in weight 1 up_from 201 up_thru 201 down_at 200 last_clean_interval [116,185) 10.0.0.102:6800/2750 10.0.1.102:6800/2750 10.0.1.102:6801/2750 10.0.0.102:6801/2750 exists,up d1a0d7f9-3c74-484f-a0f3-9adf156bf627 osd.10 up in weight 1 up_from 194 up_thru 201 down_at 193 last_clean_interval [117,185) 10.0.0.102:6821/3000 10.0.1.102:6814/3000 10.0.1.102:6815/3000 10.0.0.102:6822/3000 exists,up 8f478fa7-c1ae-416d-838f-af11722b8223 osd.11 up in weight 1 up_from 190 up_thru 201 down_at 189 last_clean_interval [116,185) 10.0.0.102:6812/2839 10.0.1.102:6808/2839 10.0.1.102:6809/2839 10.0.0.102:6813/2839 exists,up e298c892-28d9-4abc-9b23-21dbf513b893 osd.12 up in weight 1 up_from 190 up_thru 201 down_at 189 last_clean_interval [116,185) 10.0.0.102:6806/2777 10.0.1.102:6804/2777 10.0.1.102:6805/2777 10.0.0.102:6807/2777 exists,up 6a1fd468-a70a-4885-b824-6215348d813e osd.13 up in weight 1 up_from 189 up_thru 201 down_at 188 last_clean_interval [116,185) 10.0.0.102:6815/2889 10.0.1.102:6810/2889 10.0.1.102:6811/2889 10.0.0.102:6816/2889 exists,up fcb52765-9333-459f-8143-90f0513f67d0 osd.14 up in weight 1 up_from 198 up_thru 203 down_at 197 last_clean_interval [116,185) 10.0.0.102:6824/3306 10.0.1.102:6816/3306 10.0.1.102:6817/3306 10.0.0.102:6825/3306 exists,up e098fb1c-c127-4017-9e63-46cd3f11fcff osd.15 up in weight 1 up_from 190 up_thru 201 down_at 189 last_clean_interval [116,185) 10.0.0.102:6809/2819 10.0.1.102:6806/2819 10.0.1.102:6807/2819 10.0.0.102:6810/2819 exists,up 59f916d3-00ac-4493-aa72-c7d4bce39d75 osd.16 up in weight 1 up_from 187 up_thru 201 down_at 186 last_clean_interval [116,185) 10.0.0.102:6803/2753 10.0.1.102:6802/2753 10.0.1.102:6803/2753 10.0.0.102:6804/2753 exists,up 0e6d9dec-baeb-46e6-9c1d-8fed1c50bc58 osd.17 up in weight 1 up_from 191 up_thru 202 down_at 190 last_clean_interval [116,185) 10.0.0.102:6818/2950 10.0.1.102:6812/2950 10.0.1.102:6813/2950 10.0.0.102:6819/2950 exists,up 520b86b0-93de-4da8-a25f-a3b8fc15321a ceph@joceph-admin01:~$
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com