Hi, I would prob. start by figuring out exactly what pg are stuck unclean.
You can do 'ceph pg dump | grep unclean' to get that info - then if your theory holds you should be able to verify the disk(s) in question. I cannot see any _too_full so am curious what could be the cause. You can also always adjust the weights manually if needed ( http://ceph.com/docs/master/rados/operations/control/#osd-subsystem ) with the (re)weight cmd. Cheers, Martin On Mon, Feb 24, 2014 at 2:09 AM, Gautam Saxena <gsax...@i-a-inc.com> wrote: > I have 19 pages that are stuck unclean (see below result of ceph -s). This > occurred after I executed a "ceph osd reweight-by-utilization 108" to > resolve problems with "backfill_too_full" messages, which I believe > occurred because my OSDs sizes vary significantly in size (from a low of > 600GB to a high of 3 TB). How can I get ceph to get these pages out of > stuck-unclean? (And why is this occurring anyways?) My best guess of how to > fix (though I don't know why) is that I need to run: > > ceph osd crush tunables optimal. > > However, my kernel version (on a fully up-to-date Centos 6.5) is 2.6.32, > which is well below the minimum required version of 3.6 that's stated in > the documentation (http://ceph.com/docs/master/rados/operations/crush-map/ > ) -- so if I must run "ceph osd crush tunables optimal" to fix this > problem, I presume I must upgrade my kernel first, right?...Any thoughts or > am I chasing the wrong solution -- I want to avoid kernel upgrade unless > it's needed.) > > ===================== > > [root@ia2 ceph4]# ceph -s > cluster 14f78538-6085-43f9-ac80-e886ca4de119 > health HEALTH_WARN 19 pgs backfilling; 19 pgs stuck unclean; recovery > 42959/5511127 objects degraded (0.779%) > monmap e9: 3 mons at {ia1= > 192.168.1.11:6789/0,ia2=192.168.1.12:6789/0,ia3=192.168.1.13:6789/0}, > election epoch 496, quorum 0,1,2 ia1,ia2,ia3 > osdmap e7931: 23 osds: 23 up, 23 in > pgmap v1904820: 1500 pgs, 1 pools, 10531 GB data, 2670 kobjects > 18708 GB used, 26758 GB / 45467 GB avail > 42959/5511127 objects degraded (0.779%) > 1481 active+clean > 19 active+remapped+backfilling > client io 1457 B/s wr, 0 op/s > > [root@ia2 ceph4]# ceph -v > ceph version 0.72.2 (a913ded2ff138aefb8cb84d347d72164099cfd60) > > [root@ia2 ceph4]# uname -r > 2.6.32-431.3.1.el6.x86_64 > > ==== > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com