I have 19 pages that are stuck unclean (see below result of ceph -s). This
occurred after I executed a "ceph osd reweight-by-utilization 108" to
resolve problems with "backfill_too_full" messages, which I believe
occurred because my OSDs sizes vary significantly in size (from a low of
600GB to a high of 3 TB). How can I get ceph to get these pages out of
stuck-unclean? (And why is this occurring anyways?) My best guess of how to
fix (though I don't know why) is that I need to run:

ceph osd crush tunables optimal.

However, my kernel version (on a fully up-to-date Centos 6.5) is 2.6.32,
which is well below the minimum required version of 3.6 that's stated in
the documentation (http://ceph.com/docs/master/rados/operations/crush-map/) --
so if I must run "ceph osd crush tunables optimal" to fix this problem, I
presume I must upgrade my kernel first, right?...Any thoughts or am I
chasing the wrong solution -- I want to avoid kernel upgrade unless it's
needed.)

=====================

[root@ia2 ceph4]# ceph -s
    cluster 14f78538-6085-43f9-ac80-e886ca4de119
     health HEALTH_WARN 19 pgs backfilling; 19 pgs stuck unclean; recovery
42959/5511127 objects degraded (0.779%)
     monmap e9: 3 mons at {ia1=
192.168.1.11:6789/0,ia2=192.168.1.12:6789/0,ia3=192.168.1.13:6789/0},
election epoch 496, quorum 0,1,2 ia1,ia2,ia3
     osdmap e7931: 23 osds: 23 up, 23 in
      pgmap v1904820: 1500 pgs, 1 pools, 10531 GB data, 2670 kobjects
            18708 GB used, 26758 GB / 45467 GB avail
            42959/5511127 objects degraded (0.779%)
                1481 active+clean
                  19 active+remapped+backfilling
  client io 1457 B/s wr, 0 op/s

[root@ia2 ceph4]# ceph -v
ceph version 0.72.2 (a913ded2ff138aefb8cb84d347d72164099cfd60)

[root@ia2 ceph4]# uname -r
2.6.32-431.3.1.el6.x86_64

====
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to