I'm pulling my hair out with ceph. I am testing things with a 5 server 
cluster. I have 3 monitors, and two storage machines each with 4 osd's. I 
have started from scratch 4 times now, and can't seem to figure out how to 
get a clean status. Ceph health reports: 

HEALTH_WARN 34 pgs degraded; 192 pgs stuck unclean; recovery 40/60 objects 
degraded (66.667%)

ceph status reports: 

cluster 99567882-2e01-4dec-8ca5-692e439a5a47
     health HEALTH_WARN 34 pgs degraded; 192 pgs stuck unclean; recovery 
40/60 objects degraded (66.667%)
     monmap e2: 3 mons at 
{monitor01=192.168.1.200:6789/0,monitor02=192.168.1.201:6789/0,monitor03=192
.168.1.202:6789/0}, election epoch 8, quorum 0,1,2 
monitor01,monitor02,monitor03
     mdsmap e4: 1/1/1 up {0=monitor01.mydomain.com=up:active}
     osdmap e49: 8 osds: 8 up, 8 in
      pgmap v85: 192 pgs, 3 pools, 1884 bytes data, 20 objects
            297 MB used, 14856 GB / 14856 GB avail
            40/60 objects degraded (66.667%)
                   1 active
                  34 active+degraded
                 157 active+remapped


My ceph.conf contains the following:

[default]
osd_pool_default_size = 2

[global]
auth_service_required = cephx
filestore_xattr_use_omap = true
auth_client_required = cephx
auth_cluster_required = cephx
mon_host = 192.168.1.200,192.168.1.201,192.168.1.202
mon_initial_members = monitor01, monitor02, monitor03
fsid = 99567882-2e01-4dec-8ca5-692e439a5a47



Any suggestions are welcome at this point.

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to