Hi again

Looked at your ceph -s.

You have only 2 OSDs, one on each node. The default replica count is 2, the
default crush map says each replica on a different host, or may be you set
it to 2 different OSDs. Anyway, when one of your OSD goes down, Ceph can no
longer find another OSDs to host the second replica it must create.

Looking at your crushmap we would know better.

Recommendation: for testing efficiently and most options available,
functionnally speaking, deploy a cluster with 3 nodes, 3 OSDs each is my
best practice.

Or make 1 node with 3 OSDs modifying your crushmap to "choose type osd" in
your rulesets.

JC


On Saturday, April 19, 2014, Gonzalo Aguilar Delgado <
gagui...@aguilardelgado.com> wrote:

> Hi,
>
> I'm building a cluster where two nodes replicate objects inside. I found
> that shutting down just one of the nodes (the second one), makes everything
> "incomplete".
>
> I cannot find why, since crushmap looks good to me.
>
> after shutting down one node
>
>     cluster 9028f4da-0d77-462b-be9b-dbdf7fa57771
>      health HEALTH_WARN 192 pgs incomplete; 96 pgs stuck inactive; 96 pgs
> stuck unclean; 1/2 in osds are down
>      monmap e9: 1 mons at {blue-compute=172.16.0.119:6789/0}, election
> epoch 1, quorum 0 blue-compute
>      osdmap e73: 2 osds: 1 up, 2 in
>       pgmap v172: 192 pgs, 3 pools, 275 bytes data, 1 objects
>             7552 kB used, 919 GB / 921 GB avail
>                  192 incomplete
>
>
> Both nodes has WD Caviar Black 500MB disk with btrfs filesystem on it.
> Full disk used.
>
> I cannot understand why does not replicate to both nodes.
>
> Someone can help?
>
> Best regards,
>


-- 
Sent while moving
Pardon my French and any spelling &| grammar glitches
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to