disregard. i did this on a cluster of test vms and didnt bother setting
different hostnames, thus confusing ceph.

On Wed, Jul 29, 2015 at 2:24 AM pixelfairy <pixelfa...@gmail.com> wrote:

> have a small test cluster (vmware fusion, 3 mon+osd nodes) all run ubuntu
> trusty. tried rebooting all 3 nodes and this happend.
>
> root@ubuntu:~# ceph --version ceph version 0.94.2
> (5fb85614ca8f354284c713a2f9c610860720bbf3)
>
> root@ubuntu:~# ceph health 2015-07-29 02:08:31.360516 7f5bd711a700 -1
> asok(0x7f5bd0000bf0) AdminSocketConfigObs::init: failed:
> AdminSocket::bind_and_listen: failed to bind the UNIX domain socket to
> '/var/run/ceph/rbd-clients/ceph-client.admin.3282.140032308415712.asok':
> (2) No such file or directory HEALTH_WARN 64 pgs stuck unclean; recovery
> 512/1024 objects misplaced (50.000%); too few PGs per OSD (21 < min 30)
>
> the osd disks are only 50gigs but they seemed to work fine before the
> reboot.
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to