And finally works. Thanks. Now i need to see another erros. My Cluster is very problematic.
Em sáb, 19 de nov de 2016 às 19:12, Bruno Silva <bemanuel...@gmail.com> escreveu: > I put and didn't works, in the end i put an osd with id 5 in production. > > > Em sáb, 19 de nov de 2016 às 17:46, Paweł Sadowski <c...@sadziu.pl> > escreveu: > > Hi, > > Make a temporary OSD with the same ID and weight 0 to avoid putting data > on it. Cluster should contact this OSD and move forward. If not you can > also use 'ceph osd lost ID' but OSD with that ID must exists in crushmap > (and this probably not the case here). > > On 19.11.2016 13:46, Bruno Silva wrote: > > Version: Hammer > > On my cluster a pg is saying: > > "down_osds_we_would_probe": [ > > 5 > > ], > > > > But this osd was removed. How can i solve this. > > Reading on group list ceph-users they say that this could be the > > reason to my cluster is stoped. > > How can i solve this? > > > > -- > PS > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com