On Tue, Jul 1, 2014 at 11:45 AM, Gregory Farnum <g...@inktank.com> wrote:
> On Tue, Jul 1, 2014 at 11:33 AM, Brian Lovett
> <brian.lov...@prosperent.com> wrote:
>> Brian Lovett <brian.lovett@...> writes:
>>
>>
>> I restarted all of the osd's and noticed that ceph shows 2 osd's up even if
>> the servers are completely powered down:  osdmap e95: 8 osds: 2 up, 8 in
>>
>> Why would that be?
>
> The OSDs report each other down much more quickly (~30s) than the
> monitor timeout (~15 minutes). They'd get marked down eventually.
>
> On Tue, Jul 1, 2014 at 11:43 AM, Brian Lovett
> <brian.lov...@prosperent.com> wrote:
>> Gregory Farnum <greg@...> writes:
>>
>>>
>>> What's the output of "ceph osd map"?
>>>
>>> Your CRUSH map probably isn't trying to segregate properly, with 2
>>> hosts and 4 OSDs each.
>>> Software Engineer #42  <at>  http://inktank.com | http://ceph.com
>>>
>>  Is this what you are looking for?
>>
>> ceph osd map rbd ceph
>> osdmap e104 pool 'rbd' (2) object 'ceph' -> pg 2.3482c180 (2.0) -> up ([3,5],
>> p3) acting ([3,5,0], p3)
>
> Whoops, I mean "ceph osd list", sorry! (That should output a textual
> representation of how they're arranged in the CRUSH map.)

...and one more time, because apparently my brain's out to lunch today:

ceph osd tree

*sigh*
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to