"nothing to send, going to standby" isn't necessarily bad, I see it from
time to time.  It shouldn't stay like that for long though.  If it's been 5
minutes, and the cluster still isn't doing anything, I'd restart that osd.

On Fri, Nov 7, 2014 at 1:55 PM, Jan Pekař <jan.pe...@imatic.cz> wrote:

> Hi,
>
> I was testing ceph cluster map changes and I got to stuck state which
> seems to be indefinite.
> First my description what I have done.
>
> I'm testing special case with only one copy of pg's (pool size = 1).
>
> All pg's was on one osd.0. I created second osd.1 and modified cluster map
> to transfer one pool (metadata) to the newly created osd.1
> PG's started to remap and "objects degraded number" was dropping - so
> everything looked normal.
>
> During that recovery process I restarted both osd daemons.
> After that I noticed, that pg's, that should be remapped had "stale" state
> - stale+active+remapped+backfilling and other object with stale state .
> I tried to run ceph pg force_create_pg on one pg, that should be remapped,
> but nothing changed (that is 1 stuck / creating PG below in ceph health)
>
> Command rados -p metadata ls hangs so data are unavailable, but it should
> be there.
>
> What should I do in this state to get it working?
>
> ceph -s below:
>
>     cluster 93418692-8e2e-4689-a237-ed5b47f39f72
>      health HEALTH_WARN 52 pgs backfill; 1 pgs backfilling; 63 pgs stale;
> 1 pgs stuck inactive; 63 pgs stuck stale; 54 pgs stuck unclean; recovery
> 107232/1881806 objects degraded (5.698%); mon.imatic-mce low disk space
>      monmap e1: 1 mons at {imatic-mce=192.168.11.165:6789/0}, election
> epoch 1, quorum 0 imatic-mce
>      mdsmap e450: 1/1/1 up {0=imatic-mce=up:active}
>      osdmap e275: 2 osds: 2 up, 2 in
>       pgmap v51624: 448 pgs, 4 pools, 790 GB data, 1732 kobjects
>             804 GB used, 2915 GB / 3720 GB avail
>             107232/1881806 objects degraded (5.698%)
>                   52 stale+active+remapped+wait_backfill
>                    1 creating
>                    1 stale+active+remapped+backfilling
>                   10 stale+active+clean
>                  384 active+clean
>
> Last message in OSD log's:
>
> 2014-11-07 22:17:45.402791 deb4db70  0 -- 192.168.11.165:6804/29564 >>
> 192.168.11.165:6807/29939 pipe(0x9d52f00 sd=213 :53216 s=2 pgs=1 cs=1 l=0
> c=0x2c7f58c0).fault with nothing to send, going to standby
>
> Thank you for help
> With regards
> Jan Pekar, ceph fan
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to