The replication size is 3 and min_size is 2. Yes , they don't have
enough copies. Ceph by itself should recover from this state to ensure
durability .

@Tupper : In this bug, each node is hosting only three osds . In my
set up , every node has 23 osds. So this should not be the issue .



On Tue, May 3, 2016 at 7:00 PM, Varada Kari <varada.k...@sandisk.com> wrote:
> Pgs are degraded because they don't have enough copies of the data. What
> is your replication size?
>
> You can refer to
> http://docs.ceph.com/docs/master/rados/operations/pg-states/  for PG states.
>
> Varada
>
> On Tuesday 03 May 2016 06:56 PM, Gaurav Bafna wrote:
>> Also , the old PGs are not mapped to the down osd as seen from the
>> ceph health detail
>>
>> pg 5.72 is active+undersized+degraded, acting [16,49]
>> pg 5.4e is active+undersized+degraded, acting [16,38]
>> pg 5.32 is active+undersized+degraded, acting [39,19]
>> pg 5.37 is active+undersized+degraded, acting [43,1]
>> pg 5.2c is active+undersized+degraded, acting [47,18]
>> pg 5.27 is active+undersized+degraded, acting [26,19]
>> pg 6.13 is active+undersized+degraded, acting [30,16]
>> pg 4.17 is active+undersized+degraded, acting [47,20]
>> pg 7.a is active+undersized+degraded, acting [38,2]
>>
>> From pg query of 7.a
>>
>> {
>>     "state": "active+undersized+degraded",
>>     "snap_trimq": "[]",
>>     "epoch": 857,
>>     "up": [
>>         38,
>>         2
>>     ],
>>     "acting": [
>>         38,
>>         2
>>     ],
>>     "actingbackfill": [
>>         "2",
>>         "38"
>>     ],
>>     "info": {
>>         "pgid": "7.a",
>>         "last_update": "0'0",
>>         "last_complete": "0'0",
>>         "log_tail": "0'0",
>>         "last_user_version": 0,
>>         "last_backfill": "MAX",
>>         "purged_snaps": "[]",
>>         "history": {
>>             "epoch_created": 13,
>>             "last_epoch_started": 818,
>>             "last_epoch_clean": 818,
>>             "last_epoch_split": 0,
>>             "same_up_since": 817,
>>             "same_interval_since": 817,
>>
>>
>> Complete pq query info at : http://pastebin.com/ZHB6M4PQ
>>
>> On Tue, May 3, 2016 at 6:46 PM, Gaurav Bafna <baf...@gmail.com> wrote:
>>> Thanks Tupper for replying.
>>>
>>> Shouldn't the PG be remapped to other OSDs ?
>>>
>>> Yes , removing OSD from the cluster is resulting into full recovery.
>>> But that should not be needed , right ?
>>>
>>>
>>>
>>> On Tue, May 3, 2016 at 6:31 PM, Tupper Cole <tc...@redhat.com> wrote:
>>>> The degraded pgs are mapped to the down OSD and have not mapped to a new
>>>> OSD. Removing the OSD would likely result in a full recovery.
>>>>
>>>> As a note, having two monitors (or any even number of monitors) is not
>>>> recommended. If either monitor goes down you will lose quorum. The
>>>> recommended number of monitors for any cluster is at least three.
>>>>
>>>> On Tue, May 3, 2016 at 8:42 AM, Gaurav Bafna <baf...@gmail.com> wrote:
>>>>> Hi Cephers,
>>>>>
>>>>> I am running a very small cluster of 3 storage and 2 monitor nodes.
>>>>>
>>>>> After I kill 1 osd-daemon, the cluster never recovers fully. 9 PGs
>>>>> remain undersized for unknown reason.
>>>>>
>>>>> After I restart that 1 osd deamon, the cluster recovers in no time .
>>>>>
>>>>> Size of all pools are 3 and min_size is 2.
>>>>>
>>>>> Can anybody please help ?
>>>>>
>>>>> Output of  "ceph -s"
>>>>>     cluster fac04d85-db48-4564-b821-deebda046261
>>>>>      health HEALTH_WARN
>>>>>             9 pgs degraded
>>>>>             9 pgs stuck degraded
>>>>>             9 pgs stuck unclean
>>>>>             9 pgs stuck undersized
>>>>>             9 pgs undersized
>>>>>             recovery 3327/195138 objects degraded (1.705%)
>>>>>             pool .users pg_num 512 > pgp_num 8
>>>>>      monmap e2: 2 mons at
>>>>> {dssmon2=10.140.13.13:6789/0,dssmonleader1=10.140.13.11:6789/0}
>>>>>             election epoch 1038, quorum 0,1 dssmonleader1,dssmon2
>>>>>      osdmap e857: 69 osds: 68 up, 68 in
>>>>>       pgmap v106601: 896 pgs, 9 pools, 435 MB data, 65047 objects
>>>>>             279 GB used, 247 TB / 247 TB avail
>>>>>             3327/195138 objects degraded (1.705%)
>>>>>                  887 active+clean
>>>>>                    9 active+undersized+degraded
>>>>>   client io 395 B/s rd, 0 B/s wr, 0 op/s
>>>>>
>>>>> ceph health detail output :
>>>>>
>>>>> HEALTH_WARN 9 pgs degraded; 9 pgs stuck degraded; 9 pgs stuck unclean;
>>>>> 9 pgs stuck undersized; 9 pgs undersized; recovery 3327/195138 objects
>>>>> degraded (1.705%); pool .users pg_num 512 > pgp_num 8
>>>>> pg 7.a is stuck unclean for 322742.938959, current state
>>>>> active+undersized+degraded, last acting [38,2]
>>>>> pg 5.27 is stuck unclean for 322754.823455, current state
>>>>> active+undersized+degraded, last acting [26,19]
>>>>> pg 5.32 is stuck unclean for 322750.685684, current state
>>>>> active+undersized+degraded, last acting [39,19]
>>>>> pg 6.13 is stuck unclean for 322732.665345, current state
>>>>> active+undersized+degraded, last acting [30,16]
>>>>> pg 5.4e is stuck unclean for 331869.103538, current state
>>>>> active+undersized+degraded, last acting [16,38]
>>>>> pg 5.72 is stuck unclean for 331871.208948, current state
>>>>> active+undersized+degraded, last acting [16,49]
>>>>> pg 4.17 is stuck unclean for 331822.771240, current state
>>>>> active+undersized+degraded, last acting [47,20]
>>>>> pg 5.2c is stuck unclean for 323021.274535, current state
>>>>> active+undersized+degraded, last acting [47,18]
>>>>> pg 5.37 is stuck unclean for 323007.574395, current state
>>>>> active+undersized+degraded, last acting [43,1]
>>>>> pg 7.a is stuck undersized for 322487.284302, current state
>>>>> active+undersized+degraded, last acting [38,2]
>>>>> pg 5.27 is stuck undersized for 322487.287164, current state
>>>>> active+undersized+degraded, last acting [26,19]
>>>>> pg 5.32 is stuck undersized for 322487.285566, current state
>>>>> active+undersized+degraded, last acting [39,19]
>>>>> pg 6.13 is stuck undersized for 322487.287168, current state
>>>>> active+undersized+degraded, last acting [30,16]
>>>>> pg 5.4e is stuck undersized for 331351.476170, current state
>>>>> active+undersized+degraded, last acting [16,38]
>>>>> pg 5.72 is stuck undersized for 331351.475707, current state
>>>>> active+undersized+degraded, last acting [16,49]
>>>>> pg 4.17 is stuck undersized for 322487.280309, current state
>>>>> active+undersized+degraded, last acting [47,20]
>>>>> pg 5.2c is stuck undersized for 322487.286347, current state
>>>>> active+undersized+degraded, last acting [47,18]
>>>>> pg 5.37 is stuck undersized for 322487.280027, current state
>>>>> active+undersized+degraded, last acting [43,1]
>>>>> pg 7.a is stuck degraded for 322487.284340, current state
>>>>> active+undersized+degraded, last acting [38,2]
>>>>> pg 5.27 is stuck degraded for 322487.287202, current state
>>>>> active+undersized+degraded, last acting [26,19]
>>>>> pg 5.32 is stuck degraded for 322487.285604, current state
>>>>> active+undersized+degraded, last acting [39,19]
>>>>> pg 6.13 is stuck degraded for 322487.287207, current state
>>>>> active+undersized+degraded, last acting [30,16]
>>>>> pg 5.4e is stuck degraded for 331351.476209, current state
>>>>> active+undersized+degraded, last acting [16,38]
>>>>> pg 5.72 is stuck degraded for 331351.475746, current state
>>>>> active+undersized+degraded, last acting [16,49]
>>>>> pg 4.17 is stuck degraded for 322487.280348, current state
>>>>> active+undersized+degraded, last acting [47,20]
>>>>> pg 5.2c is stuck degraded for 322487.286386, current state
>>>>> active+undersized+degraded, last acting [47,18]
>>>>> pg 5.37 is stuck degraded for 322487.280066, current state
>>>>> active+undersized+degraded, last acting [43,1]
>>>>> pg 5.72 is active+undersized+degraded, acting [16,49]
>>>>> pg 5.4e is active+undersized+degraded, acting [16,38]
>>>>> pg 5.32 is active+undersized+degraded, acting [39,19]
>>>>> pg 5.37 is active+undersized+degraded, acting [43,1]
>>>>> pg 5.2c is active+undersized+degraded, acting [47,18]
>>>>> pg 5.27 is active+undersized+degraded, acting [26,19]
>>>>> pg 6.13 is active+undersized+degraded, acting [30,16]
>>>>> pg 4.17 is active+undersized+degraded, acting [47,20]
>>>>> pg 7.a is active+undersized+degraded, acting [38,2]
>>>>> recovery 3327/195138 objects degraded (1.705%)
>>>>> pool .users pg_num 512 > pgp_num 8
>>>>>
>>>>>
>>>>> My crush map is default.
>>>>>
>>>>> Ceph.conf is :
>>>>>
>>>>> [osd]
>>>>> osd mkfs type=xfs
>>>>> osd recovery threads=2
>>>>> osd disk thread ioprio class=idle
>>>>> osd disk thread ioprio priority=7
>>>>> osd journal=/var/lib/ceph/osd/ceph-$id/journal
>>>>> filestore flusher=False
>>>>> osd op num shards=3
>>>>> debug osd=5
>>>>> osd disk threads=2
>>>>> osd data=/var/lib/ceph/osd/ceph-$id
>>>>> osd op num threads per shard=5
>>>>> osd op threads=4
>>>>> keyring=/var/lib/ceph/osd/ceph-$id/keyring
>>>>> osd journal size=4096
>>>>>
>>>>>
>>>>> [global]
>>>>> filestore max sync interval=10
>>>>> auth cluster required=cephx
>>>>> osd pool default min size=3
>>>>> osd pool default size=3
>>>>> public network=10.140.13.0/26
>>>>> objecter inflight op_bytes=1073741824
>>>>> auth service required=cephx
>>>>> filestore min sync interval=1
>>>>> fsid=fac04d85-db48-4564-b821-deebda046261
>>>>> keyring=/etc/ceph/keyring
>>>>> cluster network=10.140.13.0/26
>>>>> auth client required=cephx
>>>>> filestore xattr use omap=True
>>>>> max open files=65536
>>>>> objecter inflight ops=2048
>>>>> osd pool default pg num=512
>>>>> log to syslog = true
>>>>> #err to syslog = true
>>>>>
>>>>>
>>>>> --
>>>>> Gaurav Bafna
>>>>> 9540631400
>>>>> _______________________________________________
>>>>> ceph-users mailing list
>>>>> ceph-users@lists.ceph.com
>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>
>>>>
>>>>
>>>> --
>>>>
>>>> Thanks,
>>>> Tupper Cole
>>>> Senior Storage Consultant
>>>> Global Storage Consulting, Red Hat
>>>> tc...@redhat.com
>>>> phone:  + 01 919-720-2612
>>>
>>>
>>> --
>>> Gaurav Bafna
>>> 9540631400
>>
>>
>
> PLEASE NOTE: The information contained in this electronic mail message is 
> intended only for the use of the designated recipient(s) named above. If the 
> reader of this message is not the intended recipient, you are hereby notified 
> that you have received this message in error and that any review, 
> dissemination, distribution, or copying of this message is strictly 
> prohibited. If you have received this communication in error, please notify 
> the sender by telephone or e-mail (as shown above) immediately and destroy 
> any and all copies of this message in your possession (whether hard copies or 
> electronically stored copies).



-- 
Gaurav Bafna
9540631400
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to