Hi all,

Maybe this will help:

The issue is with shards 3,4 and 5 of PG 6.3f:

LOG's of OSD's 16, 17 & 36 (the ones crashing on startup).


*Log OSD.16 (shard 4):*

2018-06-08 08:35:01.727261 7f4c585e3700 -1
bluestore(/var/lib/ceph/osd/ceph-16) _txc_add_transaction error (2) No such
file or directory not handled on operation 30 (op 0, counting from 0)
2018-06-08 08:35:01.727273 7f4c585e3700 -1
bluestore(/var/lib/ceph/osd/ceph-16) ENOENT on clone suggests osd bug
2018-06-08 08:35:01.727274 7f4c585e3700  0
bluestore(/var/lib/ceph/osd/ceph-16)  transaction dump:
{
    "ops": [
        {
            "op_num": 0,
            "op_name": "clonerange2",
            "collection": "6.3fs4_head",
            "src_oid":
"4#6:fc074663:::rbd_data.5.6c1d9574b0dc51.00000000000312db:head#903d0",
            "dst_oid":
"4#6:fc074663:::rbd_data.5.6c1d9574b0dc51.00000000000312db:head#",
            "src_offset": 950272,
            "len": 98304,
            "dst_offset": 950272
        },
        {
            "op_num": 1,
            "op_name": "remove",
            "collection": "6.3fs4_head",
            "oid":
"4#6:fc074663:::rbd_data.5.6c1d9574b0dc51.00000000000312db:head#903d0"
        },
        {
            "op_num": 2,
            "op_name": "setattrs",
            "collection": "6.3fs4_head",
            "oid":
"4#6:fc074663:::rbd_data.5.6c1d9574b0dc51.00000000000312db:head#",
            "attr_lens": {
                "_": 297,
                "hinfo_key": 18,
                "snapset": 35
            }
        },
        {
            "op_num": 3,
            "op_name": "clonerange2",
            "collection": "6.3fs4_head",
            "src_oid":
"4#6:fc074663:::rbd_data.5.6c1d9574b0dc51.00000000000312db:head#903cf",
            "dst_oid":
"4#6:fc074663:::rbd_data.5.6c1d9574b0dc51.00000000000312db:head#",
            "src_offset": 679936,
            "len": 274432,
            "dst_offset": 679936
        },
        {
            "op_num": 4,
            "op_name": "remove",
            "collection": "6.3fs4_head",
            "oid":
"4#6:fc074663:::rbd_data.5.6c1d9574b0dc51.00000000000312db:head#903cf"
        },
        {
            "op_num": 5,
            "op_name": "setattrs",
            "collection": "6.3fs4_head",
            "oid":
"4#6:fc074663:::rbd_data.5.6c1d9574b0dc51.00000000000312db:head#",
            "attr_lens": {
                "_": 297,
                "hinfo_key": 18,
                "snapset": 35
            }
        },
        {
            "op_num": 6,
            "op_name": "nop"
        },
        {
            "op_num": 7,
            "op_name": "op_omap_rmkeyrange",
            "collection": "6.3fs4_head",
            "oid": "4#6:fc000000::::head#",
            "first": "0000011124.00000000000000590799",
            "last": "4294967295.18446744073709551615"
        },
        {
            "op_num": 8,
            "op_name": "omap_setkeys",
            "collection": "6.3fs4_head",
            "oid": "4#6:fc000000::::head#",
            "attr_lens": {
                "_biginfo": 597,
                "_epoch": 4,
                "_info": 953,
                "can_rollback_to": 12,
                "rollback_info_trimmed_to": 12
            }
        }
    ]
}

2018-06-08 08:35:01.730584 7f4c585e3700 -1
/home/builder/source/ceph-12.2.2/src/os/bluestore/BlueStore.cc: In function
'void BlueStore::_txc_add_transaction(BlueStore::TransContext*,
ObjectStore::Transaction*)' thread 7f4c585e3700 time 2018-06-08
08:35:01.727379
/home/builder/source/ceph-12.2.2/src/os/bluestore/BlueStore.cc: 9363:
FAILED assert(0 == "unexpected error")

 ceph version 12.2.2 (215dd7151453fae88e6f968c975b6ce309d42dcf) luminous
(stable)
 1: (ceph::__ceph_assert_fail(char const*, char const*, int, char
const*)+0x102) [0x558e08ba4202]
 2: (BlueStore::_txc_add_transaction(BlueStore::TransContext*,
ObjectStore::Transaction*)+0x15fa) [0x558e08a55c3a]
 3: (BlueStore::queue_transactions(ObjectStore::Sequencer*,
std::vector<ObjectStore::Transaction,
std::allocator<ObjectStore::Transaction> >&,
boost::intrusive_ptr<TrackedOp>, ThreadPool::TPHandle*)+0x546)
[0x558e08a572a6]
 4: (ObjectStore::queue_transaction(ObjectStore::Sequencer*,
ObjectStore::Transaction&&, Context*, Context*, Context*,
boost::intrusive_ptr<TrackedOp>, ThreadPool::TPHandle*)+0x14f)
[0x558e085fa37f]
 5: (OSD::dispatch_context_transaction(PG::RecoveryCtx&, PG*,
ThreadPool::TPHandle*)+0x6c) [0x558e0857db5c]
 6: (OSD::process_peering_events(std::__cxx11::list<PG*,
std::allocator<PG*> > const&, ThreadPool::TPHandle&)+0x442) [0x558e085abec2]
 7: (ThreadPool::BatchWorkQueue<PG>::_void_process(void*,
ThreadPool::TPHandle&)+0x2c) [0x558e0861a91c]
 8: (ThreadPool::worker(ThreadPool::WorkThread*)+0xeb8) [0x558e08bab3a8]
 9: (ThreadPool::WorkThread::entry()+0x10) [0x558e08bac540]
 10: (()+0x7494) [0x7f4c709ca494]
 11: (clone()+0x3f) [0x7f4c6fa51aff]
 NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed
to interpret this.


*Log OSD.17 (Shard 5):*

2018-06-08 08:35:56.745249 7fe3fa687700 -1
bluestore(/var/lib/ceph/osd/ceph-17) _txc_add_transaction error (2) No such
file or directory not handled on operation 30 (op 0, counting from 0)
2018-06-08 08:35:56.745264 7fe3fa687700 -1
bluestore(/var/lib/ceph/osd/ceph-17) ENOENT on clone suggests osd bug
2018-06-08 08:35:56.745266 7fe3fa687700  0
bluestore(/var/lib/ceph/osd/ceph-17)  transaction dump:
{
    "ops": [
        {
            "op_num": 0,
            "op_name": "clonerange2",
            "collection": "6.3fs5_head",
            "src_oid":
"5#6:fc074663:::rbd_data.5.6c1d9574b0dc51.00000000000312db:head#903d0",
            "dst_oid":
"5#6:fc074663:::rbd_data.5.6c1d9574b0dc51.00000000000312db:head#",
            "src_offset": 950272,
            "len": 98304,
            "dst_offset": 950272
        },
        {
            "op_num": 1,
            "op_name": "remove",
            "collection": "6.3fs5_head",
            "oid":
"5#6:fc074663:::rbd_data.5.6c1d9574b0dc51.00000000000312db:head#903d0"
        },
        {
            "op_num": 2,
            "op_name": "setattrs",
            "collection": "6.3fs5_head",
            "oid":
"5#6:fc074663:::rbd_data.5.6c1d9574b0dc51.00000000000312db:head#",
            "attr_lens": {
                "_": 297,
                "hinfo_key": 18,
                "snapset": 35
            }
        },
        {
            "op_num": 3,
            "op_name": "clonerange2",
            "collection": "6.3fs5_head",
            "src_oid":
"5#6:fc074663:::rbd_data.5.6c1d9574b0dc51.00000000000312db:head#903cf",
            "dst_oid":
"5#6:fc074663:::rbd_data.5.6c1d9574b0dc51.00000000000312db:head#",
            "src_offset": 679936,
            "len": 274432,
            "dst_offset": 679936
        },
        {
            "op_num": 4,
            "op_name": "remove",
            "collection": "6.3fs5_head",
            "oid":
"5#6:fc074663:::rbd_data.5.6c1d9574b0dc51.00000000000312db:head#903cf"
        },
        {
            "op_num": 5,
            "op_name": "setattrs",
            "collection": "6.3fs5_head",
            "oid":
"5#6:fc074663:::rbd_data.5.6c1d9574b0dc51.00000000000312db:head#",
            "attr_lens": {
                "_": 297,
                "hinfo_key": 18,
                "snapset": 35
            }
        },
        {
            "op_num": 6,
            "op_name": "nop"
        },
        {
            "op_num": 7,
            "op_name": "op_omap_rmkeyrange",
            "collection": "6.3fs5_head",
            "oid": "5#6:fc000000::::head#",
            "first": "0000011124.00000000000000590799",
            "last": "4294967295.18446744073709551615"
        },
        {
            "op_num": 8,
            "op_name": "omap_setkeys",
            "collection": "6.3fs5_head",
            "oid": "5#6:fc000000::::head#",
            "attr_lens": {
                "_biginfo": 586,
                "_epoch": 4,
                "_info": 953,
                "can_rollback_to": 12,
                "rollback_info_trimmed_to": 12
            }
        }
    ]
}

2018-06-08 08:35:56.748436 7fe3fa687700 -1
/home/builder/source/ceph-12.2.2/src/os/bluestore/BlueStore.cc: In function
'void BlueStore::_txc_add_transaction(BlueStore::TransContext*,
ObjectStore::Transaction*)' thread 7fe3fa687700 time 2018-06-08
08:35:56.745458
/home/builder/source/ceph-12.2.2/src/os/bluestore/BlueStore.cc: 9363:
FAILED assert(0 == "unexpected error")

 ceph version 12.2.2 (215dd7151453fae88e6f968c975b6ce309d42dcf) luminous
(stable)
 1: (ceph::__ceph_assert_fail(char const*, char const*, int, char
const*)+0x102) [0x55695de1f202]
 2: (BlueStore::_txc_add_transaction(BlueStore::TransContext*,
ObjectStore::Transaction*)+0x15fa) [0x55695dcd0c3a]
 3: (BlueStore::queue_transactions(ObjectStore::Sequencer*,
std::vector<ObjectStore::Transaction,
std::allocator<ObjectStore::Transaction> >&,
boost::intrusive_ptr<TrackedOp>, ThreadPool::TPHandle*)+0x546)
[0x55695dcd22a6]
 4: (ObjectStore::queue_transaction(ObjectStore::Sequencer*,
ObjectStore::Transaction&&, Context*, Context*, Context*,
boost::intrusive_ptr<TrackedOp>, ThreadPool::TPHandle*)+0x14f)
[0x55695d87537f]
 5: (OSD::dispatch_context_transaction(PG::RecoveryCtx&, PG*,
ThreadPool::TPHandle*)+0x6c) [0x55695d7f8b5c]
 6: (OSD::process_peering_events(std::__cxx11::list<PG*,
std::allocator<PG*> > const&, ThreadPool::TPHandle&)+0x442) [0x55695d826ec2]
 7: (ThreadPool::BatchWorkQueue<PG>::_void_process(void*,
ThreadPool::TPHandle&)+0x2c) [0x55695d89591c]
 8: (ThreadPool::worker(ThreadPool::WorkThread*)+0xeb8) [0x55695de263a8]
 9: (ThreadPool::WorkThread::entry()+0x10) [0x55695de27540]
 10: (()+0x7494) [0x7fe412a6e494]
 11: (clone()+0x3f) [0x7fe411af5aff]
 NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed
to interpret this.


*Log OSD.36 (Shard 3):*

    -3> 2018-06-08 09:25:36.660055 7f43c9262700 -1
bluestore(/var/lib/ceph/osd/ceph-36) _txc_add_transaction error (2) No such
file or directory not handled on operation 30 (op 0, counting from 0)
    -2> 2018-06-08 09:25:36.660068 7f43c9262700 -1
bluestore(/var/lib/ceph/osd/ceph-36) ENOENT on clone suggests osd bug
    -1> 2018-06-08 09:25:36.660070 7f43c9262700  0
bluestore(/var/lib/ceph/osd/ceph-36)  transaction dump:
{
    "ops": [
        {
            "op_num": 0,
            "op_name": "clonerange2",
            "collection": "*6.3fs3*_head",
            "src_oid":
"3#6:fc074663:::rbd_data.5.6c1d9574b0dc51.00000000000312db:head#903d0",
            "dst_oid":
"3#6:fc074663:::rbd_data.5.6c1d9574b0dc51.00000000000312db:head#",
            "src_offset": 950272,
            "len": 98304,
            "dst_offset": 950272
        },
        {
            "op_num": 1,
            "op_name": "remove",
            "collection": "6.3fs3_head",
            "oid":
"3#6:fc074663:::rbd_data.5.6c1d9574b0dc51.00000000000312db:head#903d0"
        },
        {
            "op_num": 2,
            "op_name": "setattrs",
            "collection": "6.3fs3_head",
            "oid":
"3#6:fc074663:::rbd_data.5.6c1d9574b0dc51.00000000000312db:head#",
            "attr_lens": {
                "_": 297,
                "hinfo_key": 18,
                "snapset": 35
            }
        },
        {
            "op_num": 3,
            "op_name": "clonerange2",
            "collection": "6.3fs3_head",
            "src_oid":
"3#6:fc074663:::rbd_data.5.6c1d9574b0dc51.00000000000312db:head#903cf",
            "dst_oid":
"3#6:fc074663:::rbd_data.5.6c1d9574b0dc51.00000000000312db:head#",
            "src_offset": 679936,
            "len": 274432,
            "dst_offset": 679936
        },
        {
            "op_num": 4,
            "op_name": "remove",
            "collection": "6.3fs3_head",
            "oid":
"3#6:fc074663:::rbd_data.5.6c1d9574b0dc51.00000000000312db:head#903cf"
        },
        {
            "op_num": 5,
            "op_name": "setattrs",
            "collection": "6.3fs3_head",
            "oid":
"3#6:fc074663:::rbd_data.5.6c1d9574b0dc51.00000000000312db:head#",
            "attr_lens": {
                "_": 297,
                "hinfo_key": 18,
                "snapset": 35
            }
        },
        {
            "op_num": 6,
            "op_name": "nop"
        },
        {
            "op_num": 7,
            "op_name": "op_omap_rmkeyrange",
            "collection": "6.3fs3_head",
            "oid": "3#6:fc000000::::head#",
            "first": "0000011124.00000000000000590799",
            "last": "4294967295.18446744073709551615"
        },
        {
            "op_num": 8,
            "op_name": "omap_setkeys",
            "collection": "6.3fs3_head",
            "oid": "3#6:fc000000::::head#",
            "attr_lens": {
                "_biginfo": 743,
                "_epoch": 4,
                "_info": 953,
                "can_rollback_to": 12,
                "rollback_info_trimmed_to": 12
            }
        }
    ]
}

     0> 2018-06-08 09:25:36.663334 7f43c9262700 -1
/home/builder/source/ceph-12.2.2/src/os/bluestore/BlueStore.cc: In function
'void BlueStore::_txc_add_transaction(BlueStore::TransContext*,
ObjectStore::Transaction*)' thread 7f43c9262700 time 2018-06-08
09:25:36.660177
/home/builder/source/ceph-12.2.2/src/os/bluestore/BlueStore.cc: 9363:
FAILED assert(0 == "unexpected error")

 ceph version 12.2.2 (215dd7151453fae88e6f968c975b6ce309d42dcf) luminous
(stable)
 1: (ceph::__ceph_assert_fail(char const*, char const*, int, char
const*)+0x102) [0x557025013202]
 2: (BlueStore::_txc_add_transaction(BlueStore::TransContext*,
ObjectStore::Transaction*)+0x15fa) [0x557024ec4c3a]
 3: (BlueStore::queue_transactions(ObjectStore::Sequencer*,
std::vector<ObjectStore::Transaction,
std::allocator<ObjectStore::Transaction> >&,
boost::intrusive_ptr<TrackedOp>, ThreadPool::TPHandle*)+0x546)
[0x557024ec62a6]
 4: (ObjectStore::queue_transaction(ObjectStore::Sequencer*,
ObjectStore::Transaction&&, Context*, Context*, Context*,
boost::intrusive_ptr<TrackedOp>, ThreadPool::TPHandle*)+0x14f)
[0x557024a6937f]
 5: (OSD::dispatch_context_transaction(PG::RecoveryCtx&, PG*,
ThreadPool::TPHandle*)+0x6c) [0x5570249ecb5c]
 6: (OSD::process_peering_events(std::__cxx11::list<PG*,
std::allocator<PG*> > const&, ThreadPool::TPHandle&)+0x442) [0x557024a1aec2]
 7: (ThreadPool::BatchWorkQueue<PG>::_void_process(void*,
ThreadPool::TPHandle&)+0x2c) [0x557024a8991c]
 8: (ThreadPool::worker(ThreadPool::WorkThread*)+0xeb8) [0x55702501a3a8]
 9: (ThreadPool::WorkThread::entry()+0x10) [0x55702501b540]
 10: (()+0x7494) [0x7f43e0e48494]
 11: (clone()+0x3f) [0x7f43dfecfaff]
 NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed
to interpret this.


Kind regards,
Caspar Smit

2018-06-08 16:38 GMT+02:00 Caspar Smit <caspars...@supernas.eu>:

> Hi all,
>
> I seem to be hitting these tracker issues:
>
> https://tracker.ceph.com/issues/23145
> http://tracker.ceph.com/issues/24422
>
> PG's 6.1 and 6.3f are having the issues
>
> When i list all PG's of a down OSD with:
>
> ceph-objectstore-tool --dry-run --type bluestore --data-path
> /var/lib/ceph/osd/ceph-17/ --op list-pgs
>
> There are a lot of 'double' pgid's like (also for other pg's):
>
> 6.3fs3
> 6.3fs5
>
> Is that normal? I would assume different shards for EC would be on
> seperate OSD's
>
> We still have 4 OSD's down and 2 PG's down+remapped and i can't find any
> way to get the crashed OSD's back up.
>
>     pg 6.1 is down+remapped, acting [6,3,2147483647,29,2147483647,
> 2147483647]
>     pg 6.3f is down+remapped, acting [20,24,2147483647,2147483647,3,28]
>
> Kind regards,
> Caspar Smit
>
> 2018-06-08 8:53 GMT+02:00 Caspar Smit <caspars...@supernas.eu>:
>
>> Update:
>>
>> I've unset nodown to let it continue but now 4 osd's are down and cannot
>> be brought up again, here's what the lofgfile reads:
>>
>> 2018-06-08 08:35:01.716245 7f4c58de4700  0 log_channel(cluster) log [INF]
>> : 6.e3s0 continuing backfill to osd.37(4) from (10864'911406,11124'921472]
>> 6:c7d71bbd:::rbd_data.5.6c1d9574b0dc51.0000000000bf38b9:head to
>> 11124'921472
>> 2018-06-08 08:35:01.727261 7f4c585e3700 -1 
>> bluestore(/var/lib/ceph/osd/ceph-16)
>> _txc_add_transaction error (2) No such file or directory not handled on
>> operation 30 (op 0, counting from 0)
>> 2018-06-08 08:35:01.727273 7f4c585e3700 -1 
>> bluestore(/var/lib/ceph/osd/ceph-16)
>> ENOENT on clone suggests osd bug
>>
>> 2018-06-08 08:35:01.730584 7f4c585e3700 -1 /home/builder/source/ceph-12.2
>> .2/src/os/bluestore/BlueStore.cc: In function 'void
>> BlueStore::_txc_add_transaction(BlueStore::TransContext*,
>> ObjectStore::Transaction*)' thread 7f4c585e3700 time 2018-06-08
>> 08:35:01.727379
>> /home/builder/source/ceph-12.2.2/src/os/bluestore/BlueStore.cc: 9363:
>> FAILED assert(0 == "unexpected error")
>>
>>  ceph version 12.2.2 (215dd7151453fae88e6f968c975b6ce309d42dcf) luminous
>> (stable)
>>  1: (ceph::__ceph_assert_fail(char const*, char const*, int, char
>> const*)+0x102) [0x558e08ba4202]
>>  2: (BlueStore::_txc_add_transaction(BlueStore::TransContext*,
>> ObjectStore::Transaction*)+0x15fa) [0x558e08a55c3a]
>>  3: (BlueStore::queue_transactions(ObjectStore::Sequencer*,
>> std::vector<ObjectStore::Transaction, 
>> std::allocator<ObjectStore::Transaction>
>> >&, boost::intrusive_ptr<TrackedOp>, ThreadPool::TPHandle*)+0x546)
>> [0x558e08a572a6]
>>  4: (ObjectStore::queue_transaction(ObjectStore::Sequencer*,
>> ObjectStore::Transaction&&, Context*, Context*, Context*,
>> boost::intrusive_ptr<TrackedOp>, ThreadPool::TPHandle*)+0x14f)
>> [0x558e085fa37f]
>>  5: (OSD::dispatch_context_transaction(PG::RecoveryCtx&, PG*,
>> ThreadPool::TPHandle*)+0x6c) [0x558e0857db5c]
>>  6: (OSD::process_peering_events(std::__cxx11::list<PG*,
>> std::allocator<PG*> > const&, ThreadPool::TPHandle&)+0x442) [0x558e085abec2]
>>  7: (ThreadPool::BatchWorkQueue<PG>::_void_process(void*,
>> ThreadPool::TPHandle&)+0x2c) [0x558e0861a91c]
>>  8: (ThreadPool::worker(ThreadPool::WorkThread*)+0xeb8) [0x558e08bab3a8]
>>  9: (ThreadPool::WorkThread::entry()+0x10) [0x558e08bac540]
>>  10: (()+0x7494) [0x7f4c709ca494]
>>  11: (clone()+0x3f) [0x7f4c6fa51aff]
>>  NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed
>> to interpret this.
>>
>> Any help is highly appreciated.
>>
>> Kind regards,
>> Caspar Smit
>>
>>
>> 2018-06-08 7:57 GMT+02:00 Caspar Smit <caspars...@supernas.eu>:
>>
>>> Well i let it run with flags nodown and it looked like it would finish
>>> BUT it all went wrong somewhere:
>>>
>>> This is now the state:
>>>
>>>     health: HEALTH_ERR
>>>             nodown flag(s) set
>>>             5602396/94833780 objects misplaced (5.908%)
>>>             Reduced data availability: 143 pgs inactive, 142 pgs
>>> peering, 7 pgs stale
>>>             Degraded data redundancy: 248859/94833780 objects degraded
>>> (0.262%), 194 pgs unclean, 21 pgs degraded, 12 pgs undersized
>>>             11 stuck requests are blocked > 4096 sec
>>>
>>>     pgs:     13.965% pgs not active
>>>              248859/94833780 objects degraded (0.262%)
>>>              5602396/94833780 objects misplaced (5.908%)
>>>              830 active+clean
>>>              75  remapped+peering
>>>              66  peering
>>>              26  active+remapped+backfill_wait
>>>              6   active+undersized+degraded+remapped+backfill_wait
>>>              6   active+recovery_wait+degraded+remapped
>>>              3   active+undersized+degraded+remapped+backfilling
>>>              3   stale+active+undersized+degraded+remapped+backfill_wait
>>>              3   stale+active+remapped+backfill_wait
>>>              2   active+recovery_wait+degraded
>>>              2   active+remapped+backfilling
>>>              1   activating+degraded+remapped
>>>              1   stale+remapped+peering
>>>
>>>
>>> #ceph health detail shows:
>>>
>>> REQUEST_STUCK 11 stuck requests are blocked > 4096 sec
>>>     11 ops are blocked > 16777.2 sec
>>>     osds 4,7,23,24 have stuck requests > 16777.2 sec
>>>
>>>
>>> So what happened and what should i do now?
>>>
>>> Thank you very much for any help
>>>
>>> Kind regards,
>>> Caspar
>>>
>>>
>>> 2018-06-07 13:33 GMT+02:00 Sage Weil <s...@newdream.net>:
>>>
>>>> On Wed, 6 Jun 2018, Caspar Smit wrote:
>>>> > Hi all,
>>>> >
>>>> > We have a Luminous 12.2.2 cluster with 3 nodes and i recently added a
>>>> node
>>>> > to it.
>>>> >
>>>> > osd-max-backfills is at the default 1 so backfilling didn't go very
>>>> fast
>>>> > but that doesn't matter.
>>>> >
>>>> > Once it started backfilling everything looked ok:
>>>> >
>>>> > ~300 pgs in backfill_wait
>>>> > ~10 pgs backfilling (~number of new osd's)
>>>> >
>>>> > But i noticed the degraded objects increasing a lot. I presume a pg
>>>> that is
>>>> > in backfill_wait state doesn't accept any new writes anymore? Hence
>>>> > increasing the degraded objects?
>>>> >
>>>> > So far so good, but once a while i noticed a random OSD flapping
>>>> (they come
>>>> > back up automatically). This isn't because the disk is saturated but a
>>>> > driver/controller/kernel incompatibility which 'hangs' the disk for a
>>>> short
>>>> > time (scsi abort_task error in syslog). Investigating further i
>>>> noticed
>>>> > this was already the case before the node expansion.
>>>> >
>>>> > These OSD's flapping results in lots of pg states which are a bit
>>>> worrying:
>>>> >
>>>> >              109 active+remapped+backfill_wait
>>>> >              80  active+undersized+degraded+remapped+backfill_wait
>>>> >              51  active+recovery_wait+degraded+remapped
>>>> >              41  active+recovery_wait+degraded
>>>> >              27  active+recovery_wait+undersized+degraded+remapped
>>>> >              14  active+undersized+remapped+backfill_wait
>>>> >              4   active+undersized+degraded+remapped+backfilling
>>>> >
>>>> > I think the recovery_wait is more important then the backfill_wait,
>>>> so i
>>>> > like to prioritize these because the recovery_wait was triggered by
>>>> the
>>>> > flapping OSD's
>>>>
>>>> Just a note: this is fixed in mimic.  Previously, we would choose the
>>>> highest-priority PG to start recovery on at the time, but once recovery
>>>> had started, the appearance of a new PG with a higher priority (e.g.,
>>>> because it finished peering after the others) wouldn't preempt/cancel
>>>> the
>>>> other PG's recovery, so you would get behavior like the above.
>>>>
>>>> Mimic implements that preemption, so you should not see behavior like
>>>> this.  (If you do, then the function that assigns a priority score to a
>>>> PG needs to be tweaked.)
>>>>
>>>> sage
>>>>
>>>
>>>
>>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to