Update:
I rebuilt ceph-osd with latest PR and it started, worked for a few minutes and 
eventually failed on enospc.
After that ceph-bluestore-tool repair started to fail on enospc again. I was 
unable to collect ceph-osd log, so emailed you the most recent repair log.



> On 3.10.2018, at 13:58, Sergey Malinin <h...@newmail.com> wrote:
> 
> Repair has gone farther but failed on something different - this time it 
> appears to be related to store inconsistency rather than lack of free space. 
> Emailed log to you, beware: over 2GB uncompressed.
> 
> 
>> On 3.10.2018, at 13:15, Igor Fedotov <ifedo...@suse.de> wrote:
>> 
>> You may want to try new updates from the PR along with disabling flush on 
>> recovery for rocksdb (avoid_flush_during_recovery parameter).
>> 
>> Full cmd line might looks like:
>> 
>> CEPH_ARGS="--bluestore_rocksdb_options avoid_flush_during_recovery=1" 
>> bin/ceph-bluestore-tool --path <osd path> repair
>> 
>> 
>> To be applied for "non-expanded" OSDs where repair didn't pass.
>> 
>> Please collect a log during repair...
>> 
>> 
>> Thanks,
>> 
>> Igor
>> 
>> On 10/2/2018 4:32 PM, Sergey Malinin wrote:
>>> Repair goes through only when LVM volume has been expanded, otherwise it 
>>> fails with enospc as well as any other operation. However, expanding the 
>>> volume immediately renders bluefs unmountable with IO error.
>>> 2 of 3 OSDs got bluefs log currupted (bluestore tool segfaults at the very 
>>> end of bluefs-log-dump), I'm not sure whether corruption occurred before or 
>>> after volume expansion.
>>> 
>>> 
>>>> On 2.10.2018, at 16:07, Igor Fedotov <ifedo...@suse.de> wrote:
>>>> 
>>>> You mentioned repair had worked before, is that correct? What's the 
>>>> difference now except the applied patch? Different OSD? Anything else?
>>>> 
>>>> 
>>>> On 10/2/2018 3:52 PM, Sergey Malinin wrote:
>>>> 
>>>>> It didn't work, emailed logs to you.
>>>>> 
>>>>> 
>>>>>> On 2.10.2018, at 14:43, Igor Fedotov <ifedo...@suse.de> wrote:
>>>>>> 
>>>>>> The major change is in get_bluefs_rebalance_txn function, it lacked 
>>>>>> bluefs_rebalance_txn assignment..
>>>>>> 
>>>>>> 
>>>>>> 
>>>>>> On 10/2/2018 2:40 PM, Sergey Malinin wrote:
>>>>>>> PR doesn't seem to have changed since yesterday. Am I missing something?
>>>>>>> 
>>>>>>> 
>>>>>>>> On 2.10.2018, at 14:15, Igor Fedotov <ifedo...@suse.de> wrote:
>>>>>>>> 
>>>>>>>> Please update the patch from the PR - it didn't update bluefs extents 
>>>>>>>> list before.
>>>>>>>> 
>>>>>>>> Also please set debug bluestore 20 when re-running repair and collect 
>>>>>>>> the log.
>>>>>>>> 
>>>>>>>> If repair doesn't help - would you send repair and startup logs 
>>>>>>>> directly to me as I have some issues accessing ceph-post-file uploads.
>>>>>>>> 
>>>>>>>> 
>>>>>>>> Thanks,
>>>>>>>> 
>>>>>>>> Igor
>>>>>>>> 
>>>>>>>> 
>>>>>>>> On 10/2/2018 11:39 AM, Sergey Malinin wrote:
>>>>>>>>> Yes, I did repair all OSDs and it finished with 'repair success'. I 
>>>>>>>>> backed up OSDs so now I have more room to play.
>>>>>>>>> I posted log files using ceph-post-file with the following IDs:
>>>>>>>>> 4af9cc4d-9c73-41c9-9c38-eb6c551047a0
>>>>>>>>> 20df7df5-f0c9-4186-aa21-4e5c0172cd93
>>>>>>>>> 
>>>>>>>>> 
>>>>>>>>>> On 2.10.2018, at 11:26, Igor Fedotov <ifedo...@suse.de> wrote:
>>>>>>>>>> 
>>>>>>>>>> You did repair for any of this OSDs, didn't you? For all of them?
>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>>> Would you please provide the log for both types (failed on mount and 
>>>>>>>>>> failed with enospc) of failing OSDs. Prior to collecting please 
>>>>>>>>>> remove existing ones prior and set debug bluestore to 20.
>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>>> 
>>>>>>>>>> On 10/2/2018 2:16 AM, Sergey Malinin wrote:
>>>>>>>>>>> I was able to apply patches to mimic, but nothing changed. One osd 
>>>>>>>>>>> that I had space expanded on fails with bluefs mount IO error, 
>>>>>>>>>>> others keep failing with enospc.
>>>>>>>>>>> 
>>>>>>>>>>> 
>>>>>>>>>>>> On 1.10.2018, at 19:26, Igor Fedotov <ifedo...@suse.de> wrote:
>>>>>>>>>>>> 
>>>>>>>>>>>> So you should call repair which rebalances (i.e. allocates 
>>>>>>>>>>>> additional space) BlueFS space. Hence allowing OSD to start.
>>>>>>>>>>>> 
>>>>>>>>>>>> Thanks,
>>>>>>>>>>>> 
>>>>>>>>>>>> Igor
>>>>>>>>>>>> 
>>>>>>>>>>>> 
>>>>>>>>>>>> On 10/1/2018 7:22 PM, Igor Fedotov wrote:
>>>>>>>>>>>>> Not exactly. The rebalancing from this kv_sync_thread still might 
>>>>>>>>>>>>> be deferred due to the nature of this thread (haven't 100% sure 
>>>>>>>>>>>>> though).
>>>>>>>>>>>>> 
>>>>>>>>>>>>> Here is my PR showing the idea (still untested and perhaps 
>>>>>>>>>>>>> unfinished!!!)
>>>>>>>>>>>>> 
>>>>>>>>>>>>> https://github.com/ceph/ceph/pull/24353
>>>>>>>>>>>>> 
>>>>>>>>>>>>> 
>>>>>>>>>>>>> Igor
>>>>>>>>>>>>> 
>>>>>>>>>>>>> 
>>>>>>>>>>>>> On 10/1/2018 7:07 PM, Sergey Malinin wrote:
>>>>>>>>>>>>>> Can you please confirm whether I got this right:
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>> --- BlueStore.cc.bak    2018-10-01 18:54:45.096836419 +0300
>>>>>>>>>>>>>> +++ BlueStore.cc    2018-10-01 19:01:35.937623861 +0300
>>>>>>>>>>>>>> @@ -9049,22 +9049,17 @@
>>>>>>>>>>>>>>        throttle_bytes.put(costs);
>>>>>>>>>>>>>>          PExtentVector bluefs_gift_extents;
>>>>>>>>>>>>>> -      if (bluefs &&
>>>>>>>>>>>>>> -      after_flush - bluefs_last_balance >
>>>>>>>>>>>>>> -      cct->_conf->bluestore_bluefs_balance_interval) {
>>>>>>>>>>>>>> -    bluefs_last_balance = after_flush;
>>>>>>>>>>>>>> -    int r = _balance_bluefs_freespace(&bluefs_gift_extents);
>>>>>>>>>>>>>> -    assert(r >= 0);
>>>>>>>>>>>>>> -    if (r > 0) {
>>>>>>>>>>>>>> -      for (auto& p : bluefs_gift_extents) {
>>>>>>>>>>>>>> -        bluefs_extents.insert(p.offset, p.length);
>>>>>>>>>>>>>> -      }
>>>>>>>>>>>>>> -      bufferlist bl;
>>>>>>>>>>>>>> -      encode(bluefs_extents, bl);
>>>>>>>>>>>>>> -      dout(10) << __func__ << " bluefs_extents now 0x" << 
>>>>>>>>>>>>>> std::hex
>>>>>>>>>>>>>> -           << bluefs_extents << std::dec << dendl;
>>>>>>>>>>>>>> -      synct->set(PREFIX_SUPER, "bluefs_extents", bl);
>>>>>>>>>>>>>> +      int r = _balance_bluefs_freespace(&bluefs_gift_extents);
>>>>>>>>>>>>>> +      ceph_assert(r >= 0);
>>>>>>>>>>>>>> +      if (r > 0) {
>>>>>>>>>>>>>> +    for (auto& p : bluefs_gift_extents) {
>>>>>>>>>>>>>> +      bluefs_extents.insert(p.offset, p.length);
>>>>>>>>>>>>>>      }
>>>>>>>>>>>>>> +    bufferlist bl;
>>>>>>>>>>>>>> +    encode(bluefs_extents, bl);
>>>>>>>>>>>>>> +    dout(10) << __func__ << " bluefs_extents now 0x" << std::hex
>>>>>>>>>>>>>> +         << bluefs_extents << std::dec << dendl;
>>>>>>>>>>>>>> +    synct->set(PREFIX_SUPER, "bluefs_extents", bl);
>>>>>>>>>>>>>>        }
>>>>>>>>>>>>>>          // cleanup sync deferred keys
>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> On 1.10.2018, at 18:39, Igor Fedotov <ifedo...@suse.de> wrote:
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> So you have just a single main device per OSD....
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Then bluestore-tool wouldn't help, it's unable to expand BlueFS 
>>>>>>>>>>>>>>> partition at main device, standalone devices are supported only.
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Given that you're able to rebuild the code I can suggest to 
>>>>>>>>>>>>>>> make a patch that triggers BlueFS rebalance (see code snippet 
>>>>>>>>>>>>>>> below) on repairing.
>>>>>>>>>>>>>>>     PExtentVector bluefs_gift_extents;
>>>>>>>>>>>>>>>     int r = _balance_bluefs_freespace(&bluefs_gift_extents);
>>>>>>>>>>>>>>>     ceph_assert(r >= 0);
>>>>>>>>>>>>>>>     if (r > 0) {
>>>>>>>>>>>>>>>       for (auto& p : bluefs_gift_extents) {
>>>>>>>>>>>>>>>         bluefs_extents.insert(p.offset, p.length);
>>>>>>>>>>>>>>>       }
>>>>>>>>>>>>>>>       bufferlist bl;
>>>>>>>>>>>>>>>       encode(bluefs_extents, bl);
>>>>>>>>>>>>>>>       dout(10) << __func__ << " bluefs_extents now 0x" << 
>>>>>>>>>>>>>>> std::hex
>>>>>>>>>>>>>>>            << bluefs_extents << std::dec << dendl;
>>>>>>>>>>>>>>>       synct->set(PREFIX_SUPER, "bluefs_extents", bl);
>>>>>>>>>>>>>>>     }
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> If it waits I can probably make a corresponding PR tomorrow.
>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>> Igor
>>>>>>>>>>>>>>> On 10/1/2018 6:16 PM, Sergey Malinin wrote:
>>>>>>>>>>>>>>>> I have rebuilt the tool, but none of my OSDs no matter dead or 
>>>>>>>>>>>>>>>> alive have any symlinks other than 'block' pointing to LVM.
>>>>>>>>>>>>>>>> I adjusted main device size but it looks like it needs even 
>>>>>>>>>>>>>>>> more space for db compaction. After executing 
>>>>>>>>>>>>>>>> bluefs-bdev-expand OSD fails to start, however 'fsck' and 
>>>>>>>>>>>>>>>> 'repair' commands finished successfully.
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> 2018-10-01 18:02:39.755 7fc9226c6240  1 freelist init
>>>>>>>>>>>>>>>> 2018-10-01 18:02:39.763 7fc9226c6240  1 
>>>>>>>>>>>>>>>> bluestore(/var/lib/ceph/osd/ceph-1) _open_alloc opening 
>>>>>>>>>>>>>>>> allocation metadata
>>>>>>>>>>>>>>>> 2018-10-01 18:02:40.907 7fc9226c6240  1 
>>>>>>>>>>>>>>>> bluestore(/var/lib/ceph/osd/ceph-1) _open_alloc loaded 285 GiB 
>>>>>>>>>>>>>>>> in 2249899 extents
>>>>>>>>>>>>>>>> 2018-10-01 18:02:40.951 7fc9226c6240 -1 
>>>>>>>>>>>>>>>> bluestore(/var/lib/ceph/osd/ceph-1) 
>>>>>>>>>>>>>>>> _reconcile_bluefs_freespace bluefs extra 
>>>>>>>>>>>>>>>> 0x[6d6f000000~50c800000]
>>>>>>>>>>>>>>>> 2018-10-01 18:02:40.951 7fc9226c6240  1 stupidalloc 
>>>>>>>>>>>>>>>> 0x0x55d053fb9180 shutdown
>>>>>>>>>>>>>>>> 2018-10-01 18:02:40.963 7fc9226c6240  1 freelist shutdown
>>>>>>>>>>>>>>>> 2018-10-01 18:02:40.963 7fc9226c6240  4 rocksdb: 
>>>>>>>>>>>>>>>> [/build/ceph-13.2.2/src/rocksdb/db/db_impl.cc:252] Shutdown: 
>>>>>>>>>>>>>>>> canceling all background work
>>>>>>>>>>>>>>>> 2018-10-01 18:02:40.967 7fc9226c6240  4 rocksdb: 
>>>>>>>>>>>>>>>> [/build/ceph-13.2.2/src/rocksdb/db/db_impl.cc:397] Shutdown 
>>>>>>>>>>>>>>>> complete
>>>>>>>>>>>>>>>> 2018-10-01 18:02:40.971 7fc9226c6240  1 bluefs umount
>>>>>>>>>>>>>>>> 2018-10-01 18:02:40.975 7fc9226c6240  1 stupidalloc 
>>>>>>>>>>>>>>>> 0x0x55d053883800 shutdown
>>>>>>>>>>>>>>>> 2018-10-01 18:02:40.975 7fc9226c6240  1 bdev(0x55d053c32e00 
>>>>>>>>>>>>>>>> /var/lib/ceph/osd/ceph-1/block) close
>>>>>>>>>>>>>>>> 2018-10-01 18:02:41.267 7fc9226c6240  1 bdev(0x55d053c32a80 
>>>>>>>>>>>>>>>> /var/lib/ceph/osd/ceph-1/block) close
>>>>>>>>>>>>>>>> 2018-10-01 18:02:41.443 7fc9226c6240 -1 osd.1 0 OSD:init: 
>>>>>>>>>>>>>>>> unable to mount object store
>>>>>>>>>>>>>>>> 2018-10-01 18:02:41.443 7fc9226c6240 -1  ** ERROR: osd init 
>>>>>>>>>>>>>>>> failed: (5) Input/output error
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> On 1.10.2018, at 18:09, Igor Fedotov <ifedo...@suse.de> wrote:
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> Well, actually you can avoid bluestore-tool rebuild.
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> You'll need to edit the first chunk of blocks.db where labels 
>>>>>>>>>>>>>>>>> are stored. (Please make a backup first!!!)
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> Size label is stored at offset 0x52 and is 8 bytes long - 
>>>>>>>>>>>>>>>>> little-endian 64bit integer encoding. (Please verify that old 
>>>>>>>>>>>>>>>>> value at this offset exactly corresponds to you original 
>>>>>>>>>>>>>>>>> volume size and/or 'size' label reported by 
>>>>>>>>>>>>>>>>> ceph-bluestore-tool).
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> So you have to put new DB volume size there. Or you can send 
>>>>>>>>>>>>>>>>> the first 4K chunk (e.g. extracted with dd) along with new DB 
>>>>>>>>>>>>>>>>> volume size (in bytes) to me and I'll do that for you.
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> Igor
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>> On 10/1/2018 5:32 PM, Igor Fedotov wrote:
>>>>>>>>>>>>>>>>>> On 10/1/2018 5:03 PM, Sergey Malinin wrote:
>>>>>>>>>>>>>>>>>>> Before I received your response, I had already added 20GB 
>>>>>>>>>>>>>>>>>>> to the OSD (by epanding LV followed by bluefs-bdev-expand) 
>>>>>>>>>>>>>>>>>>> and ran "ceph-kvstore-tool bluestore-kv <path> compact", 
>>>>>>>>>>>>>>>>>>> however it still needs more space.
>>>>>>>>>>>>>>>>>>> Is that because I didn't update DB size with set-label-key?
>>>>>>>>>>>>>>>>>> In mimic you need to run both "bluefs-bdev-expand" and 
>>>>>>>>>>>>>>>>>> "set-label-key" command to commit bluefs volume expansion.
>>>>>>>>>>>>>>>>>> Unfortunately the last command doesn't handle "size" label 
>>>>>>>>>>>>>>>>>> properly. That's why you might need to backport and rebuild 
>>>>>>>>>>>>>>>>>> with the mentioned commits.
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> What exactly is the label-key that needs to be updated, as 
>>>>>>>>>>>>>>>>>>> I couldn't find which one is related to DB:
>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>> # ceph-bluestore-tool show-label --path 
>>>>>>>>>>>>>>>>>>> /var/lib/ceph/osd/ceph-1
>>>>>>>>>>>>>>>>>>> inferring bluefs devices from bluestore path
>>>>>>>>>>>>>>>>>>> {
>>>>>>>>>>>>>>>>>>>      "/var/lib/ceph/osd/ceph-1/block": {
>>>>>>>>>>>>>>>>>>>          "osd_uuid": "f8f122ee-70a6-4c54-8eb0-9b42205b1ecc",
>>>>>>>>>>>>>>>>>>>          "size": 471305551872,
>>>>>>>>>>>>>>>>>>>          "btime": "2018-07-31 03:06:43.751243",
>>>>>>>>>>>>>>>>>>>          "description": "main",
>>>>>>>>>>>>>>>>>>>          "bluefs": "1",
>>>>>>>>>>>>>>>>>>>          "ceph_fsid": 
>>>>>>>>>>>>>>>>>>> "7d320499-5b3f-453e-831f-60d4db9a4533",
>>>>>>>>>>>>>>>>>>>          "kv_backend": "rocksdb",
>>>>>>>>>>>>>>>>>>>          "magic": "ceph osd volume v026",
>>>>>>>>>>>>>>>>>>>          "mkfs_done": "yes",
>>>>>>>>>>>>>>>>>>>          "osd_key": "XXX",
>>>>>>>>>>>>>>>>>>>          "ready": "ready",
>>>>>>>>>>>>>>>>>>>          "whoami": "1"
>>>>>>>>>>>>>>>>>>>      }
>>>>>>>>>>>>>>>>>>> }
>>>>>>>>>>>>>>>>>> 'size' label but your output is for block(aka slow) device.
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> It should return labels for db/wal devices as well (block.db 
>>>>>>>>>>>>>>>>>> and block.wal symlinks respectively). It works for me in 
>>>>>>>>>>>>>>>>>> master, can't verify with mimic at the moment though..
>>>>>>>>>>>>>>>>>> Here is output for master:
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> # bin/ceph-bluestore-tool show-label --path dev/osd0
>>>>>>>>>>>>>>>>>> inferring bluefs devices from bluestore path
>>>>>>>>>>>>>>>>>> {
>>>>>>>>>>>>>>>>>>     "dev/osd0/block": {
>>>>>>>>>>>>>>>>>>         "osd_uuid": "404dcbe9-3f8d-4ef5-ac59-2582454a9a75",
>>>>>>>>>>>>>>>>>>         "size": 21474836480,
>>>>>>>>>>>>>>>>>>         "btime": "2018-09-10 15:55:09.044039",
>>>>>>>>>>>>>>>>>>         "description": "main",
>>>>>>>>>>>>>>>>>>         "bluefs": "1",
>>>>>>>>>>>>>>>>>>         "ceph_fsid": "56eddc15-11b9-4e0b-9192-e391fbae551c",
>>>>>>>>>>>>>>>>>>         "kv_backend": "rocksdb",
>>>>>>>>>>>>>>>>>>         "magic": "ceph osd volume v026",
>>>>>>>>>>>>>>>>>>         "mkfs_done": "yes",
>>>>>>>>>>>>>>>>>>         "osd_key": 
>>>>>>>>>>>>>>>>>> "AQCsaZZbYTxXJBAAe3jJI4p6WbMjvA8CBBUJbA==",
>>>>>>>>>>>>>>>>>>         "ready": "ready",
>>>>>>>>>>>>>>>>>>         "whoami": "0"
>>>>>>>>>>>>>>>>>>     },
>>>>>>>>>>>>>>>>>>     "dev/osd0/block.wal": {
>>>>>>>>>>>>>>>>>>         "osd_uuid": "404dcbe9-3f8d-4ef5-ac59-2582454a9a75",
>>>>>>>>>>>>>>>>>>         "size": 1048576000,
>>>>>>>>>>>>>>>>>>         "btime": "2018-09-10 15:55:09.044985",
>>>>>>>>>>>>>>>>>>         "description": "bluefs wal"
>>>>>>>>>>>>>>>>>>     },
>>>>>>>>>>>>>>>>>>     "dev/osd0/block.db": {
>>>>>>>>>>>>>>>>>>         "osd_uuid": "404dcbe9-3f8d-4ef5-ac59-2582454a9a75",
>>>>>>>>>>>>>>>>>>         "size": 1048576000,
>>>>>>>>>>>>>>>>>>         "btime": "2018-09-10 15:55:09.044469",
>>>>>>>>>>>>>>>>>>         "description": "bluefs db"
>>>>>>>>>>>>>>>>>>     }
>>>>>>>>>>>>>>>>>> }
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> You can try --dev option instead of --path, e.g.
>>>>>>>>>>>>>>>>>> ceph-bluestore-tool show-label --dev <path-to-block.db>
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> On 1.10.2018, at 16:48, Igor Fedotov <ifedo...@suse.de> 
>>>>>>>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> This looks like a sort of deadlock when BlueFS needs some 
>>>>>>>>>>>>>>>>>>>> additional space to replay the log left after the crash. 
>>>>>>>>>>>>>>>>>>>> Which happens during BlueFS open.
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> But such a space (at slow device as DB is full) is gifted 
>>>>>>>>>>>>>>>>>>>> in background during bluefs rebalance procedure which will 
>>>>>>>>>>>>>>>>>>>> occur after the open.
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> Hence OSDs stuck in permanent crashing..
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> The only way to recover I can suggest for now is to expand 
>>>>>>>>>>>>>>>>>>>> DB volumes. You can do that with lvm tools if you have any 
>>>>>>>>>>>>>>>>>>>> spare space for that.
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> Once resized you'll need ceph-bluestore-tool to indicate 
>>>>>>>>>>>>>>>>>>>> volume expansion to BlueFS (bluefs-bdev-expand command ) 
>>>>>>>>>>>>>>>>>>>> and finally update DB volume size label with  
>>>>>>>>>>>>>>>>>>>> set-label-key command.
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> The latter is a bit tricky for mimic - you might need to 
>>>>>>>>>>>>>>>>>>>> backport 
>>>>>>>>>>>>>>>>>>>> https://github.com/ceph/ceph/pull/22085/commits/ffac450da5d6e09cf14b8363b35f21819b48f38b
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> and rebuild ceph-bluestore-tool. Alternatively you can 
>>>>>>>>>>>>>>>>>>>> backport 
>>>>>>>>>>>>>>>>>>>> https://github.com/ceph/ceph/pull/22085/commits/71c3b58da4e7ced3422bce2b1da0e3fa9331530b
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> then bluefs expansion and label updates will occur in a 
>>>>>>>>>>>>>>>>>>>> single step.
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> I'll do these backports in upstream but this will take 
>>>>>>>>>>>>>>>>>>>> some time to pass all the procedures and get into official 
>>>>>>>>>>>>>>>>>>>> mimic release.
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> Will fire a ticket to fix the original issue as well.
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> Igor
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>> On 10/1/2018 3:28 PM, Sergey Malinin wrote:
>>>>>>>>>>>>>>>>>>>>> These are LVM bluestore NVMe SSDs created with 
>>>>>>>>>>>>>>>>>>>>> "ceph-volume --lvm prepare --bluestore /dev/nvme0n1p3" 
>>>>>>>>>>>>>>>>>>>>> i.e. without specifying wal/db devices.
>>>>>>>>>>>>>>>>>>>>> OSDs were created with bluestore_min_alloc_size_ssd=4096, 
>>>>>>>>>>>>>>>>>>>>> another modified setting is 
>>>>>>>>>>>>>>>>>>>>> bluestore_cache_kv_max=1073741824
>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>> DB/block usage collected by prometheus module for 3 
>>>>>>>>>>>>>>>>>>>>> failed and 1 survived OSDs:
>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>> ceph_bluefs_db_total_bytes{ceph_daemon="osd.0"} 
>>>>>>>>>>>>>>>>>>>>> 65493008384.0
>>>>>>>>>>>>>>>>>>>>> ceph_bluefs_db_total_bytes{ceph_daemon="osd.1"} 
>>>>>>>>>>>>>>>>>>>>> 49013587968.0
>>>>>>>>>>>>>>>>>>>>> ceph_bluefs_db_total_bytes{ceph_daemon="osd.2"} 
>>>>>>>>>>>>>>>>>>>>> 76834406400.0 --> this one has survived
>>>>>>>>>>>>>>>>>>>>> ceph_bluefs_db_total_bytes{ceph_daemon="osd.3"} 
>>>>>>>>>>>>>>>>>>>>> 63726157824.0
>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>> ceph_bluefs_db_used_bytes{ceph_daemon="osd.0"} 
>>>>>>>>>>>>>>>>>>>>> 65217232896.0
>>>>>>>>>>>>>>>>>>>>> ceph_bluefs_db_used_bytes{ceph_daemon="osd.1"} 
>>>>>>>>>>>>>>>>>>>>> 48944381952.0
>>>>>>>>>>>>>>>>>>>>> ceph_bluefs_db_used_bytes{ceph_daemon="osd.2"} 
>>>>>>>>>>>>>>>>>>>>> 68093476864.0
>>>>>>>>>>>>>>>>>>>>> ceph_bluefs_db_used_bytes{ceph_daemon="osd.3"} 
>>>>>>>>>>>>>>>>>>>>> 63632834560.0
>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>> ceph_osd_stat_bytes{ceph_daemon="osd.0"} 471305551872.0
>>>>>>>>>>>>>>>>>>>>> ceph_osd_stat_bytes{ceph_daemon="osd.1"} 471305551872.0
>>>>>>>>>>>>>>>>>>>>> ceph_osd_stat_bytes{ceph_daemon="osd.2"} 471305551872.0
>>>>>>>>>>>>>>>>>>>>> ceph_osd_stat_bytes{ceph_daemon="osd.3"} 471305551872.0
>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>> ceph_osd_stat_bytes_used{ceph_daemon="osd.0"} 
>>>>>>>>>>>>>>>>>>>>> 222328213504.0
>>>>>>>>>>>>>>>>>>>>> ceph_osd_stat_bytes_used{ceph_daemon="osd.1"} 
>>>>>>>>>>>>>>>>>>>>> 214472544256.0
>>>>>>>>>>>>>>>>>>>>> ceph_osd_stat_bytes_used{ceph_daemon="osd.2"} 
>>>>>>>>>>>>>>>>>>>>> 163603996672.0
>>>>>>>>>>>>>>>>>>>>> ceph_osd_stat_bytes_used{ceph_daemon="osd.3"} 
>>>>>>>>>>>>>>>>>>>>> 212806815744.0
>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>> First crashed OSD was doing DB compaction, others crashed 
>>>>>>>>>>>>>>>>>>>>> shortly after during backfilling. Workload was 
>>>>>>>>>>>>>>>>>>>>> "ceph-data-scan scan_inodes" filling metadata pool 
>>>>>>>>>>>>>>>>>>>>> located on these OSDs at the rate close to 10k 
>>>>>>>>>>>>>>>>>>>>> objects/second.
>>>>>>>>>>>>>>>>>>>>> Here is the log excerpt of the first crash occurrence:
>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>> 2018-10-01 03:27:12.762 7fbf16dd6700  0 
>>>>>>>>>>>>>>>>>>>>> bluestore(/var/lib/ceph/osd/ceph-1) 
>>>>>>>>>>>>>>>>>>>>> _balance_bluefs_freespace no allocate on 0x80000000 
>>>>>>>>>>>>>>>>>>>>> min_alloc_size 0x1000
>>>>>>>>>>>>>>>>>>>>> 2018-10-01 03:27:12.886 7fbf1e5e5700  4 rocksdb: 
>>>>>>>>>>>>>>>>>>>>> [/build/ceph-13.2.2/src/rocksdb/db/compaction_job.cc:1166]
>>>>>>>>>>>>>>>>>>>>>  [default] [JOB 24] Generated table #89741: 106356 keys, 
>>>>>>>>>>>>>>>>>>>>> 68110589 bytes
>>>>>>>>>>>>>>>>>>>>> 2018-10-01 03:27:12.886 7fbf1e5e5700  4 rocksdb: 
>>>>>>>>>>>>>>>>>>>>> EVENT_LOG_v1 {"time_micros": 1538353632892744, "cf_name": 
>>>>>>>>>>>>>>>>>>>>> "default", "job": 24, "event": "table_file_creation", 
>>>>>>>>>>>>>>>>>>>>> "file_number": 89741, "file_size": 68110589, 
>>>>>>>>>>>>>>>>>>>>> "table_properties": {"data_size": 67112903, "index_size": 
>>>>>>>>>>>>>>>>>>>>> 579319, "filter_size": 417316, "raw_key_size": 6733561, 
>>>>>>>>>>>>>>>>>>>>> "raw_average_key_size": 63, "raw_value_size": 60994583, 
>>>>>>>>>>>>>>>>>>>>> "raw_average_value_size": 573, "num_data_blocks": 16336, 
>>>>>>>>>>>>>>>>>>>>> "num_entries": 106356, "filter_policy_name": 
>>>>>>>>>>>>>>>>>>>>> "rocksdb.BuiltinBloomFilter", "kDeletedKeys": "14444", 
>>>>>>>>>>>>>>>>>>>>> "kMergeOperands": "0"}}
>>>>>>>>>>>>>>>>>>>>> 2018-10-01 03:27:12.934 7fbf1e5e5700  4 rocksdb: 
>>>>>>>>>>>>>>>>>>>>> [/build/ceph-13.2.2/src/rocksdb/db/compaction_job.cc:1166]
>>>>>>>>>>>>>>>>>>>>>  [default] [JOB 24] Generated table #89742: 23214 keys, 
>>>>>>>>>>>>>>>>>>>>> 16352315 bytes
>>>>>>>>>>>>>>>>>>>>> 2018-10-01 03:27:12.934 7fbf1e5e5700  4 rocksdb: 
>>>>>>>>>>>>>>>>>>>>> EVENT_LOG_v1 {"time_micros": 1538353632938670, "cf_name": 
>>>>>>>>>>>>>>>>>>>>> "default", "job": 24, "event": "table_file_creation", 
>>>>>>>>>>>>>>>>>>>>> "file_number": 89742, "file_size": 16352315, 
>>>>>>>>>>>>>>>>>>>>> "table_properties": {"data_size": 16116986, "index_size": 
>>>>>>>>>>>>>>>>>>>>> 139894, "filter_size": 94386, "raw_key_size": 1470883, 
>>>>>>>>>>>>>>>>>>>>> "raw_average_key_size": 63, "raw_value_size": 14775006, 
>>>>>>>>>>>>>>>>>>>>> "raw_average_value_size": 636, "num_data_blocks": 3928, 
>>>>>>>>>>>>>>>>>>>>> "num_entries": 23214, "filter_policy_name": 
>>>>>>>>>>>>>>>>>>>>> "rocksdb.BuiltinBloomFilter", "kDeletedKeys": "90", 
>>>>>>>>>>>>>>>>>>>>> "kMergeOperands": "0"}}
>>>>>>>>>>>>>>>>>>>>> 2018-10-01 03:27:13.042 7fbf1e5e5700  1 bluefs _allocate 
>>>>>>>>>>>>>>>>>>>>> failed to allocate 0x4100000 on bdev 1, free 0x1a00000; 
>>>>>>>>>>>>>>>>>>>>> fallback to bdev 2
>>>>>>>>>>>>>>>>>>>>> 2018-10-01 03:27:13.042 7fbf1e5e5700 -1 bluefs _allocate 
>>>>>>>>>>>>>>>>>>>>> failed to allocate 0x4100000 on bdev 2, dne
>>>>>>>>>>>>>>>>>>>>> 2018-10-01 03:27:13.042 7fbf1e5e5700 -1 bluefs 
>>>>>>>>>>>>>>>>>>>>> _flush_range allocated: 0x0 offset: 0x0 length: 0x40ea9f1
>>>>>>>>>>>>>>>>>>>>> 2018-10-01 03:27:13.046 7fbf1e5e5700 -1 
>>>>>>>>>>>>>>>>>>>>> /build/ceph-13.2.2/src/os/bluestore/BlueFS.cc: In 
>>>>>>>>>>>>>>>>>>>>> function 'int BlueFS::_flush_range(BlueFS::FileWriter*, 
>>>>>>>>>>>>>>>>>>>>> uint64_t, uint64_t)' thread 7fbf1e5e5700 time 2018-10-01 
>>>>>>>>>>>>>>>>>>>>> 03:27:13.048298
>>>>>>>>>>>>>>>>>>>>> /build/ceph-13.2.2/src/os/bluestore/BlueFS.cc: 1663: 
>>>>>>>>>>>>>>>>>>>>> FAILED assert(0 == "bluefs enospc")
>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>>   ceph version 13.2.2 
>>>>>>>>>>>>>>>>>>>>> (02899bfda814146b021136e9d8e80eba494e1126) mimic (stable)
>>>>>>>>>>>>>>>>>>>>>   1: (ceph::__ceph_assert_fail(char const*, char const*, 
>>>>>>>>>>>>>>>>>>>>> int, char const*)+0x102) [0x7fbf2d4fe5c2]
>>>>>>>>>>>>>>>>>>>>>   2: (()+0x26c787) [0x7fbf2d4fe787]
>>>>>>>>>>>>>>>>>>>>>   3: (BlueFS::_flush_range(BlueFS::FileWriter*, unsigned 
>>>>>>>>>>>>>>>>>>>>> long, unsigned long)+0x1ab4) [0x5619325114b4]
>>>>>>>>>>>>>>>>>>>>>   4: (BlueRocksWritableFile::Flush()+0x3d) 
>>>>>>>>>>>>>>>>>>>>> [0x561932527c1d]
>>>>>>>>>>>>>>>>>>>>>   5: (rocksdb::WritableFileWriter::Flush()+0x1b9) 
>>>>>>>>>>>>>>>>>>>>> [0x56193271c399]
>>>>>>>>>>>>>>>>>>>>>   6: (rocksdb::WritableFileWriter::Sync(bool)+0x3b) 
>>>>>>>>>>>>>>>>>>>>> [0x56193271d42b]
>>>>>>>>>>>>>>>>>>>>>   7: 
>>>>>>>>>>>>>>>>>>>>> (rocksdb::CompactionJob::FinishCompactionOutputFile(rocksdb::Status
>>>>>>>>>>>>>>>>>>>>>  const&, rocksdb::CompactionJob::SubcompactionState*, 
>>>>>>>>>>>>>>>>>>>>> rocksdb::RangeDelAggregator*, CompactionIterationStats*, 
>>>>>>>>>>>>>>>>>>>>> rocksdb::Slice const*)+0x3db) [0x56193276098b]
>>>>>>>>>>>>>>>>>>>>>   8: 
>>>>>>>>>>>>>>>>>>>>> (rocksdb::CompactionJob::ProcessKeyValueCompaction(rocksdb::CompactionJob::SubcompactionState*)+0x7d9)
>>>>>>>>>>>>>>>>>>>>>  [0x561932763da9]
>>>>>>>>>>>>>>>>>>>>>   9: (rocksdb::CompactionJob::Run()+0x314) 
>>>>>>>>>>>>>>>>>>>>> [0x561932765504]
>>>>>>>>>>>>>>>>>>>>>   10: (rocksdb::DBImpl::BackgroundCompaction(bool*, 
>>>>>>>>>>>>>>>>>>>>> rocksdb::JobContext*, rocksdb::LogBuffer*, 
>>>>>>>>>>>>>>>>>>>>> rocksdb::DBImpl::PrepickedCompaction*)+0xc54) 
>>>>>>>>>>>>>>>>>>>>> [0x5619325b5c44]
>>>>>>>>>>>>>>>>>>>>>   11: 
>>>>>>>>>>>>>>>>>>>>> (rocksdb::DBImpl::BackgroundCallCompaction(rocksdb::DBImpl::PrepickedCompaction*,
>>>>>>>>>>>>>>>>>>>>>  rocksdb::Env::Priority)+0x397) [0x5619325b8557]
>>>>>>>>>>>>>>>>>>>>>   12: (rocksdb::DBImpl::BGWorkCompaction(void*)+0x97) 
>>>>>>>>>>>>>>>>>>>>> [0x5619325b8cd7]
>>>>>>>>>>>>>>>>>>>>>   13: (rocksdb::ThreadPoolImpl::Impl::BGThread(unsigned 
>>>>>>>>>>>>>>>>>>>>> long)+0x266) [0x5619327a5e36]
>>>>>>>>>>>>>>>>>>>>>   14: 
>>>>>>>>>>>>>>>>>>>>> (rocksdb::ThreadPoolImpl::Impl::BGThreadWrapper(void*)+0x47)
>>>>>>>>>>>>>>>>>>>>>  [0x5619327a5fb7]
>>>>>>>>>>>>>>>>>>>>>   15: (()+0xbe733) [0x7fbf2b500733]
>>>>>>>>>>>>>>>>>>>>>   16: (()+0x76db) [0x7fbf2bbf86db]
>>>>>>>>>>>>>>>>>>>>>   17: (clone()+0x3f) [0x7fbf2abbc88f]
>>>>>>>>>>>>>>>>>>>>>   NOTE: a copy of the executable, or `objdump -rdS 
>>>>>>>>>>>>>>>>>>>>> <executable>` is needed to interpret this.
>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>>> On 1.10.2018, at 15:01, Igor Fedotov <ifedo...@suse.de> 
>>>>>>>>>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>>> Hi Sergey,
>>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>>> could you please provide more details on your OSDs ?
>>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>>> What are sizes for DB/block devices?
>>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>>> Do you have any modifications in BlueStore config 
>>>>>>>>>>>>>>>>>>>>>> settings?
>>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>>> Can you share stats you're referring to?
>>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>>> Thanks,
>>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>>> Igor
>>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>>> On 10/1/2018 12:29 PM, Sergey Malinin wrote:
>>>>>>>>>>>>>>>>>>>>>>> Hello,
>>>>>>>>>>>>>>>>>>>>>>> 3 of 4 NVME OSDs crashed at the same time on assert(0 
>>>>>>>>>>>>>>>>>>>>>>> == "bluefs enospc") and no longer start.
>>>>>>>>>>>>>>>>>>>>>>> Stats collected just before crash show that 
>>>>>>>>>>>>>>>>>>>>>>> ceph_bluefs_db_used_bytes is 100% used. Although OSDs 
>>>>>>>>>>>>>>>>>>>>>>> have over 50% of free space, it is not reallocated for 
>>>>>>>>>>>>>>>>>>>>>>> DB usage.
>>>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>>>> 2018-10-01 12:18:06.744 7f1d6a04d240  1 bluefs 
>>>>>>>>>>>>>>>>>>>>>>> _allocate failed to allocate 0x100000 on bdev 1, free 
>>>>>>>>>>>>>>>>>>>>>>> 0x0; fallback to bdev 2
>>>>>>>>>>>>>>>>>>>>>>> 2018-10-01 12:18:06.744 7f1d6a04d240 -1 bluefs 
>>>>>>>>>>>>>>>>>>>>>>> _allocate failed to allocate 0x100000 on bdev 2, dne
>>>>>>>>>>>>>>>>>>>>>>> 2018-10-01 12:18:06.744 7f1d6a04d240 -1 bluefs 
>>>>>>>>>>>>>>>>>>>>>>> _flush_range allocated: 0x0 offset: 0x0 length: 0xa8700
>>>>>>>>>>>>>>>>>>>>>>> 2018-10-01 12:18:06.748 7f1d6a04d240 -1 
>>>>>>>>>>>>>>>>>>>>>>> /build/ceph-13.2.2/src/os/bluestore/BlueFS.cc: In 
>>>>>>>>>>>>>>>>>>>>>>> function 'int BlueFS::_flush_range(BlueFS::FileWriter*, 
>>>>>>>>>>>>>>>>>>>>>>> uint64_t, uint64_t)' thread 7f1d6a04d240 time 
>>>>>>>>>>>>>>>>>>>>>>> 2018-10-01 12:18:06.746800
>>>>>>>>>>>>>>>>>>>>>>> /build/ceph-13.2.2/src/os/bluestore/BlueFS.cc: 1663: 
>>>>>>>>>>>>>>>>>>>>>>> FAILED assert(0 == "bluefs enospc")
>>>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>>>>   ceph version 13.2.2 
>>>>>>>>>>>>>>>>>>>>>>> (02899bfda814146b021136e9d8e80eba494e1126) mimic 
>>>>>>>>>>>>>>>>>>>>>>> (stable)
>>>>>>>>>>>>>>>>>>>>>>>   1: (ceph::__ceph_assert_fail(char const*, char 
>>>>>>>>>>>>>>>>>>>>>>> const*, int, char const*)+0x102) [0x7f1d6146f5c2]
>>>>>>>>>>>>>>>>>>>>>>>   2: (()+0x26c787) [0x7f1d6146f787]
>>>>>>>>>>>>>>>>>>>>>>>   3: (BlueFS::_flush_range(BlueFS::FileWriter*, 
>>>>>>>>>>>>>>>>>>>>>>> unsigned long, unsigned long)+0x1ab4) [0x5586b22684b4]
>>>>>>>>>>>>>>>>>>>>>>>   4: (BlueRocksWritableFile::Flush()+0x3d) 
>>>>>>>>>>>>>>>>>>>>>>> [0x5586b227ec1d]
>>>>>>>>>>>>>>>>>>>>>>>   5: (rocksdb::WritableFileWriter::Flush()+0x1b9) 
>>>>>>>>>>>>>>>>>>>>>>> [0x5586b2473399]
>>>>>>>>>>>>>>>>>>>>>>>   6: (rocksdb::WritableFileWriter::Sync(bool)+0x3b) 
>>>>>>>>>>>>>>>>>>>>>>> [0x5586b247442b]
>>>>>>>>>>>>>>>>>>>>>>>   7: 
>>>>>>>>>>>>>>>>>>>>>>> (rocksdb::BuildTable(std::__cxx11::basic_string<char, 
>>>>>>>>>>>>>>>>>>>>>>> std::char_traits<char>, std::allocator<char> > const&, 
>>>>>>>>>>>>>>>>>>>>>>> rocksdb::Env*, rocksdb::ImmutableCFOptions const&, 
>>>>>>>>>>>>>>>>>>>>>>> rocksdb::MutableCFOptions const&, rocksdb::EnvOptions 
>>>>>>>>>>>>>>>>>>>>>>> const&, rock
>>>>>>>>>>>>>>>>>>>>>>> sdb::TableCache*, rocksdb::InternalIterator*, 
>>>>>>>>>>>>>>>>>>>>>>> std::unique_ptr<rocksdb::InternalIterator, 
>>>>>>>>>>>>>>>>>>>>>>> std::default_delete<rocksdb::InternalIterator> >, 
>>>>>>>>>>>>>>>>>>>>>>> rocksdb::FileMetaData*, rocksdb::InternalKeyComparator 
>>>>>>>>>>>>>>>>>>>>>>> const&, std::vector<std::unique_ptr<
>>>>>>>>>>>>>>>>>>>>>>> rocksdb::IntTblPropCollectorFactory, 
>>>>>>>>>>>>>>>>>>>>>>> std::default_delete<rocksdb::IntTblPropCollectorFactory>
>>>>>>>>>>>>>>>>>>>>>>>  >, 
>>>>>>>>>>>>>>>>>>>>>>> std::allocator<std::unique_ptr<rocksdb::IntTblPropCollectorFactory,
>>>>>>>>>>>>>>>>>>>>>>>  
>>>>>>>>>>>>>>>>>>>>>>> std::default_delete<rocksdb::IntTblPropCollectorFactory>
>>>>>>>>>>>>>>>>>>>>>>>  > > > co
>>>>>>>>>>>>>>>>>>>>>>> nst*, unsigned int, std::__cxx11::basic_string<char, 
>>>>>>>>>>>>>>>>>>>>>>> std::char_traits<char>, std::allocator<char> > const&, 
>>>>>>>>>>>>>>>>>>>>>>> std::vector<unsigned long, std::allocator<unsigned 
>>>>>>>>>>>>>>>>>>>>>>> long> >, unsigned long, rocksdb::SnapshotChecker*, 
>>>>>>>>>>>>>>>>>>>>>>> rocksdb::Compression
>>>>>>>>>>>>>>>>>>>>>>> Type, rocksdb::CompressionOptions const&, bool, 
>>>>>>>>>>>>>>>>>>>>>>> rocksdb::InternalStats*, 
>>>>>>>>>>>>>>>>>>>>>>> rocksdb::TableFileCreationReason, 
>>>>>>>>>>>>>>>>>>>>>>> rocksdb::EventLogger*, int, rocksdb::Env::IOPriority, 
>>>>>>>>>>>>>>>>>>>>>>> rocksdb::TableProperties*, int, unsigned long, unsigned 
>>>>>>>>>>>>>>>>>>>>>>> long, rocksdb
>>>>>>>>>>>>>>>>>>>>>>> ::Env::WriteLifeTimeHint)+0x1e24) [0x5586b249ef94]
>>>>>>>>>>>>>>>>>>>>>>>   8: (rocksdb::DBImpl::WriteLevel0TableForRecovery(int, 
>>>>>>>>>>>>>>>>>>>>>>> rocksdb::ColumnFamilyData*, rocksdb::MemTable*, 
>>>>>>>>>>>>>>>>>>>>>>> rocksdb::VersionEdit*)+0xcb7) [0x5586b2321457]
>>>>>>>>>>>>>>>>>>>>>>>   9: 
>>>>>>>>>>>>>>>>>>>>>>> (rocksdb::DBImpl::RecoverLogFiles(std::vector<unsigned 
>>>>>>>>>>>>>>>>>>>>>>> long, std::allocator<unsigned long> > const&, unsigned 
>>>>>>>>>>>>>>>>>>>>>>> long*, bool)+0x19de) [0x5586b232373e]
>>>>>>>>>>>>>>>>>>>>>>>   10: 
>>>>>>>>>>>>>>>>>>>>>>> (rocksdb::DBImpl::Recover(std::vector<rocksdb::ColumnFamilyDescriptor,
>>>>>>>>>>>>>>>>>>>>>>>  std::allocator<rocksdb::ColumnFamilyDescriptor> > 
>>>>>>>>>>>>>>>>>>>>>>> const&, bool, bool, bool)+0x5d4) [0x5586b23242f4]
>>>>>>>>>>>>>>>>>>>>>>>   11: (rocksdb::DBImpl::Open(rocksdb::DBOptions const&, 
>>>>>>>>>>>>>>>>>>>>>>> std::__cxx11::basic_string<char, 
>>>>>>>>>>>>>>>>>>>>>>> std::char_traits<char>, std::allocator<char> > const&, 
>>>>>>>>>>>>>>>>>>>>>>> std::vector<rocksdb::ColumnFamilyDescriptor, 
>>>>>>>>>>>>>>>>>>>>>>> std::allocator<rocksdb::ColumnFamilyDescri
>>>>>>>>>>>>>>>>>>>>>>> ptor> > const&, 
>>>>>>>>>>>>>>>>>>>>>>> std::vector<rocksdb::ColumnFamilyHandle*, 
>>>>>>>>>>>>>>>>>>>>>>> std::allocator<rocksdb::ColumnFamilyHandle*> >*, 
>>>>>>>>>>>>>>>>>>>>>>> rocksdb::DB**, bool)+0x68b) [0x5586b232559b]
>>>>>>>>>>>>>>>>>>>>>>>   12: (rocksdb::DB::Open(rocksdb::DBOptions const&, 
>>>>>>>>>>>>>>>>>>>>>>> std::__cxx11::basic_string<char, 
>>>>>>>>>>>>>>>>>>>>>>> std::char_traits<char>, std::allocator<char> > const&, 
>>>>>>>>>>>>>>>>>>>>>>> std::vector<rocksdb::ColumnFamilyDescriptor, 
>>>>>>>>>>>>>>>>>>>>>>> std::allocator<rocksdb::ColumnFamilyDescriptor
>>>>>>>>>>>>>>>>>>>>>>>>> const&, std::vector<rocksdb::ColumnFamilyHandle*, 
>>>>>>>>>>>>>>>>>>>>>>>>> std::allocator<rocksdb::ColumnFamilyHandle*> >*, 
>>>>>>>>>>>>>>>>>>>>>>>>> rocksdb::DB**)+0x22) [0x5586b2326e72]
>>>>>>>>>>>>>>>>>>>>>>>   13: (RocksDBStore::do_open(std::ostream&, bool, 
>>>>>>>>>>>>>>>>>>>>>>> std::vector<KeyValueDB::ColumnFamily, 
>>>>>>>>>>>>>>>>>>>>>>> std::allocator<KeyValueDB::ColumnFamily> > 
>>>>>>>>>>>>>>>>>>>>>>> const*)+0x170c) [0x5586b220219c]
>>>>>>>>>>>>>>>>>>>>>>>   14: (BlueStore::_open_db(bool, bool)+0xd8e) 
>>>>>>>>>>>>>>>>>>>>>>> [0x5586b218ee1e]
>>>>>>>>>>>>>>>>>>>>>>>   15: (BlueStore::_mount(bool, bool)+0x4b7) 
>>>>>>>>>>>>>>>>>>>>>>> [0x5586b21bf807]
>>>>>>>>>>>>>>>>>>>>>>>   16: (OSD::init()+0x295) [0x5586b1d673c5]
>>>>>>>>>>>>>>>>>>>>>>>   17: (main()+0x268d) [0x5586b1c554ed]
>>>>>>>>>>>>>>>>>>>>>>>   18: (__libc_start_main()+0xe7) [0x7f1d5ea2db97]
>>>>>>>>>>>>>>>>>>>>>>>   19: (_start()+0x2a) [0x5586b1d1d7fa]
>>>>>>>>>>>>>>>>>>>>>>>   NOTE: a copy of the executable, or `objdump -rdS 
>>>>>>>>>>>>>>>>>>>>>>> <executable>` is needed to interpret this.
>>>>>>>>>>>>>>>>>>>>>>> 
>>>>>>>>>>>>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>>>>>>>>>>>>> ceph-users mailing list
>>>>>>>>>>>>>>>>>>>>>>> ceph-users@lists.ceph.com
>>>>>>>>>>>>>>>>>>>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>>>>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>>>>>>>> ceph-users mailing list
>>>>>>>>>>>>>>>>>> ceph-users@lists.ceph.com
>>>>>>>>>>>>>>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>>>>>>>>>>>> _______________________________________________
>>>>>>>>>>>>> ceph-users mailing list
>>>>>>>>>>>>> ceph-users@lists.ceph.com
>>>>>>>>>>>>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> 
> 

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to