I'll give that a try. Is it something like... ceph tell 'osd.*' bluestore_allocator stupid ceph tell 'osd.*' bluefs_allocator stupid
And should I expect any issues doing this? On Mon, Jul 8, 2019 at 1:04 PM Igor Fedotov <ifedo...@suse.de> wrote: > I should read call stack more carefully... It's not about lacking free > space - this is rather the bug from this ticket: > > http://tracker.ceph.com/issues/40080 > > > You should upgrade to v14.2.2 (once it's available) or temporarily switch > to stupid allocator as a workaround. > > > Thanks, > > Igor > > > > On 7/8/2019 8:00 PM, Igor Fedotov wrote: > > Hi Brett, > > looks like BlueStore is unable to allocate additional space for BlueFS at > main device. It's either lacking free space or it's too fragmented... > > Would you share osd log, please? > > Also please run "ceph-bluestore-tool --path <substitute with > path-to-osd!!!> bluefs-bdev-sizes" and share the output. > > Thanks, > > Igor > On 7/3/2019 9:59 PM, Brett Chancellor wrote: > > Hi All! Today I've had 3 OSDs stop themselves and are unable to restart, > all with the same error. These OSDs are all on different hosts. All are > running 14.2.1 > > I did try the following two commands > - ceph-kvstore-tool bluestore-kv /var/lib/ceph/osd/ceph-80 list > keys > ## This failed with the same error below > - ceph-bluestore-tool --path /var/lib/ceph/osd/ceph-80 fsck > ## After a couple of hours returned... > 2019-07-03 18:30:02.095 7fe7c1c1ef00 -1 > bluestore(/var/lib/ceph/osd/ceph-80) fsck warning: legacy statfs record > found, suggest to run store repair to get consistent statistic reports > fsck success > > > ## Error when trying to start one of the OSDs > -12> 2019-07-03 18:36:57.450 7f5e42366700 -1 *** Caught signal > (Aborted) ** > in thread 7f5e42366700 thread_name:rocksdb:low0 > > ceph version 14.2.1 (d555a9489eb35f84f2e1ef49b77e19da9d113972) nautilus > (stable) > 1: (()+0xf5d0) [0x7f5e50bd75d0] > 2: (gsignal()+0x37) [0x7f5e4f9ce207] > 3: (abort()+0x148) [0x7f5e4f9cf8f8] > 4: (ceph::__ceph_assert_fail(char const*, char const*, int, char > const*)+0x199) [0x55a7aaee96ab] > 5: (ceph::__ceph_assertf_fail(char const*, char const*, int, char const*, > char const*, ...)+0) [0x55a7aaee982a] > 6: (interval_set<unsigned long, std::map<unsigned long, unsigned long, > std::less<unsigned long>, std::allocator<std::pair<unsigned long const, > unsigned long> > > >::insert(unsigned long, unsigned long, unsigned long*, > unsigned long*)+0x3c6) [0x55a7ab212a66] > 7: (BlueStore::allocate_bluefs_freespace(unsigned long, unsigned long, > std::vector<bluestore_pextent_t, > mempool::pool_allocator<(mempool::pool_index_t)4, bluestore_pextent_t> > >*)+0x74e) [0x55a7ab48253e] > 8: (BlueFS::_expand_slow_device(unsigned long, > std::vector<bluestore_pextent_t, > mempool::pool_allocator<(mempool::pool_index_t)4, bluestore_pextent_t> > >&)+0x111) [0x55a7ab59e921] > 9: (BlueFS::_allocate(unsigned char, unsigned long, > bluefs_fnode_t*)+0x68b) [0x55a7ab59f68b] > 10: (BlueFS::_flush_range(BlueFS::FileWriter*, unsigned long, unsigned > long)+0xe5) [0x55a7ab59fce5] > 11: (BlueFS::_flush(BlueFS::FileWriter*, bool)+0x10b) [0x55a7ab5a1b4b] > 12: (BlueRocksWritableFile::Flush()+0x3d) [0x55a7ab5bf84d] > 13: (rocksdb::WritableFileWriter::Flush()+0x19e) [0x55a7abbedd0e] > 14: (rocksdb::WritableFileWriter::Sync(bool)+0x2e) [0x55a7abbedfee] > 15: (rocksdb::CompactionJob::FinishCompactionOutputFile(rocksdb::Status > const&, rocksdb::CompactionJob::SubcompactionState*, > rocksdb::RangeDelAggregator*, CompactionIterationStats*, rocksdb::Slice > const*)+0xbaa) [0x55a7abc3b73a] > 16: > (rocksdb::CompactionJob::ProcessKeyValueCompaction(rocksdb::CompactionJob::SubcompactionState*)+0x7d0) > [0x55a7abc3f150] > 17: (rocksdb::CompactionJob::Run()+0x298) [0x55a7abc40618] > 18: (rocksdb::DBImpl::BackgroundCompaction(bool*, rocksdb::JobContext*, > rocksdb::LogBuffer*, rocksdb::DBImpl::PrepickedCompaction*)+0xcb7) > [0x55a7aba7fb67] > 19: > (rocksdb::DBImpl::BackgroundCallCompaction(rocksdb::DBImpl::PrepickedCompaction*, > rocksdb::Env::Priority)+0xd0) [0x55a7aba813c0] > 20: (rocksdb::DBImpl::BGWorkCompaction(void*)+0x3a) [0x55a7aba8190a] > 21: (rocksdb::ThreadPoolImpl::Impl::BGThread(unsigned long)+0x264) > [0x55a7abc8d9c4] > 22: (rocksdb::ThreadPoolImpl::Impl::BGThreadWrapper(void*)+0x4f) > [0x55a7abc8db4f] > 23: (()+0x129dfff) [0x55a7abd1afff] > 24: (()+0x7dd5) [0x7f5e50bcfdd5] > 25: (clone()+0x6d) [0x7f5e4fa95ead] > NOTE: a copy of the executable, or `objdump -rdS <executable>` is needed > to interpret this. > > _______________________________________________ > ceph-users mailing > listceph-us...@lists.ceph.comhttp://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > _______________________________________________ > ceph-users mailing > listceph-us...@lists.ceph.comhttp://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com