SOLVED Short: I followed the procedure to replace an OSD.
Long: I reweighted the flapping OSD to 0 until it was done, then marked it out, and then unmounted it, then followed the replacement procedure[1], then restore weight. I attempted to recreate my actions from bash history from various terminals. Please excuse any mistakes: admin:$ X=0 # ID of OSD admin:$ ceph osd crush reweight osd.$X 0.0 admin:$ ceph -w # observe data migration admin:$ ceph osd out $X osd:$ systemctl stop ceph.target admin:$ ceph osd destroy $X --yes-i-really-mean-it osd:$ umount /dev/sdb1 osd:$ ceph-disk zap /dev/sdb osd:$ ceph-disk prepare --bluestore /dev/sdb --osd-id 0 --osd-uuid `uuidgen` osd:$ ceph-disk activate /dev/sdb1 admin:$ ceph osd crush reweight osd.0 9.09560 admin:$ ceph -w # observe data migration # I also found it helpful to set 'nodown' for several minutes because two of the OSDs kept marking eachother down. admin:$ ceph osd set nodown # Once all was 'active+clean', I removed the 'nodown' flag. admin:$ ceph osd unset nodown And for the first time since upgrading from Kraken to all the Luminous releases through 12.1.2, my cluster is finally HEALTH_OK. Yay! roger@desktop:~/ceph-cluster$ ceph -w cluster: id: eea7b78c-b138-40fc-9f3e-3d77afb770f0 health: HEALTH_OK services: mon: 3 daemons, quorum nuc1,nuc2,nuc3 mgr: nuc3(active), standbys: nuc1, nuc2 osd: 3 osds: 3 up, 3 in rgw: 1 daemon active data: pools: 19 pools, 372 pgs objects: 54278 objects, 71724 MB usage: 121 GB used, 27820 GB / 27941 GB avail pgs: 372 active+clean 1. http://docs.ceph.com/docs/master/rados/operations/add-or-rm-osds/#replacing-an-osd On Wed, Aug 2, 2017 at 11:08 AM Roger Brown <rogerpbr...@gmail.com> wrote: > Hi, > > My OSD's were continuously crashing in cephx_verify_authorizer() while on > Luminous v12.1.0 and v12.1.1, but the crashes stopped once I upgraded to > v12.1.2. > > Now however, one of my OSDs is continuing to crash. Looking closer, the > crash reason is different reason and started with v12.1.1. > > I've been troubleshooting with the aid of > http://docs.ceph.com/docs/master/rados/troubleshooting/troubleshooting-osd/ > . > > I'm considering reweight to 0 and then redeploy that OSD from scratch, > unless you can do a filesystem repair on bluestore/rocksdb somehow. Please > advise. > > > Data follows... > > Log: > roger@osd3:~$ sudo journalctl -u ceph-osd@0 --no-pager > ... > Aug 02 10:38:47 osd3 systemd[1]: ceph-osd@0.service: Failed with result > 'signal'. > Aug 02 10:39:07 osd3 systemd[1]: ceph-osd@0.service: Service hold-off > time over, scheduling restart. > Aug 02 10:39:07 osd3 systemd[1]: Stopped Ceph object storage daemon osd.0. > Aug 02 10:39:07 osd3 systemd[1]: Starting Ceph object storage daemon > osd.0... > Aug 02 10:39:07 osd3 systemd[1]: Started Ceph object storage daemon osd.0. > Aug 02 10:39:07 osd3 ceph-osd[7413]: starting osd.0 at - osd_data > /var/lib/ceph/osd/ceph-0 /var/lib/ceph/osd/ceph-0/journal > Aug 02 10:40:48 osd3 ceph-osd[7413]: 2017-08-02 10:40:48.583063 > 7f5262cc3e00 -1 osd.0 25924 log_to_monitors {default=true} > Aug 02 10:43:32 osd3 ceph-osd[7413]: *** Caught signal (Aborted) ** > Aug 02 10:43:32 osd3 ceph-osd[7413]: in thread 7f524861b700 > thread_name:tp_osd_tp > Aug 02 10:43:32 osd3 ceph-osd[7413]: ceph version 12.1.2 > (b661348f156f148d764b998b65b90451f096cb27) luminous (rc) > Aug 02 10:43:32 osd3 ceph-osd[7413]: 1: (()+0xa9a964) [0x5623f0a9c964] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 2: (()+0x11390) [0x7f52611a6390] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 3: (pread64()+0x33) [0x7f52611a5d43] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 4: > (KernelDevice::direct_read_unaligned(unsigned long, unsigned long, > char*)+0x81) [0x5623f0a7cfc1] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 5: > (KernelDevice::read_random(unsigned long, unsigned long, char*, > bool)+0x4f3) [0x5623f0a7da43] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 6: > (BlueFS::_read_random(BlueFS::FileReader*, unsigned long, unsigned long, > char*)+0x4fa) [0x5623f0a4d9ca] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 7: > (BlueRocksRandomAccessFile::Read(unsigned long, unsigned long, > rocksdb::Slice*, char*) const+0x20) [0x5623f0a77e10] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 8: > (rocksdb::RandomAccessFileReader::Read(unsigned long, unsigned long, > rocksdb::Slice*, char*) const+0xf8f) [0x5623f0e50acf] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 9: > (rocksdb::ReadBlockContents(rocksdb::RandomAccessFileReader*, > rocksdb::Footer const&, rocksdb::ReadOptions const&, rocksdb::BlockHandle > const&, rocksdb::BlockContents*, rocksdb::ImmutableCFOptions const&, bool, > rocksdb::Slice const&, rocksdb::PersistentCacheOptions const&)+0x5f3) > [0x5623f0e21383] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 10: (()+0xe0f7c6) [0x5623f0e117c6] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 11: > (rocksdb::BlockBasedTable::MaybeLoadDataBlockToCache(rocksdb::BlockBasedTable::Rep*, > rocksdb::ReadOptions const&, rocksdb::BlockHandle const&, rocksdb::Slice, > rocksdb::BlockBasedTable::CachableEntry<rocksdb::Block>*, bool)+0x2f8) > [0x5623f0e13928] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 12: > (rocksdb::BlockBasedTable::NewDataBlockIterator(rocksdb::BlockBasedTable::Rep*, > rocksdb::ReadOptions const&, rocksdb::BlockHandle const&, > rocksdb::BlockIter*, bool, rocksdb::Status)+0x2ac) [0x5623f0e13d2c] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 13: > (rocksdb::BlockBasedTable::BlockEntryIteratorState::NewSecondaryIterator(rocksdb::Slice > const&)+0x97) [0x5623f0e1c4a7] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 14: (()+0xe4576e) [0x5623f0e4776e] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 15: (()+0xe45836) [0x5623f0e47836] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 16: (()+0xe459b1) [0x5623f0e479b1] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 17: > (rocksdb::MergingIterator::Next()+0x449) [0x5623f0e2ab09] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 18: > (rocksdb::DBIter::FindNextUserEntryInternal(bool, bool)+0x182) > [0x5623f0ec7ed2] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 19: (rocksdb::DBIter::Next()+0x1eb) > [0x5623f0ec8c8b] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 20: > (RocksDBStore::RocksDBWholeSpaceIteratorImpl::next()+0x9a) [0x5623f09dd58a] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 21: > (BlueStore::_collection_list(BlueStore::Collection*, ghobject_t const&, > ghobject_t const&, int, std::vector<ghobject_t, std::allocator<ghobject_t> > >*, ghobject_t*)+0x1170) [0x5623f093d250] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 22: > (BlueStore::collection_list(boost::intrusive_ptr<ObjectStore::CollectionImpl>&, > ghobject_t const&, ghobject_t const&, int, std::vector<ghobject_t, > std::allocator<ghobject_t> >*, ghobject_t*)+0x25a) [0x5623f093e6ea] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 23: > (PGBackend::objects_list_range(hobject_t const&, hobject_t const&, > snapid_t, std::vector<hobject_t, std::allocator<hobject_t> >*, > std::vector<ghobject_t, std::allocator<ghobject_t> >*)+0x192) > [0x5623f0700ef2] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 24: > (PG::build_scrub_map_chunk(ScrubMap&, hobject_t, hobject_t, bool, unsigned > int, ThreadPool::TPHandle&)+0x200) [0x5623f05a8b30] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 25: > (PG::chunky_scrub(ThreadPool::TPHandle&)+0x3ea) [0x5623f05d61ca] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 26: (PG::scrub(unsigned int, > ThreadPool::TPHandle&)+0x45c) [0x5623f05d7cec] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 27: > (OSD::ShardedOpWQ::_process(unsigned int, > ceph::heartbeat_handle_d*)+0x12d0) [0x5623f05179e0] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 28: > (ShardedThreadPool::shardedthreadpool_worker(unsigned int)+0x884) > [0x5623f0ae44e4] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 29: > (ShardedThreadPool::WorkThreadSharded::entry()+0x10) [0x5623f0ae7520] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 30: (()+0x76ba) [0x7f526119c6ba] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 31: (clone()+0x6d) [0x7f52602133dd] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 2017-08-02 10:43:32.251841 > 7f524861b700 -1 *** Caught signal (Aborted) ** > Aug 02 10:43:32 osd3 ceph-osd[7413]: in thread 7f524861b700 > thread_name:tp_osd_tp > Aug 02 10:43:32 osd3 ceph-osd[7413]: ceph version 12.1.2 > (b661348f156f148d764b998b65b90451f096cb27) luminous (rc) > Aug 02 10:43:32 osd3 ceph-osd[7413]: 1: (()+0xa9a964) [0x5623f0a9c964] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 2: (()+0x11390) [0x7f52611a6390] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 3: (pread64()+0x33) [0x7f52611a5d43] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 4: > (KernelDevice::direct_read_unaligned(unsigned long, unsigned long, > char*)+0x81) [0x5623f0a7cfc1] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 5: > (KernelDevice::read_random(unsigned long, unsigned long, char*, > bool)+0x4f3) [0x5623f0a7da43] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 6: > (BlueFS::_read_random(BlueFS::FileReader*, unsigned long, unsigned long, > char*)+0x4fa) [0x5623f0a4d9ca] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 7: > (BlueRocksRandomAccessFile::Read(unsigned long, unsigned long, > rocksdb::Slice*, char*) const+0x20) [0x5623f0a77e10] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 8: > (rocksdb::RandomAccessFileReader::Read(unsigned long, unsigned long, > rocksdb::Slice*, char*) const+0xf8f) [0x5623f0e50acf] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 9: > (rocksdb::ReadBlockContents(rocksdb::RandomAccessFileReader*, > rocksdb::Footer const&, rocksdb::ReadOptions const&, rocksdb::BlockHandle > const&, rocksdb::BlockContents*, rocksdb::ImmutableCFOptions const&, bool, > rocksdb::Slice const&, rocksdb::PersistentCacheOptions const&)+0x5f3) > [0x5623f0e21383] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 10: (()+0xe0f7c6) [0x5623f0e117c6] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 11: > (rocksdb::BlockBasedTable::MaybeLoadDataBlockToCache(rocksdb::BlockBasedTable::Rep*, > rocksdb::ReadOptions const&, rocksdb::BlockHandle const&, rocksdb::Slice, > rocksdb::BlockBasedTable::CachableEntry<rocksdb::Block>*, bool)+0x2f8) > [0x5623f0e13928] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 12: > (rocksdb::BlockBasedTable::NewDataBlockIterator(rocksdb::BlockBasedTable::Rep*, > rocksdb::ReadOptions const&, rocksdb::BlockHandle const&, > rocksdb::BlockIter*, bool, rocksdb::Status)+0x2ac) [0x5623f0e13d2c] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 13: > (rocksdb::BlockBasedTable::BlockEntryIteratorState::NewSecondaryIterator(rocksdb::Slice > const&)+0x97) [0x5623f0e1c4a7] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 14: (()+0xe4576e) [0x5623f0e4776e] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 15: (()+0xe45836) [0x5623f0e47836] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 16: (()+0xe459b1) [0x5623f0e479b1] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 17: > (rocksdb::MergingIterator::Next()+0x449) [0x5623f0e2ab09] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 18: > (rocksdb::DBIter::FindNextUserEntryInternal(bool, bool)+0x182) > [0x5623f0ec7ed2] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 19: (rocksdb::DBIter::Next()+0x1eb) > [0x5623f0ec8c8b] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 20: > (RocksDBStore::RocksDBWholeSpaceIteratorImpl::next()+0x9a) [0x5623f09dd58a] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 21: > (BlueStore::_collection_list(BlueStore::Collection*, ghobject_t const&, > ghobject_t const&, int, std::vector<ghobject_t, std::allocator<ghobject_t> > >*, ghobject_t*)+0x1170) [0x5623f093d250] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 22: > (BlueStore::collection_list(boost::intrusive_ptr<ObjectStore::CollectionImpl>&, > ghobject_t const&, ghobject_t const&, int, std::vector<ghobject_t, > std::allocator<ghobject_t> >*, ghobject_t*)+0x25a) [0x5623f093e6ea] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 23: > (PGBackend::objects_list_range(hobject_t const&, hobject_t const&, > snapid_t, std::vector<hobject_t, std::allocator<hobject_t> >*, > std::vector<ghobject_t, std::allocator<ghobject_t> >*)+0x192) > [0x5623f0700ef2] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 24: > (PG::build_scrub_map_chunk(ScrubMap&, hobject_t, hobject_t, bool, unsigned > int, ThreadPool::TPHandle&)+0x200) [0x5623f05a8b30] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 25: > (PG::chunky_scrub(ThreadPool::TPHandle&)+0x3ea) [0x5623f05d61ca] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 26: (PG::scrub(unsigned int, > ThreadPool::TPHandle&)+0x45c) [0x5623f05d7cec] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 27: > (OSD::ShardedOpWQ::_process(unsigned int, > ceph::heartbeat_handle_d*)+0x12d0) [0x5623f05179e0] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 28: > (ShardedThreadPool::shardedthreadpool_worker(unsigned int)+0x884) > [0x5623f0ae44e4] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 29: > (ShardedThreadPool::WorkThreadSharded::entry()+0x10) [0x5623f0ae7520] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 30: (()+0x76ba) [0x7f526119c6ba] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 31: (clone()+0x6d) [0x7f52602133dd] > Aug 02 10:43:32 osd3 ceph-osd[7413]: NOTE: a copy of the executable, or > `objdump -rdS <executable>` is needed to interpret this. > Aug 02 10:43:32 osd3 ceph-osd[7413]: 0> 2017-08-02 10:43:32.251841 > 7f524861b700 -1 *** Caught signal (Aborted) ** > Aug 02 10:43:32 osd3 ceph-osd[7413]: in thread 7f524861b700 > thread_name:tp_osd_tp > Aug 02 10:43:32 osd3 ceph-osd[7413]: ceph version 12.1.2 > (b661348f156f148d764b998b65b90451f096cb27) luminous (rc) > Aug 02 10:43:32 osd3 ceph-osd[7413]: 1: (()+0xa9a964) [0x5623f0a9c964] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 2: (()+0x11390) [0x7f52611a6390] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 3: (pread64()+0x33) [0x7f52611a5d43] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 4: > (KernelDevice::direct_read_unaligned(unsigned long, unsigned long, > char*)+0x81) [0x5623f0a7cfc1] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 5: > (KernelDevice::read_random(unsigned long, unsigned long, char*, > bool)+0x4f3) [0x5623f0a7da43] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 6: > (BlueFS::_read_random(BlueFS::FileReader*, unsigned long, unsigned long, > char*)+0x4fa) [0x5623f0a4d9ca] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 7: > (BlueRocksRandomAccessFile::Read(unsigned long, unsigned long, > rocksdb::Slice*, char*) const+0x20) [0x5623f0a77e10] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 8: > (rocksdb::RandomAccessFileReader::Read(unsigned long, unsigned long, > rocksdb::Slice*, char*) const+0xf8f) [0x5623f0e50acf] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 9: > (rocksdb::ReadBlockContents(rocksdb::RandomAccessFileReader*, > rocksdb::Footer const&, rocksdb::ReadOptions const&, rocksdb::BlockHandle > const&, rocksdb::BlockContents*, rocksdb::ImmutableCFOptions const&, bool, > rocksdb::Slice const&, rocksdb::PersistentCacheOptions const&)+0x5f3) > [0x5623f0e21383] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 10: (()+0xe0f7c6) [0x5623f0e117c6] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 11: > (rocksdb::BlockBasedTable::MaybeLoadDataBlockToCache(rocksdb::BlockBasedTable::Rep*, > rocksdb::ReadOptions const&, rocksdb::BlockHandle const&, rocksdb::Slice, > rocksdb::BlockBasedTable::CachableEntry<rocksdb::Block>*, bool)+0x2f8) > [0x5623f0e13928] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 12: > (rocksdb::BlockBasedTable::NewDataBlockIterator(rocksdb::BlockBasedTable::Rep*, > rocksdb::ReadOptions const&, rocksdb::BlockHandle const&, > rocksdb::BlockIter*, bool, rocksdb::Status)+0x2ac) [0x5623f0e13d2c] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 13: > (rocksdb::BlockBasedTable::BlockEntryIteratorState::NewSecondaryIterator(rocksdb::Slice > const&)+0x97) [0x5623f0e1c4a7] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 14: (()+0xe4576e) [0x5623f0e4776e] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 15: (()+0xe45836) [0x5623f0e47836] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 16: (()+0xe459b1) [0x5623f0e479b1] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 17: > (rocksdb::MergingIterator::Next()+0x449) [0x5623f0e2ab09] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 18: > (rocksdb::DBIter::FindNextUserEntryInternal(bool, bool)+0x182) > [0x5623f0ec7ed2] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 19: (rocksdb::DBIter::Next()+0x1eb) > [0x5623f0ec8c8b] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 20: > (RocksDBStore::RocksDBWholeSpaceIteratorImpl::next()+0x9a) [0x5623f09dd58a] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 21: > (BlueStore::_collection_list(BlueStore::Collection*, ghobject_t const&, > ghobject_t const&, int, std::vector<ghobject_t, std::allocator<ghobject_t> > >*, ghobject_t*)+0x1170) [0x5623f093d250] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 22: > (BlueStore::collection_list(boost::intrusive_ptr<ObjectStore::CollectionImpl>&, > ghobject_t const&, ghobject_t const&, int, std::vector<ghobject_t, > std::allocator<ghobject_t> >*, ghobject_t*)+0x25a) [0x5623f093e6ea] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 23: > (PGBackend::objects_list_range(hobject_t const&, hobject_t const&, > snapid_t, std::vector<hobject_t, std::allocator<hobject_t> >*, > std::vector<ghobject_t, std::allocator<ghobject_t> >*)+0x192) > [0x5623f0700ef2] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 24: > (PG::build_scrub_map_chunk(ScrubMap&, hobject_t, hobject_t, bool, unsigned > int, ThreadPool::TPHandle&)+0x200) [0x5623f05a8b30] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 25: > (PG::chunky_scrub(ThreadPool::TPHandle&)+0x3ea) [0x5623f05d61ca] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 26: (PG::scrub(unsigned int, > ThreadPool::TPHandle&)+0x45c) [0x5623f05d7cec] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 27: > (OSD::ShardedOpWQ::_process(unsigned int, > ceph::heartbeat_handle_d*)+0x12d0) [0x5623f05179e0] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 28: > (ShardedThreadPool::shardedthreadpool_worker(unsigned int)+0x884) > [0x5623f0ae44e4] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 29: > (ShardedThreadPool::WorkThreadSharded::entry()+0x10) [0x5623f0ae7520] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 30: (()+0x76ba) [0x7f526119c6ba] > Aug 02 10:43:32 osd3 ceph-osd[7413]: 31: (clone()+0x6d) [0x7f52602133dd] > Aug 02 10:43:32 osd3 ceph-osd[7413]: NOTE: a copy of the executable, or > `objdump -rdS <executable>` is needed to interpret this. > Aug 02 10:43:32 osd3 systemd[1]: ceph-osd@0.service: Main process exited, > code=killed, status=6/ABRT > Aug 02 10:43:32 osd3 systemd[1]: ceph-osd@0.service: Unit entered failed > state. > Aug 02 10:43:32 osd3 systemd[1]: ceph-osd@0.service: Failed with result > 'signal'. > Aug 02 10:43:52 osd3 systemd[1]: ceph-osd@0.service: Service hold-off > time over, scheduling restart. > Aug 02 10:43:52 osd3 systemd[1]: Stopped Ceph object storage daemon osd.0. > Aug 02 10:43:52 osd3 systemd[1]: Starting Ceph object storage daemon > osd.0... > Aug 02 10:43:52 osd3 systemd[1]: Started Ceph object storage daemon osd.0. > Aug 02 10:43:52 osd3 ceph-osd[8322]: starting osd.0 at - osd_data > /var/lib/ceph/osd/ceph-0 /var/lib/ceph/osd/ceph-0/journal > > roger@desktop:~$ ceph -s > cluster: > id: eea7b78c-b138-40fc-9f3e-3d77afb770f0 > health: HEALTH_WARN > 1 osds down > 1 host (1 osds) down > Degraded data redundancy: 43922/162834 objects degraded > (26.973%), 300 pgs unclean, 305 pgs degraded > 114 pgs not deep-scrubbed for 86400 > 155 pgs not scrubbed for 86400 > 10 slow requests are blocked > 32 sec > > services: > mon: 3 daemons, quorum nuc1,nuc2,nuc3 > mgr: nuc3(active), standbys: nuc2, nuc1 > osd: 3 osds: 2 up, 3 in > rgw: 1 daemon active > > data: > pools: 19 pools, 372 pgs > objects: 54278 objects, 71724 MB > usage: 122 GB used, 27819 GB / 27941 GB avail > pgs: 43922/162834 objects degraded (26.973%) > 303 active+undersized+degraded > 67 active+clean > 2 active+recovery_wait+degraded > > roger@desktop:~$ ceph osd tree > ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF > -1 27.28679 root default > -5 9.09560 host osd1 > 3 hdd 9.09560 osd.3 up 1.00000 1.00000 > -6 9.09560 host osd2 > 4 hdd 9.09560 osd.4 up 1.00000 1.00000 > -2 9.09560 host osd3 > 0 hdd 9.09560 osd.0 down 1.00000 1.00000 > > roger@desktop:~$ ceph mon versions > { > "ceph version 12.1.2 (b661348f156f148d764b998b65b90451f096cb27) > luminous (rc)": 3 > } > roger@desktop:~$ ceph osd versions > { > "ceph version 12.1.2 (b661348f156f148d764b998b65b90451f096cb27) > luminous (rc)": 2 > } > > roger@osd3:~$ sudo ceph daemon osd.0 status > { > "cluster_fsid": "eea7b78c-b138-40fc-9f3e-3d77afb770f0", > "osd_fsid": "bdb31a03-e381-4bf8-82e3-18916c838308", > "whoami": 0, > "state": "waiting_for_healthy", > "oldest_map": 25389, > "newest_map": 25938, > "num_pgs": 372 > } > > roger@desktop:~$ ceph df > GLOBAL: > SIZE AVAIL RAW USED %RAW USED > 27941G 27819G 122G 0.44 > POOLS: > NAME ID USED %USED MAX AVAIL > OBJECTS > default.rgw.rgw.gc 70 0 0 8807G > 0 > default.rgw.buckets.non-ec 83 0 0 8807G > 43 > default.rgw.control 85 0 0 8807G > 8 > default.rgw.data.root 86 15601 0 8807G > 49 > default.rgw.gc 87 0 0 8807G > 32 > default.rgw.lc 88 0 0 8807G > 32 > default.rgw.log 89 0 0 8807G > 144 > default.rgw.users.uid 90 3346 0 8807G > 14 > default.rgw.users.email 91 100 0 8807G > 7 > default.rgw.users.keys 92 100 0 8807G > 7 > default.rgw.buckets.index 93 0 0 8807G > 39 > default.rgw.intent-log 95 0 0 8807G > 0 > default.rgw.meta 96 0 0 8807G > 0 > default.rgw.usage 97 0 0 8807G > 0 > default.rgw.users.swift 98 39 0 8807G > 4 > default.rgw.buckets.extra 99 0 0 8807G > 0 > .rgw.root 100 1681 0 8807G > 4 > default.rgw.reshard 101 0 0 8807G > 17 > default.rgw.buckets.data 103 71724M 0.40 17614G > 53878 > >
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com