I updated firmware and kernel, running torture tests. So far no assert, but I still noticed this on the same osd as yesterday
Oct 01 19:35:13 storage2n2-la ceph-osd-34[11188]: 2019-10-01 19:35:13.721 7f8d03150700 1 heartbeat_map is_healthy 'OSD::osd_op_tp thread 0x7f8cd05d7700' had timed out after 60 Oct 01 19:35:13 storage2n2-la ceph-osd-34[11188]: 2019-10-01 19:35:13.721 7f8d03150700 1 heartbeat_map is_healthy 'OSD::osd_op_tp thread 0x7f8cd0dd8700' had timed out after 60 Oct 01 19:35:13 storage2n2-la ceph-osd-34[11188]: 2019-10-01 19:35:13.721 7f8d03150700 1 heartbeat_map is_healthy 'OSD::osd_op_tp thread 0x7f8cd2ddc700' had timed out after 60 Oct 01 19:35:13 storage2n2-la ceph-osd-34[11188]: 2019-10-01 19:35:13.721 7f8d03150700 1 heartbeat_map is_healthy 'OSD::osd_op_tp thread 0x7f8cd35dd700' had timed out after 60 Oct 01 19:35:13 storage2n2-la ceph-osd-34[11188]: 2019-10-01 19:35:13.721 7f8d03150700 1 heartbeat_map is_healthy 'OSD::osd_op_tp thread 0x7f8cd3dde700' had timed out after 60 The spike of latency on this OSD is 6 seconds at that time. Any ideas? On Tue, Oct 1, 2019 at 8:03 AM Sasha Litvak <alexander.v.lit...@gmail.com> wrote: > It was hardware indeed. Dell server reported a disk being reset with > power on. Checking the usual suspects i.e. controller firmware, controller > event log (if I can get one), drive firmware. > I will report more when I get a better idea > > Thank you! > > On Tue, Oct 1, 2019 at 2:33 AM Brad Hubbard <bhubb...@redhat.com> wrote: > >> Removed ceph-de...@vger.kernel.org and added d...@ceph.io >> >> On Tue, Oct 1, 2019 at 4:26 PM Alex Litvak <alexander.v.lit...@gmail.com> >> wrote: >> > >> > Hellow everyone, >> > >> > Can you shed the line on the cause of the crash? Could actually client >> request trigger it? >> > >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 2019-09-30 >> 22:52:58.867 7f093d71e700 -1 bdev(0x55b72c156000 >> /var/lib/ceph/osd/ceph-17/block) aio_submit retries 16 >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 2019-09-30 >> 22:52:58.867 7f093d71e700 -1 bdev(0x55b72c156000 >> /var/lib/ceph/osd/ceph-17/block) aio submit got (11) Resource temporarily >> unavailable >> >> The KernelDevice::aio_submit function has tried to submit Io 16 times >> (a hard coded limit) and received an error each time causing it to >> assert. Can you check the status of the underlying device(s)? >> >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: >> > >> /home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos7/DIST/centos7/MACHINE_SIZE/huge/release/14.2.2/rpm/el7/BUILD/ceph-14.2.2/src/os/bluestore/KernelDevice.cc: >> > In fun >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: >> > >> /home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos7/DIST/centos7/MACHINE_SIZE/huge/release/14.2.2/rpm/el7/BUILD/ceph-14.2.2/src/os/bluestore/KernelDevice.cc: >> > 757: F >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: ceph version 14.2.2 >> (4f8fa0a0024755aae7d95567c63f11d6862d55be) nautilus (stable) >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 1: >> (ceph::__ceph_assert_fail(char const*, char const*, int, char >> const*)+0x14a) [0x55b71f668cf4] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 2: >> (ceph::__ceph_assertf_fail(char const*, char const*, int, char const*, char >> const*, ...)+0) [0x55b71f668ec2] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 3: >> (KernelDevice::aio_submit(IOContext*)+0x701) [0x55b71fd61ca1] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 4: >> (BlueStore::_txc_aio_submit(BlueStore::TransContext*)+0x42) [0x55b71fc29892] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 5: >> (BlueStore::_txc_state_proc(BlueStore::TransContext*)+0x42b) >> [0x55b71fc496ab] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 6: >> (BlueStore::queue_transactions(boost::intrusive_ptr<ObjectStore::CollectionImpl>&, >> std::vector<ObjectStore::Transaction, >> > std::allocator<ObjectStore::Transaction> >&, >> boost::intrusive_ptr<TrackedOp>, ThreadPool::T >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 7: (non-virtual >> thunk to >> PrimaryLogPG::queue_transactions(std::vector<ObjectStore::Transaction, >> std::allocator<ObjectStore::Transaction> >&, >> > boost::intrusive_ptr<OpRequest>)+0x54) [0x55b71f9b1b84] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 8: >> (ReplicatedBackend::submit_transaction(hobject_t const&, object_stat_sum_t >> const&, eversion_t const&, std::unique_ptr<PGTransaction, >> > std::default_delete<PGTransaction> >&&, eversion_t const&, eversion_t >> const&, s >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 9: >> (PrimaryLogPG::issue_repop(PrimaryLogPG::RepGather*, >> PrimaryLogPG::OpContext*)+0xf12) [0x55b71f90e322] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 10: >> (PrimaryLogPG::execute_ctx(PrimaryLogPG::OpContext*)+0xfae) [0x55b71f969b7e] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 11: >> (PrimaryLogPG::do_op(boost::intrusive_ptr<OpRequest>&)+0x3965) >> [0x55b71f96de15] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 12: >> (PrimaryLogPG::do_request(boost::intrusive_ptr<OpRequest>&, >> ThreadPool::TPHandle&)+0xbd4) [0x55b71f96f8a4] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 13: >> (OSD::dequeue_op(boost::intrusive_ptr<PG>, boost::intrusive_ptr<OpRequest>, >> ThreadPool::TPHandle&)+0x1a9) [0x55b71f7a9ea9] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 14: >> (PGOpItem::run(OSD*, OSDShard*, boost::intrusive_ptr<PG>&, >> ThreadPool::TPHandle&)+0x62) [0x55b71fa475d2] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 15: >> (OSD::ShardedOpWQ::_process(unsigned int, ceph::heartbeat_handle_d*)+0x9f4) >> [0x55b71f7c6ef4] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 16: >> (ShardedThreadPool::shardedthreadpool_worker(unsigned int)+0x433) >> [0x55b71fdc5ce3] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 17: >> (ShardedThreadPool::WorkThreadSharded::entry()+0x10) [0x55b71fdc8d80] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 18: (()+0x7dd5) >> [0x7f0971da9dd5] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 19: (clone()+0x6d) >> [0x7f0970c7002d] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 2019-09-30 >> 22:52:58.879 7f093d71e700 -1 >> > >> /home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos7/DIST/centos7/MACHINE_SIZE/huge/release/14.2.2/rpm/el7/BUILD/ceph-14.2.2/ >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: >> > >> /home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos7/DIST/centos7/MACHINE_SIZE/huge/release/14.2.2/rpm/el7/BUILD/ceph-14.2.2/src/os/bluestore/KernelDevice.cc: >> > 757: F >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: ceph version 14.2.2 >> (4f8fa0a0024755aae7d95567c63f11d6862d55be) nautilus (stable) >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 1: >> (ceph::__ceph_assert_fail(char const*, char const*, int, char >> const*)+0x14a) [0x55b71f668cf4] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 2: >> (ceph::__ceph_assertf_fail(char const*, char const*, int, char const*, char >> const*, ...)+0) [0x55b71f668ec2] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 3: >> (KernelDevice::aio_submit(IOContext*)+0x701) [0x55b71fd61ca1] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 4: >> (BlueStore::_txc_aio_submit(BlueStore::TransContext*)+0x42) [0x55b71fc29892] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 5: >> (BlueStore::_txc_state_proc(BlueStore::TransContext*)+0x42b) >> [0x55b71fc496ab] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 6: >> (BlueStore::queue_transactions(boost::intrusive_ptr<ObjectStore::CollectionImpl>&, >> std::vector<ObjectStore::Transaction, >> > std::allocator<ObjectStore::Transaction> >&, >> boost::intrusive_ptr<TrackedOp>, ThreadPool::T >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 7: (non-virtual >> thunk to >> PrimaryLogPG::queue_transactions(std::vector<ObjectStore::Transaction, >> std::allocator<ObjectStore::Transaction> >&, >> > boost::intrusive_ptr<OpRequest>)+0x54) [0x55b71f9b1b84] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 8: >> (ReplicatedBackend::submit_transaction(hobject_t const&, object_stat_sum_t >> const&, eversion_t const&, std::unique_ptr<PGTransaction, >> > std::default_delete<PGTransaction> >&&, eversion_t const&, eversion_t >> const&, s >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 9: >> (PrimaryLogPG::issue_repop(PrimaryLogPG::RepGather*, >> PrimaryLogPG::OpContext*)+0xf12) [0x55b71f90e322] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 10: >> (PrimaryLogPG::execute_ctx(PrimaryLogPG::OpContext*)+0xfae) [0x55b71f969b7e] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 11: >> (PrimaryLogPG::do_op(boost::intrusive_ptr<OpRequest>&)+0x3965) >> [0x55b71f96de15] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 12: >> (PrimaryLogPG::do_request(boost::intrusive_ptr<OpRequest>&, >> ThreadPool::TPHandle&)+0xbd4) [0x55b71f96f8a4] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 13: >> (OSD::dequeue_op(boost::intrusive_ptr<PG>, boost::intrusive_ptr<OpRequest>, >> ThreadPool::TPHandle&)+0x1a9) [0x55b71f7a9ea9] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 14: >> (PGOpItem::run(OSD*, OSDShard*, boost::intrusive_ptr<PG>&, >> ThreadPool::TPHandle&)+0x62) [0x55b71fa475d2] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 15: >> (OSD::ShardedOpWQ::_process(unsigned int, ceph::heartbeat_handle_d*)+0x9f4) >> [0x55b71f7c6ef4] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 16: >> (ShardedThreadPool::shardedthreadpool_worker(unsigned int)+0x433) >> [0x55b71fdc5ce3] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 17: >> (ShardedThreadPool::WorkThreadSharded::entry()+0x10) [0x55b71fdc8d80] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 18: (()+0x7dd5) >> [0x7f0971da9dd5] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 19: (clone()+0x6d) >> [0x7f0970c7002d] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: *** Caught signal >> (Aborted) ** >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: in thread >> 7f093d71e700 thread_name:tp_osd_tp >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: ceph version 14.2.2 >> (4f8fa0a0024755aae7d95567c63f11d6862d55be) nautilus (stable) >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 1: (()+0xf5d0) >> [0x7f0971db15d0] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 2: (gsignal()+0x37) >> [0x7f0970ba82c7] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 3: (abort()+0x148) >> [0x7f0970ba99b8] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 4: >> (ceph::__ceph_assert_fail(char const*, char const*, int, char >> const*)+0x199) [0x55b71f668d43] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 5: >> (ceph::__ceph_assertf_fail(char const*, char const*, int, char const*, char >> const*, ...)+0) [0x55b71f668ec2] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 6: >> (KernelDevice::aio_submit(IOContext*)+0x701) [0x55b71fd61ca1] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 7: >> (BlueStore::_txc_aio_submit(BlueStore::TransContext*)+0x42) [0x55b71fc29892] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 8: >> (BlueStore::_txc_state_proc(BlueStore::TransContext*)+0x42b) >> [0x55b71fc496ab] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 9: >> (BlueStore::queue_transactions(boost::intrusive_ptr<ObjectStore::CollectionImpl>&, >> std::vector<ObjectStore::Transaction, >> > std::allocator<ObjectStore::Transaction> >&, >> boost::intrusive_ptr<TrackedOp>, ThreadPool::T >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 10: (non-virtual >> thunk to >> PrimaryLogPG::queue_transactions(std::vector<ObjectStore::Transaction, >> std::allocator<ObjectStore::Transaction> >&, >> > boost::intrusive_ptr<OpRequest>)+0x54) [0x55b71f9b1b84] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 11: >> (ReplicatedBackend::submit_transaction(hobject_t const&, object_stat_sum_t >> const&, eversion_t const&, std::unique_ptr<PGTransaction, >> > std::default_delete<PGTransaction> >&&, eversion_t const&, eversion_t >> const&, >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 12: >> (PrimaryLogPG::issue_repop(PrimaryLogPG::RepGather*, >> PrimaryLogPG::OpContext*)+0xf12) [0x55b71f90e322] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 13: >> (PrimaryLogPG::execute_ctx(PrimaryLogPG::OpContext*)+0xfae) [0x55b71f969b7e] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 14: >> (PrimaryLogPG::do_op(boost::intrusive_ptr<OpRequest>&)+0x3965) >> [0x55b71f96de15] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 15: >> (PrimaryLogPG::do_request(boost::intrusive_ptr<OpRequest>&, >> ThreadPool::TPHandle&)+0xbd4) [0x55b71f96f8a4] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 16: >> (OSD::dequeue_op(boost::intrusive_ptr<PG>, boost::intrusive_ptr<OpRequest>, >> ThreadPool::TPHandle&)+0x1a9) [0x55b71f7a9ea9] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 17: >> (PGOpItem::run(OSD*, OSDShard*, boost::intrusive_ptr<PG>&, >> ThreadPool::TPHandle&)+0x62) [0x55b71fa475d2] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 18: >> (OSD::ShardedOpWQ::_process(unsigned int, ceph::heartbeat_handle_d*)+0x9f4) >> [0x55b71f7c6ef4] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 19: >> (ShardedThreadPool::shardedthreadpool_worker(unsigned int)+0x433) >> [0x55b71fdc5ce3] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 20: >> (ShardedThreadPool::WorkThreadSharded::entry()+0x10) [0x55b71fdc8d80] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 21: (()+0x7dd5) >> [0x7f0971da9dd5] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 22: (clone()+0x6d) >> [0x7f0970c7002d] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 2019-09-30 >> 22:52:58.883 7f093d71e700 -1 *** Caught signal (Aborted) ** >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: in thread >> 7f093d71e700 thread_name:tp_osd_tp >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: ceph version 14.2.2 >> (4f8fa0a0024755aae7d95567c63f11d6862d55be) nautilus (stable) >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 1: (()+0xf5d0) >> [0x7f0971db15d0] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 2: (gsignal()+0x37) >> [0x7f0970ba82c7] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 3: (abort()+0x148) >> [0x7f0970ba99b8] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 4: >> (ceph::__ceph_assert_fail(char const*, char const*, int, char >> const*)+0x199) [0x55b71f668d43] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 5: >> (ceph::__ceph_assertf_fail(char const*, char const*, int, char const*, char >> const*, ...)+0) [0x55b71f668ec2] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 6: >> (KernelDevice::aio_submit(IOContext*)+0x701) [0x55b71fd61ca1] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 7: >> (BlueStore::_txc_aio_submit(BlueStore::TransContext*)+0x42) [0x55b71fc29892] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 8: >> (BlueStore::_txc_state_proc(BlueStore::TransContext*)+0x42b) >> [0x55b71fc496ab] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 9: >> (BlueStore::queue_transactions(boost::intrusive_ptr<ObjectStore::CollectionImpl>&, >> std::vector<ObjectStore::Transaction, >> > std::allocator<ObjectStore::Transaction> >&, >> boost::intrusive_ptr<TrackedOp>, ThreadPool::T >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 10: (non-virtual >> thunk to >> PrimaryLogPG::queue_transactions(std::vector<ObjectStore::Transaction, >> std::allocator<ObjectStore::Transaction> >&, >> > boost::intrusive_ptr<OpRequest>)+0x54) [0x55b71f9b1b84] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 11: >> (ReplicatedBackend::submit_transaction(hobject_t const&, object_stat_sum_t >> const&, eversion_t const&, std::unique_ptr<PGTransaction, >> > std::default_delete<PGTransaction> >&&, eversion_t const&, eversion_t >> const&, >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 12: >> (PrimaryLogPG::issue_repop(PrimaryLogPG::RepGather*, >> PrimaryLogPG::OpContext*)+0xf12) [0x55b71f90e322] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 13: >> (PrimaryLogPG::execute_ctx(PrimaryLogPG::OpContext*)+0xfae) [0x55b71f969b7e] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 14: >> (PrimaryLogPG::do_op(boost::intrusive_ptr<OpRequest>&)+0x3965) >> [0x55b71f96de15] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 15: >> (PrimaryLogPG::do_request(boost::intrusive_ptr<OpRequest>&, >> ThreadPool::TPHandle&)+0xbd4) [0x55b71f96f8a4] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 16: >> (OSD::dequeue_op(boost::intrusive_ptr<PG>, boost::intrusive_ptr<OpRequest>, >> ThreadPool::TPHandle&)+0x1a9) [0x55b71f7a9ea9] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 17: >> (PGOpItem::run(OSD*, OSDShard*, boost::intrusive_ptr<PG>&, >> ThreadPool::TPHandle&)+0x62) [0x55b71fa475d2] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 18: >> (OSD::ShardedOpWQ::_process(unsigned int, ceph::heartbeat_handle_d*)+0x9f4) >> [0x55b71f7c6ef4] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 19: >> (ShardedThreadPool::shardedthreadpool_worker(unsigned int)+0x433) >> [0x55b71fdc5ce3] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 20: >> (ShardedThreadPool::WorkThreadSharded::entry()+0x10) [0x55b71fdc8d80] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 21: (()+0x7dd5) >> [0x7f0971da9dd5] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 22: (clone()+0x6d) >> [0x7f0970c7002d] >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: NOTE: a copy of the >> executable, or `objdump -rdS <executable>` is needed to interpret this. >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: --- begin dump of >> recent events --- >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: -9999> 2019-09-30 >> 20:46:02.076 7f0937f13700 5 osd.17 6328 heartbeat >> osd_stat(store_statfs(0x1a485594000/0x40000000/0x1bf00000000, data >> > 0x19cfcadc2a/0x1a3aa68000, compress 0x0/0x0/0x0, omap 0xcdefc78, meta >> 0x3321038 >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: -9998> 2019-09-30 >> 20:46:03.776 7f0937f13700 5 osd.17 6328 heartbeat >> osd_stat(store_statfs(0x1a485594000/0x40000000/0x1bf00000000, data >> > 0x19cfcadc2a/0x1a3aa68000, compress 0x0/0x0/0x0, omap 0xcdefc78, meta >> 0x3321038 >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: -9997> 2019-09-30 >> 20:46:04.277 7f0937f13700 5 osd.17 6328 heartbeat >> osd_stat(store_statfs(0x1a485594000/0x40000000/0x1bf00000000, data >> > 0x19cfcadc2a/0x1a3aa68000, compress 0x0/0x0/0x0, omap 0xcdefc78, meta >> 0x3321038 >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: -9996> 2019-09-30 >> 20:46:04.777 7f0937f13700 5 osd.17 6328 heartbeat >> osd_stat(store_statfs(0x1a485594000/0x40000000/0x1bf00000000, data >> > 0x19cfcadc2a/0x1a3aa68000, compress 0x0/0x0/0x0, omap 0xcdefc78, meta >> 0x3321038 >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: -9995> 2019-09-30 >> 20:46:04.905 7f0950744700 5 bluestore.MempoolThread(0x55b72c210a88) >> _tune_cache_size target: 8485076992 heap: 398680064 unmapped: >> > 9035776 mapped: 389644288 old cache_size: 5064831794 new cache siz >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: -9994> 2019-09-30 >> 20:46:04.905 7f0950744700 5 bluestore.MempoolThread(0x55b72c210a88) >> _trim_shards cache_size: 5064831794 kv_alloc: 1979711488 >> > kv_used: 120635251 meta_alloc: 1979711488 meta_used: 117486636 data_all >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: -9993> 2019-09-30 >> 20:46:05.813 7f094ccd2700 10 monclient: tick >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: -9992> 2019-09-30 >> 20:46:05.813 7f094ccd2700 10 monclient: _check_auth_rotating have uptodate >> secrets (they expire after 2019-09-30 20:45:35.819641) >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: -9991> 2019-09-30 >> 20:46:09.913 7f0950744700 5 bluestore.MempoolThread(0x55b72c210a88) >> _tune_cache_size target: 8485076992 heap: 398680064 unmapped: >> > 9035776 mapped: 389644288 old cache_size: 5064831794 new cache siz >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: -9990> 2019-09-30 >> 20:46:09.913 7f0950744700 5 bluestore.MempoolThread(0x55b72c210a88) >> _trim_shards cache_size: 5064831794 kv_alloc: 1979711488 >> > kv_used: 120635251 meta_alloc: 1979711488 meta_used: 117486836 data_all >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: -9989> 2019-09-30 >> 20:46:10.681 7f0937f13700 5 osd.17 6328 heartbeat >> osd_stat(store_statfs(0x1a485594000/0x40000000/0x1bf00000000, data >> > 0x19cfcadc2a/0x1a3aa68000, compress 0x0/0x0/0x0, omap 0xcdefc78, meta >> 0x3321038 >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: -9988> 2019-09-30 >> 20:46:14.569 7f094dcd4700 2 osd.17 6328 ms_handle_reset con 0x55b72c797000 >> session 0x55b7401f1800 >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: -9987> 2019-09-30 >> 20:46:14.569 7f096da92700 10 monclient: handle_auth_request added challenge >> on 0x55b733bf0400 >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: -9986> 2019-09-30 >> 20:46:14.917 7f0950744700 5 bluestore.MempoolThread(0x55b72c210a88) >> _tune_cache_size target: 8485076992 heap: 398680064 unmapped: >> > 9035776 mapped: 389644288 old cache_size: 5064831794 new cache siz >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: -9985> 2019-09-30 >> 20:46:14.917 7f0950744700 5 bluestore.MempoolThread(0x55b72c210a88) >> _trim_shards cache_size: 5064831794 kv_alloc: 1979711488 >> > kv_used: 120635251 meta_alloc: 1979711488 meta_used: 117487036 data_all >> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: -9984> 2019-09-30 >> 20:46:15.381 7f0937f13700 5 osd.17 6328 heartbeat >> osd_stat(store_statfs(0x1a485594000/0x40000000/0x1bf00000000, data >> > 0x19cfcadc2a/0x1a3aa68000, compress 0x0/0x0/0x0, omap 0xcdefc78, meta >> 0x3321038 >> > >> > _______________________________________________ >> > ceph-users mailing list >> > ceph-users@lists.ceph.com >> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >> >> >> -- >> Cheers, >> Brad >> >>
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com