I updated firmware and kernel, running torture tests.  So far no assert,
but I still noticed this on the same osd as yesterday

Oct 01 19:35:13 storage2n2-la ceph-osd-34[11188]: 2019-10-01 19:35:13.721
7f8d03150700  1 heartbeat_map is_healthy 'OSD::osd_op_tp thread
0x7f8cd05d7700' had timed out after 60
Oct 01 19:35:13 storage2n2-la ceph-osd-34[11188]: 2019-10-01 19:35:13.721
7f8d03150700  1 heartbeat_map is_healthy 'OSD::osd_op_tp thread
0x7f8cd0dd8700' had timed out after 60
Oct 01 19:35:13 storage2n2-la ceph-osd-34[11188]: 2019-10-01 19:35:13.721
7f8d03150700  1 heartbeat_map is_healthy 'OSD::osd_op_tp thread
0x7f8cd2ddc700' had timed out after 60
Oct 01 19:35:13 storage2n2-la ceph-osd-34[11188]: 2019-10-01 19:35:13.721
7f8d03150700  1 heartbeat_map is_healthy 'OSD::osd_op_tp thread
0x7f8cd35dd700' had timed out after 60
Oct 01 19:35:13 storage2n2-la ceph-osd-34[11188]: 2019-10-01 19:35:13.721
7f8d03150700  1 heartbeat_map is_healthy 'OSD::osd_op_tp thread
0x7f8cd3dde700' had timed out after 60

The spike of latency on this OSD is 6 seconds at that time.  Any ideas?

On Tue, Oct 1, 2019 at 8:03 AM Sasha Litvak <alexander.v.lit...@gmail.com>
wrote:

> It was hardware indeed.  Dell server reported a disk being reset with
> power on.  Checking the usual suspects i.e. controller firmware, controller
> event log (if I can get one), drive firmware.
> I will report more when I get a better idea
>
> Thank you!
>
> On Tue, Oct 1, 2019 at 2:33 AM Brad Hubbard <bhubb...@redhat.com> wrote:
>
>> Removed ceph-de...@vger.kernel.org and added d...@ceph.io
>>
>> On Tue, Oct 1, 2019 at 4:26 PM Alex Litvak <alexander.v.lit...@gmail.com>
>> wrote:
>> >
>> > Hellow everyone,
>> >
>> > Can you shed the line on the cause of the crash?  Could actually client
>> request trigger it?
>> >
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 2019-09-30
>> 22:52:58.867 7f093d71e700 -1 bdev(0x55b72c156000
>> /var/lib/ceph/osd/ceph-17/block) aio_submit retries 16
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 2019-09-30
>> 22:52:58.867 7f093d71e700 -1 bdev(0x55b72c156000
>> /var/lib/ceph/osd/ceph-17/block)  aio submit got (11) Resource temporarily
>> unavailable
>>
>> The KernelDevice::aio_submit function has tried to submit Io 16 times
>> (a hard coded limit) and received an error each time causing it to
>> assert. Can you check the status of the underlying device(s)?
>>
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:
>> >
>> /home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos7/DIST/centos7/MACHINE_SIZE/huge/release/14.2.2/rpm/el7/BUILD/ceph-14.2.2/src/os/bluestore/KernelDevice.cc:
>> > In fun
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:
>> >
>> /home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos7/DIST/centos7/MACHINE_SIZE/huge/release/14.2.2/rpm/el7/BUILD/ceph-14.2.2/src/os/bluestore/KernelDevice.cc:
>> > 757: F
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  ceph version 14.2.2
>> (4f8fa0a0024755aae7d95567c63f11d6862d55be) nautilus (stable)
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  1:
>> (ceph::__ceph_assert_fail(char const*, char const*, int, char
>> const*)+0x14a) [0x55b71f668cf4]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  2:
>> (ceph::__ceph_assertf_fail(char const*, char const*, int, char const*, char
>> const*, ...)+0) [0x55b71f668ec2]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  3:
>> (KernelDevice::aio_submit(IOContext*)+0x701) [0x55b71fd61ca1]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  4:
>> (BlueStore::_txc_aio_submit(BlueStore::TransContext*)+0x42) [0x55b71fc29892]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  5:
>> (BlueStore::_txc_state_proc(BlueStore::TransContext*)+0x42b)
>> [0x55b71fc496ab]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  6:
>> (BlueStore::queue_transactions(boost::intrusive_ptr<ObjectStore::CollectionImpl>&,
>> std::vector<ObjectStore::Transaction,
>> > std::allocator<ObjectStore::Transaction> >&,
>> boost::intrusive_ptr<TrackedOp>, ThreadPool::T
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  7: (non-virtual
>> thunk to
>> PrimaryLogPG::queue_transactions(std::vector<ObjectStore::Transaction,
>> std::allocator<ObjectStore::Transaction> >&,
>> > boost::intrusive_ptr<OpRequest>)+0x54) [0x55b71f9b1b84]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  8:
>> (ReplicatedBackend::submit_transaction(hobject_t const&, object_stat_sum_t
>> const&, eversion_t const&, std::unique_ptr<PGTransaction,
>> > std::default_delete<PGTransaction> >&&, eversion_t const&, eversion_t
>> const&, s
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  9:
>> (PrimaryLogPG::issue_repop(PrimaryLogPG::RepGather*,
>> PrimaryLogPG::OpContext*)+0xf12) [0x55b71f90e322]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  10:
>> (PrimaryLogPG::execute_ctx(PrimaryLogPG::OpContext*)+0xfae) [0x55b71f969b7e]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  11:
>> (PrimaryLogPG::do_op(boost::intrusive_ptr<OpRequest>&)+0x3965)
>> [0x55b71f96de15]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  12:
>> (PrimaryLogPG::do_request(boost::intrusive_ptr<OpRequest>&,
>> ThreadPool::TPHandle&)+0xbd4) [0x55b71f96f8a4]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  13:
>> (OSD::dequeue_op(boost::intrusive_ptr<PG>, boost::intrusive_ptr<OpRequest>,
>> ThreadPool::TPHandle&)+0x1a9) [0x55b71f7a9ea9]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  14:
>> (PGOpItem::run(OSD*, OSDShard*, boost::intrusive_ptr<PG>&,
>> ThreadPool::TPHandle&)+0x62) [0x55b71fa475d2]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  15:
>> (OSD::ShardedOpWQ::_process(unsigned int, ceph::heartbeat_handle_d*)+0x9f4)
>> [0x55b71f7c6ef4]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  16:
>> (ShardedThreadPool::shardedthreadpool_worker(unsigned int)+0x433)
>> [0x55b71fdc5ce3]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  17:
>> (ShardedThreadPool::WorkThreadSharded::entry()+0x10) [0x55b71fdc8d80]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  18: (()+0x7dd5)
>> [0x7f0971da9dd5]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  19: (clone()+0x6d)
>> [0x7f0970c7002d]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 2019-09-30
>> 22:52:58.879 7f093d71e700 -1
>> >
>> /home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos7/DIST/centos7/MACHINE_SIZE/huge/release/14.2.2/rpm/el7/BUILD/ceph-14.2.2/
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:
>> >
>> /home/jenkins-build/build/workspace/ceph-build/ARCH/x86_64/AVAILABLE_ARCH/x86_64/AVAILABLE_DIST/centos7/DIST/centos7/MACHINE_SIZE/huge/release/14.2.2/rpm/el7/BUILD/ceph-14.2.2/src/os/bluestore/KernelDevice.cc:
>> > 757: F
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  ceph version 14.2.2
>> (4f8fa0a0024755aae7d95567c63f11d6862d55be) nautilus (stable)
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  1:
>> (ceph::__ceph_assert_fail(char const*, char const*, int, char
>> const*)+0x14a) [0x55b71f668cf4]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  2:
>> (ceph::__ceph_assertf_fail(char const*, char const*, int, char const*, char
>> const*, ...)+0) [0x55b71f668ec2]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  3:
>> (KernelDevice::aio_submit(IOContext*)+0x701) [0x55b71fd61ca1]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  4:
>> (BlueStore::_txc_aio_submit(BlueStore::TransContext*)+0x42) [0x55b71fc29892]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  5:
>> (BlueStore::_txc_state_proc(BlueStore::TransContext*)+0x42b)
>> [0x55b71fc496ab]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  6:
>> (BlueStore::queue_transactions(boost::intrusive_ptr<ObjectStore::CollectionImpl>&,
>> std::vector<ObjectStore::Transaction,
>> > std::allocator<ObjectStore::Transaction> >&,
>> boost::intrusive_ptr<TrackedOp>, ThreadPool::T
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  7: (non-virtual
>> thunk to
>> PrimaryLogPG::queue_transactions(std::vector<ObjectStore::Transaction,
>> std::allocator<ObjectStore::Transaction> >&,
>> > boost::intrusive_ptr<OpRequest>)+0x54) [0x55b71f9b1b84]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  8:
>> (ReplicatedBackend::submit_transaction(hobject_t const&, object_stat_sum_t
>> const&, eversion_t const&, std::unique_ptr<PGTransaction,
>> > std::default_delete<PGTransaction> >&&, eversion_t const&, eversion_t
>> const&, s
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  9:
>> (PrimaryLogPG::issue_repop(PrimaryLogPG::RepGather*,
>> PrimaryLogPG::OpContext*)+0xf12) [0x55b71f90e322]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  10:
>> (PrimaryLogPG::execute_ctx(PrimaryLogPG::OpContext*)+0xfae) [0x55b71f969b7e]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  11:
>> (PrimaryLogPG::do_op(boost::intrusive_ptr<OpRequest>&)+0x3965)
>> [0x55b71f96de15]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  12:
>> (PrimaryLogPG::do_request(boost::intrusive_ptr<OpRequest>&,
>> ThreadPool::TPHandle&)+0xbd4) [0x55b71f96f8a4]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  13:
>> (OSD::dequeue_op(boost::intrusive_ptr<PG>, boost::intrusive_ptr<OpRequest>,
>> ThreadPool::TPHandle&)+0x1a9) [0x55b71f7a9ea9]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  14:
>> (PGOpItem::run(OSD*, OSDShard*, boost::intrusive_ptr<PG>&,
>> ThreadPool::TPHandle&)+0x62) [0x55b71fa475d2]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  15:
>> (OSD::ShardedOpWQ::_process(unsigned int, ceph::heartbeat_handle_d*)+0x9f4)
>> [0x55b71f7c6ef4]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  16:
>> (ShardedThreadPool::shardedthreadpool_worker(unsigned int)+0x433)
>> [0x55b71fdc5ce3]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  17:
>> (ShardedThreadPool::WorkThreadSharded::entry()+0x10) [0x55b71fdc8d80]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  18: (()+0x7dd5)
>> [0x7f0971da9dd5]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  19: (clone()+0x6d)
>> [0x7f0970c7002d]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: *** Caught signal
>> (Aborted) **
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  in thread
>> 7f093d71e700 thread_name:tp_osd_tp
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  ceph version 14.2.2
>> (4f8fa0a0024755aae7d95567c63f11d6862d55be) nautilus (stable)
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  1: (()+0xf5d0)
>> [0x7f0971db15d0]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  2: (gsignal()+0x37)
>> [0x7f0970ba82c7]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  3: (abort()+0x148)
>> [0x7f0970ba99b8]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  4:
>> (ceph::__ceph_assert_fail(char const*, char const*, int, char
>> const*)+0x199) [0x55b71f668d43]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  5:
>> (ceph::__ceph_assertf_fail(char const*, char const*, int, char const*, char
>> const*, ...)+0) [0x55b71f668ec2]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  6:
>> (KernelDevice::aio_submit(IOContext*)+0x701) [0x55b71fd61ca1]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  7:
>> (BlueStore::_txc_aio_submit(BlueStore::TransContext*)+0x42) [0x55b71fc29892]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  8:
>> (BlueStore::_txc_state_proc(BlueStore::TransContext*)+0x42b)
>> [0x55b71fc496ab]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  9:
>> (BlueStore::queue_transactions(boost::intrusive_ptr<ObjectStore::CollectionImpl>&,
>> std::vector<ObjectStore::Transaction,
>> > std::allocator<ObjectStore::Transaction> >&,
>> boost::intrusive_ptr<TrackedOp>, ThreadPool::T
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  10: (non-virtual
>> thunk to
>> PrimaryLogPG::queue_transactions(std::vector<ObjectStore::Transaction,
>> std::allocator<ObjectStore::Transaction> >&,
>> > boost::intrusive_ptr<OpRequest>)+0x54) [0x55b71f9b1b84]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  11:
>> (ReplicatedBackend::submit_transaction(hobject_t const&, object_stat_sum_t
>> const&, eversion_t const&, std::unique_ptr<PGTransaction,
>> > std::default_delete<PGTransaction> >&&, eversion_t const&, eversion_t
>> const&,
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  12:
>> (PrimaryLogPG::issue_repop(PrimaryLogPG::RepGather*,
>> PrimaryLogPG::OpContext*)+0xf12) [0x55b71f90e322]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  13:
>> (PrimaryLogPG::execute_ctx(PrimaryLogPG::OpContext*)+0xfae) [0x55b71f969b7e]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  14:
>> (PrimaryLogPG::do_op(boost::intrusive_ptr<OpRequest>&)+0x3965)
>> [0x55b71f96de15]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  15:
>> (PrimaryLogPG::do_request(boost::intrusive_ptr<OpRequest>&,
>> ThreadPool::TPHandle&)+0xbd4) [0x55b71f96f8a4]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  16:
>> (OSD::dequeue_op(boost::intrusive_ptr<PG>, boost::intrusive_ptr<OpRequest>,
>> ThreadPool::TPHandle&)+0x1a9) [0x55b71f7a9ea9]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  17:
>> (PGOpItem::run(OSD*, OSDShard*, boost::intrusive_ptr<PG>&,
>> ThreadPool::TPHandle&)+0x62) [0x55b71fa475d2]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  18:
>> (OSD::ShardedOpWQ::_process(unsigned int, ceph::heartbeat_handle_d*)+0x9f4)
>> [0x55b71f7c6ef4]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  19:
>> (ShardedThreadPool::shardedthreadpool_worker(unsigned int)+0x433)
>> [0x55b71fdc5ce3]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  20:
>> (ShardedThreadPool::WorkThreadSharded::entry()+0x10) [0x55b71fdc8d80]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  21: (()+0x7dd5)
>> [0x7f0971da9dd5]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  22: (clone()+0x6d)
>> [0x7f0970c7002d]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: 2019-09-30
>> 22:52:58.883 7f093d71e700 -1 *** Caught signal (Aborted) **
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  in thread
>> 7f093d71e700 thread_name:tp_osd_tp
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  ceph version 14.2.2
>> (4f8fa0a0024755aae7d95567c63f11d6862d55be) nautilus (stable)
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  1: (()+0xf5d0)
>> [0x7f0971db15d0]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  2: (gsignal()+0x37)
>> [0x7f0970ba82c7]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  3: (abort()+0x148)
>> [0x7f0970ba99b8]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  4:
>> (ceph::__ceph_assert_fail(char const*, char const*, int, char
>> const*)+0x199) [0x55b71f668d43]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  5:
>> (ceph::__ceph_assertf_fail(char const*, char const*, int, char const*, char
>> const*, ...)+0) [0x55b71f668ec2]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  6:
>> (KernelDevice::aio_submit(IOContext*)+0x701) [0x55b71fd61ca1]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  7:
>> (BlueStore::_txc_aio_submit(BlueStore::TransContext*)+0x42) [0x55b71fc29892]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  8:
>> (BlueStore::_txc_state_proc(BlueStore::TransContext*)+0x42b)
>> [0x55b71fc496ab]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  9:
>> (BlueStore::queue_transactions(boost::intrusive_ptr<ObjectStore::CollectionImpl>&,
>> std::vector<ObjectStore::Transaction,
>> > std::allocator<ObjectStore::Transaction> >&,
>> boost::intrusive_ptr<TrackedOp>, ThreadPool::T
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  10: (non-virtual
>> thunk to
>> PrimaryLogPG::queue_transactions(std::vector<ObjectStore::Transaction,
>> std::allocator<ObjectStore::Transaction> >&,
>> > boost::intrusive_ptr<OpRequest>)+0x54) [0x55b71f9b1b84]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  11:
>> (ReplicatedBackend::submit_transaction(hobject_t const&, object_stat_sum_t
>> const&, eversion_t const&, std::unique_ptr<PGTransaction,
>> > std::default_delete<PGTransaction> >&&, eversion_t const&, eversion_t
>> const&,
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  12:
>> (PrimaryLogPG::issue_repop(PrimaryLogPG::RepGather*,
>> PrimaryLogPG::OpContext*)+0xf12) [0x55b71f90e322]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  13:
>> (PrimaryLogPG::execute_ctx(PrimaryLogPG::OpContext*)+0xfae) [0x55b71f969b7e]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  14:
>> (PrimaryLogPG::do_op(boost::intrusive_ptr<OpRequest>&)+0x3965)
>> [0x55b71f96de15]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  15:
>> (PrimaryLogPG::do_request(boost::intrusive_ptr<OpRequest>&,
>> ThreadPool::TPHandle&)+0xbd4) [0x55b71f96f8a4]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  16:
>> (OSD::dequeue_op(boost::intrusive_ptr<PG>, boost::intrusive_ptr<OpRequest>,
>> ThreadPool::TPHandle&)+0x1a9) [0x55b71f7a9ea9]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  17:
>> (PGOpItem::run(OSD*, OSDShard*, boost::intrusive_ptr<PG>&,
>> ThreadPool::TPHandle&)+0x62) [0x55b71fa475d2]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  18:
>> (OSD::ShardedOpWQ::_process(unsigned int, ceph::heartbeat_handle_d*)+0x9f4)
>> [0x55b71f7c6ef4]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  19:
>> (ShardedThreadPool::shardedthreadpool_worker(unsigned int)+0x433)
>> [0x55b71fdc5ce3]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  20:
>> (ShardedThreadPool::WorkThreadSharded::entry()+0x10) [0x55b71fdc8d80]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  21: (()+0x7dd5)
>> [0x7f0971da9dd5]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  22: (clone()+0x6d)
>> [0x7f0970c7002d]
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  NOTE: a copy of the
>> executable, or `objdump -rdS <executable>` is needed to interpret this.
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]: --- begin dump of
>> recent events ---
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  -9999> 2019-09-30
>> 20:46:02.076 7f0937f13700  5 osd.17 6328 heartbeat
>> osd_stat(store_statfs(0x1a485594000/0x40000000/0x1bf00000000, data
>> > 0x19cfcadc2a/0x1a3aa68000, compress 0x0/0x0/0x0, omap 0xcdefc78, meta
>> 0x3321038
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  -9998> 2019-09-30
>> 20:46:03.776 7f0937f13700  5 osd.17 6328 heartbeat
>> osd_stat(store_statfs(0x1a485594000/0x40000000/0x1bf00000000, data
>> > 0x19cfcadc2a/0x1a3aa68000, compress 0x0/0x0/0x0, omap 0xcdefc78, meta
>> 0x3321038
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  -9997> 2019-09-30
>> 20:46:04.277 7f0937f13700  5 osd.17 6328 heartbeat
>> osd_stat(store_statfs(0x1a485594000/0x40000000/0x1bf00000000, data
>> > 0x19cfcadc2a/0x1a3aa68000, compress 0x0/0x0/0x0, omap 0xcdefc78, meta
>> 0x3321038
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  -9996> 2019-09-30
>> 20:46:04.777 7f0937f13700  5 osd.17 6328 heartbeat
>> osd_stat(store_statfs(0x1a485594000/0x40000000/0x1bf00000000, data
>> > 0x19cfcadc2a/0x1a3aa68000, compress 0x0/0x0/0x0, omap 0xcdefc78, meta
>> 0x3321038
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  -9995> 2019-09-30
>> 20:46:04.905 7f0950744700  5 bluestore.MempoolThread(0x55b72c210a88)
>> _tune_cache_size target: 8485076992 heap: 398680064 unmapped:
>> > 9035776 mapped: 389644288 old cache_size: 5064831794 new cache siz
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  -9994> 2019-09-30
>> 20:46:04.905 7f0950744700  5 bluestore.MempoolThread(0x55b72c210a88)
>> _trim_shards cache_size: 5064831794 kv_alloc: 1979711488
>> > kv_used: 120635251 meta_alloc: 1979711488 meta_used: 117486636 data_all
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  -9993> 2019-09-30
>> 20:46:05.813 7f094ccd2700 10 monclient: tick
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  -9992> 2019-09-30
>> 20:46:05.813 7f094ccd2700 10 monclient: _check_auth_rotating have uptodate
>> secrets (they expire after 2019-09-30 20:45:35.819641)
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  -9991> 2019-09-30
>> 20:46:09.913 7f0950744700  5 bluestore.MempoolThread(0x55b72c210a88)
>> _tune_cache_size target: 8485076992 heap: 398680064 unmapped:
>> > 9035776 mapped: 389644288 old cache_size: 5064831794 new cache siz
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  -9990> 2019-09-30
>> 20:46:09.913 7f0950744700  5 bluestore.MempoolThread(0x55b72c210a88)
>> _trim_shards cache_size: 5064831794 kv_alloc: 1979711488
>> > kv_used: 120635251 meta_alloc: 1979711488 meta_used: 117486836 data_all
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  -9989> 2019-09-30
>> 20:46:10.681 7f0937f13700  5 osd.17 6328 heartbeat
>> osd_stat(store_statfs(0x1a485594000/0x40000000/0x1bf00000000, data
>> > 0x19cfcadc2a/0x1a3aa68000, compress 0x0/0x0/0x0, omap 0xcdefc78, meta
>> 0x3321038
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  -9988> 2019-09-30
>> 20:46:14.569 7f094dcd4700  2 osd.17 6328 ms_handle_reset con 0x55b72c797000
>> session 0x55b7401f1800
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  -9987> 2019-09-30
>> 20:46:14.569 7f096da92700 10 monclient: handle_auth_request added challenge
>> on 0x55b733bf0400
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  -9986> 2019-09-30
>> 20:46:14.917 7f0950744700  5 bluestore.MempoolThread(0x55b72c210a88)
>> _tune_cache_size target: 8485076992 heap: 398680064 unmapped:
>> > 9035776 mapped: 389644288 old cache_size: 5064831794 new cache siz
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  -9985> 2019-09-30
>> 20:46:14.917 7f0950744700  5 bluestore.MempoolThread(0x55b72c210a88)
>> _trim_shards cache_size: 5064831794 kv_alloc: 1979711488
>> > kv_used: 120635251 meta_alloc: 1979711488 meta_used: 117487036 data_all
>> > Sep 30 22:52:58 storage2n2-la ceph-osd-17[10770]:  -9984> 2019-09-30
>> 20:46:15.381 7f0937f13700  5 osd.17 6328 heartbeat
>> osd_stat(store_statfs(0x1a485594000/0x40000000/0x1bf00000000, data
>> > 0x19cfcadc2a/0x1a3aa68000, compress 0x0/0x0/0x0, omap 0xcdefc78, meta
>> 0x3321038
>> >
>> > _______________________________________________
>> > ceph-users mailing list
>> > ceph-users@lists.ceph.com
>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>>
>>
>> --
>> Cheers,
>> Brad
>>
>>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to