Adding the dev list since it seems like a bug in 14.2.5.

I was able to capture the output from perf top:

  21.58%  libceph-common.so.0               [.] 
ceph::buffer::v14_2_0::list::append
  20.90%  libstdc++.so.6.0.19               [.] std::getline<char, 
std::char_traits<char>, std::allocator<char> >
  13.25%  libceph-common.so.0               [.] 
ceph::buffer::v14_2_0::list::append
  10.11%  libstdc++.so.6.0.19               [.] std::istream::sentry::sentry
   8.94%  libstdc++.so.6.0.19               [.] std::basic_ios<char, 
std::char_traits<char> >::clear
   3.24%  libceph-common.so.0               [.] 
ceph::buffer::v14_2_0::ptr::unused_tail_length
   1.69%  libceph-common.so.0               [.] std::getline<char, 
std::char_traits<char>, std::allocator<char> >@plt
   1.63%  libstdc++.so.6.0.19               [.] std::istream::sentry::sentry@plt
   1.21%  [kernel]                          [k] __do_softirq
   0.77%  libpython2.7.so.1.0               [.] PyEval_EvalFrameEx
   0.55%  [kernel]                          [k] _raw_spin_unlock_irqrestore

I increased mon debugging to 20 and nothing stuck out to me.

Bryan

> On Dec 12, 2019, at 4:46 PM, Bryan Stillwell <bstillw...@godaddy.com> wrote:
> 
> On our test cluster after upgrading to 14.2.5 I'm having problems with the 
> mons pegging a CPU core while moving data around.  I'm currently converting 
> the OSDs from FileStore to BlueStore by marking the OSDs out in multiple 
> nodes, destroying the OSDs, and then recreating them with ceph-volume lvm 
> batch.  This seems too get the ceph-mon process into a state where it pegs a 
> CPU core on one of the mons:
> 
> 1764450 ceph      20   0 4802412   2.1g  16980 S 100.0 28.1   4:54.72 ceph-mon
> 
> Has anyone else run into this with 14.2.5 yet?  I didn't see this problem 
> while the cluster was running 14.2.4.
> 
> Thanks,
> Bryan
_______________________________________________
ceph-users mailing list -- ceph-users@ceph.io
To unsubscribe send an email to ceph-users-le...@ceph.io

Reply via email to