We had a similar issue in Firefly, where we had a very large number (about 1.500.000) of buckets for a single RGW user. We observed a number of slow requests in day-to-day use, but did not think much of it at the time.

At one point the primary OSD managing the list of buckets for that user crashed and could not restart, because processing the tremendous amount of buckets on startup - which also seemed to be single-threaded, judging by to 100% CPU usage we could see - took longer than the suicide-timeout. That lead to this OSD crashing again, and again. Eventually, it would be marked out and the secondary tried to process the list with the same result, leading to a cascading failure.

While I am quite certain it is a different code path in your case (you speak about a handful of buckets), it certainly sounds like the a very similar issue. Do you have lots of objects in those few buckets, or are they few, but large in size to reach the 30TB? Worst case you might be in for a similar procedure as we had to take: Take load off the cluster, increase the timeouts to ridiculous levels and copy the data over into a more evenly distributed set of buckets (users in our case). Fortunately as long as we did not try to write to the problematic buckets, we could still read from them.

Please notice that this is only a guess, I could be completely wrong.

Daniel

On 2015-11-03 13:33:19 +0000, Gerd Jakobovitsch said:

Dear all,

I have a cluster running hammer (0.94.5), with 5 nodes. The main usage is for S3-compatible object storage. I am getting to a very troublesome problem at a ceph cluster. A single object in the .rgw.buckets.index is not responding to request and takes a very long time while recovering after an osd restart. During this time, the OSDs where this object is mapped got heavily loaded, with high cpu as well as memory usage. At the same time, the directory /var/lib/ceph/osd/ceph-XX/current/omap gets a large number of entries ( > 10000), that won't decrease.

Very frequently, I get >100 blocked requests for this object, and the main OSD that stores it ends up accepting no other requests. Very frequently the OSD ends up crashing due to filestore timeout, and getting it up again is very troublesome - it usually has to run alone in the node for a long time, until the object gets recovered, somehow.

At the OSD logs, there are several entries like these:
 -7051> 2015-11-03 10:46:08.339283 7f776974f700 10 log_client  logged 2015-11-03 10:46:02.942023 osd.63 10.17.0.9:6857/2002 41 : cluster [WRN] slow re quest 120.003081 seconds old, received at 2015-11-03 10:43:56.472825: osd_repop(osd.53.236531:7 34.7 8a7482ff/.dir.default.198764998.1/head//34 v 2369
84'22) currently commit_sent


2015-11-03 10:28:32.405265 7f0035982700  0 log_channel(cluster) log [WRN] : 97 slow requests, 1 included below; oldest blocked for > 2046.502848 secs 2015-11-03 10:28:32.405269 7f0035982700  0 log_channel(cluster) log [WRN] : slow request 1920.676998 seconds old, received at 2015-11-03 09:56:31.7282 24: osd_op(client.210508702.0:14696798 .dir.default.198764998.1 [call rgw.bucket_prepare_op] 15.8a7482ff ondisk+write+known_if_redirected e236956) cur
rently waiting for blocked object

Is there any way to go deeper into this problem, or to rebuild the .rgw index without loosing data? I currently have 30 TB of data in the cluster - most of it concentrated in a handful of buckets - that I can't loose.

Regards.
--
 
 
 
 
 
 
 
 

--

As informações contidas nesta mensagem são CONFIDENCIAIS, protegidas pelo sigilo legal e por direitos autorais. A divulgação, distribuição, reprodução ou qualquer forma de utilização do teor deste documento depende de autorização do emissor, sujeitando-se o infrator às sanções legais. Caso esta comunicação tenha sido recebida por engano, favor avisar imediatamente, respondendo esta mensagem.

_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com


--
--
Daniel Schneller
Principal Cloud Engineer

CenterDevice GmbH
https://www.centerdevice.de
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to