We (my customer ) are trying to test at Jewell now but I can say that the
above behavior was also observed by my customer at Infernalis. After 300
million or so objects in a single bucket the cluster basically fell down as
described above. Few hundred osds in this cluster. We are concerned that
this may not be remedied by a hundreds of buckets approach as well. Testing
continues.
On Mon, May 23, 2016 at 7:35 PM Vickey Singh <vickey.singh22...@gmail.com>
wrote:

> Hello Guys
>
> Is several millions of object with Ceph ( for RGW use case ) still an
> issue ?  Or is it fixed ?
>
> Thnx
> Vickey
>
> On Thu, Jan 28, 2016 at 12:55 AM, Krzysztof Księżyk <kksie...@gmail.com>
> wrote:
>
>> Stefan Rogge <stefan.ceph@...> writes:
>>
>> >
>> >
>> > Hi,
>> > we are using the Ceph with RadosGW and S3 setting.
>> > With more and more objects in the storage the writing speed slows down
>> significantly. With 5 million object in the storage we had a writing speed
>> of 10MS/s. With 10 million objects in the storage its only 5MB/s.
>> > Is this a common issue?
>> > Is the RadosGW suitable for a large amount of objects or would you
>> recommend to not use the RadosGW with these amount of objects?
>> >
>> > Thank you.
>> >
>> > Stefan
>> >
>> > I found also a ticket at the ceph tracker with the same issue:
>> >
>> >
>> http://tracker.ceph.com/projects/ceph/wiki/Rgw_-_bucket_index_scalability
>> >
>> > _______________________________________________
>> > ceph-users mailing list
>> > ceph-users@...
>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> >
>>
>> Hi,
>>
>> I'm struggling with the same issue on Ceph 9.2.0. Unfortunately I wasn't
>> aware of it and now the only way to improve things is create new bucket
>> with bucket index shrading or change way our apps store data into buckets.
>> And of course copy tons of data :( In my case also sth happened to leveldb
>> files and now I cannot even run some radosgw-admin commands like:
>>
>> radosgw-admin bucket check -b ....
>>
>> what causes osd daemon flapping and process timeout messages in logs. PGS
>> containing  .rgw.bucket.index  can't be even backfilled to other osd as
>> osd
>> process dies with messages:
>>
>> [...]
>> > 2016-01-25 15:47:22.700737 7f79fc66d700  1 heartbeat_map is_healthy
>> 'OSD::osd_op_tp thread 0x7f7992c86700' had suicide timed out after 150
>> > 2016-01-25 15:47:22.702619 7f79fc66d700 -1 common/HeartbeatMap.cc: In
>> function 'bool ceph::HeartbeatMap::_check(const ceph::heartbeat_handle_d*,
>> const char*, time_t)' thread 7f79fc66d700 time 2016-01-25 15:47:22.700751
>> > common/HeartbeatMap.cc: 81: FAILED assert(0 == "hit suicide timeout")
>> >
>> >  ceph version 9.2.0 (bb2ecea240f3a1d525bcb35670cb07bd1f0ca299)
>> >  1: (ceph::__ceph_assert_fail(char const*, char const*, int, char
>> const*)+0x85) [0x7f7a019f4be5]
>> >  2: (ceph::HeartbeatMap::_check(ceph::heartbeat_handle_d const*, char
>> const*, long)+0x2d9) [0x7f7a019343b9]
>> >  3: (ceph::HeartbeatMap::is_healthy()+0xd6) [0x7f7a01934bf6]
>> >  4: (ceph::HeartbeatMap::check_touch_file()+0x2c) [0x7f7a019353bc]
>> >  5: (CephContextServiceThread::entry()+0x15b) [0x7f7a01a10dcb]
>> >  6: (()+0x7df5) [0x7f79ffa8fdf5]
>> >  7: (clone()+0x6d) [0x7f79fe3381ad]
>> >
>> >
>> I don't know - maybe it's because number of leveldb files in omap folder
>> (total 5.1GB). Read somewhere that things can be improved by setting
>> 'leveldb_compression' to false and leveldb_compact_on_mount to true but I
>> don't know if these options have any effect in 9.2.0 as they are not
>> documented for this release. Tried with 'leveldb_compression' but without
>> visible effect and wasn't brave enough with trying
>> leveldb_compact_on_mount
>> on production env. But setting it to true on my test 0.94.5 makes osd
>> failing on restart.
>>
>> Kind regards -
>> Krzysztof Księżyk
>>
>>
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to