Hello Guys

Is several millions of object with Ceph ( for RGW use case ) still an issue
?  Or is it fixed ?

Thnx
Vickey

On Thu, Jan 28, 2016 at 12:55 AM, Krzysztof Księżyk <kksie...@gmail.com>
wrote:

> Stefan Rogge <stefan.ceph@...> writes:
>
> >
> >
> > Hi,
> > we are using the Ceph with RadosGW and S3 setting.
> > With more and more objects in the storage the writing speed slows down
> significantly. With 5 million object in the storage we had a writing speed
> of 10MS/s. With 10 million objects in the storage its only 5MB/s.
> > Is this a common issue?
> > Is the RadosGW suitable for a large amount of objects or would you
> recommend to not use the RadosGW with these amount of objects?
> >
> > Thank you.
> >
> > Stefan
> >
> > I found also a ticket at the ceph tracker with the same issue:
> >
> >
> http://tracker.ceph.com/projects/ceph/wiki/Rgw_-_bucket_index_scalability
> >
> > _______________________________________________
> > ceph-users mailing list
> > ceph-users@...
> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
> >
>
> Hi,
>
> I'm struggling with the same issue on Ceph 9.2.0. Unfortunately I wasn't
> aware of it and now the only way to improve things is create new bucket
> with bucket index shrading or change way our apps store data into buckets.
> And of course copy tons of data :( In my case also sth happened to leveldb
> files and now I cannot even run some radosgw-admin commands like:
>
> radosgw-admin bucket check -b ....
>
> what causes osd daemon flapping and process timeout messages in logs. PGS
> containing  .rgw.bucket.index  can't be even backfilled to other osd as osd
> process dies with messages:
>
> [...]
> > 2016-01-25 15:47:22.700737 7f79fc66d700  1 heartbeat_map is_healthy
> 'OSD::osd_op_tp thread 0x7f7992c86700' had suicide timed out after 150
> > 2016-01-25 15:47:22.702619 7f79fc66d700 -1 common/HeartbeatMap.cc: In
> function 'bool ceph::HeartbeatMap::_check(const ceph::heartbeat_handle_d*,
> const char*, time_t)' thread 7f79fc66d700 time 2016-01-25 15:47:22.700751
> > common/HeartbeatMap.cc: 81: FAILED assert(0 == "hit suicide timeout")
> >
> >  ceph version 9.2.0 (bb2ecea240f3a1d525bcb35670cb07bd1f0ca299)
> >  1: (ceph::__ceph_assert_fail(char const*, char const*, int, char
> const*)+0x85) [0x7f7a019f4be5]
> >  2: (ceph::HeartbeatMap::_check(ceph::heartbeat_handle_d const*, char
> const*, long)+0x2d9) [0x7f7a019343b9]
> >  3: (ceph::HeartbeatMap::is_healthy()+0xd6) [0x7f7a01934bf6]
> >  4: (ceph::HeartbeatMap::check_touch_file()+0x2c) [0x7f7a019353bc]
> >  5: (CephContextServiceThread::entry()+0x15b) [0x7f7a01a10dcb]
> >  6: (()+0x7df5) [0x7f79ffa8fdf5]
> >  7: (clone()+0x6d) [0x7f79fe3381ad]
> >
> >
> I don't know - maybe it's because number of leveldb files in omap folder
> (total 5.1GB). Read somewhere that things can be improved by setting
> 'leveldb_compression' to false and leveldb_compact_on_mount to true but I
> don't know if these options have any effect in 9.2.0 as they are not
> documented for this release. Tried with 'leveldb_compression' but without
> visible effect and wasn't brave enough with trying leveldb_compact_on_mount
> on production env. But setting it to true on my test 0.94.5 makes osd
> failing on restart.
>
> Kind regards -
> Krzysztof Księżyk
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to