Re: [ceph-users] Huge memory usage spike in OSD on hammer/giant

2015-09-07 Thread
Yeh, There is bug which would use huge memory. It be triggered when osd down or add into cluster and do recovery/backfilling. The patch https://github.com/ceph/ceph/pull/5656 https://github.com/ceph/ceph/pull/5451 merged into master would fix it, and it would be backport. I think ceph v0.93 or ne

Re: [ceph-users] RAM usage only very slowly decreases after cluster recovery

2015-08-28 Thread
You could use 'ceph tell osd.* heap release' to release memory cached by tcmalloc. 2015-08-28 12:51 GMT+08:00 Somnath Roy : > Slow memory release could also be because of tcmalloc. Tcmalloc doesn't > release the memory the moment application issue a 'delete' but it cached it > inside for future

Re: [ceph-users] injectargs not working?

2015-07-29 Thread
hi, ceph osd set noscrub(or nodeep-scrub) would stop the scrub forever. and ceph osd unset noscrub would continue to reschedule scrub. so maybe you could use this two command in crontab to schedule scrub manually. 2015-07-30 7:59 GMT+08:00 Quentin Hartman : > well, that would certainly do it. I _

[ceph-users] ec pool history objects

2015-06-15 Thread
hi, all: when I use ec poll, I see there are some object history for object xx. Such as: xx__head_610951D6__2_fe1_2, xx__head_610951D6__2_fe2_2 xx__head_610951D6__2__2 I think this object is used for roll_back when not all shards have written object to

[ceph-users] ec pool history objects

2015-06-15 Thread
hi, all: when I use ec poll, I see there are some object history for object xx. [root@node3 2.1d6s2_head]# ll -R | grep xx -rw-r--r--. 1 root root 65536 Jun 15 17:41 xx__head_610951D6__2_fe1_2 -rw-r--r--. 1 root root 65536 Jun 15 17:41 xx__head_610951D6__2_fe2_2 -rw-r--r--.

Re: [ceph-users] deep scrubbing causes osd down

2015-04-13 Thread
Sorry, I am not sure whether it is look ok in your production environment. Maybe you could use the command: ceph tell osd.0 injectargs "-osd_scrub_sleep 0.5" . This command would affect only one osd. If it works fine for some days, you could set for all osd. This is just a suggestion. 2015-04-1

Re: [ceph-users] deep scrubbing causes osd down

2015-04-12 Thread
hi, Loic: Do you think it is patch https://github.com/ceph/ceph/pull/3318 worth of backport to firely and giant? 2015-04-13 14:00 GMT+08:00 Christian Balzer : > > On Mon, 13 Apr 2015 13:42:39 +0800 池信泽 wrote: > > I knew the scheduler was in the pipeline, good to see i

Re: [ceph-users] deep scrubbing causes osd down

2015-04-12 Thread
Sorry, 0.87 is giant. BTW, you could also set osd_scrub_sleep to your cluster. ceph would sleep some time as you defined when it has scrub some objects. But I am not sure whether is could works good to you. Thanks. 2015-04-13 13:30 GMT+08:00 池信泽 : > hi, you could restrict scrub to certain ti

Re: [ceph-users] deep scrubbing causes osd down

2015-04-12 Thread
hi, you could restrict scrub to certain times of day based on https://github.com/ceph/ceph/pull/3318. You could set osd_scrub_begin_hour and osd_scrub_begin_hour which are suitable for you. This feature is available since 0.93. But it has not been backport to 0.87 (hammer). 2015-04-13 12:55 GMT+

[ceph-users] more human readable log to track request or using mapreduce for data statistics

2015-03-26 Thread
hi,ceph: Currently, the command ”ceph --admin-daemon /var/run/ceph/ceph-osd.0.asok dump_historic_ops“ may return as below: { "description": "osd_op(client.4436.1:11617 rb.0.1153.6b8b4567.0192 [] 2.8eb4757c ondisk+write e92)", "received_at": "2015-03-25 19:41:47.146145",

Re: [ceph-users] scubbing for a long time and not finished

2015-03-17 Thread
On 周二, 3月 17, 2015 at 10:01 上午, Xinze Chi wrote:hi,all:     I find a pg on my test cluster in doing scrubbing for a long time and not finish. there are not some useful scrubbing log. scrubs_active is 1, so inc_scrubs_pending return false. I think the reason is that some sc

[ceph-users] perf counter reset

2014-11-26 Thread
Hi, cepher: How to reset the perf counter. Such as I want to reset the journal_queue_ops 0. Is there a comand to reset it? Thanks. ___ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Re: [ceph-users] why the erasure code pool not support random write?

2014-10-20 Thread
le (in order > to redo the parity calculations). So we simply disallow them in EC pools, > which works fine for the target use cases right now. > -Greg > > > On Monday, October 20, 2014, 池信泽 wrote: > >> hi, cephers: >> >> When I look into the ceph sour

[ceph-users] why the erasure code pool not support random write?

2014-10-20 Thread
hi, cephers: When I look into the ceph source code, I found the erasure code pool not support the random write, it only support the append write. Why? Is that random write of is erasure code high cost and the performance of the deep scrub is very poor? Thanks. _

[ceph-users] ceph data consistency

2014-09-09 Thread
hi, everyone: when I read the filestore.cc, I find the ceph use crc the check the data. Why should check the data? In my knowledge, the disk has error-correcting code (ECC) for each sector. Looking at wiki: http://en.wikipedia.org/wiki/Dis