I tend to set it whenever I don't want to be bothered by storage
performance woes (nights I value sleep, etc).
This cluster is bounded by relentless small writes (it has a couple
dozen rbd volumes backing video surveillance DVRs). Some of the software
we run is completely unaffected whereas other software falls apart
during periods of deep-scrubs. I theorize it has to do with the
individual software's attitude about flushing to disk / buffering.
- Mike
On 5/20/2014 8:31 PM, Aaron Ten Clay wrote:
For what it's worth, version 0.79 has different headers, and the awk
command needs $19 instead of $20. But here is the output I have on a
small cluster that I recently rebuilt:
$ ceph pg dump all | grep active | awk '{ print $19}' | sort -k1 | uniq -c
dumped all in format plain
1 2014-05-15
2 2014-05-17
19 2014-05-18
193 2014-05-19
105 2014-05-20
I have set noscrub and nodeep-scrub, as well as noout and nodown off and
on while I performed various maintenance, but that hasn't (apparently)
impeded the regular schedule.
With what frequency are you setting the nodeep-scrub flag?
-Aaron
On Tue, May 20, 2014 at 5:21 PM, Mike Dawson <mike.daw...@cloudapt.com
<mailto:mike.daw...@cloudapt.com>> wrote:
Today I noticed that deep-scrub is consistently missing some of my
Placement Groups, leaving me with the following distribution of PGs
and the last day they were successfully deep-scrubbed.
# ceph pg dump all | grep active | awk '{ print $20}' | sort -k1 |
uniq -c
5 2013-11-06
221 2013-11-20
1 2014-02-17
25 2014-02-19
60 2014-02-20
4 2014-03-06
3 2014-04-03
6 2014-04-04
6 2014-04-05
13 2014-04-06
4 2014-04-08
3 2014-04-10
2 2014-04-11
50 2014-04-12
28 2014-04-13
14 2014-04-14
3 2014-04-15
78 2014-04-16
44 2014-04-17
8 2014-04-18
1 2014-04-20
16 2014-05-02
69 2014-05-04
140 2014-05-05
569 2014-05-06
9231 2014-05-07
103 2014-05-08
514 2014-05-09
1593 2014-05-10
393 2014-05-16
2563 2014-05-17
1283 2014-05-18
1640 2014-05-19
1979 2014-05-20
I have been running the default "osd deep scrub interval" of once
per week, but have disabled deep-scrub on several occasions in an
attempt to avoid the associated degraded cluster performance I have
written about before.
To get the PGs longest in need of a deep-scrub started, I set the
nodeep-scrub flag, and wrote a script to manually kick off
deep-scrub according to age. It is processing as expected.
Do you consider this a feature request or a bug? Perhaps the code
that schedules PGs to deep-scrub could be improved to prioritize PGs
that have needed a deep-scrub the longest.
Thanks,
Mike Dawson
_________________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com <mailto:ceph-users@lists.ceph.com>
http://lists.ceph.com/__listinfo.cgi/ceph-users-ceph.__com
<http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com