Yeah.
I'm monitoring such issue reports for a while and it looks like
something is definitely wrong with response times under certain
circumstances. Mpt sure if all these reports have the same root cause
though.
Scrubbing seems to be one of the trigger.
Perhaps we need more low-level detection/warning for high response times
from HW and/or DB.
Planning to look how feasible is that warning means shortly.
Thanks,
Igor
On 2/15/2019 3:24 PM, Denny Kreische wrote:
Hi Igor,
Thanks for your reply.
I can verify, discard is disabled in our cluster:
10:03 root@node106b [fra]:~# ceph daemon osd.417 config show | grep discard
"bdev_async_discard": "false",
"bdev_enable_discard": "false",
[...]
So there must be something else causing the problems.
Thanks,
Denny
Am 15.02.2019 um 12:41 schrieb Igor Fedotov <ifedo...@suse.de>:
Hi Denny,
Do not remember exactly when discards appeared in BlueStore but they are
disabled by default:
See bdev_enable_discard option.
Thanks,
Igor
On 2/15/2019 2:12 PM, Denny Kreische wrote:
Hi,
two weeks ago we upgraded one of our ceph clusters from luminous 12.2.8 to
mimic 13.2.4, cluster is SSD-only, bluestore-only, 68 nodes, 408 OSDs.
somehow we see strange behaviour since then. Single OSDs seem to block for
around 5 minutes and this causes the whole cluster and connected applications
to hang. This happened 5 times during the last 10 days at irregular times, it
didn't happen before the upgrade.
OSD log shows something like this (more log here:
https://pastebin.com/6BYam5r4):
[...]
2019-02-14 23:53:39.754 7f379a368700 -1 osd.417 340516 get_health_metrics
reporting 3 slow ops, oldest is osd_op(client.84226977.0:5112539976 0.dff
0.1d783dff (undecoded) ondisk+read+known_if_redirected e340516)
2019-02-14 23:53:40.706 7f379a368700 -1 osd.417 340516 get_health_metrics
reporting 7 slow ops, oldest is osd_op(client.84226977.0:5112539976 0.dff
0.1d783dff (undecoded) ondisk+read+known_if_redirected e340516)
[...]
In this example osd.417 seems to have a problem. I can see same log line in
other osd logs with placement groups related to osd.417.
I assume that all placement groups related to osd.417 are hanging or blocked
when osd.417 is blocked.
How can I see in detail what might cause a certain OSD to stop working?
The cluster consists of 3 different SSD vendors (micron, samsung, intel), but
only micron disks are affected until now. we earlier had problems with micron
SSDs with filestore (xfs), it was fstrim to cause single OSDs to block for
several minutes. we migrated to bluestore about a year ago. just in case, is
there any kind of ssd trim/discard happening in bluestore since mimic?
Thanks,
Denny
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com