Re: [ceph-users] incomplete PG for erasure coding pool after OSD failure

2018-06-26 Thread Caspar Smit
Hi Anton, With Erasure Coding the min_size (minimum number of shards/replicas needed to allow IO) of a pool is K+1 (in your case 4) so a single OSD failure already triggers an IO freeze (because k=3 m=1) if you have 5 equal hosts ceph 'should' get back to HEALTH_OK automatically (it will be backfi

[ceph-users] incomplete PG for erasure coding pool after OSD failure

2018-06-26 Thread Anton Aleksandrov
Hello, We have small cluster, initially on 4 hosts (1 osd per host, 8tb each) with erasure-coding for data-pool (k=3 m=1). After some time I have added one more small host (1 osd, 2tb). Ceph has synced fine. Then I have powered off one of first 8tb hosts and terminated it. Also removed fro