[ceph-users] Re: How to check available storage with EC and different sized OSD's ?

2022-11-09 Thread Paweł Kowalski
I don't need to redistribute data after OSD failure. All I want to do in this test setup is to keep data safe in RO after such failure. Paweł W dniu 9.11.2022 o 17:09, Danny Webb pisze: With a 3 osd pool it's not possible for data to be redistributed on failure of an OSD. with a K=2,M=1 va

[ceph-users] Re: How to check available storage with EC and different sized OSD's ?

2022-11-09 Thread Paweł Kowalski
If I start to use all available space that pool can offer (4.5T) and first OSD (2.7T) fails, I'm sure I'll end up with lost data since it's not possible to fit 4.5T on 2 remaining drives with total raw capacity of 3.6T. I'm wondering why ceph isn't complaining now. I thought it should place d

[ceph-users] Re: How to check available storage with EC and different sized OSD's ?

2022-11-08 Thread Paweł Kowalski
I guess it's not an issue in larger scenarios, but I hope there's some feature to inform user that pool is not safe. And what is the general rule? If k+m = #OSDs than do not use disks of different size? P. W dniu 8.11.2022 o 15:25, Paweł Kowalski pisze: Hi, I've set

[ceph-users] How to check available storage with EC and different sized OSD's ?

2022-11-08 Thread Paweł Kowalski
Hi, I've set up a minimal EC setup - 3 OSDs, k=2, m=1: root@skarb:~# ceph osd df ID  CLASS    WEIGHT   REWEIGHT  SIZE RAW USE  DATA OMAP META AVAIL    %USE   VAR   PGS  STATUS [...]  9  low_hdd  2.72849   1.0  2.7 TiB  632 GiB  631 GiB  121 KiB  1.6 GiB  2.1 TiB  22.62  0.

[ceph-users] Re: PG inactive - why?

2022-11-02 Thread Paweł Kowalski
I had to check logs once again to find out what I exactly did... 1. Destroyed 2 OSDs from host pirat and recreated them, but backfilling was still in progress: 2022-10-26T13:22:13.744545+0200 mgr.skarb (mgr.40364478) 93039 : cluster [DBG] pgmap v94205: 285 pgs: 2 active+undersized+degraded+

[ceph-users] Re: PG inactive - why?

2022-11-02 Thread Paweł Kowalski
for the affected pool? Your osd tree could help as well to figure out what happened. Zitat von Paweł Kowalski : Hello, 2 PG's were stuck in "inactive state". I'm trying to find out why this happened. Here's what I found in logs (only what I think is important): 2

[ceph-users] PG inactive - why?

2022-10-28 Thread Paweł Kowalski
Hello, 2 PG's were stuck in "inactive state". I'm trying to find out why this happened. Here's what I found in logs (only what I think is important): 2022-10-27T22:39:29.707362+0200 mgr.skarb (mgr.40364478) 152930 : cluster [DBG] pgmap v154892: 410 pgs: 1 active+remapped+backfilling, 2 activ