Does anyone know if this is also respecting an nearfull values? Thank you in advice Mehmet
Am 14. Januar 2020 15:20:39 MEZ schrieb Stephan Mueller <smuel...@suse.com>: >Hi, >I sent out this message on the 19th of December and somehow it didn't >got into the list and I just noticed it now. Sorry for the delay. >I tried to resend it but it just returned the same error that mail was >not deliverable to the ceph mailing list. I will send the message >beneath as soon it's finally possible, but for now this should help you >out. > >Stephan > >------ > >Hi, > >if "MAX AVAIL" displays the wrong data, the bug is just made more >visible through the dashboard, as the calculation is correct. > >To get the right percentage you have to divide the used space through >the total, and the total can only consist of two states used and not >used space, so both states will be added together to get the total. > >Or in short: > >used / (avail + used) > >Just looked into the C++ code - Max avail will be calculated the >following way: > >avail_res = avail / raw_used_rate ( >https://github.com/ceph/ceph/blob/nautilus/src/mon/PGMap.cc#L905) > >raw_used_rate *= (sum.num_object_copies - sum.num_objects_degraded) / >sum.num_object_copies >(https://github.com/ceph/ceph/blob/nautilus/src/mon/PGMap.cc#L892) > > >Am Dienstag, den 17.12.2019, 07:07 +0100 schrieb c...@elchaka.de: >> I have observed this in the ceph nautilus dashboard too - and Think >> it is a Display Bug... but sometimes it Shows tue right values >> >> >> Which nautilus u use? >> >> >> Am 10. Dezember 2019 14:31:05 MEZ schrieb "David Majchrzak, ODERLAND >> Webbhotell AB" <da...@oderland.se>: >> > Hi! >> > >> > While browsing /#/pool in nautilus ceph dashboard I noticed it said >> > 93% >> > used on the single pool we have (3x replica). >> > >> > ceph df detal however shows 81% used on the pool and 67% raw >> > useage. >> > >> > # ceph df detail >> > RAW STORAGE: >> > CLASS SIZE AVAIL USED RAW USED %RAW >> > USED >> > ssd 478 TiB 153 TiB 324 TiB 325 >> > TiB 67.96 >> > TOTAL 478 TiB 153 TiB 324 TiB 325 >> > TiB 67.96 >> > >> > POOLS: >> > POOL ID STORED OBJECTS USED %USED >> > >> > MAX AVAIL QUOTA OBJECTS QUOTA BYTES DIRTY USED >> > COMPR UNDER COMPR >> > echo 3 108 TiB 29.49M 324 >> > TiB 81.61 24 >> > TiB N/A N/A 29.49M 0 >> > B 0 B > >I manually calculated the used percentage to get "avail" in your case >it seems to be 73 TiB. That means the the total space available for >your pool would be 397 TiB. >I'm not sure why that is, but it's what the math behind those >calculations say. >(Found a thread regarding that on the new mailing list (ceph- >us...@ceph.io) -> > > >https://lists.ceph.io/hyperkitty/list/ceph-us...@ceph.io/thread/NH2LMMX5KVRWCURI3BARRUAETKE2T2QN/#JDHXOQKWF6NZLQMOGEPAQCLI44KB54A3 > ) > >0.8161 = used (324) / total => total = 397 > >Than I looked at the remaining calculations: > >raw_used_rate *= (sum.num_object_copies - sum.num_objects_degraded) / >sum.num_object_copies > >and > >avail_res = avail / raw_used_rate > >First I looked up the init value for "raw_used_rate" for replicated >pools. It's their size so we can put in 3 here and for "avail_res" is >24. > >So I first calculated the final "raw_used_rate" which is 3.042. That >means that you have around 4.2% degraded pg's in your pool. > >> > >> > >> > I know we're looking at the most full OSD (210PGs, 79% used, 1.17 >> > VAR) >> > and count max avail from that. But where's the 93% full from in >> > dashboard? > >As said above the calculation is right but the data is wrong... As it >uses the real data that can be put in the selected pool, but it uses >everywhere else sizes that consider all pool replicas. > >I created an issue to fix this https://tracker.ceph.com/issues/43384 > >> > >> > My guess is that is comes from calculating: >> > >> > 1 - Max Avail / (Used + Max avail) = 0.93 >> > >> > >> > Kind Regards, >> > >> > David Majchrzak >> > >> > _______________________________________________ >> > ceph-users mailing list >> > ceph-users@lists.ceph.com >> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com >> >> _______________________________________________ >> ceph-users mailing list >> ceph-users@lists.ceph.com >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > >Hope I could clarify some things and thanks for your feedback :) > >BTW this problem currently still exists as there wasn't any change to >these mentioned lines after the nautilus release. > >Stephan
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com