Hi,
    This's just a test cluster so, I'm just testing the relationship
between these ratios.

I did changes as such --
failsafe_full = 1 (osd failsafe full ratio = 1 in ceph.conf)
backfillfull = 0.99
nearfull = 0.95
full = 0.96

But the ceph health detail output shows a different story (different from
what I set) --
OSD_OUT_OF_ORDER_FULL full ratio(s) out of order
    full_ratio (0.96) < backfillfull_ratio (0.99), increased
    osd_failsafe_full_ratio (0.97) < full_ratio (0.99), increased

Also as per the documentation
<http://docs.ceph.com/docs/master/rados/operations/health-checks/#osd-out-of-order-full>,
the expected order must be, backfillfull < nearfull.

Thanks of the response!

On Fri, Sep 15, 2017 at 1:27 AM, David Turner <drakonst...@gmail.com> wrote:

> The warning you are seeing is because those settings are out of order and
> it's showing you which ones are greater than the ones they should be.
>  backfillfull_ratio is supposed to be higher than nearfull_ratio and
> osd_failsafe_full_ratio is supposed to be higher than full_ratio.
>  nearfull_ratio is a warning that shows up in your ceph status, but doesn't
> prevent anything from happening; backfillfull_ratio prevents backfilling
> from happening; and full_ratio prevents any IO from happening at all.
>
> That is the answer to your question, but below is addressing the
> ridiculous values you are trying to set those to.
>
> Why are you using such high ratios?  By default 5% of the disk is reserved
> by root for root and nobody but root.  I think that can be adjusted when
> you create the filesystem, but I am unaware if ceph-deploy does that or
> not.  But if that is the setting and if you're running your OSDs as user
> ceph (Jewel or later), then they will cap out at 95% full and the OS will
> fail to write to the OSD disk.
>
> (assuming you set your ratios in the proper order) You are leaving
> yourself no room for your cluster to recover from any sort of down osds or
> failed osds.  I don't know what disks you're using, but I don't know of any
> that are guaranteed not to fail.  If your disks can't perform any
> backfilling, then you can't recover from anything... including just
> restarting an osd daemon or a node...  Based on 97% nearfull being your
> setting... you're giving yourself a 2% warning period to add more storage
> before your cluster is incapable of receiving reads or writes.  BUT you
> also set your cluster to not be able to backfill anything if the OSD is
> over 98% full.  Those settings pretty much guarantee that you will be 100%
> stuck and unable to even add more storage to your cluster if you wait until
> your nearfull_ratio is triggered.
>
> I'm just going to say it... DON'T RUN WITH THESE SETTINGS EVER.  DON'T
> EVEN COME CLOSE TO THESE SETTINGS, THEY ARE TERRIBLE!!!
>
> 90% full_ratio is good (95% is the default) because it is a setting you
> can change and if you get into a situation where you need to recover your
> cluster and your cluster is full because of a failed node or anything, then
> you can change the full_ratio and have a chance to still recover your
> cluster.
>
> 80% nearfull_ratio is good (85% is the default) because it gives you 10%
> usable disk space for you to add more storage to your cluster or clean up
> cruft in your cluster that you don't need.  If it takes you a long time to
> get new hardware or find things to delete in your cluster, consider a lower
> number for this warning.
>
> 85% backfillfull_ratio is good (90% is the default) because of the same
> reason as full_ratio.  You can increase it if you need to for a critical
> recovery.  But with these setting a backfilling operation won't bring you
> too close to your full_ratio that you are in a high danger of blocking all
> IO to your cluster.
>
> Even if you stick with the defaults you're in a good enough situation
> where you will be most likely able to recover from most failures in your
> cluster.  But don't push them up unless you are in the middle of a
> catastrophic failure and you're doing it specifically to recover after you
> have your game-plan resolution in place.
>
>
>
> On Thu, Sep 14, 2017 at 10:03 AM Ronny Aasen <ronny+ceph-us...@aasen.cx>
> wrote:
>
>> On 14. sep. 2017 11:58, dE . wrote:
>> > Hi,
>> >      I got a ceph cluster where I'm getting a OSD_OUT_OF_ORDER_FULL
>> > health error, even though it appears that it is in order --
>> >
>> > full_ratio 0.99
>> > backfillfull_ratio 0.97
>> > nearfull_ratio 0.98
>> >
>> > These don't seem like a mistake to me but ceph is complaining --
>> > OSD_OUT_OF_ORDER_FULL full ratio(s) out of order
>> >      backfillfull_ratio (0.97) < nearfull_ratio (0.98), increased
>> >      osd_failsafe_full_ratio (0.97) < full_ratio (0.99), increased
>> >
>> >
>> > _______________________________________________
>> > ceph-users mailing list
>> > ceph-users@lists.ceph.com
>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> >
>>
>>
>>
>> post output from
>>
>> ceph osd df
>> _______________________________________________
>> ceph-users mailing list
>> ceph-users@lists.ceph.com
>> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to