>That should be a config option, since reading while writes still block is also 
>a danger. Multiple clients could read the same object, >perform a in-memory 
>change and their write will block.
>Now, which client will 'win' after the full flag has been removed?

>That could lead to data corruption.

Read ops may not do the any issue...but, I agree with you - that write
IO is an issue and its blocked.

>Just make sure you have proper monitoring on your Ceph cluster. At nearfull it 
>goes into WARN and you should act on that.

Yes..

Thanks
Swami

On Tue, Jul 19, 2016 at 4:36 PM, Wido den Hollander <w...@42on.com> wrote:
>
>> Op 19 juli 2016 om 12:37 schreef M Ranga Swami Reddy <swamire...@gmail.com>:
>>
>>
>> Thanks for the correction...so even one OSD reaches to 95% full, the
>> total ceph cluster IO (R/W) will be blocked...Ideally read IO should
>> work...
>
> That should be a config option, since reading while writes still block is 
> also a danger. Multiple clients could read the same object, perform a 
> in-memory change and their write will block.
>
> Now, which client will 'win' after the full flag has been removed?
>
> That could lead to data corruption.
>
> Just make sure you have proper monitoring on your Ceph cluster. At nearfull 
> it goes into WARN and you should act on that.
>
> Wido
>
>>
>> Thanks
>> Swami
>>
>> On Tue, Jul 19, 2016 at 3:41 PM, Wido den Hollander <w...@42on.com> wrote:
>> >
>> >> Op 19 juli 2016 om 11:55 schreef M Ranga Swami Reddy 
>> >> <swamire...@gmail.com>:
>> >>
>> >>
>> >> Thanks for detail...
>> >> When an OSD is 95% full, then that specific OSD's write IO blocked.
>> >>
>> >
>> > No, the *whole* cluster will block. In the OSDMap the flag 'full' is set 
>> > which causes all I/O to stop (even read!) until you make sure the OSD 
>> > drops below 95%.
>> >
>> > Wido
>> >
>> >> Thanks
>> >> Swami
>> >>
>> >> On Tue, Jul 19, 2016 at 3:07 PM, Christian Balzer <ch...@gol.com> wrote:
>> >> >
>> >> > Hello,
>> >> >
>> >> > On Tue, 19 Jul 2016 14:23:32 +0530 M Ranga Swami Reddy wrote:
>> >> >
>> >> >> >> Using ceph cluster with 100+ OSDs and cluster is filled with 60% 
>> >> >> >> data.
>> >> >> >> One of the OSD is 95% full.
>> >> >> >> If an OSD is 95% full, is it impact the any storage operation? Is 
>> >> >> >> this
>> >> >> >> impacts on VM/Instance?
>> >> >>
>> >> >> >Yes, one OSD will impact whole cluster. It will block write 
>> >> >> >operations to the cluster
>> >> >>
>> >> >> Thanks for clarification. Really?? Is this(OSD 95%) full designed to
>> >> >> block write I/O of ceph cluster?
>> >> >>
>> >> > Really.
>> >> > To be more precise, any I/O that touches any PG on that OSD will block.
>> >> > So with a sufficiently large cluster you may have some, few, I/Os still 
>> >> > go
>> >> > through as they don't use that OSD at all.
>> >> >
>> >> > That's why:
>> >> >
>> >> > 1. Ceph has the near-full warning (which of course may need to be
>> >> > adjusted to correctly reflect things, especially with smaller clusters).
>> >> > Once you get that warning, you NEED to take action immediately.
>> >> >
>> >> > 2. You want to graph the space utilization of all your OSDs with 
>> >> > something
>> >> > like graphite. That allows you to spot trends of uneven data 
>> >> > distribution
>> >> > early and thus react early to it.
>> >> > I re-weight (CRUSH re-weight, as this is permanent and my clusters 
>> >> > aren't
>> >> > growing frequently) OSDs so they they are at least within 10% of each
>> >> > other.
>> >> >
>> >> > Christian
>> >> >> Because I have around 251 OSDs out which one OSD is 95% full, but
>> >> >> other 250 OSDs not in near full also...
>> >> >>
>> >> >> Thanks
>> >> >> Swami
>> >> >>
>> >> >>
>> >> >> On Tue, Jul 19, 2016 at 2:17 PM, Henrik Korkuc <li...@kirneh.eu> wrote:
>> >> >> > On 16-07-19 11:44, M Ranga Swami Reddy wrote:
>> >> >> >>
>> >> >> >> Hi,
>> >> >> >> Using ceph cluster with 100+ OSDs and cluster is filled with 60% 
>> >> >> >> data.
>> >> >> >> One of the OSD is 95% full.
>> >> >> >> If an OSD is 95% full, is it impact the any storage operation? Is 
>> >> >> >> this
>> >> >> >> impacts on VM/Instance?
>> >> >> >
>> >> >> > Yes, one OSD will impact whole cluster. It will block write 
>> >> >> > operations to
>> >> >> > the cluster
>> >> >> >>
>> >> >> >> Immediately I have reduced the OSD weight, which was filled with 95 
>> >> >> >> %
>> >> >> >> data. After re-weight, data rebalanaced and OSD came to normal state
>> >> >> >> (ie < 80%) with 1 hour time frame.
>> >> >> >>
>> >> >> >>
>> >> >> >> Thanks
>> >> >> >> Swami
>> >> >> >> _______________________________________________
>> >> >> >> ceph-users mailing list
>> >> >> >> ceph-users@lists.ceph.com
>> >> >> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> >> >> >
>> >> >> >
>> >> >> >
>> >> >> > _______________________________________________
>> >> >> > ceph-users mailing list
>> >> >> > ceph-users@lists.ceph.com
>> >> >> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> >> >> _______________________________________________
>> >> >> ceph-users mailing list
>> >> >> ceph-users@lists.ceph.com
>> >> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> >> >>
>> >> >
>> >> >
>> >> > --
>> >> > Christian Balzer        Network/Systems Engineer
>> >> > ch...@gol.com           Global OnLine Japan/Rakuten Communications
>> >> > http://www.gol.com/
>> >> _______________________________________________
>> >> ceph-users mailing list
>> >> ceph-users@lists.ceph.com
>> >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to