Thanks, Gregory!

My Ceph version is 0.94.1. What I'm trying to test is the worst situation
when the node is loosing network or becomes inresponsive. So what i do is
"killall -9 ceph-osd", then reboot.

Well, I also tried to do a clean reboot several times (just a "reboot"
command), but i saw no difference - there is always an IO freeze for about
30 seconds. Btw, i'm using Fedora 20 on all nodes.

Ok, I will play with timeouts more.

Thanks again!

On Wed, May 13, 2015 at 10:46 AM, Gregory Farnum <g...@gregs42.com> wrote:

> On Tue, May 12, 2015 at 11:39 PM, Vasiliy Angapov <anga...@gmail.com>
> wrote:
> > Hi, colleagues!
> >
> > I'm testing a simple Ceph cluster in order to use it in production
> > environment. I have 8 OSDs (1Tb SATA  drives) which are evenly
> distributed
> > between 4 nodes.
> >
> > I'v mapped rbd image on the client node and started writing a lot of
> data to
> > it. Then I just reboot one node and see what's happening. What happens is
> > very sad. I have a write freeze for about 20-30 seconds which is enough
> for
> > ext4 filesystem to switch to RO.
> >
> > I wonder, if there is any way to minimize this lag? AFAIK, ext
> filesystems
> > have 5 seconds timeout before switching to RO. So is there any way to get
> > that lag beyond 5 secs? I've tried lowering different osd timeouts, but
> it
> > doesn't seem to help.
> >
> > How do you deal with such a situations? 20 seconds of downtime is not
> > tolerable in production.
>
> What version of Ceph are you running, and how are you rebooting it?
> Any newish version that gets a clean reboot will notify the cluster
> that it's shutting down, so you shouldn't witness blocked rights
> really at all.
>
> If you're doing a reboot that involves just ending the daemon, you
> will have to wait through the timeout period before the OSD gets
> marked down, which defaults to 30 seconds. This is adjustable (look
> for docs on the "osd heartbeat grace" config option), although if you
> set it too low you'll need to change a bunch of other timeouts which I
> don't know off-hand...
> -Greg
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to