What version of ovirt and gluster? Sounds like something I just saw with gluster 3.12.x, are you using libgfapi or just fuse mounts?
> From: Sahina Bose <[email protected]> > Subject: Re: [ovirt-users] gluster self-heal takes cluster offline > Date: March 23, 2018 at 1:26:01 AM CDT > To: Jim Kusznir > Cc: Ravishankar Narayanankutty; users > > > > On Fri, Mar 16, 2018 at 2:45 AM, Jim Kusznir <[email protected] > <mailto:[email protected]>> wrote: > Hi all: > > I'm trying to understand why/how (and most importantly, how to fix) an > substantial issue I had last night. This happened one other time, but I > didn't know/understand all the parts associated with it until last night. > > I have a 3 node hyperconverged (self-hosted engine, Gluster on each node) > cluster. Gluster is Replica 2 + arbitrar. Current network configuration is > 2x GigE on load balance ("LAG Group" on switch), plus one GigE from each > server on a separate vlan, intended for Gluster (but not used). Server > hardware is Dell R610's, each server as an SSD in it. Server 1 and 2 have > the full replica, server 3 is the arbitrar. > > I put server 2 into maintence so I can work on the hardware, including turn > it off and such. In the course of the work, I found that I needed to > reconfigure the SSD's partitioning somewhat, and it resulted in wiping the > data partition (storing VM images). I figure, its no big deal, gluster will > rebuild that in short order. I did take care of the extended attr settings > and the like, and when I booted it up, gluster came up as expected and began > rebuilding the disk. > > How big was the data on this partition? What was the shard size set on the > gluster volume? > Out of curiosity, how long did it take to heal and come back to operational? > > > The problem is that suddenly my entire cluster got very sluggish. The entine > was marking nodes and VMs failed and unfaling them throughout the system, > fairly randomly. It didn't matter what node the engine or VM was on. At one > point, it power cycled server 1 for "non-responsive" (even though everything > was running on it, and the gluster rebuild was working on it). As a result > of this, about 6 VMs were killed and my entire gluster system went down hard > (suspending all remaining VMs and the engine), as there were no remaining > full copies of the data. After several minutes (these are Dell servers, > after all...), server 1 came back up, and gluster resumed the rebuild, and > came online on the cluster. I had to manually (virtsh command) unpause the > engine, and then struggle through trying to get critical VMs back up. > Everything was super slow, and load averages on the servers were often seen > in excess of 80 (these are 8 core / 16 thread boxes). Actual CPU usage > (reported by top) was rarely above 40% (inclusive of all CPUs) for any one > server. Glusterfs was often seen using 180%-350% of a CPU on server 1 and 2. > > I ended up putting the cluster in global HA maintence mode and disabling > power fencing on the nodes until the process finished. It appeared on at > least two occasions a functional node was marked bad and had the fencing not > been disabled, a node would have rebooted, just further exacerbating the > problem. > > Its clear that the gluster rebuild overloaded things and caused the problem. > I don't know why the load was so high (even IOWait was low), but load > averages were definately tied to the glusterfs cpu utilization %. At no > point did I have any problems pinging any machine (host or VM) unless the > engine decided it was dead and killed it. > > Why did my system bite it so hard with the rebuild? I baby'ed it along until > the rebuild was complete, after which it returned to normal operation. > > As of this event, all networking (host/engine management, gluster, and VM > network) were on the same vlan. I'd love to move things off, but so far any > attempt to do so breaks my cluster. How can I move my management interfaces > to a separate VLAN/IP Space? I also want to move Gluster to its own private > space, but it seems if I change anything in the peers file, the entire > gluster cluster goes down. The dedicated gluster network is listed as a > secondary hostname for all peers already. > > Will the above network reconfigurations be enough? I got the impression that > the issue may not have been purely network based, but possibly server IO > overload. Is this likely / right? > > I appreciate input. I don't think gluster's recovery is supposed to do as > much damage as it did the last two or three times any healing was required. > > Thanks! > --Jim > > _______________________________________________ > Users mailing list > [email protected] <mailto:[email protected]> > http://lists.ovirt.org/mailman/listinfo/users > <http://lists.ovirt.org/mailman/listinfo/users> > > > _______________________________________________ > Users mailing list > [email protected] > http://lists.ovirt.org/mailman/listinfo/users
_______________________________________________ Users mailing list [email protected] http://lists.ovirt.org/mailman/listinfo/users

