Oh, fence agent works fine if I select ilo4, Thank you for your help! On Sun, Apr 16, 2017 at 8:22 PM Dan Yasny <[email protected]> wrote:
> On Sun, Apr 16, 2017 at 11:19 PM, Konstantin Raskoshnyi < > [email protected]> wrote: > >> Makes sense. >> I was trying to set it up, but doesn't work with our staging hardware. >> We have old ilo100, I'll try again. >> Thanks! >> >> > It is absolutely necessary for any HA to work properly. There's of course > the "confirm host has been shutdown" option, which serves as an override > for the fence command, but it's manual > > >> On Sun, Apr 16, 2017 at 8:18 PM Dan Yasny <[email protected]> wrote: >> >>> On Sun, Apr 16, 2017 at 11:15 PM, Konstantin Raskoshnyi < >>> [email protected]> wrote: >>> >>>> Fence agent under each node? >>>> >>> >>> When you configure a host, there's the power management tab, where you >>> need to enter the bmc details for the host. If you don't have fencing >>> enabled, how do you expect the system to make sure a host running a service >>> is actually down (and it is safe to start HA services elsewhere), and not, >>> for example, just unreachable by the engine? How do you avoid a splitbraid >>> -> SBA ? >>> >>> >>>> >>>> On Sun, Apr 16, 2017 at 8:14 PM Dan Yasny <[email protected]> wrote: >>>> >>>>> On Sun, Apr 16, 2017 at 11:13 PM, Konstantin Raskoshnyi < >>>>> [email protected]> wrote: >>>>> >>>>>> "Corner cases"? >>>>>> I tried to simulate crash of SPM server and ovirt kept trying to >>>>>> reistablished connection to the failed node. >>>>>> >>>>> >>>>> Did you configure fencing? >>>>> >>>>> >>>>>> >>>>>> >>>>>> On Sun, Apr 16, 2017 at 8:10 PM Dan Yasny <[email protected]> wrote: >>>>>> >>>>>>> On Sun, Apr 16, 2017 at 7:29 AM, Nir Soffer <[email protected]> >>>>>>> wrote: >>>>>>> >>>>>>>> On Sun, Apr 16, 2017 at 2:05 PM Dan Yasny <[email protected]> >>>>>>>> wrote: >>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> On Apr 16, 2017 7:01 AM, "Nir Soffer" <[email protected]> wrote: >>>>>>>>> >>>>>>>>> On Sun, Apr 16, 2017 at 4:17 AM Dan Yasny <[email protected]> >>>>>>>>> wrote: >>>>>>>>> >>>>>>>>>> When you set up a storage domain, you need to specify a host to >>>>>>>>>> perform the initial storage operations, but once the SD is defined, >>>>>>>>>> it's >>>>>>>>>> details are in the engine database, and all the hosts get connected >>>>>>>>>> to it >>>>>>>>>> directly. If the first host you used to define the SD goes down, all >>>>>>>>>> other >>>>>>>>>> hosts will still remain connected and work. SPM is an HA service, >>>>>>>>>> and if >>>>>>>>>> the current SPM host goes down, SPM gets started on another host in >>>>>>>>>> the DC. >>>>>>>>>> In short, unless your actual NFS exporting host goes down, there is >>>>>>>>>> no >>>>>>>>>> outage. >>>>>>>>>> >>>>>>>>> >>>>>>>>> There is no storage outage, but if you shutdown the spm host, the >>>>>>>>> spm host >>>>>>>>> will not move to a new host until the spm host is online again, or >>>>>>>>> you confirm >>>>>>>>> manually that the spm host was rebooted. >>>>>>>>> >>>>>>>>> >>>>>>>>> In a properly configured setup the SBA should take care of that. >>>>>>>>> That's the whole point of HA services >>>>>>>>> >>>>>>>> >>>>>>>> In some cases like power loss or hardware failure, there is no way >>>>>>>> to start >>>>>>>> the spm host, and the system cannot recover automatically. >>>>>>>> >>>>>>> >>>>>>> There are always corner cases, no doubt. But in a normal situation. >>>>>>> where an SPM host goes down because of a hardware failure, it gets >>>>>>> fenced, >>>>>>> other hosts contend for SPM and start it. No surprises there. >>>>>>> >>>>>>> >>>>>>>> >>>>>>>> Nir >>>>>>>> >>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> Nir >>>>>>>>> >>>>>>>>> >>>>>>>>>> >>>>>>>>>> On Sat, Apr 15, 2017 at 1:53 PM, Konstantin Raskoshnyi < >>>>>>>>>> [email protected]> wrote: >>>>>>>>>> >>>>>>>>>>> Hi Fernando, >>>>>>>>>>> I see each host has direct connection nfs mount, but yes, if >>>>>>>>>>> main host to which I connected nfs storage going down the storage >>>>>>>>>>> becomes >>>>>>>>>>> unavailable and all vms are down >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>>> On Sat, Apr 15, 2017 at 10:37 AM FERNANDO FREDIANI < >>>>>>>>>>> [email protected]> wrote: >>>>>>>>>>> >>>>>>>>>>>> Hello Konstantin. >>>>>>>>>>>> >>>>>>>>>>>> That doesn`t make much sense make a whole cluster depend on a >>>>>>>>>>>> single host. From what I know any host talk directly to NFS >>>>>>>>>>>> Storage Array >>>>>>>>>>>> or whatever other Shared Storage you have. >>>>>>>>>>>> Have you tested that host going down if that affects the other >>>>>>>>>>>> with the NFS mounted directlly in a NFS Storage array ? >>>>>>>>>>>> >>>>>>>>>>>> Fernando >>>>>>>>>>>> >>>>>>>>>>>> 2017-04-15 12:42 GMT-03:00 Konstantin Raskoshnyi < >>>>>>>>>>>> [email protected]>: >>>>>>>>>>>> >>>>>>>>>>>>> In ovirt you have to attach storage through specific host. >>>>>>>>>>>>> If host goes down storage is not available. >>>>>>>>>>>>> >>>>>>>>>>>>> On Sat, Apr 15, 2017 at 7:31 AM FERNANDO FREDIANI < >>>>>>>>>>>>> [email protected]> wrote: >>>>>>>>>>>>> >>>>>>>>>>>>>> Well, make it not go through host1 and dedicate a storage >>>>>>>>>>>>>> server for running NFS and make both hosts connect to it. >>>>>>>>>>>>>> In my view NFS is much easier to manage than any other type >>>>>>>>>>>>>> of storage, specially FC and iSCSI and performance is pretty >>>>>>>>>>>>>> much the same, >>>>>>>>>>>>>> so you won`t get better results other than management going to >>>>>>>>>>>>>> other type. >>>>>>>>>>>>>> >>>>>>>>>>>>>> Fernando >>>>>>>>>>>>>> >>>>>>>>>>>>>> 2017-04-15 5:25 GMT-03:00 Konstantin Raskoshnyi < >>>>>>>>>>>>>> [email protected]>: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> Hi guys, >>>>>>>>>>>>>>> I have one nfs storage, >>>>>>>>>>>>>>> it's connected through host1. >>>>>>>>>>>>>>> host2 also has access to it, I can easily migrate >>>>>>>>>>>>>>> vms between them. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> The question is - if host1 is down - all infrastructure is >>>>>>>>>>>>>>> down, since all traffic goes through host1, >>>>>>>>>>>>>>> is there any way in oVirt to use redundant storage? >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Only glusterfs? >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Thanks >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>>> Users mailing list >>>>>>>>>>>>>>> [email protected] >>>>>>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> >>>>>>>>>>>> >>>>>>>>>>> _______________________________________________ >>>>>>>>>>> Users mailing list >>>>>>>>>>> [email protected] >>>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>>> _______________________________________________ >>>>>>>>>> Users mailing list >>>>>>>>>> [email protected] >>>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>>>>>>> >>>>>>>>> >>>>>>>>> _______________________________________________ >>>>>>>>> Users mailing list >>>>>>>>> [email protected] >>>>>>>>> http://lists.ovirt.org/mailman/listinfo/users >>>>>>>>> >>>>>>>>> >>>>>>>>>
_______________________________________________ Users mailing list [email protected] http://lists.ovirt.org/mailman/listinfo/users

