Hello, I'm setting up a 3 nodes OpenNebula[1] cluster on Debian Wheezy using a SAN for shared storage and KVM as hypervisor.
The OpenNebula fontend is a VM for HA[2]. I had some quorum issues when the node running the fontend die as the two other nodes loose quorum, so I added a pure quorum node in standby="on" mode. My physical hosts are fenced using stonith:external/ipmi, which works great, one stonith device per node with a anti-location on itself. I have more troubles fencing the VMs since they can move. I try to define a stonith device per VM and colocate it with the VM itslef like this: #+begin_src primitive ONE-Frontend ocf:heartbeat:VirtualDomain \ params config="/var/lib/one/datastores/one/one.xml" \ op start interval="0" timeout="90" \ op stop interval="0" timeout="100" \ meta target-role="Stopped" primitive Quorum-Node ocf:heartbeat:VirtualDomain \ params config="/var/lib/one/datastores/one/quorum.xml" \ op start interval="0" timeout="90" \ op stop interval="0" timeout="100" \ meta target-role="Started" is-managed="true" primitive Stonith-Quorum-Node stonith:external/libvirt \ params hostlist="quorum" hypervisor_uri="qemu:///system" pcmk_host_list="quorum" pcmk_host_check="static-list" \ op monitor interval="30m" \ meta target-role="Started" location ONE-Fontend-fenced-by-hypervisor Stonith-ONE-Frontend \ rule $id="ONE-Fontend-fenced-by-hypervisor-rule" inf: #uname ne quorum or #uname ne one location ONE-Frontend-run-on-hypervisor ONE-Frontend \ rule $id="ONE-Frontend-run-on-hypervisor-rule" 20: #uname eq nebula1 \ rule $id="ONE-Frontend-run-on-hypervisor-rule-0" 30: #uname eq nebula2 \ rule $id="ONE-Frontend-run-on-hypervisor-rule-1" 40: #uname eq nebula3 location Quorum-Node-fenced-by-hypervisor Stonith-Quorum-Node \ rule $id="Quorum-Node-fenced-by-hypervisor-rule" inf: #uname ne quorum or #uname ne one location Quorum-Node-run-on-hypervisor Quorum-Node \ rule $id="Quorum-Node-run-on-hypervisor-rule" 50: #uname eq nebula1 \ rule $id="Quorum-Node-run-on-hypervisor-rule-0" 40: #uname eq nebula2 \ rule $id="Quorum-Node-run-on-hypervisor-rule-1" 30: #uname eq nebula3 colocation Fence-ONE-Frontend-on-its-hypervisor inf: ONE-Frontend Stonith-ONE-Frontend colocation Fence-Quorum-Node-on-its-hypervisor inf: Quorum-Node Stonith-Quorum-Node property $id="cib-bootstrap-options" \ dc-version="1.1.7-ee0730e13d124c3d58f00016c3376a1de5323cff" \ cluster-infrastructure="openais" \ expected-quorum-votes="5" \ stonith-enabled="true" \ last-lrm-refresh="1412242734" \ stonith-timeout="30" \ symmetric-cluster="false" #+end_src But, I can not start the Quorum-Node resource, I get the following in logs: #+begin_src info: can_fence_host_with_device: Stonith-nebula2-IPMILAN can not fence quorum: static-list #+end_src All the examples I found describe a configuration where each VM stay on a single hypervisor, in which case libvirt is configured to listen on TCP and the “hypervisor_uri” point to it. Does someone have ideas on configuring stonith:external/libvirt for movable VMs? Regards. Footnotes: [1] http://opennebula.org/ [2] http://docs.opennebula.org/4.8/advanced_administration/high_availability/oneha.html -- Daniel Dehennin Récupérer ma clef GPG: gpg --recv-keys 0xCC1E9E5B7A6FE2DF Fingerprint: 3E69 014E 5C23 50E8 9ED6 2AAD CC1E 9E5B 7A6F E2DF
signature.asc
Description: PGP signature
_______________________________________________ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org