08.01.2014, 06:15, "Andrew Beekhof" <and...@beekhof.net>: > On 27 Nov 2013, at 12:26 am, Andrey Groshev <gre...@yandex.ru> wrote: > >> Hi, ALL. >> >> I want to clarify two more questions. >> After stonith reboot - this node hangs with status "pending". >> The logs found string ..... >> >> info: rsc_merge_weights: pgsql:1: Breaking dependency loop at >> msPostgresql >> info: rsc_merge_weights: pgsql:2: Breaking dependency loop at >> msPostgresql >> >> This means that breaking search the depends, because they are no more. >> Or interrupted by an infinite loop for search the dependency? > > The second one, but it has nothing to do with a node being in the "pending" > state. > Where did you see this?
Ok, I've already understood this the problem. I have "location" for right promote|demote resource. And too same logic trough "collocation"/"order". As I thought, they do the same thing and collisions should not happen. At least on the old cluster it works :) Now I have removed all unnecessary. > >> And two. >> Do I need to clone the stonith resource now (In PCMK 1.1.11)? > > No. > >> On the one hand, I see this resource on all nodes through command. >> # cibadmin -Q|grep stonith >> <nvpair name="stonith-enabled" value="true" >> id="cib-bootstrap-options-stonith-enabled"/> >> <primitive id="st1" class="stonith" type="external/sshbykey"> >> <lrm_resource id="st1" type="external/sshbykey" class="stonith"> >> <lrm_resource id="st1" type="external/sshbykey" class="stonith"> >> <lrm_resource id="st1" type="external/sshbykey" class="stonith"> >> (without pending node) > > Like all resources, we check all nodes at startup to see if it is already > active. > >> On the other hand, another command I see only one instance on a particular >> node. >> # crm_verify -LVVVV >> info: main: =#=#=#=#= Getting XML =#=#=#=#= >> info: main: Reading XML from: live cluster >> info: validate_with_relaxng: Creating RNG parser context >> info: determine_online_status_fencing: Node dev-cluster2-node4 is >> active >> info: determine_online_status: Node dev-cluster2-node4 is online >> info: determine_online_status_fencing: - Node dev-cluster2-node1 is >> not ready to run resources >> info: determine_online_status_fencing: Node dev-cluster2-node2 is >> active >> info: determine_online_status: Node dev-cluster2-node2 is online >> info: determine_online_status_fencing: Node dev-cluster2-node3 is >> active >> info: determine_online_status: Node dev-cluster2-node3 is online >> info: determine_op_status: Operation monitor found resource pingCheck:0 >> active on dev-cluster2-node4 >> info: native_print: VirtualIP (ocf::heartbeat:IPaddr2): >> Started dev-cluster2-node4 >> info: clone_print: Master/Slave Set: msPostgresql [pgsql] >> info: short_print: Masters: [ dev-cluster2-node4 ] >> info: short_print: Slaves: [ dev-cluster2-node2 dev-cluster2-node3 >> ] >> info: short_print: Stopped: [ dev-cluster2-node1 ] >> info: clone_print: Clone Set: clnPingCheck [pingCheck] >> info: short_print: Started: [ dev-cluster2-node2 >> dev-cluster2-node3 dev-cluster2-node4 ] >> info: short_print: Stopped: [ dev-cluster2-node1 ] >> info: native_print: st1 (stonith:external/sshbykey): >> Started dev-cluster2-node4 >> info: native_color: Resource pingCheck:3 cannot run anywhere >> info: native_color: Resource pgsql:3 cannot run anywhere >> info: rsc_merge_weights: pgsql:1: Breaking dependency loop at >> msPostgresql >> info: rsc_merge_weights: pgsql:2: Breaking dependency loop at >> msPostgresql >> info: master_color: Promoting pgsql:0 (Master dev-cluster2-node4) >> info: master_color: msPostgresql: Promoted 1 instances of a >> possible 1 to master >> info: LogActions: Leave VirtualIP (Started dev-cluster2-node4) >> info: LogActions: Leave pgsql:0 (Master dev-cluster2-node4) >> info: LogActions: Leave pgsql:1 (Slave dev-cluster2-node2) >> info: LogActions: Leave pgsql:2 (Slave dev-cluster2-node3) >> info: LogActions: Leave pgsql:3 (Stopped) >> info: LogActions: Leave pingCheck:0 (Started dev-cluster2-node4) >> info: LogActions: Leave pingCheck:1 (Started dev-cluster2-node2) >> info: LogActions: Leave pingCheck:2 (Started dev-cluster2-node3) >> info: LogActions: Leave pingCheck:3 (Stopped) >> info: LogActions: Leave st1 (Started dev-cluster2-node4) >> >> However, if I do a "clone" - it turns out the same garbage. >> >> _______________________________________________ >> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >> >> Project Home: http://www.clusterlabs.org >> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >> Bugs: http://bugs.clusterlabs.org > > , > _______________________________________________ > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org _______________________________________________ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org