Dear all,

unfortunately my stonith does not work on my pacemaker cluster. If I do ifdown 
on the two cluster interconnect interfaces of server sv2827 the server sv2828 
want to fence the server sv2827, but the messages log says:    error: 
remote_op_done: Operation reboot of sv2827-p1 by sv2828-p1 for 
crmd.7979@sv2828-p1.076062f0: No such device
Can somebody please help me?


Jul  9 12:42:49 sv2828 corosync[7749]:   [CMAN  ] quorum lost, blocking activity
Jul  9 12:42:49 sv2828 corosync[7749]:   [QUORUM] This node is within the 
non-primary component and will NOT provide any services.
Jul  9 12:42:49 sv2828 corosync[7749]:   [QUORUM] Members[1]: 1
Jul  9 12:42:49 sv2828 corosync[7749]:   [TOTEM ] A processor joined or left 
the membership and a new membership was formed.
Jul  9 12:42:49 sv2828 crmd[7979]:   notice: cman_event_callback: Membership 
1492: quorum lost
Jul  9 12:42:49 sv2828 crmd[7979]:   notice: crm_update_peer_state: 
cman_event_callback: Node sv2827-p1[2] - state is now lost (was member)
Jul  9 12:42:49 sv2828 crmd[7979]:  warning: match_down_event: No match for 
shutdown action on sv2827-p1
Jul  9 12:42:49 sv2828 crmd[7979]:   notice: peer_update_callback: 
Stonith/shutdown of sv2827-p1 not matched
Jul  9 12:42:49 sv2828 kernel: dlm: closing connection to node 2
Jul  9 12:42:49 sv2828 corosync[7749]:   [CPG   ] chosen downlist: sender r(0) 
ip(192.168.2.28) r(1) ip(192.168.3.28) ; members(old:2 left:1)
Jul  9 12:42:49 sv2828 crmd[7979]:   notice: do_state_transition: State 
transition S_IDLE -> S_INTEGRATION [ input=I_NODE_JOIN cause=C_FSA_INTERNAL 
origin=check_join_state ]
Jul  9 12:42:49 sv2828 corosync[7749]:   [MAIN  ] Completed service 
synchronization, ready to provide service.
Jul  9 12:42:49 sv2828 crmd[7979]:  warning: match_down_event: No match for 
shutdown action on sv2827-p1
Jul  9 12:42:49 sv2828 crmd[7979]:   notice: peer_update_callback: 
Stonith/shutdown of sv2827-p1 not matched
Jul  9 12:42:49 sv2828 attrd[7977]:   notice: attrd_local_callback: Sending 
full refresh (origin=crmd)
Jul  9 12:42:49 sv2828 attrd[7977]:   notice: attrd_trigger_update: Sending 
flush op to all hosts for: last-failure-MYSQLFS (1404902183)
Jul  9 12:42:49 sv2828 attrd[7977]:   notice: attrd_trigger_update: Sending 
flush op to all hosts for: probe_complete (true)
Jul  9 12:42:49 sv2828 attrd[7977]:   notice: attrd_trigger_update: Sending 
flush op to all hosts for: last-failure-MYSQL (1404901921)
Jul  9 12:42:49 sv2828 pengine[7978]:   notice: unpack_config: On loss of CCM 
Quorum: Ignore
Jul  9 12:42:49 sv2828 pengine[7978]:  warning: pe_fence_node: Node sv2827-p1 
will be fenced because the node is no longer part of the cluster
Jul  9 12:42:49 sv2828 pengine[7978]:  warning: determine_online_status: Node 
sv2827-p1 is unclean
Jul  9 12:42:49 sv2828 pengine[7978]:  warning: custom_action: Action 
ipmi-fencing-sv2828_stop_0 on sv2827-p1 is unrunnable (offline)
Jul  9 12:42:49 sv2828 pengine[7978]:  warning: stage6: Scheduling Node 
sv2827-p1 for STONITH
Jul  9 12:42:49 sv2828 pengine[7978]:   notice: LogActions: Move    
ipmi-fencing-sv2828#011(Started sv2827-p1 -> sv2828-p1)
Jul  9 12:42:49 sv2828 pengine[7978]:  warning: process_pe_message: Calculated 
Transition 38: /var/lib/pacemaker/pengine/pe-warn-28.bz2
Jul  9 12:42:49 sv2828 crmd[7979]:   notice: te_fence_node: Executing reboot 
fencing operation (20) on sv2827-p1 (timeout=60000)
Jul  9 12:42:49 sv2828 stonith-ng[7975]:   notice: handle_request: Client 
crmd.7979.6c35e3f1 wants to fence (reboot) 'sv2827-p1' with device '(any)'
Jul  9 12:42:49 sv2828 stonith-ng[7975]:   notice: initiate_remote_stonith_op: 
Initiating remote operation reboot for sv2827-p1: 
076062f0-eff3-4798-a504-16c5c5666a5b (0)
Jul  9 12:42:49 sv2828 stonith-ng[7975]:   notice: can_fence_host_with_device: 
ipmi-fencing-sv2827 can not fence sv2827-p1: static-list
Jul  9 12:42:49 sv2828 stonith-ng[7975]:   notice: can_fence_host_with_device: 
ipmi-fencing-sv2828 can not fence sv2827-p1: static-list
Jul  9 12:42:49 sv2828 stonith-ng[7975]:    error: remote_op_done: Operation 
reboot of sv2827-p1 by sv2828-p1 for crmd.7979@sv2828-p1.076062f0: No such 
device
Jul  9 12:42:49 sv2828 crmd[7979]:   notice: tengine_stonith_callback: Stonith 
operation 8/20:38:0:71703806-8a7c-447f-a033-e3c26abd607c: No such device (-19)
Jul  9 12:42:49 sv2828 crmd[7979]:   notice: run_graph: Transition 38 
(Complete=1, Pending=0, Fired=0, Skipped=5, Incomplete=0, 
Source=/var/lib/pacemaker/pengine/pe-warn-28.bz2): Stopped

With the ipmitool I could test the correct work of the ipmi like power cycle.

pcs status
Cluster name: mysql-int-prod
Last updated: Wed Jul  9 12:46:43 2014
Last change: Wed Jul  9 12:41:14 2014 via crm_resource on sv2828-p1
Stack: cman
Current DC: sv2828-p1 - partition with quorum
Version: 1.1.10-1.el6_4.4-368c726
2 Nodes configured
5 Resources configured
Online: [ sv2827-p1 sv2828-p1 ]
Full list of resources:
ipmi-fencing-sv2827    (stonith:fence_ipmilan):            Started sv2828-p1
 ipmi-fencing-sv2828    (stonith:fence_ipmilan):            Started sv2827-p1
 ClusterIP            (ocf::heartbeat:IPaddr2):            Started sv2828-p1
 MYSQLFS            (ocf::baader:MYSQLFS):               Started sv2828-p1
 MYSQL (ocf::baader:MYSQL):    Started sv2828-p1


pcs property
Cluster Properties:
cluster-infrastructure: cman
dc-version: 1.1.10-1.el6_4.4-368c726
last-lrm-refresh: 1404902348
no-quorum-policy: ignore
stonith-enabled: true


Best regards
Andreas Dvorak

_______________________________________________
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org

Reply via email to