Hi all,
I'm kinda new to Pacemaker and have some questions I hope some of you could 
help me with, any idea(s) would be highly appreciated, thank you.
I'am building the HA cluster with corosync2.3.3 + cluster-glue + 
pacemaker-1.1.11-rc5. The problem is, when I try to remove a node from HA 
cluster with "crm_node -f -R nodeID", it seems to remove the wrong node.

My setup is as follows(and it can be reproduced):
1. The HA is ok with 2 resource and 2 nodes, and with stonith-enabled="false":
root@h1:/opt/bin# crm status
Last updated: Sat Apr 12 16:00:51 2014
Last change: Sat Apr 12 16:00:43 2014 via cibadmin on h1
Stack: corosync
Current DC: h1 (1084752017) - partition with quorum
Version: 1.1.10-33f9d09
2 Nodes configured
2 Resources configured


Online: [ h0 h1 ]

 VMdos-1        (ocf::heartbeat:VirtualDomain): Started h1
root@h1:/opt/bin#

root@h1:/opt/bin# crm_node --list
1084752017 h1
1084752016 h0
root@h1:/opt/bin#

2. remove node h0 from HA
(1). Stop pacemakerd and corosync service on node h0 (2). Run the follow cmd on 
node h1 to Remove node info from cib:
cibadmin --delete --obj_type status --crm_xml "<node_state id=\"1084752016\"/>"
cibadmin --delete --obj_type nodes --crm_xml "<node id=\"1084752016\"/>"

at this time; the result of "crm status" shows that node " h0 " is offline and 
there's no information of node h0 in CIB. But the "crm_node --list " still 
include node" h0"

3. crm_node -f -R 1084752016
After this cmd, node " h1" is removed, and result of "crm status " is as 
follow(with 0 node):

root@h1:/opt/bin# crm status
Last updated: Sat Apr 12 15:59:42 2014
Last change: Sat Apr 12 15:59:37 2014 via crm_node on h1
Stack: corosync
Current DC: NONE
0 Nodes configured
2 Resources configured
root@h1:/opt/bin


The corosync.conf is as follow:
.............
       quorum {
           provider: corosync_votequorum
           expected_votes: 2
           allow_downscale: 1
           two_node: 1
       }
...........


The corosync.log is as follow, it seems node " h1" is removed due to fencing!
Apr 12 16:22:35 [25003] h1 corosync debug   [QB    ] ringbuffer.c:299 Free'ing 
ringbuffer: /dev/shm/qb-cmap-request-25004-14152-27-header
Apr 12 16:22:35 [25026] h1        cib:    debug: activateCibXml:        
Triggering CIB write for cib_delete op
Apr 12 16:22:35 [25031] h1       crmd:   notice: crm_reap_dead_member:  
Removing h0/1084752016 from the membership list
Apr 12 16:22:35 [25031] h1       crmd:   notice: reap_crm_member:       Purged 
1 peers with id=1084752016 and/or uname=(null) from the membership cache
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: log_cib_diff:  Config update: 
Local-only Change: 0.12.1
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         -  <cib 
admin_epoch="0" epoch="11" num_updates="1">
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         -    
<configuration>
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         -      
<nodes>
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --      
 <node id="1084752017" uname="h1"/>
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         -      
</nodes>
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         -    
</configuration>
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         -  
</cib>
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         ++ <cib 
epoch="12" num_updates="1" admin_epoch="0" validate-with="pacemaker-1.2" 
crm_feature_set="3.0.8" cib-last-written="Sat Apr 12 16:22:35 2014" 
update-origin="h1" update-client="crm_node" have-quorum="1" 
dc-uuid="1084752017"/>
Apr 12 16:22:35 [25026] h1        cib:   notice: log_cib_diff:  cib:diff: 
Local-only Change: 0.12.1
Apr 12 16:22:35 [25026] h1        cib:   notice: cib:diff:      --       <node 
id="1084752017" uname="h1"/>
Apr 12 16:22:35 [25026] h1        cib:   notice: cib:diff:      ++ <cib 
epoch="12" num_updates="1" admin_epoch="0" validate-with="pacemaker-1.2" 
crm_feature_set="3.0.8" cib-last-written="Sat Apr 12 16:22:35 2014" 
update-origin="h1" update-client="crm_node" have-quorum="1" 
dc-uuid="1084752017"/>
Apr 12 16:22:35 [25026] h1        cib:     info: cib_process_request:   
Completed cib_delete operation for section nodes: OK (rc=0, 
origin=local/crm_node/2, version=0.12.1)
Apr 12 16:22:35 [25031] h1       crmd:    debug: te_update_diff:        
Processing diff (cib_delete): 0.11.1 -> 0.12.1 (S_IDLE)
Apr 12 16:22:35 [25031] h1       crmd:     info: abort_transition_graph:        
te_update_diff:126 - Triggered transition abort (complete=1, node=, tag=diff, 
id=(null), magic=NA, cib=0.12.1) : Non-status change
Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:        
Cause   <diff crm_feature_set="3.0.8" digest="3cccef06483ac4dfeadfb562f6f8293a">
Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:        
Cause     <diff-removed admin_epoch="0" epoch="11" num_updates="1">
Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:        
Cause       <cib admin_epoch="0" epoch="11" num_updates="1">
Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:        
Cause         <configuration>
Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:        
Cause           <nodes>
Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:        
Cause             <node id="1084752017" uname="h1" 
__crm_diff_marker__="removed:top"/>
Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:        
Cause           </nodes>
Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:        
Cause         </configuration>
Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:        
Cause       </cib>
Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:        
Cause     </diff-removed>
Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:        
Cause     <diff-added>
Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:        
Cause       <cib epoch="12" num_updates="1" admin_epoch="0" 
validate-with="pacemaker-1.2" crm_feature_set="3.0.8" cib-last-written="Sat Apr 
12 16:22:35 2014" update-origin="h1" update-client="crm_node" have-quorum="1" 
dc-uuid="1084752017"/>
Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:        
Cause     </diff-added>
Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:        
Cause   </diff>
Apr 12 16:22:35 [25031] h1       crmd:    debug: s_crmd_fsa:    Processing 
I_PE_CALC: [ state=S_IDLE cause=C_FSA_INTERNAL origin=abort_transition_graph ]
Apr 12 16:22:35 [25031] h1       crmd:   notice: do_state_transition:   State 
transition S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_FSA_INTERNAL 
origin=abort_transition_graph ]
Apr 12 16:22:35 [25031] h1       crmd:    debug: do_state_transition:   All 1 
cluster nodes are eligible to run resources.
Apr 12 16:22:35 [25031] h1       crmd:    debug: do_pe_invoke:  Query 38: 
Requesting the current CIB: S_POLICY_ENGINE
Apr 12 16:22:35 [25026] h1        cib:     info: cib_process_request:   
Completed cib_query operation for section 'all': OK (rc=0, 
origin=local/crmd/38, version=0.12.1)
Apr 12 16:22:35 [25031] h1       crmd:    debug: do_pe_invoke_callback:         
Invoking the PE: query=38, ref=pe_calc-dc-1397290955-21, seq=1024, quorate=1
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         Diff: 
--- 0.12.1
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         Diff: 
+++ 0.12.2 3d673c27c3c92939b41c7207edee9f46
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         -  <cib 
num_updates="1">
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         -    
<status>
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --     
<node_state id="1084752017" uname="h1" in_ccm="true" crmd="online" 
crm-debug-origin="post_cache_update" join="member" expected="member">
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --      
 <transient_attributes id="1084752017">
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --      
   <instance_attributes id="status-1084752017">
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --      
     <nvpair id="status-1084752017-shutdown" name="shutdown" value="0"/>
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --      
     <nvpair id="status-1084752017-probe_complete" name="probe_complete" 
value="true"/>
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --      
   </instance_attributes>
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --      
 </transient_attributes>
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --      
 <lrm id="1084752017">
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --      
   <lrm_resources>
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --      
     <lrm_resource id="VMdos-1" type="VirtualDomain" class="ocf" 
provider="heartbeat">
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --      
       <lrm_rsc_op id="VMdos-1_last_failure_0" 
operation_key="VMdos-1_monitor_0" operation="monitor" 
crm-debug-origin="build_active_RAs" crm_feature_set="3.0.8" 
transition-key="6:4:7:affc1dba-30eb-458d-8f86-c37d0268e52c" 
transition-magic="0:0;6:4:7:affc1dba-30eb-458d-8f86-c37d0268e52c" call-id="5" 
rc-code="0" op-status="0" interval="0" last-run="1397290607" 
last-rc-change="1397290607" exec-time="92" queue-time="0" op-digest="
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --      
       <lrm_rsc_op id="VMdos-1_monitor_30000" 
operation_key="VMdos-1_monitor_30000" operation="monitor" 
crm-debug-origin="build_active_RAs" crm_feature_set="3.0.8" 
transition-key="7:5:0:affc1dba-30eb-458d-8f86-c37d0268e52c" 
transition-magic="0:0;7:5:0:affc1dba-30eb-458d-8f86-c37d0268e52c" call-id="6" 
rc-code="0" op-status="0" interval="30000" last-rc-change="1397290607" 
exec-time="67" queue-time="0" op-digest="0874c7ce5f61a12
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --      
     </lrm_resource>
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --      
     <lrm_resource id="VMdos-2" type="VirtualDomain" class="ocf" 
provider="heartbeat">
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --      
       <lrm_rsc_op id="VMdos-2_last_0" operation_key="VMdos-2_monitor_0" 
operation="monitor" crm-debug-origin="build_active_RAs" crm_feature_set="3.0.8" 
transition-key="7:6:7:affc1dba-30eb-458d-8f86-c37d0268e52c" 
transition-magic="0:7;7:6:7:affc1dba-30eb-458d-8f86-c37d0268e52c" call-id="10" 
rc-code="7" op-status="0" interval="0" last-run="1397290608" 
last-rc-change="1397290608" exec-time="61" queue-time="0" op-digest="c7d22be
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --      
     </lrm_resource>
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --      
   </lrm_resources>
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --      
 </lrm>
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --     
</node_state>
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         -    
</status>
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         -  
</cib>
Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         ++ <cib 
epoch="12" num_updates="2" admin_epoch="0" validate-with="pacemaker-1.2" 
crm_feature_set="3.0.8" cib-last-written="Sat Apr 12 16:22:35 2014" 
update-origin="h1" update-client="crm_node" have-quorum="1" 
dc-uuid="1084752017"/>
Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_config:         STONITH 
timeout: 60000
Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_config:         STONITH 
of failed nodes is disabled
Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_config:         Stop 
all active resources: false
Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_config:         Cluster 
is symmetric - resources can run anywhere by default
Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_config:         Default 
stickiness: 0
Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_config:         On loss 
of CCM Quorum: Stop ALL resources
Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_config:         Node 
scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_domains:        
Unpacking domains
Apr 12 16:22:35 [25030] h1    pengine:  warning: unpack_status:         Node h1 
in status section no longer exists
Apr 12 16:22:35 [25030] h1    pengine:     info: unpack_status:         Node 
1084752017 is unknown
Apr 12 16:22:35 [25030] h1    pengine:     info: native_print:  VMdos-1 
(ocf::heartbeat:VirtualDomain): Stopped
Apr 12 16:22:35 [25030] h1    pengine:     info: native_print:  VMdos-2 
(ocf::heartbeat:VirtualDomain): Stopped
Apr 12 16:22:35 [25030] h1    pengine:    debug: native_assign_node:    Could 
not allocate a node for VMdos-1
Apr 12 16:22:35 [25030] h1    pengine:     info: native_color:  Resource 
VMdos-1 cannot run anywhere
Apr 12 16:22:35 [25030] h1    pengine:    debug: native_assign_node:    Could 
not allocate a node for VMdos-2
Apr 12 16:22:35 [25030] h1    pengine:     info: native_color:  Resource 
VMdos-2 cannot run anywhere
Apr 12 16:22:35 [25030] h1    pengine:     info: LogActions:    Leave   VMdos-1 
(Stopped)
Apr 12 16:22:35 [25030] h1    pengine:     info: LogActions:    Leave   VMdos-2 
(Stopped)
Apr 12 16:22:35 [25030] h1    pengine:   notice: process_pe_message:    
Calculated Transition 4: /var/lib/pacemaker/pengine/pe-input-4.bz2
Apr 12 16:22:35 [25031] h1       crmd:    debug: s_crmd_fsa:    Processing 
I_PE_SUCCESS: [ state=S_POLICY_ENGINE cause=C_IPC_MESSAGE 
origin=handle_response ]
Apr 12 16:22:35 [25031] h1       crmd:     info: do_state_transition:   State 
transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS 
cause=C_IPC_MESSAGE origin=handle_response ]
Apr 12 16:22:35 [25031] h1       crmd:    debug: unpack_graph:  Unpacked 
transition 4: 0 actions in 0 synapses
Apr 12 16:22:35 [25031] h1       crmd:     info: do_te_invoke:  Processing 
graph 4 (ref=pe_calc-dc-1397290955-21) derived from 
/var/lib/pacemaker/pengine/pe-input-4.bz2
Apr 12 16:22:35 [25031] h1       crmd:    debug: print_graph:   Empty 
transition graph
Apr 12 16:22:35 [25026] h1        cib:     info: cib_process_request:   
Completed cib_delete operation for section status: OK (rc=0, 
origin=local/crm_node/3, version=0.12.2)
Apr 12 16:22:35 [25031] h1       crmd:    debug: te_update_diff:        
Processing diff (cib_delete): 0.12.1 -> 0.12.2 (S_TRANSITION_ENGINE)
Apr 12 16:22:35 [25031] h1       crmd:     info: abort_transition_graph:        
te_update_diff:188 - Triggered transition abort (complete=0, node=h1, 
tag=transient_attributes, id=1084752017, magic=NA, cib=0.12.2) : Transient 
attribute: removal
Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:        
Cause   <transient_attributes id="1084752017">
Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:        
Cause     <instance_attributes id="status-1084752017">
Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:        
Cause       <nvpair id="status-1084752017-shutdown" name="shutdown" value="0"/>
Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:        
Cause       <nvpair id="status-1084752017-probe_complete" name="probe_complete" 
value="true"/>
Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:        
Cause     </instance_attributes>
Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:        
Cause   </transient_attributes>
Apr 12 16:22:35 [25031] h1       crmd:    debug: update_abort_priority:         
Abort priority upgraded from 0 to 1000000
Apr 12 16:22:35 [25031] h1       crmd:    debug: update_abort_priority:         
Abort action done superceeded by restart
Apr 12 16:22:35 [25031] h1       crmd:   notice: run_graph:     Transition 4 
(Complete=0, Pending=0, Fired=0, Skipped=0, Incomplete=0, 
Source=/var/lib/pacemaker/pengine/pe-input-4.bz2): Complete
Apr 12 16:22:35 [25031] h1       crmd:    debug: print_graph:   Empty 
transition graph
Apr 12 16:22:35 [25031] h1       crmd:    debug: te_graph_trigger:      
Transition 4 is now complete
Apr 12 16:22:35 [25031] h1       crmd:    debug: notify_crmd:   Processing 
transition completion in state S_TRANSITION_ENGINE
Apr 12 16:22:35 [25031] h1       crmd:    debug: notify_crmd:   Transition 4 
status: restart - Transient attribute: removal
Apr 12 16:22:35 [25031] h1       crmd:    debug: s_crmd_fsa:    Processing 
I_PE_CALC: [ state=S_TRANSITION_ENGINE cause=C_FSA_INTERNAL origin=notify_crmd ]
Apr 12 16:22:35 [25031] h1       crmd:     info: do_state_transition:   State 
transition S_TRANSITION_ENGINE -> S_POLICY_ENGINE [ input=I_PE_CALC 
cause=C_FSA_INTERNAL origin=notify_crmd ]
Apr 12 16:22:35 [25031] h1       crmd:    debug: do_state_transition:   All 1 
cluster nodes are eligible to run resources.
Apr 12 16:22:35 [25031] h1       crmd:    debug: do_pe_invoke:  Query 39: 
Requesting the current CIB: S_POLICY_ENGINE
Apr 12 16:22:35 [25024] h1 pacemakerd:     info: crm_client_new:        
Connecting 0x25b4ea0 for uid=0 gid=0 pid=14152 
id=f3612e17-0806-4355-a3fc-2cf1feda1e6d
Apr 12 16:22:35 [25024] h1 pacemakerd:    debug: handle_new_connection:         
IPC credentials authenticated (25024-14152-10)
Apr 12 16:22:35 [25024] h1 pacemakerd:    debug: qb_ipcs_shm_connect:   
connecting to client [14152]
Apr 12 16:22:35 [25026] h1        cib:     info: cib_process_request:   
Completed cib_query operation for section 'all': OK (rc=0, 
origin=local/crmd/39, version=0.12.2)
Apr 12 16:22:35 [25024] h1 pacemakerd:    debug: qb_rb_open_2:  shm 
size:131085; real_size:135168; rb->word_size:33792
Apr 12 16:22:35 [25026] h1        cib:    debug: 
qb_ipcs_dispatch_connection_request:   HUP conn (25026-14152-13)
Apr 12 16:22:35 [25026] h1        cib:    debug: qb_ipcs_disconnect:    
qb_ipcs_disconnect(25026-14152-13) state:2
Apr 12 16:22:35 [25026] h1        cib:     info: crm_client_destroy:    
Destroying 0 events
Apr 12 16:22:35 [25026] h1        cib:    debug: qb_rb_close:   Free'ing 
ringbuffer: /dev/shm/qb-cib_rw-response-25026-14152-13-header
Apr 12 16:22:35 [25026] h1        cib:    debug: qb_rb_close:   Free'ing 
ringbuffer: /dev/shm/qb-cib_rw-event-25026-14152-13-header
Apr 12 16:22:35 [25024] h1 pacemakerd:    debug: qb_rb_open_2:  shm 
size:131085; real_size:135168; rb->word_size:33792
Apr 12 16:22:35 [25024] h1 pacemakerd:    debug: qb_rb_open_2:  shm 
size:131085; real_size:135168; rb->word_size:33792
Apr 12 16:22:35 [25003] h1 corosync debug   [QB    ] ipc_setup.c:478 IPC 
credentials authenticated (25004-14152-27)
Apr 12 16:22:35 [25003] h1 corosync debug   [QB    ] ipc_shm.c:294 connecting 
to client [14152]
Apr 12 16:22:35 [25003] h1 corosync debug   [QB    ] ringbuffer.c:236 shm 
size:1048589; real_size:1052672; rb->word_size:263168
Apr 12 16:22:35 [25031] h1       crmd:    debug: do_pe_invoke_callback:         
Invoking the PE: query=39, ref=pe_calc-dc-1397290955-22, seq=1024, quorate=1
Apr 12 16:22:35 [25031] h1       crmd:    debug: 
qb_ipcs_dispatch_connection_request:   HUP conn (25031-14152-14)
Apr 12 16:22:35 [25031] h1       crmd:    debug: qb_ipcs_disconnect:    
qb_ipcs_disconnect(25031-14152-14) state:2
Apr 12 16:22:35 [25031] h1       crmd:     info: crm_client_destroy:    
Destroying 0 events
Apr 12 16:22:35 [25031] h1       crmd:    debug: qb_rb_close:   Free'ing 
ringbuffer: /dev/shm/qb-crmd-response-25031-14152-14-header
...
Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_config:         STONITH 
timeout: 60000
Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_config:         STONITH 
of failed nodes is disabled
Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_config:         Stop 
all active resources: false
Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_config:         Cluster 
is symmetric - resources can run anywhere by default
Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_config:         Default 
stickiness: 0
Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_config:         On loss 
of CCM Quorum: Stop ALL resources
Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_config:         Node 
scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_domains:        
Unpacking domains
Apr 12 16:22:35 [25030] h1    pengine:     info: native_print:  VMdos-1 
(ocf::heartbeat:VirtualDomain): Stopped
Apr 12 16:22:35 [25030] h1    pengine:     info: native_print:  VMdos-2 
(ocf::heartbeat:VirtualDomain): Stopped
Apr 12 16:22:35 [25030] h1    pengine:    debug: native_assign_node:    Could 
not allocate a node for VMdos-1
Apr 12 16:22:35 [25030] h1    pengine:     info: native_color:  Resource 
VMdos-1 cannot run anywhere
Apr 12 16:22:35 [25030] h1    pengine:    debug: native_assign_node:    Could 
not allocate a node for VMdos-2
Apr 12 16:22:35 [25030] h1    pengine:     info: native_color:  Resource 
VMdos-2 cannot run anywhere
Apr 12 16:22:35 [25030] h1    pengine:     info: LogActions:    Leave   VMdos-1 
(Stopped)
Apr 12 16:22:35 [25030] h1    pengine:     info: LogActions:    Leave   VMdos-2 
(Stopped)
Apr 12 16:22:35 [25030] h1    pengine:   notice: process_pe_message:    
Calculated Transition 5: /var/lib/pacemaker/pengine/pe-input-5.bz2
Apr 12 16:22:35 [25031] h1       crmd:    debug: s_crmd_fsa:    Processing 
I_PE_SUCCESS: [ state=S_POLICY_ENGINE cause=C_IPC_MESSAGE 
origin=handle_response ]
Apr 12 16:22:35 [25031] h1       crmd:     info: do_state_transition:   State 
transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS 
cause=C_IPC_MESSAGE origin=handle_response ]
Apr 12 16:22:35 [25031] h1       crmd:    debug: unpack_graph:  Unpacked 
transition 5: 0 actions in 0 synapses
Apr 12 16:22:35 [25031] h1       crmd:     info: do_te_invoke:  Processing 
graph 5 (ref=pe_calc-dc-1397290955-22) derived from 
/var/lib/pacemaker/pengine/pe-input-5.bz2
Apr 12 16:22:35 [25031] h1       crmd:    debug: print_graph:   Empty 
transition graph
Apr 12 16:22:35 [25031] h1       crmd:   notice: run_graph:     Transition 5 
(Complete=0, Pending=0, Fired=0, Skipped=0, Incomplete=0, 
Source=/var/lib/pacemaker/pengine/pe-input-5.bz2): Complete
Apr 12 16:22:35 [25031] h1       crmd:    debug: print_graph:   Empty 
transition graph
Apr 12 16:22:35 [25031] h1       crmd:    debug: te_graph_trigger:      
Transition 5 is now complete
Apr 12 16:22:35 [25031] h1       crmd:    debug: notify_crmd:   Processing 
transition completion in state S_TRANSITION_ENGINE
Apr 12 16:22:35 [25031] h1       crmd:    debug: notify_crmd:   Transition 5 
status: done - <null>
Apr 12 16:22:35 [25031] h1       crmd:    debug: s_crmd_fsa:    Processing 
I_TE_SUCCESS: [ state=S_TRANSITION_ENGINE cause=C_FSA_INTERNAL 
origin=notify_crmd ]
Apr 12 16:22:35 [25031] h1       crmd:     info: do_log:        FSA: Input 
I_TE_SUCCESS from notify_crmd() received in state S_TRANSITION_ENGINE
Apr 12 16:22:35 [25031] h1       crmd:   notice: do_state_transition:   State 
transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS 
cause=C_FSA_INTERNAL origin=notify_crmd ]
Apr 12 16:22:35 [25031] h1       crmd:    debug: do_state_transition:   
Starting PEngine Recheck Timer
Apr 12 16:22:35 [25031] h1       crmd:    debug: crm_timer_start:       Started 
PEngine Recheck Timer (I_PE_CALC:900000ms), src=62
Apr 12 16:22:35 [25003] h1 corosync debug   [QB    ] ringbuffer.c:236 shm 
size:1048589; real_size:1052672; rb->word_size:263168
Apr 12 16:22:35 [25026] h1        cib:     info: write_cib_contents:    
Archived previous version as /var/lib/pacemaker/cib/cib-9.raw
Apr 12 16:22:35 [25026] h1        cib:    debug: write_cib_contents:    Writing 
CIB to disk
Apr 12 16:22:35 [25026] h1        cib:    debug: qb_rb_close:   Free'ing 
ringbuffer: /dev/shm/qb-cib_rw-request-25026-14152-13-header
Apr 12 16:22:35 [25026] h1        cib:     info: write_cib_contents:    Wrote 
version 0.12.0 of the CIB to disk (digest: bd7d26226d6aa75f28b9eb670a67e944)
Apr 12 16:22:35 [25003] h1 corosync debug   [QB    ] ringbuffer.c:236 shm 
size:1048589; real_size:1052672; rb->word_size:263168
Apr 12 16:22:35 [25003] h1 corosync debug   [MAIN  ] ipc_glue.c:272 connection 
created
Apr 12 16:22:35 [25003] h1 corosync debug   [QB    ] cmap.c:306 lib_init_fn: 
conn=0x7f32d0f4f4b0
Apr 12 16:22:35 [25024] h1 pacemakerd:   notice: pcmk_ipc_dispatch:     
Instructing peers to remove references to node (null)/1084752016
Apr 12 16:22:35 [25026] h1        cib:     info: crm_client_new:        
Connecting 0xa23770 for uid=0 gid=0 pid=14152 
id=701dfa3c-d585-49d2-bd1e-f44636823e1b
Apr 12 16:22:35 [25026] h1        cib:    debug: handle_new_connection:         
IPC credentials authenticated (25026-14152-13)
Apr 12 16:22:35 [25026] h1        cib:    debug: qb_ipcs_shm_connect:   
connecting to client [14152]
Apr 12 16:22:35 [25026] h1        cib:    debug: write_cib_contents:    Wrote 
digest bd7d26226d6aa75f28b9eb670a67e944 to disk
Apr 12 16:22:35 [25026] h1        cib:     info: retrieveCib:   Reading cluster 
configuration from: /var/lib/pacemaker/cib/cib.y6wnAg (digest: 
/var/lib/pacemaker/cib/cib.ncXFvs)
...
Apr 12 16:22:35 [25026] h1        cib:    debug: write_cib_contents:    
Activating /var/lib/pacemaker/cib/cib.y6wnAg
Apr 12 16:22:35 [25026] h1        cib:     info: cib_process_request:   
Completed cib_delete operation for section nodes: OK (rc=0, 
origin=local/crm_node/2, version=0.12.2)
Apr 12 16:22:35 [25003] h1 corosync debug   [QB    ] ipcs.c:757 HUP conn 
(25004-14152-27)
Apr 12 16:22:35 [25003] h1 corosync debug   [QB    ] ipcs.c:605 
qb_ipcs_disconnect(25004-14152-27) state:2
Apr 12 16:22:35 [25003] h1 corosync debug   [QB    ] loop_poll_epoll.c:117 
epoll_ctl(del): Bad file descriptor (9)
Apr 12 16:22:35 [25003] h1 corosync debug   [MAIN  ] ipc_glue.c:417 
cs_ipcs_connection_closed()
Apr 12 16:22:35 [25003] h1 corosync debug   [QB    ] cmap.c:325 exit_fn for 
conn=0x7f32d0f4f4b0
Apr 12 16:22:35 [25003] h1 corosync debug   [MAIN  ] ipc_glue.c:390 
cs_ipcs_connection_destroyed()
Apr 12 16:22:35 [25026] h1        cib:     info: cib_process_request:   
Completed cib_delete operation for section status: OK (rc=0, 
origin=local/crm_node/3, version=0.12.2)
...
Apr 12 16:22:35 [25024] h1 pacemakerd:    debug: 
qb_ipcs_dispatch_connection_request:   HUP conn (25024-14152-10)
Apr 12 16:22:35 [25024] h1 pacemakerd:    debug: qb_ipcs_disconnect:    
qb_ipcs_disconnect(25024-14152-10) state:2
Apr 12 16:22:35 [25024] h1 pacemakerd:     info: crm_client_destroy:    
Destroying 0 events
...
Apr 12 16:22:35 [25026] h1        cib:    debug: 
qb_ipcs_dispatch_connection_request:   HUP conn (25026-14152-13)
Apr 12 16:22:35 [25026] h1        cib:    debug: qb_ipcs_disconnect:    
qb_ipcs_disconnect(25026-14152-13) state:2
Apr 12 16:22:35 [25026] h1        cib:     info: crm_client_destroy:    
Destroying 0 events
...
Apr 12 16:22:35 [25024] h1 pacemakerd:   notice: crm_reap_dead_member:  
Removing h0/1084752016 from the membership list
Apr 12 16:22:35 [25024] h1 pacemakerd:   notice: reap_crm_member:       Purged 
1 peers with id=1084752016 and/or uname=(null) from the membership cache


-------------------------------------------------------------------------------------------------------------------------------------
本邮件及其附件含有杭州华三通信技术有限公司的保密信息,仅限于发送给上面地址中列出
的个人或群组。禁止任何其他人以任何形式使用(包括但不限于全部或部分地泄露、复制、
或散发)本邮件中的信息。如果您错收了本邮件,请您立即电话或邮件通知发件人并删除本
邮件!
This e-mail and its attachments contain confidential information from H3C, 
which is
intended only for the person or entity whose address is listed above. Any use 
of the
information contained herein in any way (including, but not limited to, total 
or partial
disclosure, reproduction, or dissemination) by persons other than the intended
recipient(s) is prohibited. If you receive this e-mail in error, please notify 
the sender
by phone or email immediately and delete it!
_______________________________________________
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org

Reply via email to