On 13 Apr 2014, at 11:36 am, Caizhifeng <bluewin...@h3c.com> wrote:

> Hi all,
> I'm kinda new to Pacemaker and have some questions I hope some of you could 
> help me with, any idea(s) would be highly appreciated, thank you.
> I'am building the HA cluster with corosync2.3.3 + cluster-glue + 
> pacemaker-1.1.11-rc5. The problem is, when I try to remove a node from HA 
> cluster with "crm_node -f -R nodeID", it seems to remove the wrong node.
> 
> My setup is as follows(and it can be reproduced):
> 1. The HA is ok with 2 resource and 2 nodes, and with stonith-enabled="false":
> root@h1:/opt/bin# crm status
> Last updated: Sat Apr 12 16:00:51 2014
> Last change: Sat Apr 12 16:00:43 2014 via cibadmin on h1
> Stack: corosync
> Current DC: h1 (1084752017) - partition with quorum
> Version: 1.1.10-33f9d09
> 2 Nodes configured
> 2 Resources configured
> 
> 
> Online: [ h0 h1 ]
> 
> VMdos-1        (ocf::heartbeat:VirtualDomain): Started h1
> root@h1:/opt/bin#
> 
> root@h1:/opt/bin# crm_node --list
> 1084752017 h1
> 1084752016 h0
> root@h1:/opt/bin#
> 
> 2. remove node h0 from HA
> (1). Stop pacemakerd and corosync service on node h0 (2). Run the follow cmd 
> on node h1 to Remove node info from cib:
> cibadmin --delete --obj_type status --crm_xml "<node_state 
> id=\"1084752016\"/>"
> cibadmin --delete --obj_type nodes --crm_xml "<node id=\"1084752016\"/>"
> 
> at this time; the result of "crm status" shows that node " h0 " is offline 
> and there's no information of node h0 in CIB. But the "crm_node --list " 
> still include node" h0"
> 
> 3. crm_node -f -R 1084752016

Could you re-run this command but add -VVVVVVVV please?
I suspect the fault lies in this binary.

> After this cmd, node " h1" is removed, and result of "crm status " is as 
> follow(with 0 node):
> 
> root@h1:/opt/bin# crm status
> Last updated: Sat Apr 12 15:59:42 2014
> Last change: Sat Apr 12 15:59:37 2014 via crm_node on h1
> Stack: corosync
> Current DC: NONE
> 0 Nodes configured
> 2 Resources configured
> root@h1:/opt/bin
> 
> 
> The corosync.conf is as follow:
> .............
>       quorum {
>           provider: corosync_votequorum
>           expected_votes: 2
>           allow_downscale: 1
>           two_node: 1
>       }
> ...........
> 
> 
> The corosync.log is as follow, it seems node " h1" is removed due to fencing!
> Apr 12 16:22:35 [25003] h1 corosync debug   [QB    ] ringbuffer.c:299 
> Free'ing ringbuffer: /dev/shm/qb-cmap-request-25004-14152-27-header
> Apr 12 16:22:35 [25026] h1        cib:    debug: activateCibXml:        
> Triggering CIB write for cib_delete op
> Apr 12 16:22:35 [25031] h1       crmd:   notice: crm_reap_dead_member:  
> Removing h0/1084752016 from the membership list
> Apr 12 16:22:35 [25031] h1       crmd:   notice: reap_crm_member:       
> Purged 1 peers with id=1084752016 and/or uname=(null) from the membership 
> cache
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: log_cib_diff:  Config 
> update: Local-only Change: 0.12.1
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         -  
> <cib admin_epoch="0" epoch="11" num_updates="1">
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         -    
> <configuration>
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         -     
>  <nodes>
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --    
>    <node id="1084752017" uname="h1"/>
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         -     
>  </nodes>
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         -    
> </configuration>
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         -  
> </cib>
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         ++ 
> <cib epoch="12" num_updates="1" admin_epoch="0" validate-with="pacemaker-1.2" 
> crm_feature_set="3.0.8" cib-last-written="Sat Apr 12 16:22:35 2014" 
> update-origin="h1" update-client="crm_node" have-quorum="1" 
> dc-uuid="1084752017"/>
> Apr 12 16:22:35 [25026] h1        cib:   notice: log_cib_diff:  cib:diff: 
> Local-only Change: 0.12.1
> Apr 12 16:22:35 [25026] h1        cib:   notice: cib:diff:      --       
> <node id="1084752017" uname="h1"/>
> Apr 12 16:22:35 [25026] h1        cib:   notice: cib:diff:      ++ <cib 
> epoch="12" num_updates="1" admin_epoch="0" validate-with="pacemaker-1.2" 
> crm_feature_set="3.0.8" cib-last-written="Sat Apr 12 16:22:35 2014" 
> update-origin="h1" update-client="crm_node" have-quorum="1" 
> dc-uuid="1084752017"/>
> Apr 12 16:22:35 [25026] h1        cib:     info: cib_process_request:   
> Completed cib_delete operation for section nodes: OK (rc=0, 
> origin=local/crm_node/2, version=0.12.1)
> Apr 12 16:22:35 [25031] h1       crmd:    debug: te_update_diff:        
> Processing diff (cib_delete): 0.11.1 -> 0.12.1 (S_IDLE)
> Apr 12 16:22:35 [25031] h1       crmd:     info: abort_transition_graph:      
>   te_update_diff:126 - Triggered transition abort (complete=1, node=, 
> tag=diff, id=(null), magic=NA, cib=0.12.1) : Non-status change
> Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:      
>   Cause   <diff crm_feature_set="3.0.8" 
> digest="3cccef06483ac4dfeadfb562f6f8293a">
> Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:      
>   Cause     <diff-removed admin_epoch="0" epoch="11" num_updates="1">
> Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:      
>   Cause       <cib admin_epoch="0" epoch="11" num_updates="1">
> Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:      
>   Cause         <configuration>
> Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:      
>   Cause           <nodes>
> Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:      
>   Cause             <node id="1084752017" uname="h1" 
> __crm_diff_marker__="removed:top"/>
> Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:      
>   Cause           </nodes>
> Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:      
>   Cause         </configuration>
> Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:      
>   Cause       </cib>
> Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:      
>   Cause     </diff-removed>
> Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:      
>   Cause     <diff-added>
> Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:      
>   Cause       <cib epoch="12" num_updates="1" admin_epoch="0" 
> validate-with="pacemaker-1.2" crm_feature_set="3.0.8" cib-last-written="Sat 
> Apr 12 16:22:35 2014" update-origin="h1" update-client="crm_node" 
> have-quorum="1" dc-uuid="1084752017"/>
> Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:      
>   Cause     </diff-added>
> Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:      
>   Cause   </diff>
> Apr 12 16:22:35 [25031] h1       crmd:    debug: s_crmd_fsa:    Processing 
> I_PE_CALC: [ state=S_IDLE cause=C_FSA_INTERNAL origin=abort_transition_graph ]
> Apr 12 16:22:35 [25031] h1       crmd:   notice: do_state_transition:   State 
> transition S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_FSA_INTERNAL 
> origin=abort_transition_graph ]
> Apr 12 16:22:35 [25031] h1       crmd:    debug: do_state_transition:   All 1 
> cluster nodes are eligible to run resources.
> Apr 12 16:22:35 [25031] h1       crmd:    debug: do_pe_invoke:  Query 38: 
> Requesting the current CIB: S_POLICY_ENGINE
> Apr 12 16:22:35 [25026] h1        cib:     info: cib_process_request:   
> Completed cib_query operation for section 'all': OK (rc=0, 
> origin=local/crmd/38, version=0.12.1)
> Apr 12 16:22:35 [25031] h1       crmd:    debug: do_pe_invoke_callback:       
>   Invoking the PE: query=38, ref=pe_calc-dc-1397290955-21, seq=1024, quorate=1
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         Diff: 
> --- 0.12.1
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         Diff: 
> +++ 0.12.2 3d673c27c3c92939b41c7207edee9f46
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         -  
> <cib num_updates="1">
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         -    
> <status>
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --    
>  <node_state id="1084752017" uname="h1" in_ccm="true" crmd="online" 
> crm-debug-origin="post_cache_update" join="member" expected="member">
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --    
>    <transient_attributes id="1084752017">
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --    
>      <instance_attributes id="status-1084752017">
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --    
>        <nvpair id="status-1084752017-shutdown" name="shutdown" value="0"/>
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --    
>        <nvpair id="status-1084752017-probe_complete" name="probe_complete" 
> value="true"/>
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --    
>      </instance_attributes>
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --    
>    </transient_attributes>
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --    
>    <lrm id="1084752017">
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --    
>      <lrm_resources>
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --    
>        <lrm_resource id="VMdos-1" type="VirtualDomain" class="ocf" 
> provider="heartbeat">
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --    
>          <lrm_rsc_op id="VMdos-1_last_failure_0" 
> operation_key="VMdos-1_monitor_0" operation="monitor" 
> crm-debug-origin="build_active_RAs" crm_feature_set="3.0.8" 
> transition-key="6:4:7:affc1dba-30eb-458d-8f86-c37d0268e52c" 
> transition-magic="0:0;6:4:7:affc1dba-30eb-458d-8f86-c37d0268e52c" call-id="5" 
> rc-code="0" op-status="0" interval="0" last-run="1397290607" 
> last-rc-change="1397290607" exec-time="92" queue-time="0" op-digest="
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --    
>          <lrm_rsc_op id="VMdos-1_monitor_30000" 
> operation_key="VMdos-1_monitor_30000" operation="monitor" 
> crm-debug-origin="build_active_RAs" crm_feature_set="3.0.8" 
> transition-key="7:5:0:affc1dba-30eb-458d-8f86-c37d0268e52c" 
> transition-magic="0:0;7:5:0:affc1dba-30eb-458d-8f86-c37d0268e52c" call-id="6" 
> rc-code="0" op-status="0" interval="30000" last-rc-change="1397290607" 
> exec-time="67" queue-time="0" op-digest="0874c7ce5f61a12
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --    
>        </lrm_resource>
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --    
>        <lrm_resource id="VMdos-2" type="VirtualDomain" class="ocf" 
> provider="heartbeat">
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --    
>          <lrm_rsc_op id="VMdos-2_last_0" operation_key="VMdos-2_monitor_0" 
> operation="monitor" crm-debug-origin="build_active_RAs" 
> crm_feature_set="3.0.8" 
> transition-key="7:6:7:affc1dba-30eb-458d-8f86-c37d0268e52c" 
> transition-magic="0:7;7:6:7:affc1dba-30eb-458d-8f86-c37d0268e52c" 
> call-id="10" rc-code="7" op-status="0" interval="0" last-run="1397290608" 
> last-rc-change="1397290608" exec-time="61" queue-time="0" op-digest="c7d22be
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --    
>        </lrm_resource>
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --    
>      </lrm_resources>
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --    
>    </lrm>
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         --    
>  </node_state>
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         -    
> </status>
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         -  
> </cib>
> Apr 12 16:22:35 [25027] h1 stonith-ng:    debug: Config update:         ++ 
> <cib epoch="12" num_updates="2" admin_epoch="0" validate-with="pacemaker-1.2" 
> crm_feature_set="3.0.8" cib-last-written="Sat Apr 12 16:22:35 2014" 
> update-origin="h1" update-client="crm_node" have-quorum="1" 
> dc-uuid="1084752017"/>
> Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_config:         
> STONITH timeout: 60000
> Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_config:         
> STONITH of failed nodes is disabled
> Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_config:         Stop 
> all active resources: false
> Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_config:         
> Cluster is symmetric - resources can run anywhere by default
> Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_config:         
> Default stickiness: 0
> Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_config:         On 
> loss of CCM Quorum: Stop ALL resources
> Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_config:         Node 
> scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
> Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_domains:        
> Unpacking domains
> Apr 12 16:22:35 [25030] h1    pengine:  warning: unpack_status:         Node 
> h1 in status section no longer exists
> Apr 12 16:22:35 [25030] h1    pengine:     info: unpack_status:         Node 
> 1084752017 is unknown
> Apr 12 16:22:35 [25030] h1    pengine:     info: native_print:  VMdos-1 
> (ocf::heartbeat:VirtualDomain): Stopped
> Apr 12 16:22:35 [25030] h1    pengine:     info: native_print:  VMdos-2 
> (ocf::heartbeat:VirtualDomain): Stopped
> Apr 12 16:22:35 [25030] h1    pengine:    debug: native_assign_node:    Could 
> not allocate a node for VMdos-1
> Apr 12 16:22:35 [25030] h1    pengine:     info: native_color:  Resource 
> VMdos-1 cannot run anywhere
> Apr 12 16:22:35 [25030] h1    pengine:    debug: native_assign_node:    Could 
> not allocate a node for VMdos-2
> Apr 12 16:22:35 [25030] h1    pengine:     info: native_color:  Resource 
> VMdos-2 cannot run anywhere
> Apr 12 16:22:35 [25030] h1    pengine:     info: LogActions:    Leave   
> VMdos-1 (Stopped)
> Apr 12 16:22:35 [25030] h1    pengine:     info: LogActions:    Leave   
> VMdos-2 (Stopped)
> Apr 12 16:22:35 [25030] h1    pengine:   notice: process_pe_message:    
> Calculated Transition 4: /var/lib/pacemaker/pengine/pe-input-4.bz2
> Apr 12 16:22:35 [25031] h1       crmd:    debug: s_crmd_fsa:    Processing 
> I_PE_SUCCESS: [ state=S_POLICY_ENGINE cause=C_IPC_MESSAGE 
> origin=handle_response ]
> Apr 12 16:22:35 [25031] h1       crmd:     info: do_state_transition:   State 
> transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS 
> cause=C_IPC_MESSAGE origin=handle_response ]
> Apr 12 16:22:35 [25031] h1       crmd:    debug: unpack_graph:  Unpacked 
> transition 4: 0 actions in 0 synapses
> Apr 12 16:22:35 [25031] h1       crmd:     info: do_te_invoke:  Processing 
> graph 4 (ref=pe_calc-dc-1397290955-21) derived from 
> /var/lib/pacemaker/pengine/pe-input-4.bz2
> Apr 12 16:22:35 [25031] h1       crmd:    debug: print_graph:   Empty 
> transition graph
> Apr 12 16:22:35 [25026] h1        cib:     info: cib_process_request:   
> Completed cib_delete operation for section status: OK (rc=0, 
> origin=local/crm_node/3, version=0.12.2)
> Apr 12 16:22:35 [25031] h1       crmd:    debug: te_update_diff:        
> Processing diff (cib_delete): 0.12.1 -> 0.12.2 (S_TRANSITION_ENGINE)
> Apr 12 16:22:35 [25031] h1       crmd:     info: abort_transition_graph:      
>   te_update_diff:188 - Triggered transition abort (complete=0, node=h1, 
> tag=transient_attributes, id=1084752017, magic=NA, cib=0.12.2) : Transient 
> attribute: removal
> Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:      
>   Cause   <transient_attributes id="1084752017">
> Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:      
>   Cause     <instance_attributes id="status-1084752017">
> Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:      
>   Cause       <nvpair id="status-1084752017-shutdown" name="shutdown" 
> value="0"/>
> Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:      
>   Cause       <nvpair id="status-1084752017-probe_complete" 
> name="probe_complete" value="true"/>
> Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:      
>   Cause     </instance_attributes>
> Apr 12 16:22:35 [25031] h1       crmd:    debug: abort_transition_graph:      
>   Cause   </transient_attributes>
> Apr 12 16:22:35 [25031] h1       crmd:    debug: update_abort_priority:       
>   Abort priority upgraded from 0 to 1000000
> Apr 12 16:22:35 [25031] h1       crmd:    debug: update_abort_priority:       
>   Abort action done superceeded by restart
> Apr 12 16:22:35 [25031] h1       crmd:   notice: run_graph:     Transition 4 
> (Complete=0, Pending=0, Fired=0, Skipped=0, Incomplete=0, 
> Source=/var/lib/pacemaker/pengine/pe-input-4.bz2): Complete
> Apr 12 16:22:35 [25031] h1       crmd:    debug: print_graph:   Empty 
> transition graph
> Apr 12 16:22:35 [25031] h1       crmd:    debug: te_graph_trigger:      
> Transition 4 is now complete
> Apr 12 16:22:35 [25031] h1       crmd:    debug: notify_crmd:   Processing 
> transition completion in state S_TRANSITION_ENGINE
> Apr 12 16:22:35 [25031] h1       crmd:    debug: notify_crmd:   Transition 4 
> status: restart - Transient attribute: removal
> Apr 12 16:22:35 [25031] h1       crmd:    debug: s_crmd_fsa:    Processing 
> I_PE_CALC: [ state=S_TRANSITION_ENGINE cause=C_FSA_INTERNAL 
> origin=notify_crmd ]
> Apr 12 16:22:35 [25031] h1       crmd:     info: do_state_transition:   State 
> transition S_TRANSITION_ENGINE -> S_POLICY_ENGINE [ input=I_PE_CALC 
> cause=C_FSA_INTERNAL origin=notify_crmd ]
> Apr 12 16:22:35 [25031] h1       crmd:    debug: do_state_transition:   All 1 
> cluster nodes are eligible to run resources.
> Apr 12 16:22:35 [25031] h1       crmd:    debug: do_pe_invoke:  Query 39: 
> Requesting the current CIB: S_POLICY_ENGINE
> Apr 12 16:22:35 [25024] h1 pacemakerd:     info: crm_client_new:        
> Connecting 0x25b4ea0 for uid=0 gid=0 pid=14152 
> id=f3612e17-0806-4355-a3fc-2cf1feda1e6d
> Apr 12 16:22:35 [25024] h1 pacemakerd:    debug: handle_new_connection:       
>   IPC credentials authenticated (25024-14152-10)
> Apr 12 16:22:35 [25024] h1 pacemakerd:    debug: qb_ipcs_shm_connect:   
> connecting to client [14152]
> Apr 12 16:22:35 [25026] h1        cib:     info: cib_process_request:   
> Completed cib_query operation for section 'all': OK (rc=0, 
> origin=local/crmd/39, version=0.12.2)
> Apr 12 16:22:35 [25024] h1 pacemakerd:    debug: qb_rb_open_2:  shm 
> size:131085; real_size:135168; rb->word_size:33792
> Apr 12 16:22:35 [25026] h1        cib:    debug: 
> qb_ipcs_dispatch_connection_request:   HUP conn (25026-14152-13)
> Apr 12 16:22:35 [25026] h1        cib:    debug: qb_ipcs_disconnect:    
> qb_ipcs_disconnect(25026-14152-13) state:2
> Apr 12 16:22:35 [25026] h1        cib:     info: crm_client_destroy:    
> Destroying 0 events
> Apr 12 16:22:35 [25026] h1        cib:    debug: qb_rb_close:   Free'ing 
> ringbuffer: /dev/shm/qb-cib_rw-response-25026-14152-13-header
> Apr 12 16:22:35 [25026] h1        cib:    debug: qb_rb_close:   Free'ing 
> ringbuffer: /dev/shm/qb-cib_rw-event-25026-14152-13-header
> Apr 12 16:22:35 [25024] h1 pacemakerd:    debug: qb_rb_open_2:  shm 
> size:131085; real_size:135168; rb->word_size:33792
> Apr 12 16:22:35 [25024] h1 pacemakerd:    debug: qb_rb_open_2:  shm 
> size:131085; real_size:135168; rb->word_size:33792
> Apr 12 16:22:35 [25003] h1 corosync debug   [QB    ] ipc_setup.c:478 IPC 
> credentials authenticated (25004-14152-27)
> Apr 12 16:22:35 [25003] h1 corosync debug   [QB    ] ipc_shm.c:294 connecting 
> to client [14152]
> Apr 12 16:22:35 [25003] h1 corosync debug   [QB    ] ringbuffer.c:236 shm 
> size:1048589; real_size:1052672; rb->word_size:263168
> Apr 12 16:22:35 [25031] h1       crmd:    debug: do_pe_invoke_callback:       
>   Invoking the PE: query=39, ref=pe_calc-dc-1397290955-22, seq=1024, quorate=1
> Apr 12 16:22:35 [25031] h1       crmd:    debug: 
> qb_ipcs_dispatch_connection_request:   HUP conn (25031-14152-14)
> Apr 12 16:22:35 [25031] h1       crmd:    debug: qb_ipcs_disconnect:    
> qb_ipcs_disconnect(25031-14152-14) state:2
> Apr 12 16:22:35 [25031] h1       crmd:     info: crm_client_destroy:    
> Destroying 0 events
> Apr 12 16:22:35 [25031] h1       crmd:    debug: qb_rb_close:   Free'ing 
> ringbuffer: /dev/shm/qb-crmd-response-25031-14152-14-header
> ...
> Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_config:         
> STONITH timeout: 60000
> Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_config:         
> STONITH of failed nodes is disabled
> Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_config:         Stop 
> all active resources: false
> Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_config:         
> Cluster is symmetric - resources can run anywhere by default
> Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_config:         
> Default stickiness: 0
> Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_config:         On 
> loss of CCM Quorum: Stop ALL resources
> Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_config:         Node 
> scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
> Apr 12 16:22:35 [25030] h1    pengine:    debug: unpack_domains:        
> Unpacking domains
> Apr 12 16:22:35 [25030] h1    pengine:     info: native_print:  VMdos-1 
> (ocf::heartbeat:VirtualDomain): Stopped
> Apr 12 16:22:35 [25030] h1    pengine:     info: native_print:  VMdos-2 
> (ocf::heartbeat:VirtualDomain): Stopped
> Apr 12 16:22:35 [25030] h1    pengine:    debug: native_assign_node:    Could 
> not allocate a node for VMdos-1
> Apr 12 16:22:35 [25030] h1    pengine:     info: native_color:  Resource 
> VMdos-1 cannot run anywhere
> Apr 12 16:22:35 [25030] h1    pengine:    debug: native_assign_node:    Could 
> not allocate a node for VMdos-2
> Apr 12 16:22:35 [25030] h1    pengine:     info: native_color:  Resource 
> VMdos-2 cannot run anywhere
> Apr 12 16:22:35 [25030] h1    pengine:     info: LogActions:    Leave   
> VMdos-1 (Stopped)
> Apr 12 16:22:35 [25030] h1    pengine:     info: LogActions:    Leave   
> VMdos-2 (Stopped)
> Apr 12 16:22:35 [25030] h1    pengine:   notice: process_pe_message:    
> Calculated Transition 5: /var/lib/pacemaker/pengine/pe-input-5.bz2
> Apr 12 16:22:35 [25031] h1       crmd:    debug: s_crmd_fsa:    Processing 
> I_PE_SUCCESS: [ state=S_POLICY_ENGINE cause=C_IPC_MESSAGE 
> origin=handle_response ]
> Apr 12 16:22:35 [25031] h1       crmd:     info: do_state_transition:   State 
> transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS 
> cause=C_IPC_MESSAGE origin=handle_response ]
> Apr 12 16:22:35 [25031] h1       crmd:    debug: unpack_graph:  Unpacked 
> transition 5: 0 actions in 0 synapses
> Apr 12 16:22:35 [25031] h1       crmd:     info: do_te_invoke:  Processing 
> graph 5 (ref=pe_calc-dc-1397290955-22) derived from 
> /var/lib/pacemaker/pengine/pe-input-5.bz2
> Apr 12 16:22:35 [25031] h1       crmd:    debug: print_graph:   Empty 
> transition graph
> Apr 12 16:22:35 [25031] h1       crmd:   notice: run_graph:     Transition 5 
> (Complete=0, Pending=0, Fired=0, Skipped=0, Incomplete=0, 
> Source=/var/lib/pacemaker/pengine/pe-input-5.bz2): Complete
> Apr 12 16:22:35 [25031] h1       crmd:    debug: print_graph:   Empty 
> transition graph
> Apr 12 16:22:35 [25031] h1       crmd:    debug: te_graph_trigger:      
> Transition 5 is now complete
> Apr 12 16:22:35 [25031] h1       crmd:    debug: notify_crmd:   Processing 
> transition completion in state S_TRANSITION_ENGINE
> Apr 12 16:22:35 [25031] h1       crmd:    debug: notify_crmd:   Transition 5 
> status: done - <null>
> Apr 12 16:22:35 [25031] h1       crmd:    debug: s_crmd_fsa:    Processing 
> I_TE_SUCCESS: [ state=S_TRANSITION_ENGINE cause=C_FSA_INTERNAL 
> origin=notify_crmd ]
> Apr 12 16:22:35 [25031] h1       crmd:     info: do_log:        FSA: Input 
> I_TE_SUCCESS from notify_crmd() received in state S_TRANSITION_ENGINE
> Apr 12 16:22:35 [25031] h1       crmd:   notice: do_state_transition:   State 
> transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS 
> cause=C_FSA_INTERNAL origin=notify_crmd ]
> Apr 12 16:22:35 [25031] h1       crmd:    debug: do_state_transition:   
> Starting PEngine Recheck Timer
> Apr 12 16:22:35 [25031] h1       crmd:    debug: crm_timer_start:       
> Started PEngine Recheck Timer (I_PE_CALC:900000ms), src=62
> Apr 12 16:22:35 [25003] h1 corosync debug   [QB    ] ringbuffer.c:236 shm 
> size:1048589; real_size:1052672; rb->word_size:263168
> Apr 12 16:22:35 [25026] h1        cib:     info: write_cib_contents:    
> Archived previous version as /var/lib/pacemaker/cib/cib-9.raw
> Apr 12 16:22:35 [25026] h1        cib:    debug: write_cib_contents:    
> Writing CIB to disk
> Apr 12 16:22:35 [25026] h1        cib:    debug: qb_rb_close:   Free'ing 
> ringbuffer: /dev/shm/qb-cib_rw-request-25026-14152-13-header
> Apr 12 16:22:35 [25026] h1        cib:     info: write_cib_contents:    Wrote 
> version 0.12.0 of the CIB to disk (digest: bd7d26226d6aa75f28b9eb670a67e944)
> Apr 12 16:22:35 [25003] h1 corosync debug   [QB    ] ringbuffer.c:236 shm 
> size:1048589; real_size:1052672; rb->word_size:263168
> Apr 12 16:22:35 [25003] h1 corosync debug   [MAIN  ] ipc_glue.c:272 
> connection created
> Apr 12 16:22:35 [25003] h1 corosync debug   [QB    ] cmap.c:306 lib_init_fn: 
> conn=0x7f32d0f4f4b0
> Apr 12 16:22:35 [25024] h1 pacemakerd:   notice: pcmk_ipc_dispatch:     
> Instructing peers to remove references to node (null)/1084752016
> Apr 12 16:22:35 [25026] h1        cib:     info: crm_client_new:        
> Connecting 0xa23770 for uid=0 gid=0 pid=14152 
> id=701dfa3c-d585-49d2-bd1e-f44636823e1b
> Apr 12 16:22:35 [25026] h1        cib:    debug: handle_new_connection:       
>   IPC credentials authenticated (25026-14152-13)
> Apr 12 16:22:35 [25026] h1        cib:    debug: qb_ipcs_shm_connect:   
> connecting to client [14152]
> Apr 12 16:22:35 [25026] h1        cib:    debug: write_cib_contents:    Wrote 
> digest bd7d26226d6aa75f28b9eb670a67e944 to disk
> Apr 12 16:22:35 [25026] h1        cib:     info: retrieveCib:   Reading 
> cluster configuration from: /var/lib/pacemaker/cib/cib.y6wnAg (digest: 
> /var/lib/pacemaker/cib/cib.ncXFvs)
> ...
> Apr 12 16:22:35 [25026] h1        cib:    debug: write_cib_contents:    
> Activating /var/lib/pacemaker/cib/cib.y6wnAg
> Apr 12 16:22:35 [25026] h1        cib:     info: cib_process_request:   
> Completed cib_delete operation for section nodes: OK (rc=0, 
> origin=local/crm_node/2, version=0.12.2)
> Apr 12 16:22:35 [25003] h1 corosync debug   [QB    ] ipcs.c:757 HUP conn 
> (25004-14152-27)
> Apr 12 16:22:35 [25003] h1 corosync debug   [QB    ] ipcs.c:605 
> qb_ipcs_disconnect(25004-14152-27) state:2
> Apr 12 16:22:35 [25003] h1 corosync debug   [QB    ] loop_poll_epoll.c:117 
> epoll_ctl(del): Bad file descriptor (9)
> Apr 12 16:22:35 [25003] h1 corosync debug   [MAIN  ] ipc_glue.c:417 
> cs_ipcs_connection_closed()
> Apr 12 16:22:35 [25003] h1 corosync debug   [QB    ] cmap.c:325 exit_fn for 
> conn=0x7f32d0f4f4b0
> Apr 12 16:22:35 [25003] h1 corosync debug   [MAIN  ] ipc_glue.c:390 
> cs_ipcs_connection_destroyed()
> Apr 12 16:22:35 [25026] h1        cib:     info: cib_process_request:   
> Completed cib_delete operation for section status: OK (rc=0, 
> origin=local/crm_node/3, version=0.12.2)
> ...
> Apr 12 16:22:35 [25024] h1 pacemakerd:    debug: 
> qb_ipcs_dispatch_connection_request:   HUP conn (25024-14152-10)
> Apr 12 16:22:35 [25024] h1 pacemakerd:    debug: qb_ipcs_disconnect:    
> qb_ipcs_disconnect(25024-14152-10) state:2
> Apr 12 16:22:35 [25024] h1 pacemakerd:     info: crm_client_destroy:    
> Destroying 0 events
> ...
> Apr 12 16:22:35 [25026] h1        cib:    debug: 
> qb_ipcs_dispatch_connection_request:   HUP conn (25026-14152-13)
> Apr 12 16:22:35 [25026] h1        cib:    debug: qb_ipcs_disconnect:    
> qb_ipcs_disconnect(25026-14152-13) state:2
> Apr 12 16:22:35 [25026] h1        cib:     info: crm_client_destroy:    
> Destroying 0 events
> ...
> Apr 12 16:22:35 [25024] h1 pacemakerd:   notice: crm_reap_dead_member:  
> Removing h0/1084752016 from the membership list
> Apr 12 16:22:35 [25024] h1 pacemakerd:   notice: reap_crm_member:       
> Purged 1 peers with id=1084752016 and/or uname=(null) from the membership 
> cache
> 
> 
> -------------------------------------------------------------------------------------------------------------------------------------
> 本邮件及其附件含有杭州华三通信技术有限公司的保密信息,仅限于发送给上面地址中列出
> 的个人或群组。禁止任何其他人以任何形式使用(包括但不限于全部或部分地泄露、复制、
> 或散发)本邮件中的信息。如果您错收了本邮件,请您立即电话或邮件通知发件人并删除本
> 邮件!
> This e-mail and its attachments contain confidential information from H3C, 
> which is
> intended only for the person or entity whose address is listed above. Any use 
> of the
> information contained herein in any way (including, but not limited to, total 
> or partial
> disclosure, reproduction, or dissemination) by persons other than the intended
> recipient(s) is prohibited. If you receive this e-mail in error, please 
> notify the sender
> by phone or email immediately and delete it!
> _______________________________________________
> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
> 
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org

Attachment: signature.asc
Description: Message signed with OpenPGP using GPGMail

_______________________________________________
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org

Reply via email to