Hm, I might have misinterpreted something here. That's an Etch 64bit
system, and Heartbeat comes from the backports repository.
But you are right, it also says version 2.1.3:

aptitude show -v heartbeat
Package: heartbeat
State: installed
Automatically installed: no
Version: 2.1.3-6~bpo40+2
Priority: optional
Section: admin
Maintainer: Simon Horman <[email protected]>
Filename: pool/main/h/heartbeat/heartbeat_2.1.3-6~bpo40+2_amd64.deb
Archive: etch-backports, now
Provided by: heartbeat-2 (2.0.7-2)

It might also be due to the constant updates in these backport
repositories: perhaps my working installation is different from what's
right now in the repository.
Hm, right:
crmadmin --version
HA Version 2.1.3, CRM Version 2.0 (CIB feature set 2.0)

 Thanks,
Thomas

Andrew Beekhof wrote:
> On Tue, Jan 13, 2009 at 15:46, Thomas Roth <[email protected]> wrote:
>> Hi all,
>>
>> we are using external/ipmi on a drbd8 / heartbeat-2.0.7-2 as stonith
> 
> According to the cib you sent through, you actually have something
> based on 2.1.3
> However I can't find the changeset it lists
> (552305612591183b1628baa5bc6e903e0f1e26a3) anywhere and some of the PE
> logs look significantly newer than 2.1.3.
> 
> Where did you get the heartbeat package from?  It helps to know what
> I'm looking at before commenting.
> 
>> device. While the stonith itself works well, there are some errors which
>> recur and eventually lead to unwanted reboots. And mayby there are no
>> real causes but just my lack of understanding...
>> Anyhow, in the ha-debug log of server1 I see
>>> tengine[4459]: 2009/01/13_03:40:38 info: process_graph_event: Detected
>> action server1-fencing_monitor_120000 from a different transition: 29 vs. 31
>>> tengine[4459]: 2009/01/13_03:40:38 info: update_abort_priority: Abort
>> priority upgraded to 1000000
>>> tengine[4459]: 2009/01/13_03:40:38 WARN: update_failcount: Updating
>> failcount for server1-fencing on ebffe771-505c-4e40-b0b3-d70903ed37bc
>> after failed monitor: rc=14
>>
>> Then pengine goes on about that problem:
>>> pengine[4460]: 2009/01/13_03:40:38 info: determine_online_status: Node
>> server2 is online
>>> pengine[4460]: 2009/01/13_03:40:38 ERROR: unpack_rsc_op: Remapping
>> server1-fencing_monitor_120000 (rc=14) on server2 to an ERROR
>>> pengine[4460]: 2009/01/13_03:40:38 WARN: unpack_rsc_op: Processing
>> failed op server1-fencing_monitor_120000 on server2: Error
>>> pengine[4460]: 2009/01/13_03:40:38 info: determine_online_status: Node
>> server1 is online
>>> pengine[4460]: 2009/01/13_03:40:38 ERROR: unpack_rsc_op: Remapping
>> server2-fencing_start_0 (rc=1) on server1 to an ERROR
>>> pengine[4460]: 2009/01/13_03:40:38 WARN: unpack_rsc_op: Processing
>> failed op server2-fencing_start_0 on server1: Error
>>> pengine[4460]: 2009/01/13_03:40:38 WARN: unpack_rsc_op: Compatability
>> handling for failed op server2-fencing_start_0 on server1
>>
>> This was already the second instance of this problem last night, the
>> first occured 5 hours earlier, having no further consequences.
>> Now, however, pengine after a walk through all the resources  says
>>> pengine[4460]: 2009/01/13_03:40:38 WARN: stage6: Scheduling Node
>> server2 for STONITH
>>
>> There are some repetitions of this, until finally stonithd gives up:
>>> stonithd[3621]: 2009/01/13_03:42:19 ERROR: Failed to STONITH the node
>> server2: optype=RESET, op_result=TIMEOUT
>>
>>
>> The other machine, server2,  did not show any problems whatsoever that
>> should have triggered the action.
>> The intended victim this time was the slave. However, during the last
>> weekend both machines showed very strange behavior, among them rebooting
>> each other ( with a days break, there was no stonith-war going on, but
>> still...)
>>
>>
>> Now in the cib.xml, we have the following two primitives
>>
>>       <primitive class="stonith" type="external/ipmi"
>> provider="heartbeat" id="server1-fencing">
>>         <operations>
>>           <op id="server1-fencing-monitor" name="monitor"
>> interval="120s" timeout="70s" prereq="nothing" start_delay="0"
>> disabled="false" role="Started" on_fail="fence"/>
>>           <op id="server1-fencing-start" name="start" timeout="40s"
>> prereq="nothing" start_delay="0" disabled="false" role="Started"/>
>>         </operations>
>>         <instance_attributes id="server1-fencing-ia">
>>           <attributes>
>>             <nvpair id="server1-fencing-hostname" name="hostname"
>> value="server1"/>
>>             <nvpair id="server1-fencing-ipaddr" name="ipaddr"
>> value="1.1.1.1"/>
>>             <nvpair id="server1-fencing-userid" name="userid"
>> value="USER"/>
>>             <nvpair id="server1-fencing-passwd" name="passwd"
>> value="PASSWD"/>
>>           </attributes>
>>         </instance_attributes>
>>       </primitive>
>>       <primitive id="server2-fencing" class="stonith"
>> type="external/ipmi" provider="heartbeat">
>>         <operations>
>>           <op id="server2-fencing-monitor" name="monitor"
>> interval="120s" timeout="70s" prereq="nothing"/>
>>           <op id="server2-fencing-start" name="start" timeout="40s"
>> prereq="nothing"/>
>>         </operations>
>>         <instance_attributes id="server2-fencing-ia">
>>           <attributes>
>>             <nvpair id="server2-fencing-hostname" name="hostname"
>> value="server2"/>
>>             <nvpair id="server2-fencing-ipaddr" name="ipaddr"
>> value="1.1.1.2"/>
>>             <nvpair id="server2-fencing-userid" name="userid"
>> value="USER"/>
>>             <nvpair id="server2-fencing-passwd" name="passwd"
>> value="PASSWD"/>
>>           </attributes>
>>         </instance_attributes>
>>       </primitive>
>>
>>
>> The intent of this is to reboot one machine in case of a failover which
>> cannot be completed because failing heartbeat cannot release one of its
>> resources (Lustre mounts of drbd disks).
>>
>> So my question is of course whether there is something fundamentally
>> wrong with this configuration (I'm attaching the entire cib.xml and the
>> ha-debug for last night, btw). Or whether there should really be an
>> error somewhere else (hardware).
>>
>> Thanks a lot,
>> Thomas
>>
>>
>>
>> tengine[4459]: 2009/01/12_22:34:28 info: process_graph_event: Action 
>> server2-fencing_monitor_120000 arrived after a completed transition
>> tengine[4459]: 2009/01/12_22:34:28 info: update_abort_priority: Abort 
>> priority upgraded to 1000000
>> tengine[4459]: 2009/01/12_22:34:28 WARN: update_failcount: Updating 
>> failcount for server2-fencing on ebffe771-505c-4e40-b0b3-d70903ed37bc after 
>> failed monitor: rc=14
>> crmd[3623]: 2009/01/12_22:34:28 info: do_state_transition: State transition 
>> S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_IPC_MESSAGE 
>> origin=route_message ]
>> crmd[3623]: 2009/01/12_22:34:28 info: do_state_transition: All 2 cluster 
>> nodes are eligible to run resources.
>> tengine[4459]: 2009/01/12_22:34:28 info: extract_event: Aborting on 
>> transient_attributes changes for ebffe771-505c-4e40-b0b3-d70903ed37bc
>> pengine[4460]: 2009/01/12_22:34:28 info: determine_online_status: Node 
>> server2 is online
>> pengine[4460]: 2009/01/12_22:34:28 ERROR: unpack_rsc_op: Remapping 
>> server2-fencing_monitor_120000 (rc=14) on server2 to an ERROR
>> pengine[4460]: 2009/01/12_22:34:28 WARN: unpack_rsc_op: Processing failed op 
>> server2-fencing_monitor_120000 on server2: Error
>> pengine[4460]: 2009/01/12_22:34:28 info: determine_online_status: Node 
>> server1 is online
>> pengine[4460]: 2009/01/12_22:34:28 ERROR: unpack_rsc_op: Remapping 
>> server2-fencing_start_0 (rc=1) on server1 to an ERROR
>> pengine[4460]: 2009/01/12_22:34:28 WARN: unpack_rsc_op: Processing failed op 
>> server2-fencing_start_0 on server1: Error
>> pengine[4460]: 2009/01/12_22:34:28 WARN: unpack_rsc_op: Compatability 
>> handling for failed op server2-fencing_start_0 on server1
>> pengine[4460]: 2009/01/12_22:34:28 notice: group_print: Resource Group: 
>> group_1
>> pengine[4460]: 2009/01/12_22:34:28 notice: native_print:     commonIP   
>> (heartbeat::ocf:IPaddr):        Started server1
>> pengine[4460]: 2009/01/12_22:34:28 notice: native_print:     drbddisk_2 
>> (heartbeat:drbddisk):   Started server1
>> pengine[4460]: 2009/01/12_22:34:28 notice: native_print:     Filesystem_3    
>>    (heartbeat::ocf:Filesystem):    Started server1
>> pengine[4460]: 2009/01/12_22:34:28 notice: native_print:     drbddisk_4 
>> (heartbeat:drbddisk):   Started server1
>> pengine[4460]: 2009/01/12_22:34:28 notice: native_print:     Filesystem_5    
>>    (heartbeat::ocf:Filesystem):    Started server1
>> pengine[4460]: 2009/01/12_22:34:28 notice: native_print:     MailTo_6   
>> (heartbeat::ocf:MailTo):        Started server1
>> pengine[4460]: 2009/01/12_22:34:28 notice: native_print: server1-fencing     
>>    (stonith:external/ipmi):        Started server2
>> pengine[4460]: 2009/01/12_22:34:28 notice: native_print: server2-fencing     
>>    (stonith:external/ipmi):        Started server2 FAILED
>> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Leave resource 
>> commonIP        (server1)
>> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Leave resource 
>> drbddisk_2      (server1)
>> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Leave resource 
>> Filesystem_3    (server1)
>> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Leave resource 
>> drbddisk_4      (server1)
>> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Leave resource 
>> Filesystem_5    (server1)
>> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Leave resource 
>> MailTo_6        (server1)
>> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Leave resource 
>> server1-fencing (server2)
>> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Recover resource 
>> server2-fencing       (server2)
>> pengine[4460]: 2009/01/12_22:34:28 notice: StopRsc:   server2   Stop 
>> server2-fencing
>> pengine[4460]: 2009/01/12_22:34:28 notice: StartRsc:  server2   Start 
>> server2-fencing
>> pengine[4460]: 2009/01/12_22:34:28 notice: RecurringOp: server2    
>> server2-fencing_monitor_120000
>> pengine[4460]: 2009/01/12_22:34:28 info: process_pe_message: Transition 30: 
>> PEngine Input stored in: /var/lib/heartbeat/pengine/pe-input-406.bz2
>> pengine[4460]: 2009/01/12_22:34:28 info: determine_online_status: Node 
>> server2 is online
>> pengine[4460]: 2009/01/12_22:34:28 ERROR: unpack_rsc_op: Remapping 
>> server2-fencing_monitor_120000 (rc=14) on server2 to an ERROR
>> pengine[4460]: 2009/01/12_22:34:28 WARN: unpack_rsc_op: Processing failed op 
>> server2-fencing_monitor_120000 on server2: Error
>> pengine[4460]: 2009/01/12_22:34:28 info: determine_online_status: Node 
>> server1 is online
>> pengine[4460]: 2009/01/12_22:34:28 ERROR: unpack_rsc_op: Remapping 
>> server2-fencing_start_0 (rc=1) on server1 to an ERROR
>> pengine[4460]: 2009/01/12_22:34:28 WARN: unpack_rsc_op: Processing failed op 
>> server2-fencing_start_0 on server1: Error
>> pengine[4460]: 2009/01/12_22:34:28 WARN: unpack_rsc_op: Compatability 
>> handling for failed op server2-fencing_start_0 on server1
>> pengine[4460]: 2009/01/12_22:34:28 notice: group_print: Resource Group: 
>> group_1
>> pengine[4460]: 2009/01/12_22:34:28 notice: native_print:     commonIP   
>> (heartbeat::ocf:IPaddr):        Started server1
>> pengine[4460]: 2009/01/12_22:34:28 notice: native_print:     drbddisk_2 
>> (heartbeat:drbddisk):   Started server1
>> pengine[4460]: 2009/01/12_22:34:28 notice: native_print:     Filesystem_3    
>>    (heartbeat::ocf:Filesystem):    Started server1
>> pengine[4460]: 2009/01/12_22:34:28 notice: native_print:     drbddisk_4 
>> (heartbeat:drbddisk):   Started server1
>> pengine[4460]: 2009/01/12_22:34:28 notice: native_print:     Filesystem_5    
>>    (heartbeat::ocf:Filesystem):    Started server1
>> pengine[4460]: 2009/01/12_22:34:28 notice: native_print:     MailTo_6   
>> (heartbeat::ocf:MailTo):        Started server1
>> pengine[4460]: 2009/01/12_22:34:28 notice: native_print: server1-fencing     
>>    (stonith:external/ipmi):        Started server2
>> pengine[4460]: 2009/01/12_22:34:28 notice: native_print: server2-fencing     
>>    (stonith:external/ipmi):        Started server2 FAILED
>> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Leave resource 
>> commonIP        (server1)
>> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Leave resource 
>> drbddisk_2      (server1)
>> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Leave resource 
>> Filesystem_3    (server1)
>> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Leave resource 
>> drbddisk_4      (server1)
>> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Leave resource 
>> Filesystem_5    (server1)
>> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Leave resource 
>> MailTo_6        (server1)
>> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Leave resource 
>> server1-fencing (server2)
>> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Recover resource 
>> server2-fencing       (server2)
>> pengine[4460]: 2009/01/12_22:34:28 notice: StopRsc:   server2   Stop 
>> server2-fencing
>> pengine[4460]: 2009/01/12_22:34:28 notice: StartRsc:  server2   Start 
>> server2-fencing
>> pengine[4460]: 2009/01/12_22:34:28 notice: RecurringOp: server2    
>> server2-fencing_monitor_120000
>> crmd[3623]: 2009/01/12_22:34:28 info: do_state_transition: State transition 
>> S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS 
>> cause=C_IPC_MESSAGE origin=route_message ]
>> tengine[4459]: 2009/01/12_22:34:28 info: unpack_graph: Unpacked transition 
>> 31: 4 actions in 4 synapses
>> tengine[4459]: 2009/01/12_22:34:28 info: te_pseudo_action: Pseudo action 13 
>> fired and confirmed
>> tengine[4459]: 2009/01/12_22:34:28 info: send_rsc_command: Initiating action 
>> 3: server2-fencing_stop_0 on server2
>> pengine[4460]: 2009/01/12_22:34:28 info: process_pe_message: Transition 31: 
>> PEngine Input stored in: /var/lib/heartbeat/pengine/pe-input-407.bz2
>> tengine[4459]: 2009/01/12_22:34:29 info: match_graph_event: Action 
>> server2-fencing_stop_0 (3) confirmed on server2 (rc=0)
>> tengine[4459]: 2009/01/12_22:34:29 info: send_rsc_command: Initiating action 
>> 32: server2-fencing_start_0 on server2
>> tengine[4459]: 2009/01/12_22:34:31 info: match_graph_event: Action 
>> server2-fencing_start_0 (32) confirmed on server2 (rc=0)
>> tengine[4459]: 2009/01/12_22:34:31 info: send_rsc_command: Initiating action 
>> 2: server2-fencing_monitor_120000 on server2
>> tengine[4459]: 2009/01/12_22:34:33 info: match_graph_event: Action 
>> server2-fencing_monitor_120000 (2) confirmed on server2 (rc=0)
>> tengine[4459]: 2009/01/12_22:34:33 info: run_graph: Transition 31: 
>> (Complete=4, Pending=0, Fired=0, Skipped=0, Incomplete=0)
>> tengine[4459]: 2009/01/12_22:34:33 info: notify_crmd: Transition 31 status: 
>> te_complete - <null>
>> crmd[3623]: 2009/01/12_22:34:33 info: do_state_transition: State transition 
>> S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS cause=C_IPC_MESSAGE 
>> origin=route_message ]
>> cib[3619]: 2009/01/12_22:42:19 info: cib_stats: Processed 49 operations 
>> (4285.00us average, 0% utilization) in the last 10min
>> cib[3619]: 2009/01/12_22:52:19 info: cib_stats: Processed 39 operations 
>> (4102.00us average, 0% utilization) in the last 10min
>> cib[3619]: 2009/01/12_23:02:19 info: cib_stats: Processed 40 operations 
>> (6000.00us average, 0% utilization) in the last 10min
>> lrmd[3620]: 2009/01/12_23:08:37 WARN: G_SIG_dispatch: Dispatch function for 
>> SIGCHLD was delayed 1000 ms (> 100 ms) before being called (GSource: 
>> 0x514c58)
>> lrmd[3620]: 2009/01/12_23:08:37 info: G_SIG_dispatch: started at 1731484574 
>> should have started at 1731484474
>> cib[3619]: 2009/01/12_23:12:19 info: cib_stats: Processed 40 operations 
>> (4000.00us average, 0% utilization) in the last 10min
>> cib[3619]: 2009/01/12_23:22:19 info: cib_stats: Processed 40 operations 
>> (5000.00us average, 0% utilization) in the last 10min
>> cib[3619]: 2009/01/12_23:32:19 info: cib_stats: Processed 39 operations 
>> (7435.00us average, 0% utilization) in the last 10min
>> cib[3619]: 2009/01/12_23:42:19 info: cib_stats: Processed 40 operations 
>> (5250.00us average, 0% utilization) in the last 10min
>> cib[3619]: 2009/01/12_23:52:19 info: cib_stats: Processed 40 operations 
>> (6500.00us average, 0% utilization) in the last 10min
>> cib[3619]: 2009/01/13_00:02:19 info: cib_stats: Processed 39 operations 
>> (5641.00us average, 0% utilization) in the last 10min
>> cib[3619]: 2009/01/13_00:12:19 info: cib_stats: Processed 40 operations 
>> (6000.00us average, 0% utilization) in the last 10min
>> cib[3619]: 2009/01/13_00:22:19 info: cib_stats: Processed 40 operations 
>> (4750.00us average, 0% utilization) in the last 10min
>> cib[3619]: 2009/01/13_00:32:19 info: cib_stats: Processed 40 operations 
>> (4500.00us average, 0% utilization) in the last 10min
>> cib[3619]: 2009/01/13_00:42:19 info: cib_stats: Processed 39 operations 
>> (5128.00us average, 0% utilization) in the last 10min
>> cib[3619]: 2009/01/13_00:52:19 info: cib_stats: Processed 40 operations 
>> (4000.00us average, 0% utilization) in the last 10min
>> cib[3619]: 2009/01/13_01:02:19 info: cib_stats: Processed 40 operations 
>> (5500.00us average, 0% utilization) in the last 10min
>> cib[3619]: 2009/01/13_01:12:19 info: cib_stats: Processed 40 operations 
>> (4500.00us average, 0% utilization) in the last 10min
>> cib[3619]: 2009/01/13_01:22:19 info: cib_stats: Processed 39 operations 
>> (4358.00us average, 0% utilization) in the last 10min
>> cib[3619]: 2009/01/13_01:32:19 info: cib_stats: Processed 40 operations 
>> (4750.00us average, 0% utilization) in the last 10min
>> cib[3619]: 2009/01/13_01:42:19 info: cib_stats: Processed 40 operations 
>> (7000.00us average, 0% utilization) in the last 10min
>> cib[3619]: 2009/01/13_01:52:19 info: cib_stats: Processed 40 operations 
>> (5500.00us average, 0% utilization) in the last 10min
>> cib[3619]: 2009/01/13_02:02:19 info: cib_stats: Processed 39 operations 
>> (5384.00us average, 0% utilization) in the last 10min
>> cib[3619]: 2009/01/13_02:12:19 info: cib_stats: Processed 40 operations 
>> (5000.00us average, 0% utilization) in the last 10min
>> cib[3619]: 2009/01/13_02:22:19 info: cib_stats: Processed 40 operations 
>> (6750.00us average, 0% utilization) in the last 10min
>> cib[3619]: 2009/01/13_02:32:19 info: cib_stats: Processed 39 operations 
>> (5897.00us average, 0% utilization) in the last 10min
>> cib[3619]: 2009/01/13_02:42:19 info: cib_stats: Processed 40 operations 
>> (3750.00us average, 0% utilization) in the last 10min
>> cib[3619]: 2009/01/13_02:52:19 info: cib_stats: Processed 40 operations 
>> (5000.00us average, 0% utilization) in the last 10min
>> cib[3619]: 2009/01/13_03:02:19 info: cib_stats: Processed 40 operations 
>> (5500.00us average, 0% utilization) in the last 10min
>> cib[3619]: 2009/01/13_03:12:19 info: cib_stats: Processed 39 operations 
>> (6153.00us average, 0% utilization) in the last 10min
>> cib[3619]: 2009/01/13_03:22:19 info: cib_stats: Processed 40 operations 
>> (4750.00us average, 0% utilization) in the last 10min
>> cib[3619]: 2009/01/13_03:32:19 info: cib_stats: Processed 40 operations 
>> (4750.00us average, 0% utilization) in the last 10min
>> tengine[4459]: 2009/01/13_03:40:38 info: process_graph_event: Detected 
>> action server1-fencing_monitor_120000 from a different transition: 29 vs. 31
>> tengine[4459]: 2009/01/13_03:40:38 info: update_abort_priority: Abort 
>> priority upgraded to 1000000
>> tengine[4459]: 2009/01/13_03:40:38 WARN: update_failcount: Updating 
>> failcount for server1-fencing on ebffe771-505c-4e40-b0b3-d70903ed37bc after 
>> failed monitor: rc=14
>> crmd[3623]: 2009/01/13_03:40:38 info: do_state_transition: State transition 
>> S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_IPC_MESSAGE 
>> origin=route_message ]
>> crmd[3623]: 2009/01/13_03:40:38 info: do_state_transition: All 2 cluster 
>> nodes are eligible to run resources.
>> pengine[4460]: 2009/01/13_03:40:38 info: determine_online_status: Node 
>> server2 is online
>> pengine[4460]: 2009/01/13_03:40:38 ERROR: unpack_rsc_op: Remapping 
>> server1-fencing_monitor_120000 (rc=14) on server2 to an ERROR
>> pengine[4460]: 2009/01/13_03:40:38 WARN: unpack_rsc_op: Processing failed op 
>> server1-fencing_monitor_120000 on server2: Error
>> pengine[4460]: 2009/01/13_03:40:38 info: determine_online_status: Node 
>> server1 is online
>> pengine[4460]: 2009/01/13_03:40:38 ERROR: unpack_rsc_op: Remapping 
>> server2-fencing_start_0 (rc=1) on server1 to an ERROR
>> pengine[4460]: 2009/01/13_03:40:38 WARN: unpack_rsc_op: Processing failed op 
>> server2-fencing_start_0 on server1: Error
>> pengine[4460]: 2009/01/13_03:40:38 WARN: unpack_rsc_op: Compatability 
>> handling for failed op server2-fencing_start_0 on server1
>> pengine[4460]: 2009/01/13_03:40:38 notice: group_print: Resource Group: 
>> group_1
>> pengine[4460]: 2009/01/13_03:40:38 notice: native_print:     commonIP   
>> (heartbeat::ocf:IPaddr):        Started server1
>> pengine[4460]: 2009/01/13_03:40:38 notice: native_print:     drbddisk_2 
>> (heartbeat:drbddisk):   Started server1
>> pengine[4460]: 2009/01/13_03:40:38 notice: native_print:     Filesystem_3    
>>    (heartbeat::ocf:Filesystem):    Started server1
>> pengine[4460]: 2009/01/13_03:40:38 notice: native_print:     drbddisk_4 
>> (heartbeat:drbddisk):   Started server1
>> pengine[4460]: 2009/01/13_03:40:38 notice: native_print:     Filesystem_5    
>>    (heartbeat::ocf:Filesystem):    Started server1
>> pengine[4460]: 2009/01/13_03:40:38 notice: native_print:     MailTo_6   
>> (heartbeat::ocf:MailTo):        Started server1
>> pengine[4460]: 2009/01/13_03:40:38 notice: native_print: server1-fencing     
>>    (stonith:external/ipmi):        Started server2 FAILED
>> pengine[4460]: 2009/01/13_03:40:38 notice: native_print: server2-fencing     
>>    (stonith:external/ipmi):        Started server2
>> pengine[4460]: 2009/01/13_03:40:38 notice: NoRoleChange: Leave resource 
>> commonIP        (server1)
>> pengine[4460]: 2009/01/13_03:40:38 notice: NoRoleChange: Leave resource 
>> drbddisk_2      (server1)
>> pengine[4460]: 2009/01/13_03:40:38 notice: NoRoleChange: Leave resource 
>> Filesystem_3    (server1)
>> pengine[4460]: 2009/01/13_03:40:38 notice: NoRoleChange: Leave resource 
>> drbddisk_4      (server1)
>> pengine[4460]: 2009/01/13_03:40:38 notice: NoRoleChange: Leave resource 
>> Filesystem_5    (server1)
>> pengine[4460]: 2009/01/13_03:40:38 notice: NoRoleChange: Leave resource 
>> MailTo_6        (server1)
>> pengine[4460]: 2009/01/13_03:40:38 notice: NoRoleChange: Recover resource 
>> server1-fencing       (server1)
>> pengine[4460]: 2009/01/13_03:40:38 notice: StopRsc:   server2   Stop 
>> server1-fencing
>> pengine[4460]: 2009/01/13_03:40:38 notice: StartRsc:  server1   Start 
>> server1-fencing
>> pengine[4460]: 2009/01/13_03:40:38 notice: RecurringOp: server1    
>> server1-fencing_monitor_120000
>> pengine[4460]: 2009/01/13_03:40:38 WARN: native_color: Resource 
>> server2-fencing cannot run anywhere
>> pengine[4460]: 2009/01/13_03:40:38 notice: StopRsc:   server2   Stop 
>> server2-fencing
>> pengine[4460]: 2009/01/13_03:40:38 WARN: stage6: Scheduling Node server2 for 
>> STONITH
>> pengine[4460]: 2009/01/13_03:40:38 WARN: native_stop_constraints: Stop of 
>> failed resource server1-fencing is implicit after server2 is fenced
>> pengine[4460]: 2009/01/13_03:40:38 info: native_stop_constraints: 
>> server2-fencing_stop_0 is implicit after server2 is fenced
>> crmd[3623]: 2009/01/13_03:40:38 info: do_state_transition: State transition 
>> S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS 
>> cause=C_IPC_MESSAGE origin=route_message ]
>> tengine[4459]: 2009/01/13_03:40:38 info: unpack_graph: Unpacked transition 
>> 32: 7 actions in 7 synapses
>> tengine[4459]: 2009/01/13_03:40:38 info: te_pseudo_action: Pseudo action 2 
>> fired and confirmed
>> tengine[4459]: 2009/01/13_03:40:38 info: send_rsc_command: Initiating action 
>> 29: server1-fencing_start_0 on server1
>> tengine[4459]: 2009/01/13_03:40:38 info: te_pseudo_action: Pseudo action 31 
>> fired and confirmed
>> crmd[3623]: 2009/01/13_03:40:38 info: do_lrm_rsc_op: Performing 
>> op=server1-fencing_start_0 key=29:32:8686f9af-9ced-43ab-bf20-be6e8437abc0)
>> lrmd[3620]: 2009/01/13_03:40:38 info: rsc:server1-fencing: start
>> lrmd[27785]: 2009/01/13_03:40:38 info: Try to start STONITH resource 
>> <rsc_id=server1-fencing> : Device=external/ipmi
>> pengine[4460]: 2009/01/13_03:40:38 WARN: process_pe_message: Transition 32: 
>> WARNINGs found during PE processing. PEngine Input stored in: 
>> /var/lib/heartbeat/pengine/pe-warn-6.bz2
>> pengine[4460]: 2009/01/13_03:40:38 info: process_pe_message: Configuration 
>> WARNINGs found during PE processing.  Please run "crm_verify -L" to identify 
>> issues.
>> crmd[3623]: 2009/01/13_03:40:39 info: process_lrm_event: LRM operation 
>> server1-fencing_start_0 (call=51, rc=0) complete
>> tengine[4459]: 2009/01/13_03:40:39 info: match_graph_event: Action 
>> server1-fencing_start_0 (29) confirmed on server1 (rc=0)
>> tengine[4459]: 2009/01/13_03:40:39 info: send_rsc_command: Initiating action 
>> 30: server1-fencing_monitor_120000 on server1
>> tengine[4459]: 2009/01/13_03:40:39 info: te_pseudo_action: Pseudo action 32 
>> fired and confirmed
>> tengine[4459]: 2009/01/13_03:40:39 info: te_fence_node: Executing reboot 
>> fencing operation (33) on server2 (timeout=100000)
>> crmd[3623]: 2009/01/13_03:40:39 info: do_lrm_rsc_op: Performing 
>> op=server1-fencing_monitor_120000 
>> key=30:32:8686f9af-9ced-43ab-bf20-be6e8437abc0)
>> stonithd[3621]: 2009/01/13_03:40:39 info: client tengine [pid: 4459] want a 
>> STONITH operation RESET to node server2.
>> stonithd[3621]: 2009/01/13_03:40:39 info: Broadcasting the message 
>> succeeded: require others to stonith node server2.
>> crmd[3623]: 2009/01/13_03:40:39 info: process_lrm_event: LRM operation 
>> server1-fencing_monitor_120000 (call=52, rc=0) complete
>> tengine[4459]: 2009/01/13_03:40:39 info: match_graph_event: Action 
>> server1-fencing_monitor_120000 (30) confirmed on server1 (rc=0)
>> stonithd[3621]: 2009/01/13_03:42:19 ERROR: Failed to STONITH the node 
>> server2: optype=RESET, op_result=TIMEOUT
>> tengine[4459]: 2009/01/13_03:42:19 info: tengine_stonith_callback: call=-2, 
>> optype=1, node_name=server2, result=2, node_list=, 
>> action=33:32:8686f9af-9ced-43ab-bf20-be6e8437abc0
>> tengine[4459]: 2009/01/13_03:42:19 ERROR: tengine_stonith_callback: Stonith 
>> of server2 failed (2)... aborting transition.
>> tengine[4459]: 2009/01/13_03:42:19 info: update_abort_priority: Abort 
>> priority upgraded to 1000000
>> tengine[4459]: 2009/01/13_03:42:19 info: update_abort_priority: Abort action 
>> 0 superceeded by 2
>> tengine[4459]: 2009/01/13_03:42:19 info: run_graph: 
>> ====================================================
>> tengine[4459]: 2009/01/13_03:42:19 notice: run_graph: Transition 32: 
>> (Complete=6, Pending=0, Fired=0, Skipped=1, Incomplete=0)
>> crmd[3623]: 2009/01/13_03:42:19 info: do_state_transition: State transition 
>> S_TRANSITION_ENGINE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_IPC_MESSAGE 
>> origin=route_message ]
>> crmd[3623]: 2009/01/13_03:42:19 info: do_state_transition: All 2 cluster 
>> nodes are eligible to run resources.
>> pengine[4460]: 2009/01/13_03:42:19 info: determine_online_status: Node 
>> server2 is online
>> pengine[4460]: 2009/01/13_03:42:19 ERROR: unpack_rsc_op: Remapping 
>> server1-fencing_monitor_120000 (rc=14) on server2 to an ERROR
>> pengine[4460]: 2009/01/13_03:42:19 WARN: unpack_rsc_op: Processing failed op 
>> server1-fencing_monitor_120000 on server2: Error
>> pengine[4460]: 2009/01/13_03:42:19 info: determine_online_status: Node 
>> server1 is online
>> pengine[4460]: 2009/01/13_03:42:19 ERROR: native_add_running: Resource 
>> stonith::external/ipmi:server1-fencing appears to be active on 2 nodes.
>> pengine[4460]: 2009/01/13_03:42:19 ERROR: See 
>> http://linux-ha.org/v2/faq/resource_too_active for more information.
>> pengine[4460]: 2009/01/13_03:42:19 ERROR: unpack_rsc_op: Remapping 
>> server2-fencing_start_0 (rc=1) on server1 to an ERROR
>> pengine[4460]: 2009/01/13_03:42:19 WARN: unpack_rsc_op: Processing failed op 
>> server2-fencing_start_0 on server1: Error
>> pengine[4460]: 2009/01/13_03:42:19 WARN: unpack_rsc_op: Compatability 
>> handling for failed op server2-fencing_start_0 on server1
>> pengine[4460]: 2009/01/13_03:42:19 notice: group_print: Resource Group: 
>> group_1
>> pengine[4460]: 2009/01/13_03:42:19 notice: native_print:     commonIP   
>> (heartbeat::ocf:IPaddr):        Started server1
>> pengine[4460]: 2009/01/13_03:42:19 notice: native_print:     drbddisk_2 
>> (heartbeat:drbddisk):   Started server1
>> pengine[4460]: 2009/01/13_03:42:19 notice: native_print:     Filesystem_3    
>>    (heartbeat::ocf:Filesystem):    Started server1
>> pengine[4460]: 2009/01/13_03:42:19 notice: native_print:     drbddisk_4 
>> (heartbeat:drbddisk):   Started server1
>> pengine[4460]: 2009/01/13_03:42:19 notice: native_print:     Filesystem_5    
>>    (heartbeat::ocf:Filesystem):    Started server1
>> pengine[4460]: 2009/01/13_03:42:19 notice: native_print:     MailTo_6   
>> (heartbeat::ocf:MailTo):        Started server1
>> pengine[4460]: 2009/01/13_03:42:19 notice: native_print: server1-fencing     
>>    (stonith:external/ipmi)
>> pengine[4460]: 2009/01/13_03:42:19 notice: native_print:        0 : server2
>> pengine[4460]: 2009/01/13_03:42:19 notice: native_print:        1 : server1
>> pengine[4460]: 2009/01/13_03:42:19 notice: native_print: server2-fencing     
>>    (stonith:external/ipmi):        Started server2
>> pengine[4460]: 2009/01/13_03:42:19 notice: NoRoleChange: Leave resource 
>> commonIP        (server1)
>> pengine[4460]: 2009/01/13_03:42:19 notice: NoRoleChange: Leave resource 
>> drbddisk_2      (server1)
>> pengine[4460]: 2009/01/13_03:42:19 notice: NoRoleChange: Leave resource 
>> Filesystem_3    (server1)
>> pengine[4460]: 2009/01/13_03:42:19 notice: NoRoleChange: Leave resource 
>> drbddisk_4      (server1)
>> pengine[4460]: 2009/01/13_03:42:19 notice: NoRoleChange: Leave resource 
>> Filesystem_5    (server1)
>> pengine[4460]: 2009/01/13_03:42:19 notice: NoRoleChange: Leave resource 
>> MailTo_6        (server1)
>> pengine[4460]: 2009/01/13_03:42:19 ERROR: native_create_actions: Attempting 
>> recovery of resource server1-fencing
>> pengine[4460]: 2009/01/13_03:42:19 notice: StopRsc:   server2   Stop 
>> server1-fencing
>> pengine[4460]: 2009/01/13_03:42:19 notice: StopRsc:   server1   Stop 
>> server1-fencing
>> pengine[4460]: 2009/01/13_03:42:19 notice: StartRsc:  server1   Start 
>> server1-fencing
>> pengine[4460]: 2009/01/13_03:42:19 notice: RecurringOp: server1    
>> server1-fencing_monitor_120000
>> pengine[4460]: 2009/01/13_03:42:19 WARN: native_color: Resource 
>> server2-fencing cannot run anywhere
>> pengine[4460]: 2009/01/13_03:42:19 notice: StopRsc:   server2   Stop 
>> server2-fencing
>> pengine[4460]: 2009/01/13_03:42:19 WARN: stage6: Scheduling Node server2 for 
>> STONITH
>> pengine[4460]: 2009/01/13_03:42:19 WARN: native_stop_constraints: Stop of 
>> failed resource server1-fencing is implicit after server2 is fenced
>> pengine[4460]: 2009/01/13_03:42:19 info: native_stop_constraints: 
>> server2-fencing_stop_0 is implicit after server2 is fenced
>> crmd[3623]: 2009/01/13_03:42:19 info: do_state_transition: State transition 
>> S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS 
>> cause=C_IPC_MESSAGE origin=route_message ]
>> tengine[4459]: 2009/01/13_03:42:19 info: unpack_graph: Unpacked transition 
>> 33: 8 actions in 8 synapses
>> tengine[4459]: 2009/01/13_03:42:19 info: te_pseudo_action: Pseudo action 2 
>> fired and confirmed
>> tengine[4459]: 2009/01/13_03:42:19 info: te_pseudo_action: Pseudo action 32 
>> fired and confirmed
>> tengine[4459]: 2009/01/13_03:42:19 notice: run_graph: 
>> ====================================================
>> tengine[4459]: 2009/01/13_03:42:19 WARN: run_graph: Transition 33: 
>> (Complete=2, Pending=0, Fired=0, Skipped=0, Incomplete=6)
>> tengine[4459]: 2009/01/13_03:42:19 ERROR: te_graph_trigger: Transition 
>> failed: terminated
>> tengine[4459]: 2009/01/13_03:42:19 WARN: print_graph: Graph 33 (8 actions in 
>> 8 synapses): batch-limit=30 jobs, network-delay=60000ms
>> tengine[4459]: 2009/01/13_03:42:19 WARN: print_graph: Synapse 0 was 
>> confirmed (priority: 0)
>> tengine[4459]: 2009/01/13_03:42:19 WARN: print_graph: Synapse 1 is pending 
>> (priority: 0)
>> tengine[4459]: 2009/01/13_03:42:19 WARN: print_elem:     [Action 8]: Pending 
>> (id: server1-fencing_monitor_120000, loc: server1, priority: 0)
>> tengine[4459]: 2009/01/13_03:42:19 WARN: print_elem:      * [Input 31]: 
>> Pending (id: server1-fencing_start_0, loc: server1, priority: 0)
>> tengine[4459]: 2009/01/13_03:42:19 WARN: print_graph: Synapse 2 is pending 
>> (priority: 0)
>> tengine[4459]: 2009/01/13_03:42:19 WARN: print_elem:     [Action 30]: 
>> Pending (id: server1-fencing_stop_0, loc: server1, priority: 0)
>> tengine[4459]: 2009/01/13_03:42:19 WARN: print_elem:      * [Input 13]: 
>> Pending (id: all_stopped, type: pseduo, priority: 0)
>> tengine[4459]: 2009/01/13_03:42:19 WARN: print_graph: Synapse 3 is pending 
>> (priority: 0)
>> tengine[4459]: 2009/01/13_03:42:19 WARN: print_elem:     [Action 31]: 
>> Pending (id: server1-fencing_start_0, loc: server1, priority: 0)
>> tengine[4459]: 2009/01/13_03:42:19 WARN: print_elem:      * [Input 2]: 
>> Completed (id: server1-fencing_stop_0, type: pseduo, priority: 0)
>> tengine[4459]: 2009/01/13_03:42:19 WARN: print_elem:      * [Input 30]: 
>> Pending (id: server1-fencing_stop_0, loc: server1, priority: 0)
>> tengine[4459]: 2009/01/13_03:42:19 WARN: print_graph: Synapse 4 was 
>> confirmed (priority: 0)
>> tengine[4459]: 2009/01/13_03:42:19 WARN: print_graph: Synapse 5 is pending 
>> (priority: 0)
>> tengine[4459]: 2009/01/13_03:42:19 WARN: print_elem:     [Action 13]: 
>> Pending (id: all_stopped, type: pseduo, priority: 0)
>> tengine[4459]: 2009/01/13_03:42:19 WARN: print_elem:      * [Input 34]: 
>> Pending (id: stonith, loc: server2, type: crm, priority: 0)
>> tengine[4459]: 2009/01/13_03:42:19 WARN: print_graph: Synapse 6 is pending 
>> (priority: 0)
>> tengine[4459]: 2009/01/13_03:42:19 WARN: print_elem:     [Action 33]: 
>> Pending (id: stonith_up, type: pseduo, priority: 0)
>> tengine[4459]: 2009/01/13_03:42:19 WARN: print_elem:      * [Input 31]: 
>> Pending (id: server1-fencing_start_0, loc: server1, priority: 0)
>> tengine[4459]: 2009/01/13_03:42:19 WARN: print_graph: Synapse 7 is pending 
>> (priority: 0)
>> tengine[4459]: 2009/01/13_03:42:19 WARN: print_elem:     [Action 34]: 
>> Pending (id: stonith, loc: server2, type: crm, priority: 0)
>> tengine[4459]: 2009/01/13_03:42:19 WARN: print_elem:      * [Input 33]: 
>> Pending (id: stonith_up, type: pseduo, priority: 0)
>> tengine[4459]: 2009/01/13_03:42:19 info: notify_crmd: Transition 33 status: 
>> te_complete - <null>
>> crmd[3623]: 2009/01/13_03:42:19 info: do_state_transition: State transition 
>> S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS cause=C_IPC_MESSAGE 
>> origin=route_message ]
>> pengine[4460]: 2009/01/13_03:42:19 ERROR: process_pe_message: Transition 33: 
>> ERRORs found during PE processing. PEngine Input stored in: 
>> /var/lib/heartbeat/pengine/pe-error-22.bz2
>> pengine[4460]: 2009/01/13_03:42:19 info: process_pe_message: Configuration 
>> WARNINGs found during PE processing.  Please run "crm_verify -L" to identify 
>> issues.
>> cib[3619]: 2009/01/13_03:42:19 info: cib_stats: Processed 47 operations 
>> (4680.00us average, 0% utilization) in the last 10min
>> tengine[4459]: 2009/01/13_03:42:38 info: process_graph_event: Detected 
>> action server1-fencing_monitor_120000 from a different transition: 29 vs. 33
>> tengine[4459]: 2009/01/13_03:42:38 info: update_abort_priority: Abort 
>> priority upgraded to 1000000
>> crmd[3623]: 2009/01/13_03:42:38 info: do_state_transition: State transition 
>> S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_IPC_MESSAGE 
>> origin=route_message ]
>> crmd[3623]: 2009/01/13_03:42:38 info: do_state_transition: All 2 cluster 
>> nodes are eligible to run resources.
>> pengine[4460]: 2009/01/13_03:42:38 info: determine_online_status: Node 
>> server2 is online
>> pengine[4460]: 2009/01/13_03:42:38 info: determine_online_status: Node 
>> server1 is online
>> pengine[4460]: 2009/01/13_03:42:38 ERROR: native_add_running: Resource 
>> stonith::external/ipmi:server1-fencing appears to be active on 2 nodes.
>> pengine[4460]: 2009/01/13_03:42:38 ERROR: See 
>> http://linux-ha.org/v2/faq/resource_too_active for more information.
>> pengine[4460]: 2009/01/13_03:42:38 ERROR: unpack_rsc_op: Remapping 
>> server2-fencing_start_0 (rc=1) on server1 to an ERROR
>> pengine[4460]: 2009/01/13_03:42:38 WARN: unpack_rsc_op: Processing failed op 
>> server2-fencing_start_0 on server1: Error
>> pengine[4460]: 2009/01/13_03:42:38 WARN: unpack_rsc_op: Compatability 
>> handling for failed op server2-fencing_start_0 on server1
>> pengine[4460]: 2009/01/13_03:42:38 notice: group_print: Resource Group: 
>> group_1
>> pengine[4460]: 2009/01/13_03:42:38 notice: native_print:     commonIP   
>> (heartbeat::ocf:IPaddr):        Started server1
>> pengine[4460]: 2009/01/13_03:42:38 notice: native_print:     drbddisk_2 
>> (heartbeat:drbddisk):   Started server1
>> pengine[4460]: 2009/01/13_03:42:38 notice: native_print:     Filesystem_3    
>>    (heartbeat::ocf:Filesystem):    Started server1
>> pengine[4460]: 2009/01/13_03:42:38 notice: native_print:     drbddisk_4 
>> (heartbeat:drbddisk):   Started server1
>> pengine[4460]: 2009/01/13_03:42:38 notice: native_print:     Filesystem_5    
>>    (heartbeat::ocf:Filesystem):    Started server1
>> pengine[4460]: 2009/01/13_03:42:38 notice: native_print:     MailTo_6   
>> (heartbeat::ocf:MailTo):        Started server1
>> pengine[4460]: 2009/01/13_03:42:38 notice: native_print: server1-fencing     
>>    (stonith:external/ipmi)
>> pengine[4460]: 2009/01/13_03:42:38 notice: native_print:        0 : server2
>> pengine[4460]: 2009/01/13_03:42:38 notice: native_print:        1 : server1
>> pengine[4460]: 2009/01/13_03:42:38 notice: native_print: server2-fencing     
>>    (stonith:external/ipmi):        Started server2
>> pengine[4460]: 2009/01/13_03:42:38 notice: NoRoleChange: Leave resource 
>> commonIP        (server1)
>> pengine[4460]: 2009/01/13_03:42:38 notice: NoRoleChange: Leave resource 
>> drbddisk_2      (server1)
>> pengine[4460]: 2009/01/13_03:42:38 notice: NoRoleChange: Leave resource 
>> Filesystem_3    (server1)
>> pengine[4460]: 2009/01/13_03:42:38 notice: NoRoleChange: Leave resource 
>> drbddisk_4      (server1)
>> pengine[4460]: 2009/01/13_03:42:38 notice: NoRoleChange: Leave resource 
>> Filesystem_5    (server1)
>> pengine[4460]: 2009/01/13_03:42:38 notice: NoRoleChange: Leave resource 
>> MailTo_6        (server1)
>> pengine[4460]: 2009/01/13_03:42:38 ERROR: native_create_actions: Attempting 
>> recovery of resource server1-fencing
>> pengine[4460]: 2009/01/13_03:42:38 notice: StopRsc:   server2   Stop 
>> server1-fencing
>> pengine[4460]: 2009/01/13_03:42:38 notice: StopRsc:   server1   Stop 
>> server1-fencing
>> pengine[4460]: 2009/01/13_03:42:38 notice: StartRsc:  server2   Start 
>> server1-fencing
>> pengine[4460]: 2009/01/13_03:42:38 notice: RecurringOp: server2    
>> server1-fencing_monitor_120000
>> pengine[4460]: 2009/01/13_03:42:38 notice: NoRoleChange: Leave resource 
>> server2-fencing (server2)
>> crmd[3623]: 2009/01/13_03:42:38 info: do_state_transition: State transition 
>> S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS 
>> cause=C_IPC_MESSAGE origin=route_message ]
>> tengine[4459]: 2009/01/13_03:42:38 info: unpack_graph: Unpacked transition 
>> 34: 5 actions in 5 synapses
>> tengine[4459]: 2009/01/13_03:42:38 info: te_pseudo_action: Pseudo action 13 
>> fired and confirmed
>> tengine[4459]: 2009/01/13_03:42:38 info: send_rsc_command: Initiating action 
>> 30: server1-fencing_stop_0 on server2
>> tengine[4459]: 2009/01/13_03:42:38 info: send_rsc_command: Initiating action 
>> 31: server1-fencing_stop_0 on server1
>> crmd[3623]: 2009/01/13_03:42:38 info: do_lrm_rsc_op: Performing 
>> op=server1-fencing_stop_0 key=31:34:8686f9af-9ced-43ab-bf20-be6e8437abc0)
>> lrmd[3620]: 2009/01/13_03:42:38 info: rsc:server1-fencing: stop
>> crmd[3623]: 2009/01/13_03:42:38 info: process_lrm_event: LRM operation 
>> server1-fencing_monitor_120000 (call=52, rc=-2) Cancelled
>> pengine[4460]: 2009/01/13_03:42:38 ERROR: process_pe_message: Transition 34: 
>> ERRORs found during PE processing. PEngine Input stored in: 
>> /var/lib/heartbeat/pengine/pe-error-23.bz2
>> lrmd[28014]: 2009/01/13_03:42:38 info: Try to stop STONITH resource 
>> <rsc_id=server1-fencing> : Device=external/ipmi
>> crmd[3623]: 2009/01/13_03:42:38 info: process_lrm_event: LRM operation 
>> server1-fencing_stop_0 (call=53, rc=0) complete
>> tengine[4459]: 2009/01/13_03:42:38 info: match_graph_event: Action 
>> server1-fencing_stop_0 (31) confirmed on server1 (rc=0)
>> tengine[4459]: 2009/01/13_03:42:39 info: match_graph_event: Action 
>> server1-fencing_stop_0 (30) confirmed on server2 (rc=0)
>> tengine[4459]: 2009/01/13_03:42:39 info: send_rsc_command: Initiating action 
>> 32: server1-fencing_start_0 on server2
>> tengine[4459]: 2009/01/13_03:42:41 info: match_graph_event: Action 
>> server1-fencing_start_0 (32) confirmed on server2 (rc=0)
>> tengine[4459]: 2009/01/13_03:42:41 info: send_rsc_command: Initiating action 
>> 1: server1-fencing_monitor_120000 on server2
>> tengine[4459]: 2009/01/13_03:42:42 info: match_graph_event: Action 
>> server1-fencing_monitor_120000 (1) confirmed on server2 (rc=0)
>> tengine[4459]: 2009/01/13_03:42:42 info: run_graph: Transition 34: 
>> (Complete=5, Pending=0, Fired=0, Skipped=0, Incomplete=0)
>> tengine[4459]: 2009/01/13_03:42:42 info: notify_crmd: Transition 34 status: 
>> te_complete - <null>
>> crmd[3623]: 2009/01/13_03:42:42 info: do_state_transition: State transition 
>> S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS cause=C_IPC_MESSAGE 
>> origin=route_message ]
>> cib[3619]: 2009/01/13_03:52:19 info: cib_stats: Processed 47 operations 
>> (6382.00us average, 0% utilization) in the last 10min
>> lrmd[3620]: 2009/01/13_03:59:32 WARN: G_SIG_dispatch: Dispatch function for 
>> SIGCHLD was delayed 1000 ms (> 100 ms) before being called (GSource: 
>> 0x514c58)
>> lrmd[3620]: 2009/01/13_03:59:32 info: G_SIG_dispatch: started at 1733230017 
>> should have started at 1733229917
>>
>> _______________________________________________
>> Linux-HA mailing list
>> [email protected]
>> http://lists.linux-ha.org/mailman/listinfo/linux-ha
>> See also: http://linux-ha.org/ReportingProblems
>>
> _______________________________________________
> Linux-HA mailing list
> [email protected]
> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> See also: http://linux-ha.org/ReportingProblems

-- 
--------------------------------------------------------------------
Thomas Roth
Department: Informationstechnologie
Location: SB3 1.262
Phone: +49-6159-71 1453  Fax: +49-6159-71 2986

GSI Helmholtzzentrum für Schwerionenforschung GmbH
Planckstraße 1
D-64291 Darmstadt
www.gsi.de

Gesellschaft mit beschränkter Haftung
Sitz der Gesellschaft: Darmstadt
Handelsregister: Amtsgericht Darmstadt, HRB 1528

Geschäftsführer: Professor Dr. Horst Stöcker

Vorsitzende des Aufsichtsrates: Dr. Beatrix Vierkorn-Rudolph,
Stellvertreter: Ministerialdirigent Dr. Rolf Bernhardt
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to