On Tue, Jan 13, 2009 at 15:46, Thomas Roth <[email protected]> wrote:
> Hi all,
>
> we are using external/ipmi on a drbd8 / heartbeat-2.0.7-2 as stonith

According to the cib you sent through, you actually have something
based on 2.1.3
However I can't find the changeset it lists
(552305612591183b1628baa5bc6e903e0f1e26a3) anywhere and some of the PE
logs look significantly newer than 2.1.3.

Where did you get the heartbeat package from?  It helps to know what
I'm looking at before commenting.

> device. While the stonith itself works well, there are some errors which
> recur and eventually lead to unwanted reboots. And mayby there are no
> real causes but just my lack of understanding...
> Anyhow, in the ha-debug log of server1 I see
>> tengine[4459]: 2009/01/13_03:40:38 info: process_graph_event: Detected
> action server1-fencing_monitor_120000 from a different transition: 29 vs. 31
>>tengine[4459]: 2009/01/13_03:40:38 info: update_abort_priority: Abort
> priority upgraded to 1000000
>>tengine[4459]: 2009/01/13_03:40:38 WARN: update_failcount: Updating
> failcount for server1-fencing on ebffe771-505c-4e40-b0b3-d70903ed37bc
> after failed monitor: rc=14
>
> Then pengine goes on about that problem:
>>pengine[4460]: 2009/01/13_03:40:38 info: determine_online_status: Node
> server2 is online
>>pengine[4460]: 2009/01/13_03:40:38 ERROR: unpack_rsc_op: Remapping
> server1-fencing_monitor_120000 (rc=14) on server2 to an ERROR
>>pengine[4460]: 2009/01/13_03:40:38 WARN: unpack_rsc_op: Processing
> failed op server1-fencing_monitor_120000 on server2: Error
>>pengine[4460]: 2009/01/13_03:40:38 info: determine_online_status: Node
> server1 is online
>>pengine[4460]: 2009/01/13_03:40:38 ERROR: unpack_rsc_op: Remapping
> server2-fencing_start_0 (rc=1) on server1 to an ERROR
>>pengine[4460]: 2009/01/13_03:40:38 WARN: unpack_rsc_op: Processing
> failed op server2-fencing_start_0 on server1: Error
>>pengine[4460]: 2009/01/13_03:40:38 WARN: unpack_rsc_op: Compatability
> handling for failed op server2-fencing_start_0 on server1
>
> This was already the second instance of this problem last night, the
> first occured 5 hours earlier, having no further consequences.
> Now, however, pengine after a walk through all the resources  says
>>pengine[4460]: 2009/01/13_03:40:38 WARN: stage6: Scheduling Node
> server2 for STONITH
>
> There are some repetitions of this, until finally stonithd gives up:
>>stonithd[3621]: 2009/01/13_03:42:19 ERROR: Failed to STONITH the node
> server2: optype=RESET, op_result=TIMEOUT
>
>
> The other machine, server2,  did not show any problems whatsoever that
> should have triggered the action.
> The intended victim this time was the slave. However, during the last
> weekend both machines showed very strange behavior, among them rebooting
> each other ( with a days break, there was no stonith-war going on, but
> still...)
>
>
> Now in the cib.xml, we have the following two primitives
>
>       <primitive class="stonith" type="external/ipmi"
> provider="heartbeat" id="server1-fencing">
>         <operations>
>           <op id="server1-fencing-monitor" name="monitor"
> interval="120s" timeout="70s" prereq="nothing" start_delay="0"
> disabled="false" role="Started" on_fail="fence"/>
>           <op id="server1-fencing-start" name="start" timeout="40s"
> prereq="nothing" start_delay="0" disabled="false" role="Started"/>
>         </operations>
>         <instance_attributes id="server1-fencing-ia">
>           <attributes>
>             <nvpair id="server1-fencing-hostname" name="hostname"
> value="server1"/>
>             <nvpair id="server1-fencing-ipaddr" name="ipaddr"
> value="1.1.1.1"/>
>             <nvpair id="server1-fencing-userid" name="userid"
> value="USER"/>
>             <nvpair id="server1-fencing-passwd" name="passwd"
> value="PASSWD"/>
>           </attributes>
>         </instance_attributes>
>       </primitive>
>       <primitive id="server2-fencing" class="stonith"
> type="external/ipmi" provider="heartbeat">
>         <operations>
>           <op id="server2-fencing-monitor" name="monitor"
> interval="120s" timeout="70s" prereq="nothing"/>
>           <op id="server2-fencing-start" name="start" timeout="40s"
> prereq="nothing"/>
>         </operations>
>         <instance_attributes id="server2-fencing-ia">
>           <attributes>
>             <nvpair id="server2-fencing-hostname" name="hostname"
> value="server2"/>
>             <nvpair id="server2-fencing-ipaddr" name="ipaddr"
> value="1.1.1.2"/>
>             <nvpair id="server2-fencing-userid" name="userid"
> value="USER"/>
>             <nvpair id="server2-fencing-passwd" name="passwd"
> value="PASSWD"/>
>           </attributes>
>         </instance_attributes>
>       </primitive>
>
>
> The intent of this is to reboot one machine in case of a failover which
> cannot be completed because failing heartbeat cannot release one of its
> resources (Lustre mounts of drbd disks).
>
> So my question is of course whether there is something fundamentally
> wrong with this configuration (I'm attaching the entire cib.xml and the
> ha-debug for last night, btw). Or whether there should really be an
> error somewhere else (hardware).
>
> Thanks a lot,
> Thomas
>
>
>
> tengine[4459]: 2009/01/12_22:34:28 info: process_graph_event: Action 
> server2-fencing_monitor_120000 arrived after a completed transition
> tengine[4459]: 2009/01/12_22:34:28 info: update_abort_priority: Abort 
> priority upgraded to 1000000
> tengine[4459]: 2009/01/12_22:34:28 WARN: update_failcount: Updating failcount 
> for server2-fencing on ebffe771-505c-4e40-b0b3-d70903ed37bc after failed 
> monitor: rc=14
> crmd[3623]: 2009/01/12_22:34:28 info: do_state_transition: State transition 
> S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_IPC_MESSAGE 
> origin=route_message ]
> crmd[3623]: 2009/01/12_22:34:28 info: do_state_transition: All 2 cluster 
> nodes are eligible to run resources.
> tengine[4459]: 2009/01/12_22:34:28 info: extract_event: Aborting on 
> transient_attributes changes for ebffe771-505c-4e40-b0b3-d70903ed37bc
> pengine[4460]: 2009/01/12_22:34:28 info: determine_online_status: Node 
> server2 is online
> pengine[4460]: 2009/01/12_22:34:28 ERROR: unpack_rsc_op: Remapping 
> server2-fencing_monitor_120000 (rc=14) on server2 to an ERROR
> pengine[4460]: 2009/01/12_22:34:28 WARN: unpack_rsc_op: Processing failed op 
> server2-fencing_monitor_120000 on server2: Error
> pengine[4460]: 2009/01/12_22:34:28 info: determine_online_status: Node 
> server1 is online
> pengine[4460]: 2009/01/12_22:34:28 ERROR: unpack_rsc_op: Remapping 
> server2-fencing_start_0 (rc=1) on server1 to an ERROR
> pengine[4460]: 2009/01/12_22:34:28 WARN: unpack_rsc_op: Processing failed op 
> server2-fencing_start_0 on server1: Error
> pengine[4460]: 2009/01/12_22:34:28 WARN: unpack_rsc_op: Compatability 
> handling for failed op server2-fencing_start_0 on server1
> pengine[4460]: 2009/01/12_22:34:28 notice: group_print: Resource Group: 
> group_1
> pengine[4460]: 2009/01/12_22:34:28 notice: native_print:     commonIP   
> (heartbeat::ocf:IPaddr):        Started server1
> pengine[4460]: 2009/01/12_22:34:28 notice: native_print:     drbddisk_2 
> (heartbeat:drbddisk):   Started server1
> pengine[4460]: 2009/01/12_22:34:28 notice: native_print:     Filesystem_3     
>   (heartbeat::ocf:Filesystem):    Started server1
> pengine[4460]: 2009/01/12_22:34:28 notice: native_print:     drbddisk_4 
> (heartbeat:drbddisk):   Started server1
> pengine[4460]: 2009/01/12_22:34:28 notice: native_print:     Filesystem_5     
>   (heartbeat::ocf:Filesystem):    Started server1
> pengine[4460]: 2009/01/12_22:34:28 notice: native_print:     MailTo_6   
> (heartbeat::ocf:MailTo):        Started server1
> pengine[4460]: 2009/01/12_22:34:28 notice: native_print: server1-fencing      
>   (stonith:external/ipmi):        Started server2
> pengine[4460]: 2009/01/12_22:34:28 notice: native_print: server2-fencing      
>   (stonith:external/ipmi):        Started server2 FAILED
> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Leave resource 
> commonIP        (server1)
> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Leave resource 
> drbddisk_2      (server1)
> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Leave resource 
> Filesystem_3    (server1)
> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Leave resource 
> drbddisk_4      (server1)
> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Leave resource 
> Filesystem_5    (server1)
> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Leave resource 
> MailTo_6        (server1)
> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Leave resource 
> server1-fencing (server2)
> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Recover resource 
> server2-fencing       (server2)
> pengine[4460]: 2009/01/12_22:34:28 notice: StopRsc:   server2   Stop 
> server2-fencing
> pengine[4460]: 2009/01/12_22:34:28 notice: StartRsc:  server2   Start 
> server2-fencing
> pengine[4460]: 2009/01/12_22:34:28 notice: RecurringOp: server2    
> server2-fencing_monitor_120000
> pengine[4460]: 2009/01/12_22:34:28 info: process_pe_message: Transition 30: 
> PEngine Input stored in: /var/lib/heartbeat/pengine/pe-input-406.bz2
> pengine[4460]: 2009/01/12_22:34:28 info: determine_online_status: Node 
> server2 is online
> pengine[4460]: 2009/01/12_22:34:28 ERROR: unpack_rsc_op: Remapping 
> server2-fencing_monitor_120000 (rc=14) on server2 to an ERROR
> pengine[4460]: 2009/01/12_22:34:28 WARN: unpack_rsc_op: Processing failed op 
> server2-fencing_monitor_120000 on server2: Error
> pengine[4460]: 2009/01/12_22:34:28 info: determine_online_status: Node 
> server1 is online
> pengine[4460]: 2009/01/12_22:34:28 ERROR: unpack_rsc_op: Remapping 
> server2-fencing_start_0 (rc=1) on server1 to an ERROR
> pengine[4460]: 2009/01/12_22:34:28 WARN: unpack_rsc_op: Processing failed op 
> server2-fencing_start_0 on server1: Error
> pengine[4460]: 2009/01/12_22:34:28 WARN: unpack_rsc_op: Compatability 
> handling for failed op server2-fencing_start_0 on server1
> pengine[4460]: 2009/01/12_22:34:28 notice: group_print: Resource Group: 
> group_1
> pengine[4460]: 2009/01/12_22:34:28 notice: native_print:     commonIP   
> (heartbeat::ocf:IPaddr):        Started server1
> pengine[4460]: 2009/01/12_22:34:28 notice: native_print:     drbddisk_2 
> (heartbeat:drbddisk):   Started server1
> pengine[4460]: 2009/01/12_22:34:28 notice: native_print:     Filesystem_3     
>   (heartbeat::ocf:Filesystem):    Started server1
> pengine[4460]: 2009/01/12_22:34:28 notice: native_print:     drbddisk_4 
> (heartbeat:drbddisk):   Started server1
> pengine[4460]: 2009/01/12_22:34:28 notice: native_print:     Filesystem_5     
>   (heartbeat::ocf:Filesystem):    Started server1
> pengine[4460]: 2009/01/12_22:34:28 notice: native_print:     MailTo_6   
> (heartbeat::ocf:MailTo):        Started server1
> pengine[4460]: 2009/01/12_22:34:28 notice: native_print: server1-fencing      
>   (stonith:external/ipmi):        Started server2
> pengine[4460]: 2009/01/12_22:34:28 notice: native_print: server2-fencing      
>   (stonith:external/ipmi):        Started server2 FAILED
> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Leave resource 
> commonIP        (server1)
> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Leave resource 
> drbddisk_2      (server1)
> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Leave resource 
> Filesystem_3    (server1)
> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Leave resource 
> drbddisk_4      (server1)
> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Leave resource 
> Filesystem_5    (server1)
> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Leave resource 
> MailTo_6        (server1)
> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Leave resource 
> server1-fencing (server2)
> pengine[4460]: 2009/01/12_22:34:28 notice: NoRoleChange: Recover resource 
> server2-fencing       (server2)
> pengine[4460]: 2009/01/12_22:34:28 notice: StopRsc:   server2   Stop 
> server2-fencing
> pengine[4460]: 2009/01/12_22:34:28 notice: StartRsc:  server2   Start 
> server2-fencing
> pengine[4460]: 2009/01/12_22:34:28 notice: RecurringOp: server2    
> server2-fencing_monitor_120000
> crmd[3623]: 2009/01/12_22:34:28 info: do_state_transition: State transition 
> S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS 
> cause=C_IPC_MESSAGE origin=route_message ]
> tengine[4459]: 2009/01/12_22:34:28 info: unpack_graph: Unpacked transition 
> 31: 4 actions in 4 synapses
> tengine[4459]: 2009/01/12_22:34:28 info: te_pseudo_action: Pseudo action 13 
> fired and confirmed
> tengine[4459]: 2009/01/12_22:34:28 info: send_rsc_command: Initiating action 
> 3: server2-fencing_stop_0 on server2
> pengine[4460]: 2009/01/12_22:34:28 info: process_pe_message: Transition 31: 
> PEngine Input stored in: /var/lib/heartbeat/pengine/pe-input-407.bz2
> tengine[4459]: 2009/01/12_22:34:29 info: match_graph_event: Action 
> server2-fencing_stop_0 (3) confirmed on server2 (rc=0)
> tengine[4459]: 2009/01/12_22:34:29 info: send_rsc_command: Initiating action 
> 32: server2-fencing_start_0 on server2
> tengine[4459]: 2009/01/12_22:34:31 info: match_graph_event: Action 
> server2-fencing_start_0 (32) confirmed on server2 (rc=0)
> tengine[4459]: 2009/01/12_22:34:31 info: send_rsc_command: Initiating action 
> 2: server2-fencing_monitor_120000 on server2
> tengine[4459]: 2009/01/12_22:34:33 info: match_graph_event: Action 
> server2-fencing_monitor_120000 (2) confirmed on server2 (rc=0)
> tengine[4459]: 2009/01/12_22:34:33 info: run_graph: Transition 31: 
> (Complete=4, Pending=0, Fired=0, Skipped=0, Incomplete=0)
> tengine[4459]: 2009/01/12_22:34:33 info: notify_crmd: Transition 31 status: 
> te_complete - <null>
> crmd[3623]: 2009/01/12_22:34:33 info: do_state_transition: State transition 
> S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS cause=C_IPC_MESSAGE 
> origin=route_message ]
> cib[3619]: 2009/01/12_22:42:19 info: cib_stats: Processed 49 operations 
> (4285.00us average, 0% utilization) in the last 10min
> cib[3619]: 2009/01/12_22:52:19 info: cib_stats: Processed 39 operations 
> (4102.00us average, 0% utilization) in the last 10min
> cib[3619]: 2009/01/12_23:02:19 info: cib_stats: Processed 40 operations 
> (6000.00us average, 0% utilization) in the last 10min
> lrmd[3620]: 2009/01/12_23:08:37 WARN: G_SIG_dispatch: Dispatch function for 
> SIGCHLD was delayed 1000 ms (> 100 ms) before being called (GSource: 0x514c58)
> lrmd[3620]: 2009/01/12_23:08:37 info: G_SIG_dispatch: started at 1731484574 
> should have started at 1731484474
> cib[3619]: 2009/01/12_23:12:19 info: cib_stats: Processed 40 operations 
> (4000.00us average, 0% utilization) in the last 10min
> cib[3619]: 2009/01/12_23:22:19 info: cib_stats: Processed 40 operations 
> (5000.00us average, 0% utilization) in the last 10min
> cib[3619]: 2009/01/12_23:32:19 info: cib_stats: Processed 39 operations 
> (7435.00us average, 0% utilization) in the last 10min
> cib[3619]: 2009/01/12_23:42:19 info: cib_stats: Processed 40 operations 
> (5250.00us average, 0% utilization) in the last 10min
> cib[3619]: 2009/01/12_23:52:19 info: cib_stats: Processed 40 operations 
> (6500.00us average, 0% utilization) in the last 10min
> cib[3619]: 2009/01/13_00:02:19 info: cib_stats: Processed 39 operations 
> (5641.00us average, 0% utilization) in the last 10min
> cib[3619]: 2009/01/13_00:12:19 info: cib_stats: Processed 40 operations 
> (6000.00us average, 0% utilization) in the last 10min
> cib[3619]: 2009/01/13_00:22:19 info: cib_stats: Processed 40 operations 
> (4750.00us average, 0% utilization) in the last 10min
> cib[3619]: 2009/01/13_00:32:19 info: cib_stats: Processed 40 operations 
> (4500.00us average, 0% utilization) in the last 10min
> cib[3619]: 2009/01/13_00:42:19 info: cib_stats: Processed 39 operations 
> (5128.00us average, 0% utilization) in the last 10min
> cib[3619]: 2009/01/13_00:52:19 info: cib_stats: Processed 40 operations 
> (4000.00us average, 0% utilization) in the last 10min
> cib[3619]: 2009/01/13_01:02:19 info: cib_stats: Processed 40 operations 
> (5500.00us average, 0% utilization) in the last 10min
> cib[3619]: 2009/01/13_01:12:19 info: cib_stats: Processed 40 operations 
> (4500.00us average, 0% utilization) in the last 10min
> cib[3619]: 2009/01/13_01:22:19 info: cib_stats: Processed 39 operations 
> (4358.00us average, 0% utilization) in the last 10min
> cib[3619]: 2009/01/13_01:32:19 info: cib_stats: Processed 40 operations 
> (4750.00us average, 0% utilization) in the last 10min
> cib[3619]: 2009/01/13_01:42:19 info: cib_stats: Processed 40 operations 
> (7000.00us average, 0% utilization) in the last 10min
> cib[3619]: 2009/01/13_01:52:19 info: cib_stats: Processed 40 operations 
> (5500.00us average, 0% utilization) in the last 10min
> cib[3619]: 2009/01/13_02:02:19 info: cib_stats: Processed 39 operations 
> (5384.00us average, 0% utilization) in the last 10min
> cib[3619]: 2009/01/13_02:12:19 info: cib_stats: Processed 40 operations 
> (5000.00us average, 0% utilization) in the last 10min
> cib[3619]: 2009/01/13_02:22:19 info: cib_stats: Processed 40 operations 
> (6750.00us average, 0% utilization) in the last 10min
> cib[3619]: 2009/01/13_02:32:19 info: cib_stats: Processed 39 operations 
> (5897.00us average, 0% utilization) in the last 10min
> cib[3619]: 2009/01/13_02:42:19 info: cib_stats: Processed 40 operations 
> (3750.00us average, 0% utilization) in the last 10min
> cib[3619]: 2009/01/13_02:52:19 info: cib_stats: Processed 40 operations 
> (5000.00us average, 0% utilization) in the last 10min
> cib[3619]: 2009/01/13_03:02:19 info: cib_stats: Processed 40 operations 
> (5500.00us average, 0% utilization) in the last 10min
> cib[3619]: 2009/01/13_03:12:19 info: cib_stats: Processed 39 operations 
> (6153.00us average, 0% utilization) in the last 10min
> cib[3619]: 2009/01/13_03:22:19 info: cib_stats: Processed 40 operations 
> (4750.00us average, 0% utilization) in the last 10min
> cib[3619]: 2009/01/13_03:32:19 info: cib_stats: Processed 40 operations 
> (4750.00us average, 0% utilization) in the last 10min
> tengine[4459]: 2009/01/13_03:40:38 info: process_graph_event: Detected action 
> server1-fencing_monitor_120000 from a different transition: 29 vs. 31
> tengine[4459]: 2009/01/13_03:40:38 info: update_abort_priority: Abort 
> priority upgraded to 1000000
> tengine[4459]: 2009/01/13_03:40:38 WARN: update_failcount: Updating failcount 
> for server1-fencing on ebffe771-505c-4e40-b0b3-d70903ed37bc after failed 
> monitor: rc=14
> crmd[3623]: 2009/01/13_03:40:38 info: do_state_transition: State transition 
> S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_IPC_MESSAGE 
> origin=route_message ]
> crmd[3623]: 2009/01/13_03:40:38 info: do_state_transition: All 2 cluster 
> nodes are eligible to run resources.
> pengine[4460]: 2009/01/13_03:40:38 info: determine_online_status: Node 
> server2 is online
> pengine[4460]: 2009/01/13_03:40:38 ERROR: unpack_rsc_op: Remapping 
> server1-fencing_monitor_120000 (rc=14) on server2 to an ERROR
> pengine[4460]: 2009/01/13_03:40:38 WARN: unpack_rsc_op: Processing failed op 
> server1-fencing_monitor_120000 on server2: Error
> pengine[4460]: 2009/01/13_03:40:38 info: determine_online_status: Node 
> server1 is online
> pengine[4460]: 2009/01/13_03:40:38 ERROR: unpack_rsc_op: Remapping 
> server2-fencing_start_0 (rc=1) on server1 to an ERROR
> pengine[4460]: 2009/01/13_03:40:38 WARN: unpack_rsc_op: Processing failed op 
> server2-fencing_start_0 on server1: Error
> pengine[4460]: 2009/01/13_03:40:38 WARN: unpack_rsc_op: Compatability 
> handling for failed op server2-fencing_start_0 on server1
> pengine[4460]: 2009/01/13_03:40:38 notice: group_print: Resource Group: 
> group_1
> pengine[4460]: 2009/01/13_03:40:38 notice: native_print:     commonIP   
> (heartbeat::ocf:IPaddr):        Started server1
> pengine[4460]: 2009/01/13_03:40:38 notice: native_print:     drbddisk_2 
> (heartbeat:drbddisk):   Started server1
> pengine[4460]: 2009/01/13_03:40:38 notice: native_print:     Filesystem_3     
>   (heartbeat::ocf:Filesystem):    Started server1
> pengine[4460]: 2009/01/13_03:40:38 notice: native_print:     drbddisk_4 
> (heartbeat:drbddisk):   Started server1
> pengine[4460]: 2009/01/13_03:40:38 notice: native_print:     Filesystem_5     
>   (heartbeat::ocf:Filesystem):    Started server1
> pengine[4460]: 2009/01/13_03:40:38 notice: native_print:     MailTo_6   
> (heartbeat::ocf:MailTo):        Started server1
> pengine[4460]: 2009/01/13_03:40:38 notice: native_print: server1-fencing      
>   (stonith:external/ipmi):        Started server2 FAILED
> pengine[4460]: 2009/01/13_03:40:38 notice: native_print: server2-fencing      
>   (stonith:external/ipmi):        Started server2
> pengine[4460]: 2009/01/13_03:40:38 notice: NoRoleChange: Leave resource 
> commonIP        (server1)
> pengine[4460]: 2009/01/13_03:40:38 notice: NoRoleChange: Leave resource 
> drbddisk_2      (server1)
> pengine[4460]: 2009/01/13_03:40:38 notice: NoRoleChange: Leave resource 
> Filesystem_3    (server1)
> pengine[4460]: 2009/01/13_03:40:38 notice: NoRoleChange: Leave resource 
> drbddisk_4      (server1)
> pengine[4460]: 2009/01/13_03:40:38 notice: NoRoleChange: Leave resource 
> Filesystem_5    (server1)
> pengine[4460]: 2009/01/13_03:40:38 notice: NoRoleChange: Leave resource 
> MailTo_6        (server1)
> pengine[4460]: 2009/01/13_03:40:38 notice: NoRoleChange: Recover resource 
> server1-fencing       (server1)
> pengine[4460]: 2009/01/13_03:40:38 notice: StopRsc:   server2   Stop 
> server1-fencing
> pengine[4460]: 2009/01/13_03:40:38 notice: StartRsc:  server1   Start 
> server1-fencing
> pengine[4460]: 2009/01/13_03:40:38 notice: RecurringOp: server1    
> server1-fencing_monitor_120000
> pengine[4460]: 2009/01/13_03:40:38 WARN: native_color: Resource 
> server2-fencing cannot run anywhere
> pengine[4460]: 2009/01/13_03:40:38 notice: StopRsc:   server2   Stop 
> server2-fencing
> pengine[4460]: 2009/01/13_03:40:38 WARN: stage6: Scheduling Node server2 for 
> STONITH
> pengine[4460]: 2009/01/13_03:40:38 WARN: native_stop_constraints: Stop of 
> failed resource server1-fencing is implicit after server2 is fenced
> pengine[4460]: 2009/01/13_03:40:38 info: native_stop_constraints: 
> server2-fencing_stop_0 is implicit after server2 is fenced
> crmd[3623]: 2009/01/13_03:40:38 info: do_state_transition: State transition 
> S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS 
> cause=C_IPC_MESSAGE origin=route_message ]
> tengine[4459]: 2009/01/13_03:40:38 info: unpack_graph: Unpacked transition 
> 32: 7 actions in 7 synapses
> tengine[4459]: 2009/01/13_03:40:38 info: te_pseudo_action: Pseudo action 2 
> fired and confirmed
> tengine[4459]: 2009/01/13_03:40:38 info: send_rsc_command: Initiating action 
> 29: server1-fencing_start_0 on server1
> tengine[4459]: 2009/01/13_03:40:38 info: te_pseudo_action: Pseudo action 31 
> fired and confirmed
> crmd[3623]: 2009/01/13_03:40:38 info: do_lrm_rsc_op: Performing 
> op=server1-fencing_start_0 key=29:32:8686f9af-9ced-43ab-bf20-be6e8437abc0)
> lrmd[3620]: 2009/01/13_03:40:38 info: rsc:server1-fencing: start
> lrmd[27785]: 2009/01/13_03:40:38 info: Try to start STONITH resource 
> <rsc_id=server1-fencing> : Device=external/ipmi
> pengine[4460]: 2009/01/13_03:40:38 WARN: process_pe_message: Transition 32: 
> WARNINGs found during PE processing. PEngine Input stored in: 
> /var/lib/heartbeat/pengine/pe-warn-6.bz2
> pengine[4460]: 2009/01/13_03:40:38 info: process_pe_message: Configuration 
> WARNINGs found during PE processing.  Please run "crm_verify -L" to identify 
> issues.
> crmd[3623]: 2009/01/13_03:40:39 info: process_lrm_event: LRM operation 
> server1-fencing_start_0 (call=51, rc=0) complete
> tengine[4459]: 2009/01/13_03:40:39 info: match_graph_event: Action 
> server1-fencing_start_0 (29) confirmed on server1 (rc=0)
> tengine[4459]: 2009/01/13_03:40:39 info: send_rsc_command: Initiating action 
> 30: server1-fencing_monitor_120000 on server1
> tengine[4459]: 2009/01/13_03:40:39 info: te_pseudo_action: Pseudo action 32 
> fired and confirmed
> tengine[4459]: 2009/01/13_03:40:39 info: te_fence_node: Executing reboot 
> fencing operation (33) on server2 (timeout=100000)
> crmd[3623]: 2009/01/13_03:40:39 info: do_lrm_rsc_op: Performing 
> op=server1-fencing_monitor_120000 
> key=30:32:8686f9af-9ced-43ab-bf20-be6e8437abc0)
> stonithd[3621]: 2009/01/13_03:40:39 info: client tengine [pid: 4459] want a 
> STONITH operation RESET to node server2.
> stonithd[3621]: 2009/01/13_03:40:39 info: Broadcasting the message succeeded: 
> require others to stonith node server2.
> crmd[3623]: 2009/01/13_03:40:39 info: process_lrm_event: LRM operation 
> server1-fencing_monitor_120000 (call=52, rc=0) complete
> tengine[4459]: 2009/01/13_03:40:39 info: match_graph_event: Action 
> server1-fencing_monitor_120000 (30) confirmed on server1 (rc=0)
> stonithd[3621]: 2009/01/13_03:42:19 ERROR: Failed to STONITH the node 
> server2: optype=RESET, op_result=TIMEOUT
> tengine[4459]: 2009/01/13_03:42:19 info: tengine_stonith_callback: call=-2, 
> optype=1, node_name=server2, result=2, node_list=, 
> action=33:32:8686f9af-9ced-43ab-bf20-be6e8437abc0
> tengine[4459]: 2009/01/13_03:42:19 ERROR: tengine_stonith_callback: Stonith 
> of server2 failed (2)... aborting transition.
> tengine[4459]: 2009/01/13_03:42:19 info: update_abort_priority: Abort 
> priority upgraded to 1000000
> tengine[4459]: 2009/01/13_03:42:19 info: update_abort_priority: Abort action 
> 0 superceeded by 2
> tengine[4459]: 2009/01/13_03:42:19 info: run_graph: 
> ====================================================
> tengine[4459]: 2009/01/13_03:42:19 notice: run_graph: Transition 32: 
> (Complete=6, Pending=0, Fired=0, Skipped=1, Incomplete=0)
> crmd[3623]: 2009/01/13_03:42:19 info: do_state_transition: State transition 
> S_TRANSITION_ENGINE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_IPC_MESSAGE 
> origin=route_message ]
> crmd[3623]: 2009/01/13_03:42:19 info: do_state_transition: All 2 cluster 
> nodes are eligible to run resources.
> pengine[4460]: 2009/01/13_03:42:19 info: determine_online_status: Node 
> server2 is online
> pengine[4460]: 2009/01/13_03:42:19 ERROR: unpack_rsc_op: Remapping 
> server1-fencing_monitor_120000 (rc=14) on server2 to an ERROR
> pengine[4460]: 2009/01/13_03:42:19 WARN: unpack_rsc_op: Processing failed op 
> server1-fencing_monitor_120000 on server2: Error
> pengine[4460]: 2009/01/13_03:42:19 info: determine_online_status: Node 
> server1 is online
> pengine[4460]: 2009/01/13_03:42:19 ERROR: native_add_running: Resource 
> stonith::external/ipmi:server1-fencing appears to be active on 2 nodes.
> pengine[4460]: 2009/01/13_03:42:19 ERROR: See 
> http://linux-ha.org/v2/faq/resource_too_active for more information.
> pengine[4460]: 2009/01/13_03:42:19 ERROR: unpack_rsc_op: Remapping 
> server2-fencing_start_0 (rc=1) on server1 to an ERROR
> pengine[4460]: 2009/01/13_03:42:19 WARN: unpack_rsc_op: Processing failed op 
> server2-fencing_start_0 on server1: Error
> pengine[4460]: 2009/01/13_03:42:19 WARN: unpack_rsc_op: Compatability 
> handling for failed op server2-fencing_start_0 on server1
> pengine[4460]: 2009/01/13_03:42:19 notice: group_print: Resource Group: 
> group_1
> pengine[4460]: 2009/01/13_03:42:19 notice: native_print:     commonIP   
> (heartbeat::ocf:IPaddr):        Started server1
> pengine[4460]: 2009/01/13_03:42:19 notice: native_print:     drbddisk_2 
> (heartbeat:drbddisk):   Started server1
> pengine[4460]: 2009/01/13_03:42:19 notice: native_print:     Filesystem_3     
>   (heartbeat::ocf:Filesystem):    Started server1
> pengine[4460]: 2009/01/13_03:42:19 notice: native_print:     drbddisk_4 
> (heartbeat:drbddisk):   Started server1
> pengine[4460]: 2009/01/13_03:42:19 notice: native_print:     Filesystem_5     
>   (heartbeat::ocf:Filesystem):    Started server1
> pengine[4460]: 2009/01/13_03:42:19 notice: native_print:     MailTo_6   
> (heartbeat::ocf:MailTo):        Started server1
> pengine[4460]: 2009/01/13_03:42:19 notice: native_print: server1-fencing      
>   (stonith:external/ipmi)
> pengine[4460]: 2009/01/13_03:42:19 notice: native_print:        0 : server2
> pengine[4460]: 2009/01/13_03:42:19 notice: native_print:        1 : server1
> pengine[4460]: 2009/01/13_03:42:19 notice: native_print: server2-fencing      
>   (stonith:external/ipmi):        Started server2
> pengine[4460]: 2009/01/13_03:42:19 notice: NoRoleChange: Leave resource 
> commonIP        (server1)
> pengine[4460]: 2009/01/13_03:42:19 notice: NoRoleChange: Leave resource 
> drbddisk_2      (server1)
> pengine[4460]: 2009/01/13_03:42:19 notice: NoRoleChange: Leave resource 
> Filesystem_3    (server1)
> pengine[4460]: 2009/01/13_03:42:19 notice: NoRoleChange: Leave resource 
> drbddisk_4      (server1)
> pengine[4460]: 2009/01/13_03:42:19 notice: NoRoleChange: Leave resource 
> Filesystem_5    (server1)
> pengine[4460]: 2009/01/13_03:42:19 notice: NoRoleChange: Leave resource 
> MailTo_6        (server1)
> pengine[4460]: 2009/01/13_03:42:19 ERROR: native_create_actions: Attempting 
> recovery of resource server1-fencing
> pengine[4460]: 2009/01/13_03:42:19 notice: StopRsc:   server2   Stop 
> server1-fencing
> pengine[4460]: 2009/01/13_03:42:19 notice: StopRsc:   server1   Stop 
> server1-fencing
> pengine[4460]: 2009/01/13_03:42:19 notice: StartRsc:  server1   Start 
> server1-fencing
> pengine[4460]: 2009/01/13_03:42:19 notice: RecurringOp: server1    
> server1-fencing_monitor_120000
> pengine[4460]: 2009/01/13_03:42:19 WARN: native_color: Resource 
> server2-fencing cannot run anywhere
> pengine[4460]: 2009/01/13_03:42:19 notice: StopRsc:   server2   Stop 
> server2-fencing
> pengine[4460]: 2009/01/13_03:42:19 WARN: stage6: Scheduling Node server2 for 
> STONITH
> pengine[4460]: 2009/01/13_03:42:19 WARN: native_stop_constraints: Stop of 
> failed resource server1-fencing is implicit after server2 is fenced
> pengine[4460]: 2009/01/13_03:42:19 info: native_stop_constraints: 
> server2-fencing_stop_0 is implicit after server2 is fenced
> crmd[3623]: 2009/01/13_03:42:19 info: do_state_transition: State transition 
> S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS 
> cause=C_IPC_MESSAGE origin=route_message ]
> tengine[4459]: 2009/01/13_03:42:19 info: unpack_graph: Unpacked transition 
> 33: 8 actions in 8 synapses
> tengine[4459]: 2009/01/13_03:42:19 info: te_pseudo_action: Pseudo action 2 
> fired and confirmed
> tengine[4459]: 2009/01/13_03:42:19 info: te_pseudo_action: Pseudo action 32 
> fired and confirmed
> tengine[4459]: 2009/01/13_03:42:19 notice: run_graph: 
> ====================================================
> tengine[4459]: 2009/01/13_03:42:19 WARN: run_graph: Transition 33: 
> (Complete=2, Pending=0, Fired=0, Skipped=0, Incomplete=6)
> tengine[4459]: 2009/01/13_03:42:19 ERROR: te_graph_trigger: Transition 
> failed: terminated
> tengine[4459]: 2009/01/13_03:42:19 WARN: print_graph: Graph 33 (8 actions in 
> 8 synapses): batch-limit=30 jobs, network-delay=60000ms
> tengine[4459]: 2009/01/13_03:42:19 WARN: print_graph: Synapse 0 was confirmed 
> (priority: 0)
> tengine[4459]: 2009/01/13_03:42:19 WARN: print_graph: Synapse 1 is pending 
> (priority: 0)
> tengine[4459]: 2009/01/13_03:42:19 WARN: print_elem:     [Action 8]: Pending 
> (id: server1-fencing_monitor_120000, loc: server1, priority: 0)
> tengine[4459]: 2009/01/13_03:42:19 WARN: print_elem:      * [Input 31]: 
> Pending (id: server1-fencing_start_0, loc: server1, priority: 0)
> tengine[4459]: 2009/01/13_03:42:19 WARN: print_graph: Synapse 2 is pending 
> (priority: 0)
> tengine[4459]: 2009/01/13_03:42:19 WARN: print_elem:     [Action 30]: Pending 
> (id: server1-fencing_stop_0, loc: server1, priority: 0)
> tengine[4459]: 2009/01/13_03:42:19 WARN: print_elem:      * [Input 13]: 
> Pending (id: all_stopped, type: pseduo, priority: 0)
> tengine[4459]: 2009/01/13_03:42:19 WARN: print_graph: Synapse 3 is pending 
> (priority: 0)
> tengine[4459]: 2009/01/13_03:42:19 WARN: print_elem:     [Action 31]: Pending 
> (id: server1-fencing_start_0, loc: server1, priority: 0)
> tengine[4459]: 2009/01/13_03:42:19 WARN: print_elem:      * [Input 2]: 
> Completed (id: server1-fencing_stop_0, type: pseduo, priority: 0)
> tengine[4459]: 2009/01/13_03:42:19 WARN: print_elem:      * [Input 30]: 
> Pending (id: server1-fencing_stop_0, loc: server1, priority: 0)
> tengine[4459]: 2009/01/13_03:42:19 WARN: print_graph: Synapse 4 was confirmed 
> (priority: 0)
> tengine[4459]: 2009/01/13_03:42:19 WARN: print_graph: Synapse 5 is pending 
> (priority: 0)
> tengine[4459]: 2009/01/13_03:42:19 WARN: print_elem:     [Action 13]: Pending 
> (id: all_stopped, type: pseduo, priority: 0)
> tengine[4459]: 2009/01/13_03:42:19 WARN: print_elem:      * [Input 34]: 
> Pending (id: stonith, loc: server2, type: crm, priority: 0)
> tengine[4459]: 2009/01/13_03:42:19 WARN: print_graph: Synapse 6 is pending 
> (priority: 0)
> tengine[4459]: 2009/01/13_03:42:19 WARN: print_elem:     [Action 33]: Pending 
> (id: stonith_up, type: pseduo, priority: 0)
> tengine[4459]: 2009/01/13_03:42:19 WARN: print_elem:      * [Input 31]: 
> Pending (id: server1-fencing_start_0, loc: server1, priority: 0)
> tengine[4459]: 2009/01/13_03:42:19 WARN: print_graph: Synapse 7 is pending 
> (priority: 0)
> tengine[4459]: 2009/01/13_03:42:19 WARN: print_elem:     [Action 34]: Pending 
> (id: stonith, loc: server2, type: crm, priority: 0)
> tengine[4459]: 2009/01/13_03:42:19 WARN: print_elem:      * [Input 33]: 
> Pending (id: stonith_up, type: pseduo, priority: 0)
> tengine[4459]: 2009/01/13_03:42:19 info: notify_crmd: Transition 33 status: 
> te_complete - <null>
> crmd[3623]: 2009/01/13_03:42:19 info: do_state_transition: State transition 
> S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS cause=C_IPC_MESSAGE 
> origin=route_message ]
> pengine[4460]: 2009/01/13_03:42:19 ERROR: process_pe_message: Transition 33: 
> ERRORs found during PE processing. PEngine Input stored in: 
> /var/lib/heartbeat/pengine/pe-error-22.bz2
> pengine[4460]: 2009/01/13_03:42:19 info: process_pe_message: Configuration 
> WARNINGs found during PE processing.  Please run "crm_verify -L" to identify 
> issues.
> cib[3619]: 2009/01/13_03:42:19 info: cib_stats: Processed 47 operations 
> (4680.00us average, 0% utilization) in the last 10min
> tengine[4459]: 2009/01/13_03:42:38 info: process_graph_event: Detected action 
> server1-fencing_monitor_120000 from a different transition: 29 vs. 33
> tengine[4459]: 2009/01/13_03:42:38 info: update_abort_priority: Abort 
> priority upgraded to 1000000
> crmd[3623]: 2009/01/13_03:42:38 info: do_state_transition: State transition 
> S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC cause=C_IPC_MESSAGE 
> origin=route_message ]
> crmd[3623]: 2009/01/13_03:42:38 info: do_state_transition: All 2 cluster 
> nodes are eligible to run resources.
> pengine[4460]: 2009/01/13_03:42:38 info: determine_online_status: Node 
> server2 is online
> pengine[4460]: 2009/01/13_03:42:38 info: determine_online_status: Node 
> server1 is online
> pengine[4460]: 2009/01/13_03:42:38 ERROR: native_add_running: Resource 
> stonith::external/ipmi:server1-fencing appears to be active on 2 nodes.
> pengine[4460]: 2009/01/13_03:42:38 ERROR: See 
> http://linux-ha.org/v2/faq/resource_too_active for more information.
> pengine[4460]: 2009/01/13_03:42:38 ERROR: unpack_rsc_op: Remapping 
> server2-fencing_start_0 (rc=1) on server1 to an ERROR
> pengine[4460]: 2009/01/13_03:42:38 WARN: unpack_rsc_op: Processing failed op 
> server2-fencing_start_0 on server1: Error
> pengine[4460]: 2009/01/13_03:42:38 WARN: unpack_rsc_op: Compatability 
> handling for failed op server2-fencing_start_0 on server1
> pengine[4460]: 2009/01/13_03:42:38 notice: group_print: Resource Group: 
> group_1
> pengine[4460]: 2009/01/13_03:42:38 notice: native_print:     commonIP   
> (heartbeat::ocf:IPaddr):        Started server1
> pengine[4460]: 2009/01/13_03:42:38 notice: native_print:     drbddisk_2 
> (heartbeat:drbddisk):   Started server1
> pengine[4460]: 2009/01/13_03:42:38 notice: native_print:     Filesystem_3     
>   (heartbeat::ocf:Filesystem):    Started server1
> pengine[4460]: 2009/01/13_03:42:38 notice: native_print:     drbddisk_4 
> (heartbeat:drbddisk):   Started server1
> pengine[4460]: 2009/01/13_03:42:38 notice: native_print:     Filesystem_5     
>   (heartbeat::ocf:Filesystem):    Started server1
> pengine[4460]: 2009/01/13_03:42:38 notice: native_print:     MailTo_6   
> (heartbeat::ocf:MailTo):        Started server1
> pengine[4460]: 2009/01/13_03:42:38 notice: native_print: server1-fencing      
>   (stonith:external/ipmi)
> pengine[4460]: 2009/01/13_03:42:38 notice: native_print:        0 : server2
> pengine[4460]: 2009/01/13_03:42:38 notice: native_print:        1 : server1
> pengine[4460]: 2009/01/13_03:42:38 notice: native_print: server2-fencing      
>   (stonith:external/ipmi):        Started server2
> pengine[4460]: 2009/01/13_03:42:38 notice: NoRoleChange: Leave resource 
> commonIP        (server1)
> pengine[4460]: 2009/01/13_03:42:38 notice: NoRoleChange: Leave resource 
> drbddisk_2      (server1)
> pengine[4460]: 2009/01/13_03:42:38 notice: NoRoleChange: Leave resource 
> Filesystem_3    (server1)
> pengine[4460]: 2009/01/13_03:42:38 notice: NoRoleChange: Leave resource 
> drbddisk_4      (server1)
> pengine[4460]: 2009/01/13_03:42:38 notice: NoRoleChange: Leave resource 
> Filesystem_5    (server1)
> pengine[4460]: 2009/01/13_03:42:38 notice: NoRoleChange: Leave resource 
> MailTo_6        (server1)
> pengine[4460]: 2009/01/13_03:42:38 ERROR: native_create_actions: Attempting 
> recovery of resource server1-fencing
> pengine[4460]: 2009/01/13_03:42:38 notice: StopRsc:   server2   Stop 
> server1-fencing
> pengine[4460]: 2009/01/13_03:42:38 notice: StopRsc:   server1   Stop 
> server1-fencing
> pengine[4460]: 2009/01/13_03:42:38 notice: StartRsc:  server2   Start 
> server1-fencing
> pengine[4460]: 2009/01/13_03:42:38 notice: RecurringOp: server2    
> server1-fencing_monitor_120000
> pengine[4460]: 2009/01/13_03:42:38 notice: NoRoleChange: Leave resource 
> server2-fencing (server2)
> crmd[3623]: 2009/01/13_03:42:38 info: do_state_transition: State transition 
> S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS 
> cause=C_IPC_MESSAGE origin=route_message ]
> tengine[4459]: 2009/01/13_03:42:38 info: unpack_graph: Unpacked transition 
> 34: 5 actions in 5 synapses
> tengine[4459]: 2009/01/13_03:42:38 info: te_pseudo_action: Pseudo action 13 
> fired and confirmed
> tengine[4459]: 2009/01/13_03:42:38 info: send_rsc_command: Initiating action 
> 30: server1-fencing_stop_0 on server2
> tengine[4459]: 2009/01/13_03:42:38 info: send_rsc_command: Initiating action 
> 31: server1-fencing_stop_0 on server1
> crmd[3623]: 2009/01/13_03:42:38 info: do_lrm_rsc_op: Performing 
> op=server1-fencing_stop_0 key=31:34:8686f9af-9ced-43ab-bf20-be6e8437abc0)
> lrmd[3620]: 2009/01/13_03:42:38 info: rsc:server1-fencing: stop
> crmd[3623]: 2009/01/13_03:42:38 info: process_lrm_event: LRM operation 
> server1-fencing_monitor_120000 (call=52, rc=-2) Cancelled
> pengine[4460]: 2009/01/13_03:42:38 ERROR: process_pe_message: Transition 34: 
> ERRORs found during PE processing. PEngine Input stored in: 
> /var/lib/heartbeat/pengine/pe-error-23.bz2
> lrmd[28014]: 2009/01/13_03:42:38 info: Try to stop STONITH resource 
> <rsc_id=server1-fencing> : Device=external/ipmi
> crmd[3623]: 2009/01/13_03:42:38 info: process_lrm_event: LRM operation 
> server1-fencing_stop_0 (call=53, rc=0) complete
> tengine[4459]: 2009/01/13_03:42:38 info: match_graph_event: Action 
> server1-fencing_stop_0 (31) confirmed on server1 (rc=0)
> tengine[4459]: 2009/01/13_03:42:39 info: match_graph_event: Action 
> server1-fencing_stop_0 (30) confirmed on server2 (rc=0)
> tengine[4459]: 2009/01/13_03:42:39 info: send_rsc_command: Initiating action 
> 32: server1-fencing_start_0 on server2
> tengine[4459]: 2009/01/13_03:42:41 info: match_graph_event: Action 
> server1-fencing_start_0 (32) confirmed on server2 (rc=0)
> tengine[4459]: 2009/01/13_03:42:41 info: send_rsc_command: Initiating action 
> 1: server1-fencing_monitor_120000 on server2
> tengine[4459]: 2009/01/13_03:42:42 info: match_graph_event: Action 
> server1-fencing_monitor_120000 (1) confirmed on server2 (rc=0)
> tengine[4459]: 2009/01/13_03:42:42 info: run_graph: Transition 34: 
> (Complete=5, Pending=0, Fired=0, Skipped=0, Incomplete=0)
> tengine[4459]: 2009/01/13_03:42:42 info: notify_crmd: Transition 34 status: 
> te_complete - <null>
> crmd[3623]: 2009/01/13_03:42:42 info: do_state_transition: State transition 
> S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS cause=C_IPC_MESSAGE 
> origin=route_message ]
> cib[3619]: 2009/01/13_03:52:19 info: cib_stats: Processed 47 operations 
> (6382.00us average, 0% utilization) in the last 10min
> lrmd[3620]: 2009/01/13_03:59:32 WARN: G_SIG_dispatch: Dispatch function for 
> SIGCHLD was delayed 1000 ms (> 100 ms) before being called (GSource: 0x514c58)
> lrmd[3620]: 2009/01/13_03:59:32 info: G_SIG_dispatch: started at 1733230017 
> should have started at 1733229917
>
> _______________________________________________
> Linux-HA mailing list
> [email protected]
> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> See also: http://linux-ha.org/ReportingProblems
>
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to