Re: [Linux-HA] error building resource agents on OSX 10.5
Hi, On Mon, Mar 29, 2010 at 05:08:34PM -0500, Alex Dean wrote: > Running into an error building resource agents on OSX 10.5. Using > glib 2.22.2 (downloaded from MacPorts, installed in /opt) and > libnet-1.1.4 (also installed in /opt). > > cd Cluster-Resource-Agents-000dd8956ad7 > ./autogen.sh > ./configure --prefix=/opt/software/wxweb/opt > make > ... snip ... > gcc -std=gnu99 -DHAVE_CONFIG_H -I. -I../include -I../include -I../ > include -I/opt/software/wxweb/opt/include/glib-2.0 -I/opt/software/ > wxweb/opt/lib/glib-2.0/include -I/opt/software/wxweb/opt/include > -g -O2 -I/opt/software/wxweb/opt/include -ggdb3 -O0 > -fstack-protector-all -Wall -Wbad-function-cast -Wcast-qual > -Wcast-align -Wdeclaration-after-statement -Wendif-labels > -Wfloat-equal -Wformat=2 -Wformat-security -Wformat-nonliteral > -Winline -Wmissing-prototypes -Wmissing-declarations > -Wmissing-format-attribute -Wnested-externs -Wno-long-long > -Wno-strict-aliasing -Wpointer-arith -Wstrict-prototypes -Wwrite- > strings -ansi -D_GNU_SOURCE -DANSI_ONLY -Werror -MT tickle_tcp.o -MD > -MP -MF .deps/tickle_tcp.Tpo -c -o tickle_tcp.o tickle_tcp.c > cc1: warnings being treated as errors > tickle_tcp.c:70: warning: ‘struct iphdr’ declared inside parameter list > tickle_tcp.c:70: warning: its scope is only this definition or > declaration, which is probably not what you want > tickle_tcp.c: In function ‘tcp_checksum’: > tickle_tcp.c:74: error: dereferencing pointer to incomplete type > tickle_tcp.c:75: error: dereferencing pointer to incomplete type > tickle_tcp.c:76: error: dereferencing pointer to incomplete type > tickle_tcp.c:77: error: dereferencing pointer to incomplete type > tickle_tcp.c:78: error: dereferencing pointer to incomplete type > tickle_tcp.c: In function ‘send_tickle_ack’: > tickle_tcp.c:217: error: field ‘ip’ has incomplete type > tickle_tcp.c:237: error: ‘struct tcphdr’ has no member named ‘source’ > tickle_tcp.c:238: error: ‘struct tcphdr’ has no member named ‘dest’ > tickle_tcp.c:239: error: ‘struct tcphdr’ has no member named ‘seq’ > tickle_tcp.c:240: error: ‘struct tcphdr’ has no member named ‘ack_seq’ > tickle_tcp.c:241: error: ‘struct tcphdr’ has no member named ‘ack’ > tickle_tcp.c:243: error: ‘struct tcphdr’ has no member named ‘rst’ > tickle_tcp.c:244: error: ‘struct tcphdr’ has no member named ‘doff’ > tickle_tcp.c:245: error: ‘struct tcphdr’ has no member named ‘window’ > tickle_tcp.c:246: error: ‘struct tcphdr’ has no member named ‘check’ > tickle_tcp.c:254: error: ‘SOL_IP’ undeclared (first use in this > function) > tickle_tcp.c:254: error: (Each undeclared identifier is reported > only once > tickle_tcp.c:254: error: for each function it appears in.) > tickle_tcp.c:282: error: ‘struct tcphdr’ has no member named ‘source’ > tickle_tcp.c:283: error: ‘struct tcphdr’ has no member named ‘dest’ > tickle_tcp.c:284: error: ‘struct tcphdr’ has no member named ‘seq’ > tickle_tcp.c:285: error: ‘struct tcphdr’ has no member named ‘ack_seq’ > tickle_tcp.c:286: error: ‘struct tcphdr’ has no member named ‘ack’ > tickle_tcp.c:288: error: ‘struct tcphdr’ has no member named ‘rst’ > tickle_tcp.c:289: error: ‘struct tcphdr’ has no member named ‘doff’ > tickle_tcp.c:290: error: ‘struct tcphdr’ has no member named ‘window’ > tickle_tcp.c:291: error: ‘struct tcphdr’ has no member named ‘check’ > gmake[2]: *** [tickle_tcp.o] Error 1 > gmake[2]: Leaving directory > `/usr/local/src/linux-ha/Cluster-Resource-Agents-000dd8956ad7/tools' > gmake[1]: *** [all-recursive] Error 1 > gmake[1]: Leaving directory > `/usr/local/src/linux-ha/Cluster-Resource-Agents-000dd8956ad7/tools' > make: *** [all-recursive] Error 1 > > Does anyone recognize this error? Obviously struct iphdr is not available in the development packages. Perhaps that's linux specific, really can't say. Anyway, tickle_tcp.c shouldn't be built if that's the case. Can you please open a bugzilla for this. Thanks, Dejan > thanks, > alex > ___ > Linux-HA mailing list > Linux-HA@lists.linux-ha.org > http://lists.linux-ha.org/mailman/listinfo/linux-ha > See also: http://linux-ha.org/ReportingProblems ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
Re: [Linux-HA] Pb with rpms for epel-5/x86_64
Hi, On Tue, Mar 30, 2010 at 07:32:29AM +0200, Alain.Moulle wrote: > Hi Dejan, > sorry but I'm not used to get efix like this ... could you > please tell me where in fact can I get the fix ? Well, there are no official fixes available here. You have these options: wait until Andrew makes new rpms or compile the pacemaker yourself. Or apply the patches to the installed files. The patch which fixes this problem is here: http://hg.clusterlabs.org/pacemaker/stable-1.0/rev/042548a451fc A patch for another shell issue: http://hg.clusterlabs.org/pacemaker/stable-1.0/rev/422fed9d8776 Thanks, Dejan > Thanks a lot. > Alain > > I do agree with the 3 first ERRORS about STONITH as I have not yet > > > configured anything and specially the stonith resources, > > > but for the last 2 ones about cib-bootstrap-options, I don't understand > > > ... > > > > > > Any help ? > > > > > > Those two options are not in the crmd metadata, so the shell > > whines. I added the two temporarily into the shell in the > > development tree. That could be backported into 1.0. > > > > Thanks, > > > > Dejan > > > > > ___ > Linux-HA mailing list > Linux-HA@lists.linux-ha.org > http://lists.linux-ha.org/mailman/listinfo/linux-ha > See also: http://linux-ha.org/ReportingProblems ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
[Linux-HA] Pb with rpms for epel-5/x86_64
Hi again Moreover, I have another problem with these releases on RHEL5 : I have configured a stonith resource with this command: crm configure primitive *restofencealain6* stonith:external/ipmi params hostname=alain6 ipaddr=12.81.1.7 userid=mylogin passwd=mypasswd interface=lan meta target-role=Stopped and it was successfull, I can see the resource in the cib.xml , but when I tried a location configuration, the id of the resource is not recognized : element rsc_location: validity error : IDREF attribute rsc references an unknown ID *"restofencealain6"* crm_verify[5353]: 2010/03/30_12:36:34 ERROR: main: CIB did not pass DTD/schema validation Errors found during check: config not valid Is there a problem ? (This was working fine with the releases delivered with fedora12) But perhaps it is due the the previous problem I described in the first message of this thread (about cib-bootstrap-options ) ? Don't know. How can I get releases of rpms working fine on RHEL5 , even if releases are older ones (but working with Pacemaker over corosync ) ? Thanks Regards Alain Moullé ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
Re: [Linux-HA] Why does mysld start run again?
Thanks for the reply Florian. I installed from tar ball so am a little unsure of the releases but looking at the READMEs I see this heartbeat-3.0.2 Pacemaker-1-0-17 (I think) They are all fairly recent, I downloaded them fro hg.linux-ha.org about 3 months ago. If you know of a file I can check to be 100% sure of the version # let me know. Here's my configuration: cib.xml: Florian Haas wrote: > Mike, > > the information given reduces us to guesswork. > > - Messaging layer? > - Pacemaker version? > - Glue and agents versions? > - crm configure show? > - Logs? > > Cheers, > Florian > > On 03/30/2010 03:48 AM, mike wrote: > >> So here's the situation: >> >> Node A (primary node) heartbeat up and running a VIP and mysqld >> Node B (secondary node) up and running but heartbeat stopped >> >> I start heartbeat on Node B and expect it to come quickly, which it >> does. I noticed in the logs on Node A that the cluster runs mysql start. >> Why would it do this when mysql is already running there? Doesn't seem >> to make sense to me. >> > > > > > ___ > Linux-HA mailing list > Linux-HA@lists.linux-ha.org > http://lists.linux-ha.org/mailman/listinfo/linux-ha > See also: http://linux-ha.org/ReportingProblems ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
Re: [Linux-HA] Pb with rpms for epel-5/x86_64
Hi, On Tue, Mar 30, 2010 at 01:19:21PM +0200, Alain.Moulle wrote: > Hi again > > Moreover, I have another problem with these releases on RHEL5 : > > I have configured a stonith resource with this command: > crm configure primitive *restofencealain6* stonith:external/ipmi params > hostname=alain6 ipaddr=12.81.1.7 userid=mylogin passwd=mypasswd > interface=lan meta target-role=Stopped > > and it was successfull, I can see the resource in the cib.xml , > > but when I tried a location configuration, the id of the resource is not > recognized : > element rsc_location: validity error : IDREF attribute rsc references an > unknown ID *"restofencealain6"* > crm_verify[5353]: 2010/03/30_12:36:34 ERROR: main: CIB did not pass > DTD/schema validation > Errors found during check: config not valid > > Is there a problem ? (This was working fine with the releases delivered > with fedora12) That's the second issue I mentioned in the previous message. > But perhaps it is due the the previous problem I described in the first > message of this thread > (about cib-bootstrap-options ) ? > > Don't know. > > How can I get releases of rpms working fine on RHEL5 , even if releases > are older ones > (but working with Pacemaker over corosync ) There are still previous releases 1.0.8-1 at http://clusterlabs.org/rpm/epel-5 Thanks, Dejan > ? > > Thanks > Regards > Alain Moullé > > ___ > Linux-HA mailing list > Linux-HA@lists.linux-ha.org > http://lists.linux-ha.org/mailman/listinfo/linux-ha > See also: http://linux-ha.org/ReportingProblems ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
Re: [Linux-HA] Why does mysld start run again?
Also noticed another oddity. I killed mysql on the primary node fully expecting it to either trigger a failover or a restart of mysql on the primary node; I wasn't 100% sure which. Well, nothing happened. I do however see a number of messages like this in the ha-log: Mar 30 08:59:27 DBSUAT1A.intranet.mydomain.com lrmd: [3297]: info: RA output: (mysqld_2:monitor:stderr) Usage: /etc/init.d/mysqld {start|stop|report|restart} mike wrote: > Thanks for the reply Florian. > I installed from tar ball so am a little unsure of the releases but > looking at the READMEs I see this > heartbeat-3.0.2 > Pacemaker-1-0-17 (I think) > > They are all fairly recent, I downloaded them fro hg.linux-ha.org about > 3 months ago. If you know of a file I can check to be 100% sure of the > version # let me know. > Here's my configuration: > cib.xml: > crm_feature_set="3.0.1" have-quorum="1" num_updates="25" > cib-last-written="Mo > n Mar 29 21:55:01 2010" dc-uuid="e99889ee-da15-4b09-bfc7-641e3ac0687f"> > > > > >name="symmetric-cluster" value="true"/> >name="no-quorum-policy" value="stop"/> >name="default-resource-stickiness" value="0"/> >id="cib-bootstrap-options-default-resource-failure-stickiness" > name="default-resource-failure-stickiness" value="0"/> >name="stonith-enabled" value="false"/> >name="stonith-action" value="reboot"/> >name="startup-fencing" value="true"/> >name="stop-orphan-resources" value="true"/> >name="stop-orphan-actions" value="true"/> >name="remove-after-stop" value="false"/> >name="short-resource-names" value="true"/> >name="transition-idle-timeout" value="5min"/> >name="default-action-timeout" value="20s"/> >name="is-managed-default" value="true"/> >name="cluster-delay" value="60s"/> >name="pe-error-series-max" value="-1"/> >name="pe-warn-series-max" value="-1"/> >name="pe-input-series-max" value="-1"/> >name="dc-version" value="1.0.6-17fe0022afda074a937d934b3eb625eccd1f01ef"/> >name="cluster-infrastructure" value="Heartbeat"/> > > > > >uname="dbsuat1b.intranet.mydomain.com" type="normal"/> >uname="dbsuat1a.intranet.mydomain.com" type="normal"/> > > > > type="IPaddr2"> > > timeout="5s"/> > > > >value="172.28.185.49"/> > > > > type="mysqld"> > > timeout="60s"/> > > > > > > > >id="prefered_location_group_1_expr" operation="eq" > value="DBSUAT1A.intranet.mydomain.com"/> > > > > > > uname="dbsuat1b.intranet.mydomain.com" ha="active" in_ccm="true" > crmd="online" join > ="member" expected="member" crm-debug-origin="do_update_resource" > shutdown="0"> > > id="status-e99889ee-da15-4b09-bfc7-641e3ac0687f"> > > id="status-e99889ee-da15-4b09-bfc7-641e3ac0687f-probe_complete" > name="probe_complete" value="true"/> > > > > > >provider="heartbeat"> > crm-debug-origin="build_active_RAs" crm_feature_set="3.0.1" transition-k > ey="4:1:7:443f1faa-26f0-4013-95b1-d0a43e4b7f6a" > transition-magic="0:7;4:1:7:443f1faa-26f0-4013-95b1-d0a43e4b7f6a" > call-id="2" rc-code="7" op- > status="0" interval="0" last-run="1269914318" > last-rc-change="1269914318" exec-time="190" queue-time="10" > op-digest="e6e4647755681224d96a4ba7 > fc1a3391"/> > crm-debug-origin="build_active_RAs" crm_feature_set="3.0.1" transition-key=" > 4:3:0:443f1faa-26f0-4013-95b1-d0a43e4b7f6a" > transition-magic="0:0;4:3:0:443f1faa-26f0-4013-95b1-d0a43e4b7f6a" > call-id="5" rc-code="0" op-stat > us="0" interval="0" last-run="1269914319" last-rc-change="1269914319" > exec-time="110" queue-time="0" op-digest="e6e4647755681224d96a4ba7fc1a3 > 391"/> > crm-debug-origin="build_active_RAs" crm_feature_set="3.0.1" transitio > n-key="5:3:0:443f1faa-26f0-4013-95b1-d0a43e4b7f6a" > transition-magic="0:0;5:3:0:443f1faa-26f0-4013-95b1-d0a43e4b7f6a" > call-id="6" rc-code="0" > op-status="0" interval="5000" last-run="1269914715" > last-rc-change="1269914319" exec-time="80" queue-time="0" > op-digest="8124f1b5e7c7c10bbbf3 > 82d3813c9b90"/> > crm-debug-origin="do_update_resource" crm_feature_set="3.0.1" > transition-key=" > 6:6:0:443f1faa-26f0-4013-95b1-d0a43e4b7f6a" > transition-magic="0:0;6:6:0:443f1faa-26f0-4013-95b1-d0a43e4b7f6a" > call-id="10" rc-code="0" op-sta > tus="0" interval="0" last-run="1269914720" last-rc-change="1269914720" > exec-time="60" queue-time="0" op-digest="e6e4647755681224d96a4ba7fc1a3 > 391"/> > > >
Re: [Linux-HA] Pb with rpms for epel-5/x86_64
Hi Dejan, finally, waiting for new releases, I 've applied both patches you provide , and it works much better ! I'll continue testing on RHEL5 with the last available releases plus your both patches ... Thanks a lot for your quick responses, really. Regards Alain ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
Re: [Linux-HA] Why does mysld start run again?
Hi, On Tue, Mar 30, 2010 at 10:24:59AM -0300, mike wrote: > Also noticed another oddity. I killed mysql on the primary node fully > expecting it to either trigger a failover or a restart of mysql on the > primary node; I wasn't 100% sure which. Well, nothing happened. I do > however see a number of messages like this in the ha-log: > > Mar 30 08:59:27 DBSUAT1A.intranet.mydomain.com lrmd: [3297]: info: RA > output: (mysqld_2:monitor:stderr) Usage: /etc/init.d/mysqld > {start|stop|report|restart} Looks like the script doesn't support the status action. If so, then it can't be used in a cluster. Thanks, Dejan > mike wrote: > > Thanks for the reply Florian. > > I installed from tar ball so am a little unsure of the releases but > > looking at the READMEs I see this > > heartbeat-3.0.2 > > Pacemaker-1-0-17 (I think) > > > > They are all fairly recent, I downloaded them fro hg.linux-ha.org about > > 3 months ago. If you know of a file I can check to be 100% sure of the > > version # let me know. > > Here's my configuration: > > cib.xml: > > > crm_feature_set="3.0.1" have-quorum="1" num_updates="25" > > cib-last-written="Mo > > n Mar 29 21:55:01 2010" dc-uuid="e99889ee-da15-4b09-bfc7-641e3ac0687f"> > > > > > > > > > >> name="symmetric-cluster" value="true"/> > >> name="no-quorum-policy" value="stop"/> > >> name="default-resource-stickiness" value="0"/> > >> id="cib-bootstrap-options-default-resource-failure-stickiness" > > name="default-resource-failure-stickiness" value="0"/> > >> name="stonith-enabled" value="false"/> > >> name="stonith-action" value="reboot"/> > >> name="startup-fencing" value="true"/> > >> name="stop-orphan-resources" value="true"/> > >> name="stop-orphan-actions" value="true"/> > >> name="remove-after-stop" value="false"/> > >> name="short-resource-names" value="true"/> > >> name="transition-idle-timeout" value="5min"/> > >> name="default-action-timeout" value="20s"/> > >> name="is-managed-default" value="true"/> > >> name="cluster-delay" value="60s"/> > >> name="pe-error-series-max" value="-1"/> > >> name="pe-warn-series-max" value="-1"/> > >> name="pe-input-series-max" value="-1"/> > >> name="dc-version" value="1.0.6-17fe0022afda074a937d934b3eb625eccd1f01ef"/> > >> name="cluster-infrastructure" value="Heartbeat"/> > > > > > > > > > >> uname="dbsuat1b.intranet.mydomain.com" type="normal"/> > >> uname="dbsuat1a.intranet.mydomain.com" type="normal"/> > > > > > > > > > type="IPaddr2"> > > > > > timeout="5s"/> > > > > > > > >> value="172.28.185.49"/> > > > > > > > > > type="mysqld"> > > > > > timeout="60s"/> > > > > > > > > > > > > > > > >> id="prefered_location_group_1_expr" operation="eq" > > value="DBSUAT1A.intranet.mydomain.com"/> > > > > > > > > > > > > > uname="dbsuat1b.intranet.mydomain.com" ha="active" in_ccm="true" > > crmd="online" join > > ="member" expected="member" crm-debug-origin="do_update_resource" > > shutdown="0"> > > > > > id="status-e99889ee-da15-4b09-bfc7-641e3ac0687f"> > > > > > id="status-e99889ee-da15-4b09-bfc7-641e3ac0687f-probe_complete" > > name="probe_complete" value="true"/> > > > > > > > > > > > >> provider="heartbeat"> > > > crm-debug-origin="build_active_RAs" crm_feature_set="3.0.1" transition-k > > ey="4:1:7:443f1faa-26f0-4013-95b1-d0a43e4b7f6a" > > transition-magic="0:7;4:1:7:443f1faa-26f0-4013-95b1-d0a43e4b7f6a" > > call-id="2" rc-code="7" op- > > status="0" interval="0" last-run="1269914318" > > last-rc-change="1269914318" exec-time="190" queue-time="10" > > op-digest="e6e4647755681224d96a4ba7 > > fc1a3391"/> > > > crm-debug-origin="build_active_RAs" crm_feature_set="3.0.1" transition-key=" > > 4:3:0:443f1faa-26f0-4013-95b1-d0a43e4b7f6a" > > transition-magic="0:0;4:3:0:443f1faa-26f0-4013-95b1-d0a43e4b7f6a" > > call-id="5" rc-code="0" op-stat > > us="0" interval="0" last-run="1269914319" last-rc-change="1269914319" > > exec-time="110" queue-time="0" op-digest="e6e4647755681224d96a4ba7fc1a3 > > 391"/> > > > crm-debug-origin="build_active_RAs" crm_feature_set="3.0.1" transitio > > n-key="5:3:0:443f1faa-26f0-4013-95b1-d0a43e4b7f6a" > > transition-magic="0:0;5:3:0:443f1faa-26f0-4013-95b1-d0a43e4b7f6a" > > call-id="6" rc-code="0" > > op-status="0" interval="5000" last-run="1269914715" > > last-rc-change="1269914319" exec-time="80" queue-time="
Re: [Linux-HA] Why does mysld start run again?
Thank you Dejan, I tried changing the script so that instead of requiring a "report" it now takes status. Specifically I changed it from this: report' ) "$mysqld_multi" report $2 ;; to this status' ) "$mysqld_multi" report $2 ;; I was hoping this would return a proper status and allow a failover. The messages disappeared in the log file so that was a good start. When I killed mysql on the primary node however there was no failover and crm_mon on both nodes seemed to indicate that mysql was still alive on the primary node. I grabbed this from my log file: Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: unpack_config: Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0 Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: determine_online_status: Node dbsuat1b.intranet.mydomain.com is online Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: unpack_rsc_op: Operation mysqld_2_monitor_0 found resource mysqld_2 active on dbsuat1b.intranet.mydomain.com Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: determine_online_status: Node dbsuat1a.intranet.mydomain.com is online Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: unpack_rsc_op: Operation mysqld_2_monitor_0 found resource mysqld_2 active on dbsuat1a.intranet.mydomain.com Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: group_print: Resource Group: group_1 Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: native_print: IPaddr2_1 (ocf::heartbeat:IPaddr2): Started dbsuat1a.intranet.mydomain.com Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: native_print: mysqld_2(lsb:mysqld): Started dbsuat1a.intranet.mydomain.com Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: LogActions: Leave resource IPaddr2_1 (Started dbsuat1a.intranet.mydomain.com) Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: LogActions: Leave resource mysqld_2(Started dbsuat1a.intranet.mydomain.com) Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: process_pe_message: Transition 7: PEngine Input stored in: /usr/var/lib/pengine/pe-input-801.bz2 Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com crmd: [3300]: info: do_state_transition: State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ] Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: process_pe_message: Configuration WARNINGs found during PE processing. Please run "crm_verify -L" to identify issues. Any ideas? Dejan Muhamedagic wrote: > Hi, > > On Tue, Mar 30, 2010 at 10:24:59AM -0300, mike wrote: > >> Also noticed another oddity. I killed mysql on the primary node fully >> expecting it to either trigger a failover or a restart of mysql on the >> primary node; I wasn't 100% sure which. Well, nothing happened. I do >> however see a number of messages like this in the ha-log: >> >> Mar 30 08:59:27 DBSUAT1A.intranet.mydomain.com lrmd: [3297]: info: RA >> output: (mysqld_2:monitor:stderr) Usage: /etc/init.d/mysqld >> {start|stop|report|restart} >> > > Looks like the script doesn't support the status action. If so, > then it can't be used in a cluster. > > Thanks, > > Dejan > > >> mike wrote: >> >>> Thanks for the reply Florian. >>> I installed from tar ball so am a little unsure of the releases but >>> looking at the READMEs I see this >>> heartbeat-3.0.2 >>> Pacemaker-1-0-17 (I think) >>> >>> They are all fairly recent, I downloaded them fro hg.linux-ha.org about >>> 3 months ago. If you know of a file I can check to be 100% sure of the >>> version # let me know. >>> Here's my configuration: >>> cib.xml: >>> >> crm_feature_set="3.0.1" have-quorum="1" num_updates="25" >>> cib-last-written="Mo >>> n Mar 29 21:55:01 2010" dc-uuid="e99889ee-da15-4b09-bfc7-641e3ac0687f"> >>> >>> >>> >>> >>> >> name="symmetric-cluster" value="true"/> >>> >> name="no-quorum-policy" value="stop"/> >>> >> name="default-resource-stickiness" value="0"/> >>> >> id="cib-bootstrap-options-default-resource-failure-stickiness" >>> name="default-resource-failure-stickiness" value="0"/> >>> >> name="stonith-enabled" value="false"/> >>> >> name="stonith-action" value="reboot"/> >>> >> name="startup-fencing" value="true"/> >>> >> name="stop-orphan-resources" value="true"/> >>> >> name="stop-orphan-actions" value="true"/> >>> >> name="remove-after-stop" value="false"/> >>> >> name="short-resource-names" value="true"/> >>> >> name="transition-idle-timeout" value="5min"/> >>> >> name="default-action-timeout" value="20s"/> >>> >> name="is-managed-defa
Re: [Linux-HA] Why does mysld start run again?
I may have missed it but have you tried the MySQL RA rather than the init script? I've had more success with it. -Original Message- From: linux-ha-boun...@lists.linux-ha.org [mailto:linux-ha-boun...@lists.linux-ha.org] On Behalf Of mike Sent: 30 March 2010 15:42 To: General Linux-HA mailing list Subject: Re: [Linux-HA] Why does mysld start run again? Thank you Dejan, I tried changing the script so that instead of requiring a "report" it now takes status. Specifically I changed it from this: report' ) "$mysqld_multi" report $2 ;; to this status' ) "$mysqld_multi" report $2 ;; I was hoping this would return a proper status and allow a failover. The messages disappeared in the log file so that was a good start. When I killed mysql on the primary node however there was no failover and crm_mon on both nodes seemed to indicate that mysql was still alive on the primary node. I grabbed this from my log file: Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: unpack_config: Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0 Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: determine_online_status: Node dbsuat1b.intranet.mydomain.com is online Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: unpack_rsc_op: Operation mysqld_2_monitor_0 found resource mysqld_2 active on dbsuat1b.intranet.mydomain.com Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: determine_online_status: Node dbsuat1a.intranet.mydomain.com is online Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: unpack_rsc_op: Operation mysqld_2_monitor_0 found resource mysqld_2 active on dbsuat1a.intranet.mydomain.com Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: group_print: Resource Group: group_1 Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: native_print: IPaddr2_1 (ocf::heartbeat:IPaddr2): Started dbsuat1a.intranet.mydomain.com Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: native_print: mysqld_2(lsb:mysqld): Started dbsuat1a.intranet.mydomain.com Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: LogActions: Leave resource IPaddr2_1 (Started dbsuat1a.intranet.mydomain.com) Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: LogActions: Leave resource mysqld_2(Started dbsuat1a.intranet.mydomain.com) Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: process_pe_message: Transition 7: PEngine Input stored in: /usr/var/lib/pengine/pe-input-801.bz2 Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com crmd: [3300]: info: do_state_transition: State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ] Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: process_pe_message: Configuration WARNINGs found during PE processing. Please run "crm_verify -L" to identify issues. Any ideas? Dejan Muhamedagic wrote: > Hi, > > On Tue, Mar 30, 2010 at 10:24:59AM -0300, mike wrote: > >> Also noticed another oddity. I killed mysql on the primary node fully >> expecting it to either trigger a failover or a restart of mysql on the >> primary node; I wasn't 100% sure which. Well, nothing happened. I do >> however see a number of messages like this in the ha-log: >> >> Mar 30 08:59:27 DBSUAT1A.intranet.mydomain.com lrmd: [3297]: info: RA >> output: (mysqld_2:monitor:stderr) Usage: /etc/init.d/mysqld >> {start|stop|report|restart} >> > > Looks like the script doesn't support the status action. If so, > then it can't be used in a cluster. > > Thanks, > > Dejan > > >> mike wrote: >> >>> Thanks for the reply Florian. >>> I installed from tar ball so am a little unsure of the releases but >>> looking at the READMEs I see this >>> heartbeat-3.0.2 >>> Pacemaker-1-0-17 (I think) >>> >>> They are all fairly recent, I downloaded them fro hg.linux-ha.org about >>> 3 months ago. If you know of a file I can check to be 100% sure of the >>> version # let me know. >>> Here's my configuration: >>> cib.xml: >>> >> crm_feature_set="3.0.1" have-quorum="1" num_updates="25" >>> cib-last-written="Mo >>> n Mar 29 21:55:01 2010" dc-uuid="e99889ee-da15-4b09-bfc7-641e3ac0687f"> >>> >>> >>> >>> >>> >> name="symmetric-cluster" value="true"/> >>> >> name="no-quorum-policy" value="stop"/> >>> >> name="default-resource-stickiness" value="0"/> >>> >> id="cib-bootstrap-options-default-resource-failure-stickiness" >>> name="default-resource-failure-stickiness" value="0"/> >>> >> name="stonith-enabled" value="false"/> >>> >> name="stonith-action" value="reboot"/> >>> >> name="startup-fencing" value="true"/> >>> >> name="stop-orphan-r
Re: [Linux-HA] UDP / DHCP / LDIRECTORD
Simon, Currently we are running RHEL 5.3 on everything kernel kernel-2.6.18-128, I would rather test on that kernel if possible, but if I have to roll my own RPM for the 2.6.33 with the patch I can do that as well.. (I know I would have to patch the 2.6.18-128 anyway but it would be easier and keep things more consistent in my env). Either way I can test whatever you need... Thanks, Brian Carpio -Original Message- From: linux-ha-boun...@lists.linux-ha.org [mailto:linux-ha-boun...@lists.linux-ha.org] On Behalf Of Simon Horman Sent: Monday, March 29, 2010 5:19 PM To: General Linux-HA mailing list Subject: Re: [Linux-HA] UDP / DHCP / LDIRECTORD On Mon, Mar 29, 2010 at 11:37:42AM -0700, Brian Carpio wrote: > Simon, > > Thank you very much, this would be very beneficial for us to get > working.. I did try the timeout value as you suggested and if I only send > one request every second (since that appears to be as low as I can make > it) then everything works fine, one DHCP packet is sent to one real > server, the next DHCP packet is sent to the next real server.. > unfortunately I am running tests at about 500 DHCP ACKs per second, so > the timeout never happens and all the packets still go to a single real > server. > > What can I do to help get "Julian Anastasov's" one packet scheduler > patches into the 2.6 kernel as this appears to be exactly what we need? If I ported the patch to 2.6.33 would you be able to test it? Testing would be a big help. ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems No virus found in this incoming message. Checked by AVG - www.avg.com Version: 9.0.791 / Virus Database: 271.1.1/2777 - Release Date: 03/29/10 00:32:00 ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
Re: [Linux-HA] Why does mysld start run again?
Thanks Darren. I'm not sure what you mean by the Mysql RA darren.mans...@opengi.co.uk wrote: > I may have missed it but have you tried the MySQL RA rather than the > init script? I've had more success with it. > > > -Original Message- > From: linux-ha-boun...@lists.linux-ha.org > [mailto:linux-ha-boun...@lists.linux-ha.org] On Behalf Of mike > Sent: 30 March 2010 15:42 > To: General Linux-HA mailing list > Subject: Re: [Linux-HA] Why does mysld start run again? > > Thank you Dejan, > I tried changing the script so that instead of requiring a "report" it > now takes status. Specifically I changed it from this: > report' ) > "$mysqld_multi" report $2 > ;; > > to this > status' ) > "$mysqld_multi" report $2 > ;; > > I was hoping this would return a proper status and allow a failover. The > > messages disappeared in the log file so that was a good start. When I > killed mysql on the primary node however there was no failover and > crm_mon on both nodes seemed to indicate that mysql was still alive on > the primary node. I grabbed this from my log file: > > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: > unpack_config: Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0 > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: > determine_online_status: Node dbsuat1b.intranet.mydomain.com is online > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: > > unpack_rsc_op: Operation mysqld_2_monitor_0 found resource mysqld_2 > active on dbsuat1b.intranet.mydomain.com > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: > determine_online_status: Node dbsuat1a.intranet.mydomain.com is online > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: > > unpack_rsc_op: Operation mysqld_2_monitor_0 found resource mysqld_2 > active on dbsuat1a.intranet.mydomain.com > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: > > group_print: Resource Group: group_1 > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: > > native_print: IPaddr2_1 (ocf::heartbeat:IPaddr2): Started > dbsuat1a.intranet.mydomain.com > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: > > native_print: mysqld_2(lsb:mysqld): Started > dbsuat1a.intranet.mydomain.com > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: > > LogActions: Leave resource IPaddr2_1 (Started > dbsuat1a.intranet.mydomain.com) > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: > > LogActions: Leave resource mysqld_2(Started > dbsuat1a.intranet.mydomain.com) > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: > process_pe_message: Transition 7: PEngine Input stored in: > /usr/var/lib/pengine/pe-input-801.bz2 > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com crmd: [3300]: info: > do_state_transition: State transition S_POLICY_ENGINE -> > S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE > origin=handle_response ] > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: > process_pe_message: Configuration WARNINGs found during PE processing. > Please run "crm_verify -L" to identify issues. > > Any ideas? > > Dejan Muhamedagic wrote: > >> Hi, >> >> On Tue, Mar 30, 2010 at 10:24:59AM -0300, mike wrote: >> >> >>> Also noticed another oddity. I killed mysql on the primary node fully >>> > > >>> expecting it to either trigger a failover or a restart of mysql on >>> > the > >>> primary node; I wasn't 100% sure which. Well, nothing happened. I do >>> however see a number of messages like this in the ha-log: >>> >>> Mar 30 08:59:27 DBSUAT1A.intranet.mydomain.com lrmd: [3297]: info: RA >>> > > >>> output: (mysqld_2:monitor:stderr) Usage: /etc/init.d/mysqld >>> {start|stop|report|restart} >>> >>> >> Looks like the script doesn't support the status action. If so, >> then it can't be used in a cluster. >> >> Thanks, >> >> Dejan >> >> >> >>> mike wrote: >>> >>> Thanks for the reply Florian. I installed from tar ball so am a little unsure of the releases but looking at the READMEs I see this heartbeat-3.0.2 Pacemaker-1-0-17 (I think) They are all fairly recent, I downloaded them fro hg.linux-ha.org > about > 3 months ago. If you know of a file I can check to be 100% sure of > the > version # let me know. Here's my configuration: cib.xml: >>> crm_feature_set="3.0.1" have-quorum="1" num_updates="25" cib-last-written="Mo n Mar 29 21:55:01 2010" > dc-uuid="e99889ee-da15-4b09-bfc7-641e3ac0687f"> > >>> name="symmetric-cluster" value="true"/> >>> name="no-qu
Re: [Linux-HA] Why does mysld start run again?
ocf:heartbeat:mysql I really need to change the examples in the DRBD User's Guide to no longer include any references to LSB agents. Cheers, Florian On 2010-03-30 17:52, mike wrote: > Thanks Darren. I'm not sure what you mean by the Mysql RA > > > darren.mans...@opengi.co.uk wrote: >> I may have missed it but have you tried the MySQL RA rather than the >> init script? I've had more success with it. signature.asc Description: OpenPGP digital signature ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
Re: [Linux-HA] Why does mysld start run again?
Hi, On Tue, Mar 30, 2010 at 12:52:42PM -0300, mike wrote: > Thanks Darren. I'm not sure what you mean by the Mysql RA RA: Resource Agent. Perhaps you need to do a bit of research. Try http://clusterlabs.org/ and http://www.linux-ha.org/ Thanks, Dejan > > darren.mans...@opengi.co.uk wrote: > > I may have missed it but have you tried the MySQL RA rather than the > > init script? I've had more success with it. > > > > > > -Original Message- > > From: linux-ha-boun...@lists.linux-ha.org > > [mailto:linux-ha-boun...@lists.linux-ha.org] On Behalf Of mike > > Sent: 30 March 2010 15:42 > > To: General Linux-HA mailing list > > Subject: Re: [Linux-HA] Why does mysld start run again? > > > > Thank you Dejan, > > I tried changing the script so that instead of requiring a "report" it > > now takes status. Specifically I changed it from this: > > report' ) > > "$mysqld_multi" report $2 > > ;; > > > > to this > > status' ) > > "$mysqld_multi" report $2 > > ;; > > > > I was hoping this would return a proper status and allow a failover. The > > > > messages disappeared in the log file so that was a good start. When I > > killed mysql on the primary node however there was no failover and > > crm_mon on both nodes seemed to indicate that mysql was still alive on > > the primary node. I grabbed this from my log file: > > > > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: > > unpack_config: Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0 > > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: > > determine_online_status: Node dbsuat1b.intranet.mydomain.com is online > > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: > > > > unpack_rsc_op: Operation mysqld_2_monitor_0 found resource mysqld_2 > > active on dbsuat1b.intranet.mydomain.com > > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: > > determine_online_status: Node dbsuat1a.intranet.mydomain.com is online > > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: > > > > unpack_rsc_op: Operation mysqld_2_monitor_0 found resource mysqld_2 > > active on dbsuat1a.intranet.mydomain.com > > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: > > > > group_print: Resource Group: group_1 > > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: > > > > native_print: IPaddr2_1 (ocf::heartbeat:IPaddr2): Started > > dbsuat1a.intranet.mydomain.com > > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: > > > > native_print: mysqld_2(lsb:mysqld): Started > > dbsuat1a.intranet.mydomain.com > > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: > > > > LogActions: Leave resource IPaddr2_1 (Started > > dbsuat1a.intranet.mydomain.com) > > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: > > > > LogActions: Leave resource mysqld_2(Started > > dbsuat1a.intranet.mydomain.com) > > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: > > process_pe_message: Transition 7: PEngine Input stored in: > > /usr/var/lib/pengine/pe-input-801.bz2 > > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com crmd: [3300]: info: > > do_state_transition: State transition S_POLICY_ENGINE -> > > S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE > > origin=handle_response ] > > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: > > process_pe_message: Configuration WARNINGs found during PE processing. > > Please run "crm_verify -L" to identify issues. > > > > Any ideas? > > > > Dejan Muhamedagic wrote: > > > >> Hi, > >> > >> On Tue, Mar 30, 2010 at 10:24:59AM -0300, mike wrote: > >> > >> > >>> Also noticed another oddity. I killed mysql on the primary node fully > >>> > > > > > >>> expecting it to either trigger a failover or a restart of mysql on > >>> > > the > > > >>> primary node; I wasn't 100% sure which. Well, nothing happened. I do > >>> however see a number of messages like this in the ha-log: > >>> > >>> Mar 30 08:59:27 DBSUAT1A.intranet.mydomain.com lrmd: [3297]: info: RA > >>> > > > > > >>> output: (mysqld_2:monitor:stderr) Usage: /etc/init.d/mysqld > >>> {start|stop|report|restart} > >>> > >>> > >> Looks like the script doesn't support the status action. If so, > >> then it can't be used in a cluster. > >> > >> Thanks, > >> > >> Dejan > >> > >> > >> > >>> mike wrote: > >>> > >>> > Thanks for the reply Florian. > I installed from tar ball so am a little unsure of the releases but > looking at the READMEs I see this > heartbeat-3.0.2 > Pacemaker-1-0-17 (I think) > > They are all fairly recent, I downloaded them fro hg.linux-ha.org > > > about > > > 3 months ago. If you know of a file I
Re: [Linux-HA] Why does mysld start run again?
I can see where I have a class of lsb mysql in my cib.xml file. How would I change this to ocf? Sorry but I'm new to this and while I have managed to get a cluster running with a VIP and mysql, the finer details like this escape me and I haven't been able to find a good example that tells me how to do this. Thank you Florian Haas wrote: > ocf:heartbeat:mysql > > I really need to change the examples in the DRBD User's Guide to no > longer include any references to LSB agents. > > Cheers, > Florian > > On 2010-03-30 17:52, mike wrote: > >> Thanks Darren. I'm not sure what you mean by the Mysql RA >> >> >> darren.mans...@opengi.co.uk wrote: >> >>> I may have missed it but have you tried the MySQL RA rather than the >>> init script? I've had more success with it. >>> > > > > > ___ > Linux-HA mailing list > Linux-HA@lists.linux-ha.org > http://lists.linux-ha.org/mailman/listinfo/linux-ha > See also: http://linux-ha.org/ReportingProblems ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
[Linux-HA] DRBD MC / Pacemaker GUI 0.7.1
Hi, This is the next DRBD MC beta release 0.7.1. DRBD MC, that is also a Pacemaker GUI, is a Java application that helps to configure DRBD/Pacemaker/Corosync/Heartbeat clusters. It is compatible with Heartbeat 2.1.3 to the Pacemaker 1.0.8 with both available communication layers and DRBD 8. In this release rather an obscure bug was fixed, where a clone-max resource default would be set to 2, while changing one of the other global options the first 20 seconds after starting the DRBD MC. It would be quite harmless, if setting this default parameter would not prevent IPaddr2 resource agent to work unless it is a clone. In my opinion it makes this default option useless and it should be fixed in Pacemaker, but anyway it shouldn't be set by GUI by mistake. Here is the changelog: VM disk views were added. Start, shutdown, destroy, suspend, resume VM actions were implemented. New icons were added. Skip button in cluster connect dialog was added. failure-timeout was changed to time field. One host cluster is possible now. Editing of global DRBD config was fixed. You can get DRBD MC here: http://www.drbd.org/mc/management-console/ http://oss.linbit.com/drbd-mc/DMC-0.7.1.jar http://oss.linbit.com/drbd-mc/drbd-mc-0.7.1.tar.gz You can start it with help of Java Web-Start or you can download it and start it with "java -Xmx512m -jar DMC-0.7.1.jar" command. Make sure you use the Java from SUN. The openjdk seems to work fine for some time now, but it still runs DRBD MC much slower than the original Java. Rasto Levrinc -- : Dipl-Ing Rastislav Levrinc : DRBD-MC http://www.drbd.org/mc/management-console/ : DRBD/HA support and consulting http://www.linbit.com/ DRBD(R) and LINBIT(R) are registered trademarks of LINBIT, Austria. ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems