Re: [Linux-HA] error building resource agents on OSX 10.5

2010-03-30 Thread Dejan Muhamedagic
Hi,

On Mon, Mar 29, 2010 at 05:08:34PM -0500, Alex Dean wrote:
> Running into an error building resource agents on OSX 10.5.  Using
> glib 2.22.2 (downloaded from MacPorts, installed in /opt) and
> libnet-1.1.4 (also installed in /opt).
> 
> cd Cluster-Resource-Agents-000dd8956ad7
> ./autogen.sh
> ./configure --prefix=/opt/software/wxweb/opt
> make
>   ... snip ...
> gcc -std=gnu99 -DHAVE_CONFIG_H -I. -I../include -I../include -I../
> include  -I/opt/software/wxweb/opt/include/glib-2.0 -I/opt/software/
> wxweb/opt/lib/glib-2.0/include -I/opt/software/wxweb/opt/include
> -g -O2 -I/opt/software/wxweb/opt/include -ggdb3 -O0
> -fstack-protector-all -Wall -Wbad-function-cast -Wcast-qual
> -Wcast-align -Wdeclaration-after-statement -Wendif-labels
> -Wfloat-equal -Wformat=2 -Wformat-security -Wformat-nonliteral
> -Winline -Wmissing-prototypes -Wmissing-declarations
> -Wmissing-format-attribute -Wnested-externs -Wno-long-long
> -Wno-strict-aliasing -Wpointer-arith -Wstrict-prototypes -Wwrite-
> strings -ansi -D_GNU_SOURCE -DANSI_ONLY -Werror -MT tickle_tcp.o -MD
> -MP -MF .deps/tickle_tcp.Tpo -c -o tickle_tcp.o tickle_tcp.c
> cc1: warnings being treated as errors
> tickle_tcp.c:70: warning: ‘struct iphdr’ declared inside parameter list
> tickle_tcp.c:70: warning: its scope is only this definition or
> declaration, which is probably not what you want
> tickle_tcp.c: In function ‘tcp_checksum’:
> tickle_tcp.c:74: error: dereferencing pointer to incomplete type
> tickle_tcp.c:75: error: dereferencing pointer to incomplete type
> tickle_tcp.c:76: error: dereferencing pointer to incomplete type
> tickle_tcp.c:77: error: dereferencing pointer to incomplete type
> tickle_tcp.c:78: error: dereferencing pointer to incomplete type
> tickle_tcp.c: In function ‘send_tickle_ack’:
> tickle_tcp.c:217: error: field ‘ip’ has incomplete type
> tickle_tcp.c:237: error: ‘struct tcphdr’ has no member named ‘source’
> tickle_tcp.c:238: error: ‘struct tcphdr’ has no member named ‘dest’
> tickle_tcp.c:239: error: ‘struct tcphdr’ has no member named ‘seq’
> tickle_tcp.c:240: error: ‘struct tcphdr’ has no member named ‘ack_seq’
> tickle_tcp.c:241: error: ‘struct tcphdr’ has no member named ‘ack’
> tickle_tcp.c:243: error: ‘struct tcphdr’ has no member named ‘rst’
> tickle_tcp.c:244: error: ‘struct tcphdr’ has no member named ‘doff’
> tickle_tcp.c:245: error: ‘struct tcphdr’ has no member named ‘window’
> tickle_tcp.c:246: error: ‘struct tcphdr’ has no member named ‘check’
> tickle_tcp.c:254: error: ‘SOL_IP’ undeclared (first use in this
> function)
> tickle_tcp.c:254: error: (Each undeclared identifier is reported
> only once
> tickle_tcp.c:254: error: for each function it appears in.)
> tickle_tcp.c:282: error: ‘struct tcphdr’ has no member named ‘source’
> tickle_tcp.c:283: error: ‘struct tcphdr’ has no member named ‘dest’
> tickle_tcp.c:284: error: ‘struct tcphdr’ has no member named ‘seq’
> tickle_tcp.c:285: error: ‘struct tcphdr’ has no member named ‘ack_seq’
> tickle_tcp.c:286: error: ‘struct tcphdr’ has no member named ‘ack’
> tickle_tcp.c:288: error: ‘struct tcphdr’ has no member named ‘rst’
> tickle_tcp.c:289: error: ‘struct tcphdr’ has no member named ‘doff’
> tickle_tcp.c:290: error: ‘struct tcphdr’ has no member named ‘window’
> tickle_tcp.c:291: error: ‘struct tcphdr’ has no member named ‘check’
> gmake[2]: *** [tickle_tcp.o] Error 1
> gmake[2]: Leaving directory
> `/usr/local/src/linux-ha/Cluster-Resource-Agents-000dd8956ad7/tools'
> gmake[1]: *** [all-recursive] Error 1
> gmake[1]: Leaving directory
> `/usr/local/src/linux-ha/Cluster-Resource-Agents-000dd8956ad7/tools'
> make: *** [all-recursive] Error 1
> 
> Does anyone recognize this error?

Obviously struct iphdr is not available in the development
packages. Perhaps that's linux specific, really can't say.
Anyway, tickle_tcp.c shouldn't be built if that's the case. Can
you please open a bugzilla for this.

Thanks,

Dejan

> thanks,
> alex



> ___
> Linux-HA mailing list
> Linux-HA@lists.linux-ha.org
> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> See also: http://linux-ha.org/ReportingProblems

___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Re: [Linux-HA] Pb with rpms for epel-5/x86_64

2010-03-30 Thread Dejan Muhamedagic
Hi,

On Tue, Mar 30, 2010 at 07:32:29AM +0200, Alain.Moulle wrote:
> Hi Dejan,
> sorry but I'm not used to get efix like this ...  could you
> please tell me where in fact can I get the fix ?

Well, there are no official fixes available here. You have these
options: wait until Andrew makes new rpms or compile the
pacemaker yourself. Or apply the patches to the installed files.
The patch which fixes this problem is here:

http://hg.clusterlabs.org/pacemaker/stable-1.0/rev/042548a451fc

A patch for another shell issue:

http://hg.clusterlabs.org/pacemaker/stable-1.0/rev/422fed9d8776

Thanks,

Dejan

> Thanks a lot.
> Alain
> > I do agree with the 3 first ERRORS about STONITH as I have not yet 
> > > configured anything and specially the stonith resources,
> > > but for the last 2 ones about cib-bootstrap-options, I don't understand 
> > > ...
> > > 
> > > Any help ?
> >   
> >
> > Those two options are not in the crmd metadata, so the shell
> > whines. I added the two temporarily into the shell in the
> > development tree. That could be backported into 1.0.
> >
> > Thanks,
> >
> > Dejan
> >
> >   
> ___
> Linux-HA mailing list
> Linux-HA@lists.linux-ha.org
> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> See also: http://linux-ha.org/ReportingProblems
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


[Linux-HA] Pb with rpms for epel-5/x86_64

2010-03-30 Thread Alain.Moulle
Hi again

Moreover, I have another problem with these releases on RHEL5 :

I have configured a stonith resource with this command:
crm configure primitive *restofencealain6* stonith:external/ipmi params 
hostname=alain6 ipaddr=12.81.1.7 userid=mylogin passwd=mypasswd 
interface=lan meta target-role=Stopped

and it was successfull, I can see the resource in the cib.xml ,

but when I tried a location configuration, the id of the resource is not 
recognized :
element rsc_location: validity error : IDREF attribute rsc references an 
unknown ID *"restofencealain6"*
crm_verify[5353]: 2010/03/30_12:36:34 ERROR: main: CIB did not pass 
DTD/schema validation
Errors found during check: config not valid

Is there a problem ? (This was working fine with the releases delivered 
with fedora12)

But perhaps it is due the the previous problem I described in the first 
message of this thread
(about cib-bootstrap-options ) ?

Don't know.

How can I get releases of rpms working fine on RHEL5 , even if releases 
are older ones
(but working with Pacemaker over corosync  )

?

Thanks
Regards
Alain Moullé

___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


Re: [Linux-HA] Why does mysld start run again?

2010-03-30 Thread mike
Thanks for the reply Florian.
I installed from tar ball so am a little unsure of the releases but 
looking at the READMEs I see this
heartbeat-3.0.2
Pacemaker-1-0-17 (I think)

They are all fairly recent, I downloaded them fro hg.linux-ha.org about 
3 months ago. If you know of a file I can check to be 100% sure of the 
version # let me know.
Here's my configuration:
cib.xml:

  

  

  
  
  
  
  
  
  
  
  
  
  
  
  
  
  
  
  
  
  
  

  


  
  


  

  

  
  

  

  


  

  

  


  

  

  

  
  

  

  

  

  
  

  




  
  




  

  


  

  




  
  



  

  
  

  

  

  

  


Florian Haas wrote:
> Mike,
>
> the information given reduces us to guesswork.
>
> - Messaging layer?
> - Pacemaker version?
> - Glue and agents versions?
> - crm configure show?
> - Logs?
>
> Cheers,
> Florian
>
> On 03/30/2010 03:48 AM, mike wrote:
>   
>> So here's the situation:
>>
>> Node A (primary node) heartbeat up and running a VIP and mysqld
>> Node B (secondary node) up and running but heartbeat stopped
>>
>> I start heartbeat on Node B and expect it to come quickly, which it 
>> does. I noticed in the logs on Node A that the cluster runs mysql start. 
>> Why would it do this when mysql is already running there? Doesn't seem 
>> to make sense to me.
>> 
>
>   
> 
>
> ___
> Linux-HA mailing list
> Linux-HA@lists.linux-ha.org
> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> See also: http://linux-ha.org/ReportingProblems

___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


Re: [Linux-HA] Pb with rpms for epel-5/x86_64

2010-03-30 Thread Dejan Muhamedagic
Hi,

On Tue, Mar 30, 2010 at 01:19:21PM +0200, Alain.Moulle wrote:
> Hi again
> 
> Moreover, I have another problem with these releases on RHEL5 :
> 
> I have configured a stonith resource with this command:
> crm configure primitive *restofencealain6* stonith:external/ipmi params 
> hostname=alain6 ipaddr=12.81.1.7 userid=mylogin passwd=mypasswd 
> interface=lan meta target-role=Stopped
> 
> and it was successfull, I can see the resource in the cib.xml ,
> 
> but when I tried a location configuration, the id of the resource is not 
> recognized :
> element rsc_location: validity error : IDREF attribute rsc references an 
> unknown ID *"restofencealain6"*
> crm_verify[5353]: 2010/03/30_12:36:34 ERROR: main: CIB did not pass 
> DTD/schema validation
> Errors found during check: config not valid
> 
> Is there a problem ? (This was working fine with the releases delivered 
> with fedora12)

That's the second issue I mentioned in the previous message.

> But perhaps it is due the the previous problem I described in the first 
> message of this thread
> (about cib-bootstrap-options ) ?
> 
> Don't know.
> 
> How can I get releases of rpms working fine on RHEL5 , even if releases 
> are older ones
> (but working with Pacemaker over corosync  )

There are still previous releases 1.0.8-1 at
http://clusterlabs.org/rpm/epel-5

Thanks,

Dejan

> ?
> 
> Thanks
> Regards
> Alain Moullé
> 
> ___
> Linux-HA mailing list
> Linux-HA@lists.linux-ha.org
> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> See also: http://linux-ha.org/ReportingProblems
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


Re: [Linux-HA] Why does mysld start run again?

2010-03-30 Thread mike
Also noticed another oddity. I killed mysql on the primary node fully 
expecting it to either trigger a failover or a restart of mysql on the 
primary node; I wasn't 100% sure which. Well, nothing happened. I do 
however see a number of messages like this in the ha-log:

Mar 30 08:59:27 DBSUAT1A.intranet.mydomain.com lrmd: [3297]: info: RA 
output: (mysqld_2:monitor:stderr) Usage: /etc/init.d/mysqld 
{start|stop|report|restart}

mike wrote:
> Thanks for the reply Florian.
> I installed from tar ball so am a little unsure of the releases but 
> looking at the READMEs I see this
> heartbeat-3.0.2
> Pacemaker-1-0-17 (I think)
>
> They are all fairly recent, I downloaded them fro hg.linux-ha.org about 
> 3 months ago. If you know of a file I can check to be 100% sure of the 
> version # let me know.
> Here's my configuration:
> cib.xml:
>  crm_feature_set="3.0.1" have-quorum="1" num_updates="25" 
> cib-last-written="Mo
> n Mar 29 21:55:01 2010" dc-uuid="e99889ee-da15-4b09-bfc7-641e3ac0687f">
>   
> 
>   
> 
>name="symmetric-cluster" value="true"/>
>name="no-quorum-policy" value="stop"/>
>name="default-resource-stickiness" value="0"/>
>id="cib-bootstrap-options-default-resource-failure-stickiness" 
> name="default-resource-failure-stickiness" value="0"/>
>name="stonith-enabled" value="false"/>
>name="stonith-action" value="reboot"/>
>name="startup-fencing" value="true"/>
>name="stop-orphan-resources" value="true"/>
>name="stop-orphan-actions" value="true"/>
>name="remove-after-stop" value="false"/>
>name="short-resource-names" value="true"/>
>name="transition-idle-timeout" value="5min"/>
>name="default-action-timeout" value="20s"/>
>name="is-managed-default" value="true"/>
>name="cluster-delay" value="60s"/>
>name="pe-error-series-max" value="-1"/>
>name="pe-warn-series-max" value="-1"/>
>name="pe-input-series-max" value="-1"/>
>name="dc-version" value="1.0.6-17fe0022afda074a937d934b3eb625eccd1f01ef"/>
>name="cluster-infrastructure" value="Heartbeat"/>
> 
>   
> 
> 
>uname="dbsuat1b.intranet.mydomain.com" type="normal"/>
>uname="dbsuat1a.intranet.mydomain.com" type="normal"/>
> 
> 
>   
>  type="IPaddr2">
>   
>  timeout="5s"/>
>   
>   
> 
>value="172.28.185.49"/>
> 
>   
> 
>  type="mysqld">
>   
>  timeout="60s"/>
>   
> 
>   
> 
> 
>   
> 
>id="prefered_location_group_1_expr" operation="eq" 
> value="DBSUAT1A.intranet.mydomain.com"/>
> 
>   
> 
>   
>   
>  uname="dbsuat1b.intranet.mydomain.com" ha="active" in_ccm="true" 
> crmd="online" join
> ="member" expected="member" crm-debug-origin="do_update_resource" 
> shutdown="0">
>   
>  id="status-e99889ee-da15-4b09-bfc7-641e3ac0687f">
>   
>  id="status-e99889ee-da15-4b09-bfc7-641e3ac0687f-probe_complete" 
> name="probe_complete" value="true"/>
>   
> 
>   
>   
> 
>provider="heartbeat">
>  crm-debug-origin="build_active_RAs" crm_feature_set="3.0.1" transition-k
> ey="4:1:7:443f1faa-26f0-4013-95b1-d0a43e4b7f6a" 
> transition-magic="0:7;4:1:7:443f1faa-26f0-4013-95b1-d0a43e4b7f6a" 
> call-id="2" rc-code="7" op-
> status="0" interval="0" last-run="1269914318" 
> last-rc-change="1269914318" exec-time="190" queue-time="10" 
> op-digest="e6e4647755681224d96a4ba7
> fc1a3391"/>
>  crm-debug-origin="build_active_RAs" crm_feature_set="3.0.1" transition-key="
> 4:3:0:443f1faa-26f0-4013-95b1-d0a43e4b7f6a" 
> transition-magic="0:0;4:3:0:443f1faa-26f0-4013-95b1-d0a43e4b7f6a" 
> call-id="5" rc-code="0" op-stat
> us="0" interval="0" last-run="1269914319" last-rc-change="1269914319" 
> exec-time="110" queue-time="0" op-digest="e6e4647755681224d96a4ba7fc1a3
> 391"/>
>  crm-debug-origin="build_active_RAs" crm_feature_set="3.0.1" transitio
> n-key="5:3:0:443f1faa-26f0-4013-95b1-d0a43e4b7f6a" 
> transition-magic="0:0;5:3:0:443f1faa-26f0-4013-95b1-d0a43e4b7f6a" 
> call-id="6" rc-code="0"
> op-status="0" interval="5000" last-run="1269914715" 
> last-rc-change="1269914319" exec-time="80" queue-time="0" 
> op-digest="8124f1b5e7c7c10bbbf3
> 82d3813c9b90"/>
>  crm-debug-origin="do_update_resource" crm_feature_set="3.0.1" 
> transition-key="
> 6:6:0:443f1faa-26f0-4013-95b1-d0a43e4b7f6a" 
> transition-magic="0:0;6:6:0:443f1faa-26f0-4013-95b1-d0a43e4b7f6a" 
> call-id="10" rc-code="0" op-sta
> tus="0" interval="0" last-run="1269914720" last-rc-change="1269914720" 
> exec-time="60" queue-time="0" op-digest="e6e4647755681224d96a4ba7fc1a3
> 391"/>
>   
>   
>  

Re: [Linux-HA] Pb with rpms for epel-5/x86_64

2010-03-30 Thread Alain.Moulle
Hi Dejan,

finally, waiting for new releases, I 've applied both patches you 
provide , and it
works much better !

I'll continue testing on RHEL5 with the last available releases plus 
your both patches ...

Thanks a lot for your quick responses, really.
Regards
Alain
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


Re: [Linux-HA] Why does mysld start run again?

2010-03-30 Thread Dejan Muhamedagic
Hi,

On Tue, Mar 30, 2010 at 10:24:59AM -0300, mike wrote:
> Also noticed another oddity. I killed mysql on the primary node fully 
> expecting it to either trigger a failover or a restart of mysql on the 
> primary node; I wasn't 100% sure which. Well, nothing happened. I do 
> however see a number of messages like this in the ha-log:
> 
> Mar 30 08:59:27 DBSUAT1A.intranet.mydomain.com lrmd: [3297]: info: RA 
> output: (mysqld_2:monitor:stderr) Usage: /etc/init.d/mysqld 
> {start|stop|report|restart}

Looks like the script doesn't support the status action. If so,
then it can't be used in a cluster.

Thanks,

Dejan

> mike wrote:
> > Thanks for the reply Florian.
> > I installed from tar ball so am a little unsure of the releases but 
> > looking at the READMEs I see this
> > heartbeat-3.0.2
> > Pacemaker-1-0-17 (I think)
> >
> > They are all fairly recent, I downloaded them fro hg.linux-ha.org about 
> > 3 months ago. If you know of a file I can check to be 100% sure of the 
> > version # let me know.
> > Here's my configuration:
> > cib.xml:
> >  > crm_feature_set="3.0.1" have-quorum="1" num_updates="25" 
> > cib-last-written="Mo
> > n Mar 29 21:55:01 2010" dc-uuid="e99889ee-da15-4b09-bfc7-641e3ac0687f">
> >   
> > 
> >   
> > 
> >> name="symmetric-cluster" value="true"/>
> >> name="no-quorum-policy" value="stop"/>
> >> name="default-resource-stickiness" value="0"/>
> >> id="cib-bootstrap-options-default-resource-failure-stickiness" 
> > name="default-resource-failure-stickiness" value="0"/>
> >> name="stonith-enabled" value="false"/>
> >> name="stonith-action" value="reboot"/>
> >> name="startup-fencing" value="true"/>
> >> name="stop-orphan-resources" value="true"/>
> >> name="stop-orphan-actions" value="true"/>
> >> name="remove-after-stop" value="false"/>
> >> name="short-resource-names" value="true"/>
> >> name="transition-idle-timeout" value="5min"/>
> >> name="default-action-timeout" value="20s"/>
> >> name="is-managed-default" value="true"/>
> >> name="cluster-delay" value="60s"/>
> >> name="pe-error-series-max" value="-1"/>
> >> name="pe-warn-series-max" value="-1"/>
> >> name="pe-input-series-max" value="-1"/>
> >> name="dc-version" value="1.0.6-17fe0022afda074a937d934b3eb625eccd1f01ef"/>
> >> name="cluster-infrastructure" value="Heartbeat"/>
> > 
> >   
> > 
> > 
> >> uname="dbsuat1b.intranet.mydomain.com" type="normal"/>
> >> uname="dbsuat1a.intranet.mydomain.com" type="normal"/>
> > 
> > 
> >   
> >  > type="IPaddr2">
> >   
> >  > timeout="5s"/>
> >   
> >   
> > 
> >> value="172.28.185.49"/>
> > 
> >   
> > 
> >  > type="mysqld">
> >   
> >  > timeout="60s"/>
> >   
> > 
> >   
> > 
> > 
> >   
> > 
> >> id="prefered_location_group_1_expr" operation="eq" 
> > value="DBSUAT1A.intranet.mydomain.com"/>
> > 
> >   
> > 
> >   
> >   
> >  > uname="dbsuat1b.intranet.mydomain.com" ha="active" in_ccm="true" 
> > crmd="online" join
> > ="member" expected="member" crm-debug-origin="do_update_resource" 
> > shutdown="0">
> >   
> >  > id="status-e99889ee-da15-4b09-bfc7-641e3ac0687f">
> >   
> >  > id="status-e99889ee-da15-4b09-bfc7-641e3ac0687f-probe_complete" 
> > name="probe_complete" value="true"/>
> >   
> > 
> >   
> >   
> > 
> >> provider="heartbeat">
> >  > crm-debug-origin="build_active_RAs" crm_feature_set="3.0.1" transition-k
> > ey="4:1:7:443f1faa-26f0-4013-95b1-d0a43e4b7f6a" 
> > transition-magic="0:7;4:1:7:443f1faa-26f0-4013-95b1-d0a43e4b7f6a" 
> > call-id="2" rc-code="7" op-
> > status="0" interval="0" last-run="1269914318" 
> > last-rc-change="1269914318" exec-time="190" queue-time="10" 
> > op-digest="e6e4647755681224d96a4ba7
> > fc1a3391"/>
> >  > crm-debug-origin="build_active_RAs" crm_feature_set="3.0.1" transition-key="
> > 4:3:0:443f1faa-26f0-4013-95b1-d0a43e4b7f6a" 
> > transition-magic="0:0;4:3:0:443f1faa-26f0-4013-95b1-d0a43e4b7f6a" 
> > call-id="5" rc-code="0" op-stat
> > us="0" interval="0" last-run="1269914319" last-rc-change="1269914319" 
> > exec-time="110" queue-time="0" op-digest="e6e4647755681224d96a4ba7fc1a3
> > 391"/>
> >  > crm-debug-origin="build_active_RAs" crm_feature_set="3.0.1" transitio
> > n-key="5:3:0:443f1faa-26f0-4013-95b1-d0a43e4b7f6a" 
> > transition-magic="0:0;5:3:0:443f1faa-26f0-4013-95b1-d0a43e4b7f6a" 
> > call-id="6" rc-code="0"
> > op-status="0" interval="5000" last-run="1269914715" 
> > last-rc-change="1269914319" exec-time="80" queue-time="

Re: [Linux-HA] Why does mysld start run again?

2010-03-30 Thread mike
Thank you Dejan,
I tried changing the script so that instead of requiring a "report" it 
now takes status. Specifically I changed it from this:
report' )
"$mysqld_multi" report $2
;;

to this
status' )
"$mysqld_multi" report $2
;;

I was hoping this would return a proper status and allow a failover. The 
messages disappeared in the log file so that was a good start. When I 
killed mysql on the primary node however there was no failover and 
crm_mon on both nodes seemed to indicate that mysql was still alive on 
the primary node. I grabbed this from my log file:

Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: 
unpack_config: Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: 
determine_online_status: Node dbsuat1b.intranet.mydomain.com is online
Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: 
unpack_rsc_op: Operation mysqld_2_monitor_0 found resource mysqld_2 
active on dbsuat1b.intranet.mydomain.com
Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: 
determine_online_status: Node dbsuat1a.intranet.mydomain.com is online
Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: 
unpack_rsc_op: Operation mysqld_2_monitor_0 found resource mysqld_2 
active on dbsuat1a.intranet.mydomain.com
Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: 
group_print:  Resource Group: group_1
Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: 
native_print:  IPaddr2_1   (ocf::heartbeat:IPaddr2):   Started 
dbsuat1a.intranet.mydomain.com
Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: 
native_print:  mysqld_2(lsb:mysqld):   Started 
dbsuat1a.intranet.mydomain.com
Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: 
LogActions: Leave resource IPaddr2_1   (Started 
dbsuat1a.intranet.mydomain.com)
Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice: 
LogActions: Leave resource mysqld_2(Started 
dbsuat1a.intranet.mydomain.com)
Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: 
process_pe_message: Transition 7: PEngine Input stored in: 
/usr/var/lib/pengine/pe-input-801.bz2
Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com crmd: [3300]: info: 
do_state_transition: State transition S_POLICY_ENGINE -> 
S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE 
origin=handle_response ]
Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: 
process_pe_message: Configuration WARNINGs found during PE processing.  
Please run "crm_verify -L" to identify issues.

Any ideas?

Dejan Muhamedagic wrote:
> Hi,
>
> On Tue, Mar 30, 2010 at 10:24:59AM -0300, mike wrote:
>   
>> Also noticed another oddity. I killed mysql on the primary node fully 
>> expecting it to either trigger a failover or a restart of mysql on the 
>> primary node; I wasn't 100% sure which. Well, nothing happened. I do 
>> however see a number of messages like this in the ha-log:
>>
>> Mar 30 08:59:27 DBSUAT1A.intranet.mydomain.com lrmd: [3297]: info: RA 
>> output: (mysqld_2:monitor:stderr) Usage: /etc/init.d/mysqld 
>> {start|stop|report|restart}
>> 
>
> Looks like the script doesn't support the status action. If so,
> then it can't be used in a cluster.
>
> Thanks,
>
> Dejan
>
>   
>> mike wrote:
>> 
>>> Thanks for the reply Florian.
>>> I installed from tar ball so am a little unsure of the releases but 
>>> looking at the READMEs I see this
>>> heartbeat-3.0.2
>>> Pacemaker-1-0-17 (I think)
>>>
>>> They are all fairly recent, I downloaded them fro hg.linux-ha.org about 
>>> 3 months ago. If you know of a file I can check to be 100% sure of the 
>>> version # let me know.
>>> Here's my configuration:
>>> cib.xml:
>>> >> crm_feature_set="3.0.1" have-quorum="1" num_updates="25" 
>>> cib-last-written="Mo
>>> n Mar 29 21:55:01 2010" dc-uuid="e99889ee-da15-4b09-bfc7-641e3ac0687f">
>>>   
>>> 
>>>   
>>> 
>>>   >> name="symmetric-cluster" value="true"/>
>>>   >> name="no-quorum-policy" value="stop"/>
>>>   >> name="default-resource-stickiness" value="0"/>
>>>   >> id="cib-bootstrap-options-default-resource-failure-stickiness" 
>>> name="default-resource-failure-stickiness" value="0"/>
>>>   >> name="stonith-enabled" value="false"/>
>>>   >> name="stonith-action" value="reboot"/>
>>>   >> name="startup-fencing" value="true"/>
>>>   >> name="stop-orphan-resources" value="true"/>
>>>   >> name="stop-orphan-actions" value="true"/>
>>>   >> name="remove-after-stop" value="false"/>
>>>   >> name="short-resource-names" value="true"/>
>>>   >> name="transition-idle-timeout" value="5min"/>
>>>   >> name="default-action-timeout" value="20s"/>
>>>   >> name="is-managed-defa

Re: [Linux-HA] Why does mysld start run again?

2010-03-30 Thread Darren.Mansell
I may have missed it but have you tried the MySQL RA rather than the
init script? I've had more success with it.


-Original Message-
From: linux-ha-boun...@lists.linux-ha.org
[mailto:linux-ha-boun...@lists.linux-ha.org] On Behalf Of mike
Sent: 30 March 2010 15:42
To: General Linux-HA mailing list
Subject: Re: [Linux-HA] Why does mysld start run again?

Thank you Dejan,
I tried changing the script so that instead of requiring a "report" it 
now takes status. Specifically I changed it from this:
report' )
"$mysqld_multi" report $2
;;

to this
status' )
"$mysqld_multi" report $2
;;

I was hoping this would return a proper status and allow a failover. The

messages disappeared in the log file so that was a good start. When I 
killed mysql on the primary node however there was no failover and 
crm_mon on both nodes seemed to indicate that mysql was still alive on 
the primary node. I grabbed this from my log file:

Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: 
unpack_config: Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: 
determine_online_status: Node dbsuat1b.intranet.mydomain.com is online
Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice:

unpack_rsc_op: Operation mysqld_2_monitor_0 found resource mysqld_2 
active on dbsuat1b.intranet.mydomain.com
Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: 
determine_online_status: Node dbsuat1a.intranet.mydomain.com is online
Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice:

unpack_rsc_op: Operation mysqld_2_monitor_0 found resource mysqld_2 
active on dbsuat1a.intranet.mydomain.com
Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice:

group_print:  Resource Group: group_1
Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice:

native_print:  IPaddr2_1   (ocf::heartbeat:IPaddr2):   Started 
dbsuat1a.intranet.mydomain.com
Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice:

native_print:  mysqld_2(lsb:mysqld):   Started 
dbsuat1a.intranet.mydomain.com
Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice:

LogActions: Leave resource IPaddr2_1   (Started 
dbsuat1a.intranet.mydomain.com)
Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice:

LogActions: Leave resource mysqld_2(Started 
dbsuat1a.intranet.mydomain.com)
Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: 
process_pe_message: Transition 7: PEngine Input stored in: 
/usr/var/lib/pengine/pe-input-801.bz2
Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com crmd: [3300]: info: 
do_state_transition: State transition S_POLICY_ENGINE -> 
S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE 
origin=handle_response ]
Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: 
process_pe_message: Configuration WARNINGs found during PE processing.  
Please run "crm_verify -L" to identify issues.

Any ideas?

Dejan Muhamedagic wrote:
> Hi,
>
> On Tue, Mar 30, 2010 at 10:24:59AM -0300, mike wrote:
>   
>> Also noticed another oddity. I killed mysql on the primary node fully

>> expecting it to either trigger a failover or a restart of mysql on
the 
>> primary node; I wasn't 100% sure which. Well, nothing happened. I do 
>> however see a number of messages like this in the ha-log:
>>
>> Mar 30 08:59:27 DBSUAT1A.intranet.mydomain.com lrmd: [3297]: info: RA

>> output: (mysqld_2:monitor:stderr) Usage: /etc/init.d/mysqld 
>> {start|stop|report|restart}
>> 
>
> Looks like the script doesn't support the status action. If so,
> then it can't be used in a cluster.
>
> Thanks,
>
> Dejan
>
>   
>> mike wrote:
>> 
>>> Thanks for the reply Florian.
>>> I installed from tar ball so am a little unsure of the releases but 
>>> looking at the READMEs I see this
>>> heartbeat-3.0.2
>>> Pacemaker-1-0-17 (I think)
>>>
>>> They are all fairly recent, I downloaded them fro hg.linux-ha.org
about 
>>> 3 months ago. If you know of a file I can check to be 100% sure of
the 
>>> version # let me know.
>>> Here's my configuration:
>>> cib.xml:
>>> >> crm_feature_set="3.0.1" have-quorum="1" num_updates="25" 
>>> cib-last-written="Mo
>>> n Mar 29 21:55:01 2010"
dc-uuid="e99889ee-da15-4b09-bfc7-641e3ac0687f">
>>>   
>>> 
>>>   
>>> 
>>>   >> name="symmetric-cluster" value="true"/>
>>>   >> name="no-quorum-policy" value="stop"/>
>>>   >> name="default-resource-stickiness" value="0"/>
>>>   >> id="cib-bootstrap-options-default-resource-failure-stickiness" 
>>> name="default-resource-failure-stickiness" value="0"/>
>>>   >> name="stonith-enabled" value="false"/>
>>>   >> name="stonith-action" value="reboot"/>
>>>   >> name="startup-fencing" value="true"/>
>>>   >> name="stop-orphan-r

Re: [Linux-HA] UDP / DHCP / LDIRECTORD

2010-03-30 Thread Brian Carpio
Simon,

Currently we are running RHEL 5.3 on everything kernel kernel-2.6.18-128, I 
would rather test on that kernel if possible, but if I have to roll my own RPM 
for the 2.6.33 with the patch I can do that as well.. (I know I would have to 
patch the 2.6.18-128 anyway but it would be easier and keep things more 
consistent in my env). 

Either way I can test whatever you need... 

Thanks,
Brian Carpio

-Original Message-
From: linux-ha-boun...@lists.linux-ha.org 
[mailto:linux-ha-boun...@lists.linux-ha.org] On Behalf Of Simon Horman
Sent: Monday, March 29, 2010 5:19 PM
To: General Linux-HA mailing list
Subject: Re: [Linux-HA] UDP / DHCP / LDIRECTORD

On Mon, Mar 29, 2010 at 11:37:42AM -0700, Brian Carpio wrote:
> Simon,
> 
> Thank you very much, this would be very beneficial for us to get
> working.. I did try the timeout value as you suggested and if I only send
> one request every second (since that appears to be as low as I can make
> it) then everything works fine, one DHCP packet is sent to one real
> server, the next DHCP packet is sent to the next real server..
> unfortunately I am running tests at about 500 DHCP ACKs per second, so
> the timeout never happens and all the packets still go to a single real
> server. 
> 
> What can I do to help get "Julian Anastasov's" one packet scheduler
> patches into the 2.6 kernel as this appears to be exactly what we need?

If I ported the patch to 2.6.33 would you be able to test it?
Testing would be a big help.

___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


No virus found in this incoming message.
Checked by AVG - www.avg.com 
Version: 9.0.791 / Virus Database: 271.1.1/2777 - Release Date: 03/29/10 
00:32:00
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


Re: [Linux-HA] Why does mysld start run again?

2010-03-30 Thread mike
Thanks Darren. I'm not sure what you mean by the Mysql RA


darren.mans...@opengi.co.uk wrote:
> I may have missed it but have you tried the MySQL RA rather than the
> init script? I've had more success with it.
>
>
> -Original Message-
> From: linux-ha-boun...@lists.linux-ha.org
> [mailto:linux-ha-boun...@lists.linux-ha.org] On Behalf Of mike
> Sent: 30 March 2010 15:42
> To: General Linux-HA mailing list
> Subject: Re: [Linux-HA] Why does mysld start run again?
>
> Thank you Dejan,
> I tried changing the script so that instead of requiring a "report" it 
> now takes status. Specifically I changed it from this:
> report' )
> "$mysqld_multi" report $2
> ;;
>
> to this
> status' )
> "$mysqld_multi" report $2
> ;;
>
> I was hoping this would return a proper status and allow a failover. The
>
> messages disappeared in the log file so that was a good start. When I 
> killed mysql on the primary node however there was no failover and 
> crm_mon on both nodes seemed to indicate that mysql was still alive on 
> the primary node. I grabbed this from my log file:
>
> Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: 
> unpack_config: Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
> Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: 
> determine_online_status: Node dbsuat1b.intranet.mydomain.com is online
> Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice:
>
> unpack_rsc_op: Operation mysqld_2_monitor_0 found resource mysqld_2 
> active on dbsuat1b.intranet.mydomain.com
> Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: 
> determine_online_status: Node dbsuat1a.intranet.mydomain.com is online
> Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice:
>
> unpack_rsc_op: Operation mysqld_2_monitor_0 found resource mysqld_2 
> active on dbsuat1a.intranet.mydomain.com
> Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice:
>
> group_print:  Resource Group: group_1
> Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice:
>
> native_print:  IPaddr2_1   (ocf::heartbeat:IPaddr2):   Started 
> dbsuat1a.intranet.mydomain.com
> Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice:
>
> native_print:  mysqld_2(lsb:mysqld):   Started 
> dbsuat1a.intranet.mydomain.com
> Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice:
>
> LogActions: Leave resource IPaddr2_1   (Started 
> dbsuat1a.intranet.mydomain.com)
> Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice:
>
> LogActions: Leave resource mysqld_2(Started 
> dbsuat1a.intranet.mydomain.com)
> Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: 
> process_pe_message: Transition 7: PEngine Input stored in: 
> /usr/var/lib/pengine/pe-input-801.bz2
> Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com crmd: [3300]: info: 
> do_state_transition: State transition S_POLICY_ENGINE -> 
> S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE 
> origin=handle_response ]
> Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: 
> process_pe_message: Configuration WARNINGs found during PE processing.  
> Please run "crm_verify -L" to identify issues.
>
> Any ideas?
>
> Dejan Muhamedagic wrote:
>   
>> Hi,
>>
>> On Tue, Mar 30, 2010 at 10:24:59AM -0300, mike wrote:
>>   
>> 
>>> Also noticed another oddity. I killed mysql on the primary node fully
>>>   
>
>   
>>> expecting it to either trigger a failover or a restart of mysql on
>>>   
> the 
>   
>>> primary node; I wasn't 100% sure which. Well, nothing happened. I do 
>>> however see a number of messages like this in the ha-log:
>>>
>>> Mar 30 08:59:27 DBSUAT1A.intranet.mydomain.com lrmd: [3297]: info: RA
>>>   
>
>   
>>> output: (mysqld_2:monitor:stderr) Usage: /etc/init.d/mysqld 
>>> {start|stop|report|restart}
>>> 
>>>   
>> Looks like the script doesn't support the status action. If so,
>> then it can't be used in a cluster.
>>
>> Thanks,
>>
>> Dejan
>>
>>   
>> 
>>> mike wrote:
>>> 
>>>   
 Thanks for the reply Florian.
 I installed from tar ball so am a little unsure of the releases but 
 looking at the READMEs I see this
 heartbeat-3.0.2
 Pacemaker-1-0-17 (I think)

 They are all fairly recent, I downloaded them fro hg.linux-ha.org
 
> about 
>   
 3 months ago. If you know of a file I can check to be 100% sure of
 
> the 
>   
 version # let me know.
 Here's my configuration:
 cib.xml:
 >>> crm_feature_set="3.0.1" have-quorum="1" num_updates="25" 
 cib-last-written="Mo
 n Mar 29 21:55:01 2010"
 
> dc-uuid="e99889ee-da15-4b09-bfc7-641e3ac0687f">
>   
   
 
   
 
   >>> name="symmetric-cluster" value="true"/>
   >>> name="no-qu

Re: [Linux-HA] Why does mysld start run again?

2010-03-30 Thread Florian Haas
ocf:heartbeat:mysql

I really need to change the examples in the DRBD User's Guide to no
longer include any references to LSB agents.

Cheers,
Florian

On 2010-03-30 17:52, mike wrote:
> Thanks Darren. I'm not sure what you mean by the Mysql RA
> 
> 
> darren.mans...@opengi.co.uk wrote:
>> I may have missed it but have you tried the MySQL RA rather than the
>> init script? I've had more success with it.



signature.asc
Description: OpenPGP digital signature
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Re: [Linux-HA] Why does mysld start run again?

2010-03-30 Thread Dejan Muhamedagic
Hi,

On Tue, Mar 30, 2010 at 12:52:42PM -0300, mike wrote:
> Thanks Darren. I'm not sure what you mean by the Mysql RA

RA: Resource Agent. Perhaps you need to do a bit of research. Try
http://clusterlabs.org/ and http://www.linux-ha.org/

Thanks,

Dejan

> 
> darren.mans...@opengi.co.uk wrote:
> > I may have missed it but have you tried the MySQL RA rather than the
> > init script? I've had more success with it.
> >
> >
> > -Original Message-
> > From: linux-ha-boun...@lists.linux-ha.org
> > [mailto:linux-ha-boun...@lists.linux-ha.org] On Behalf Of mike
> > Sent: 30 March 2010 15:42
> > To: General Linux-HA mailing list
> > Subject: Re: [Linux-HA] Why does mysld start run again?
> >
> > Thank you Dejan,
> > I tried changing the script so that instead of requiring a "report" it 
> > now takes status. Specifically I changed it from this:
> > report' )
> > "$mysqld_multi" report $2
> > ;;
> >
> > to this
> > status' )
> > "$mysqld_multi" report $2
> > ;;
> >
> > I was hoping this would return a proper status and allow a failover. The
> >
> > messages disappeared in the log file so that was a good start. When I 
> > killed mysql on the primary node however there was no failover and 
> > crm_mon on both nodes seemed to indicate that mysql was still alive on 
> > the primary node. I grabbed this from my log file:
> >
> > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: 
> > unpack_config: Node scores: 'red' = -INFINITY, 'yellow' = 0, 'green' = 0
> > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: 
> > determine_online_status: Node dbsuat1b.intranet.mydomain.com is online
> > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice:
> >
> > unpack_rsc_op: Operation mysqld_2_monitor_0 found resource mysqld_2 
> > active on dbsuat1b.intranet.mydomain.com
> > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: 
> > determine_online_status: Node dbsuat1a.intranet.mydomain.com is online
> > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice:
> >
> > unpack_rsc_op: Operation mysqld_2_monitor_0 found resource mysqld_2 
> > active on dbsuat1a.intranet.mydomain.com
> > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice:
> >
> > group_print:  Resource Group: group_1
> > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice:
> >
> > native_print:  IPaddr2_1   (ocf::heartbeat:IPaddr2):   Started 
> > dbsuat1a.intranet.mydomain.com
> > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice:
> >
> > native_print:  mysqld_2(lsb:mysqld):   Started 
> > dbsuat1a.intranet.mydomain.com
> > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice:
> >
> > LogActions: Leave resource IPaddr2_1   (Started 
> > dbsuat1a.intranet.mydomain.com)
> > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: notice:
> >
> > LogActions: Leave resource mysqld_2(Started 
> > dbsuat1a.intranet.mydomain.com)
> > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: 
> > process_pe_message: Transition 7: PEngine Input stored in: 
> > /usr/var/lib/pengine/pe-input-801.bz2
> > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com crmd: [3300]: info: 
> > do_state_transition: State transition S_POLICY_ENGINE -> 
> > S_TRANSITION_ENGINE [ input=I_PE_SUCCESS cause=C_IPC_MESSAGE 
> > origin=handle_response ]
> > Mar 30 10:20:31 DBSUAT1A.intranet.mydomain.com pengine: [15123]: info: 
> > process_pe_message: Configuration WARNINGs found during PE processing.  
> > Please run "crm_verify -L" to identify issues.
> >
> > Any ideas?
> >
> > Dejan Muhamedagic wrote:
> >   
> >> Hi,
> >>
> >> On Tue, Mar 30, 2010 at 10:24:59AM -0300, mike wrote:
> >>   
> >> 
> >>> Also noticed another oddity. I killed mysql on the primary node fully
> >>>   
> >
> >   
> >>> expecting it to either trigger a failover or a restart of mysql on
> >>>   
> > the 
> >   
> >>> primary node; I wasn't 100% sure which. Well, nothing happened. I do 
> >>> however see a number of messages like this in the ha-log:
> >>>
> >>> Mar 30 08:59:27 DBSUAT1A.intranet.mydomain.com lrmd: [3297]: info: RA
> >>>   
> >
> >   
> >>> output: (mysqld_2:monitor:stderr) Usage: /etc/init.d/mysqld 
> >>> {start|stop|report|restart}
> >>> 
> >>>   
> >> Looks like the script doesn't support the status action. If so,
> >> then it can't be used in a cluster.
> >>
> >> Thanks,
> >>
> >> Dejan
> >>
> >>   
> >> 
> >>> mike wrote:
> >>> 
> >>>   
>  Thanks for the reply Florian.
>  I installed from tar ball so am a little unsure of the releases but 
>  looking at the READMEs I see this
>  heartbeat-3.0.2
>  Pacemaker-1-0-17 (I think)
> 
>  They are all fairly recent, I downloaded them fro hg.linux-ha.org
>  
> > about 
> >   
>  3 months ago. If you know of a file I 

Re: [Linux-HA] Why does mysld start run again?

2010-03-30 Thread mike
I can see where I have a class of lsb mysql in my cib.xml file. How 
would I change this to ocf? Sorry but I'm new to this and while I have 
managed to get a cluster running with a VIP and mysql, the finer details 
like this escape me and I haven't been able to find a good example that 
tells me how to do this.

Thank you
Florian Haas wrote:
> ocf:heartbeat:mysql
>
> I really need to change the examples in the DRBD User's Guide to no
> longer include any references to LSB agents.
>
> Cheers,
> Florian
>
> On 2010-03-30 17:52, mike wrote:
>   
>> Thanks Darren. I'm not sure what you mean by the Mysql RA
>>
>>
>> darren.mans...@opengi.co.uk wrote:
>> 
>>> I may have missed it but have you tried the MySQL RA rather than the
>>> init script? I've had more success with it.
>>>   
>
>   
> 
>
> ___
> Linux-HA mailing list
> Linux-HA@lists.linux-ha.org
> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> See also: http://linux-ha.org/ReportingProblems

___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


[Linux-HA] DRBD MC / Pacemaker GUI 0.7.1

2010-03-30 Thread Rasto Levrinc
Hi,

This is the next DRBD MC beta release 0.7.1. DRBD MC, that is also a 
Pacemaker GUI, is a Java application that helps to configure 
DRBD/Pacemaker/Corosync/Heartbeat clusters. It is compatible with Heartbeat 
2.1.3 to the Pacemaker 1.0.8 with both available communication layers and 
DRBD 8.

In this release rather an obscure bug was fixed, where a clone-max resource 
default would be set to 2, while changing one of the other global options 
the first 20 seconds after starting the DRBD MC. It would be quite harmless, 
if setting this default parameter would not prevent IPaddr2 resource agent 
to work unless it is a clone. In my opinion it makes this default option 
useless and it should be fixed in Pacemaker, but anyway it shouldn't be set 
by GUI by mistake.

Here is the changelog:
VM disk views were added.
Start, shutdown, destroy, suspend, resume VM actions were implemented.
New icons were added.
Skip button in cluster connect dialog was added.
failure-timeout was changed to time field.
One host cluster is possible now.
Editing of global DRBD config was fixed.

You can get DRBD MC here:

http://www.drbd.org/mc/management-console/
http://oss.linbit.com/drbd-mc/DMC-0.7.1.jar
http://oss.linbit.com/drbd-mc/drbd-mc-0.7.1.tar.gz

You can start it with help of Java Web-Start or you can download it and 
start it with "java -Xmx512m -jar DMC-0.7.1.jar" command. Make sure you use 
the Java from SUN. The openjdk seems to work fine for some time now, but it 
still runs DRBD MC much slower than the original Java.

Rasto Levrinc

-- 
: Dipl-Ing Rastislav Levrinc
: DRBD-MC http://www.drbd.org/mc/management-console/
: DRBD/HA support and consulting http://www.linbit.com/
DRBD(R) and LINBIT(R) are registered trademarks of LINBIT, Austria.

___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems