On Tue, Oct 15, 2013 at 11:28:14PM +0200, Robert H. wrote: > Hi, > > I finally got it working. > > I had to set cluster-recheck-interval="5m" or some other value and
Right. Failure timeout is only evaluated on the next pengine run, so if nothing else happens, it takes up to recheck-interval ... > had to set failure-timeout to the same value (failure-timeout="5m"). You should set it smaller than the recheck interval, I think. Or you may need to wait *two* recheck intervals for the failure timeout to take effect. Did you try with failure-timeout < recheck-interval? Problems? Lars > This causes a "probe" after 5 minutes and then the cluster shows the > correct state and reevaluates the engine. > > So the very stripped down config is like this: > > primitive mysql-percona lsb:mysql \ > op start enabled="false" interval="0" \ > op stop enabled="false" interval="0" \ > op monitor enabled="true" timeout="20s" interval="10s" \ > op monitor enabled="true" timeout="20s" interval="11s" > role="Stopped" \ > meta migration-threshold="2" failure-timeout="5m" > is-managed="false" > clone CLONE-percona mysql-percona \ > meta clone-max="2" clone-node-max="1" is-managed="false" > property $id="cib-bootstrap-options" \ > .. many more options ... > cluster-recheck-interval="5m" > > With this config and some location constraints, the cluster moves > virtual IP's away from nodes not not having percona running > automatically. > > Thanks for the Tipps, > Robert > > > Am 14.10.2013 12:07, schrieb Robert H.: > >Hi, > > > >one more note: > > > >When I cleanup the ressource, the monitor operation is triggered and > >the result is as expected: > > > >[root@NODE2 ~]# crm_resource --resource mysql-percona --cleanup > >--node NODE2 > >Cleaning up mysql-percona:0 on NODE2 > >Waiting for 1 replies from the CRMd. OK > > > > Clone Set: CLONE-percona [mysql-percona] (unmanaged) > > mysql-percona:0 (lsb:mysql): Started NODE1 (unmanaged) > > mysql-percona:1 (lsb:mysql): Started NODE2 (unmanaged) > > > > > >I assumed that the failure-timeout="xxx" will cause the cleanup to be > >done automatically. Am I wrong ? > > > >Can I tell pacemaker to perform a "cleanup" automatically from time > >to time (I don't want to use cron...) ? > > > >Regards, > >Robert > > > > > >Am 14.10.2013 11:30, schrieb Robert H.: > >>>You probably also want to monitor even if pacemaker thinks this is > >>>supposed to be stopped. > >>> > >>> op monitor interval=11s timeout=20s role=Stopped > >>> > >> > >>I added this: > >> > >>primitive mysql-percona lsb:mysql \ > >> op start enabled="false" interval="0" \ > >> op stop enabled="false" interval="0" \ > >> op monitor enabled="true" timeout="20s" interval="10s" \ > >> op monitor enabled="true" timeout="20s" interval="11s" > >>role="Stopped" \ > >> meta migration-threshold="2" failure-timeout="30s" > >>is-managed="false" > >> > >>However after a reboot of NODE2, the resource stays at: > >> > >> Clone Set: CLONE-percona [mysql-percona] (unmanaged) > >> mysql-percona:0 (lsb:mysql): Started NODE1(unmanaged) > >> Stopped: [ mysql-percona:1 ] > >> > >>But mysql is running: > >> > >>[root@NODE2~]# /etc/init.d/mysql status > >>MySQL (Percona XtraDB Cluster) running (2619) [ OK ] > >>[root@NODE2~]# echo $? > >>0 > >> > >>.. hmm beeing confused :/ > >> > >> > >>>crm_mon reflects what is in the cib. If no-one re-populates > >>>the cib > >>>with the current state of the world, what it shows will be stale. > >> > >>How can I force this ? > >> > >>Regards, > >>Robert > > -- > -- > Robert > > _______________________________________________ > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org -- : Lars Ellenberg : LINBIT | Your Way to High Availability : DRBD/HA support and consulting http://www.linbit.com DRBD® and LINBIT® are registered trademarks of LINBIT, Austria. _______________________________________________ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org