Hi Vladislav, I used the below advisory colocation but its not working. On 3 node setup:
I have configured all 3 resources in clone mode to start only on node1 and node2 with a fail-count of only 1. +++++++++++++++++++++++++++++++++++++++ + crm configure primitive res_dummy_1 lsb::dummy_1 meta allow-migrate=false migration-threshold=1 op monitor interval=5s + crm configure clone dummy_1_clone res_dummy_1 meta clone-max=2 globally-unique=false + crm configure location dummy_1_clone_prefer_node dummy_1_clone -inf: node-3 +++++++++++++++++++++++++++++++++++++++ advisory ordering: + crm configure order 1-BEFORE-2 0: dummy_1_clone dummy_2_clone + crm configure order 2-BEFORE-3 0: dummy_2_clone dummy_3_clone +++++++++++++++++++++++++++++++++++++++ advisory colocation: # crm configure colocation node-with-apps inf: dummy_1_clone dummy_2_clone dummy_3_clone +++++++++++++++++++++++++++++++++++++++ After I killed dummy_1 on node1 , i expected the pacemaker to kill dummy_2 and dummy_3 on node1 and not to disturb the apps on node2. But with above colocation rule, it stopped the apps on node1 but it restarted dummy_2 and dummy_3 on node2. With a score of 0: it didn't stop dummy_2 and dummy_3 on node1. With a score of 500: it stopped only dummy_2 and restarted dummy_2 on node2. Thanks Eswar On Fri, Dec 6, 2013 at 12:20 PM, ESWAR RAO <eswar7...@gmail.com> wrote: > > Thanks Vladislav. > I will work on that. > > Thanks > Eswar > > On Fri, Dec 6, 2013 at 11:05 AM, Vladislav Bogdanov > <bub...@hoster-ok.com>wrote: > >> 06.12.2013 07:58, ESWAR RAO wrote: >> > Hi All, >> > >> > Can someone help me with below configuration?? >> > >> > I have a 3 node HB setup (node1, node2, node3) which runs HB+pacemaker. >> > I have 3 apps dummy1, dummy2 , dummy3 which needs to be run on only 2 >> > nodes among the 3 nodes. >> > >> > By using the below configuration, I was able to run 3 resources on 2 >> nodes. >> > >> > # crm configure primitive res_dummy1 lsb::dummy1 meta >> > allow-migrate="false" migration-threshold=3 failure-timeout="30s" >> > op monitor interval="5s" >> > # crm configure location app_prefer_node res_dummy1 -inf: node3 >> >> First that comes to mind is that you should put above line below the >> next one and refer app_clone instead of res_dummy1. >> >> > # crm configure clone app_clone res_dummy1 meta clone-max="2" >> > globally-unique="false" >> > >> > >> > I have a dependency order like dummy2 should start after dummy1 and >> > dummy3 should start only after dummy2. >> > >> > For now I am keeping a sleep in the script and starting the resources by >> > using crm. >> > >> > Is there any clean way to have the dependency on the resources so that >> > ordering is maintained while clone is run on bot the nodes?? >> > >> > I have tried with below config but couldn't succeed. >> > # crm configure order dum1-BEFORE-dum2 0: res_dummy1 res_dummy2 >> >> The same is here. >> >> > # crm configure order dum2-BEFORE-dum3 0: res_dummy2 res_dummy3 >> >> So, your example should look like: >> # crm configure primitive res_dummy1 lsb::dummy1 meta >> allow-migrate="false" migration-threshold=3 failure-timeout="30s" >> op monitor interval="5s" >> # crm configure clone app_clone res_dummy1 meta clone-max="2" >> globally-unique="false" >> # crm configure location app_prefer_node app_clone -inf: node3 >> # crm configure order dum1-BEFORE-dum2 0: app_clone res_dummy2 >> # crm configure order dum2-BEFORE-dum3 0: res_dummy2 res_dummy3 >> >> > Instead of group i used order , so that even if 1 app gets restarts >> > others will not be affected. >> >> Yep, advisory ordering is fine for that. >> >> > >> > Also is there any way so that if 1 app fails more than >> > migration-threshold times, we can stop all 3 resources on that node?? >> >> Maybe advisory colocations can do something similar (I'm not sure)? >> >> http://clusterlabs.org/doc/en-US/Pacemaker/1.1/html/Pacemaker_Explained/_advisory_placement.html >> >> You should find correct value for its score (positive or negative) >> though. crm_simulate is your friend for that. >> >> > >> > Thanks >> > Eswar >> > >> > >> > _______________________________________________ >> > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >> > http://oss.clusterlabs.org/mailman/listinfo/pacemaker >> > >> > Project Home: http://www.clusterlabs.org >> > Getting started: >> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >> > Bugs: http://bugs.clusterlabs.org >> > >> >> >> _______________________________________________ >> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >> >> Project Home: http://www.clusterlabs.org >> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >> Bugs: http://bugs.clusterlabs.org >> > >
_______________________________________________ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org