On Tue, May 11, 2010 at 11:58 AM, Dejan Muhamedagic <deja...@fastmail.fm> wrote:
> Do you see the attribute set in the status section (cibadmin -Ql > | grep -w pingd)? If not, then the problem is with the resource. [r...@ha1 ~]# cibadmin -Ql | grep -w pingd <expression attribute="pingd" id="nfs-group-with-pinggw-expression" operation="not_defined"/> <expression attribute="pingd" id="nfs-group-with-pinggw-expression-0" operation="lte" value="0"/> <nvpair id="status-ha1-pingd" name="pingd" value="100"/> <nvpair id="status-ha2-pingd" name="pingd" value="100"/> Tried to change from pacemaker:ping RA to pacemaker:pingd RA (even if I read that the former should be preferred) while still the iptables rule is in place and prevents ha1 to reach the gw [r...@ha1 ~]# crm resource stop cl-pinggw --> services go down (OK, expected) [r...@ha1 ~]# crm configure delete nfs-group-with-pinggw [r...@ha1 ~]# crm configure delete cl-pinggw [r...@ha1 ~]# crm resource delete nfs-group-with-pinggw -->services restart [r...@ha1 ~]# crm resource stop pinggw [r...@ha1 ~]# crm configure delete pinggw [r...@ha1 ~]# crm configure primitive pinggw ocf:pacemaker:pingd \ > params host_list="192.168.101.1" multiplier="100" \ > op start interval="0" timeout="90" \ > op stop interval="0" timeout="100" [r...@ha1 ~]# crm configure clone cl-pinggw pinggw meta globally-unique="false" Now I correctly have: Migration summary: * Node ha1: pingd=0 * Node ha2: pingd=100 [r...@ha1 ~]# crm configure location nfs-group-with-pinggw nfs-group rule -inf: not_defined pinggw or pinggw lte 0 stop of all But this is another problem I'm trying to solve (it seems to me that having a group where in the order there is before an IPaddr2 resource and then a linbit:drbd reosurce, then I don't get the failover.... in the sense that the node ha1 remains drbd primary and there is no demote/promote.... I will eventually post in separate e-mail) It seems from this test that the pacemaker:ping RA doesn't work for me.... I will stay at pingd for the moment. > > Probably I didn't understand correctly what described at the link: > > > http://www.clusterlabs.org/wiki/Pingd_with_resources_on_different_networks > [1] > > or it is outdated now... and instead of defining two clones it is better > > (aka works) to populate the host_list parameter as described here in case > of > > more networks connected: > > > > > http://www.clusterlabs.org/doc/en-US/Pacemaker/1.1/html/Pacemaker_Explained/ch09s03s03.html > [2] > > The former is when you need to test connectivity on different > networks. I don't know if you need that. > > Ok. In [1] above, it makes sense if I have different resources bound to different networks and I want to prevent the loss on a network to cause unnecessary failover of the other defined resource... Put a case where for some reason I have a single resource that depends on two networks, I can instead simply use [2] with only one clone resource and an extended host_list...
_______________________________________________ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf