Hello, Guess the change:-- location loc_pingd g_cluster_services rule -inf: not_defined pingd or pingdnumber:lte 0
should work -- Regards, Jayakrishnan. L Visit: www.foralllinux.blogspot.com www.jayakrishnan.bravehost.com On Wed, Oct 6, 2010 at 11:56 AM, Claus Denk <d...@us.es> wrote: > I am having a similar problem, so let's wait for the experts, But in the > meanwhile, try changing > > > location loc_pingd g_cluster_services rule -inf: not_defined p_pingd > or p_pingd lte 0 > > to > > location loc_pingd g_cluster_services rule -inf: not_defined pingd > or pingd number:lte 0 > > and see what happens. As far as I have read, it is also more recommended to > use the "ping" > resource instead of "pingd"... > > kind regards, Claus > > > > > > > On 10/06/2010 05:45 AM, Craig Hurley wrote: > >> Hello, >> >> I have a 2 node cluster, running DRBD, heartbeat and pacemaker in >> active/passive mode. On both nodes, eth0 is connected to the main >> network, eth1 is used to connect the nodes directly to each other. >> The nodes share a virtual IP address on eth0. Pacemaker is also >> controlling a custom service with an LSB compliant script in >> /etc/init.d/. All of this is working fine and I'm happy with it. >> >> I'd like to configure the nodes so that they fail over if eth0 goes >> down (or if they cannot access a particular gateway), so I tried >> adding the following (as per >> http://www.clusterlabs.org/wiki/Example_configurations#Set_up_pingd) >> >> primitive p_pingd ocf:pacemaker:pingd params host_list=172.20.0.254 op >> monitor interval=15s timeout=5s >> clone c_pingd p_pingd meta globally-unique=false >> location loc_pingd g_cluster_services rule -inf: not_defined p_pingd >> or p_pingd lte 0 >> >> ... but when I do add that, all resource are stopped and they don't >> come back up on either node. Am I making a basic mistake or do you >> need more info from me? >> >> All help is appreciated, >> Craig. >> >> >> pacemaker >> Version: 1.0.8+hg15494-2ubuntu2 >> >> heartbeat >> Version: 1:3.0.3-1ubuntu1 >> >> drbd8-utils >> Version: 2:8.3.7-1ubuntu2.1 >> >> >> r...@rpalpha:~$ sudo crm configure show >> node $id="32482293-7b0f-466e-b405-c64bcfa2747d" rpalpha >> node $id="3f2aac12-05aa-4ac7-b91f-c47fa28efb44" rpbravo >> primitive p_drbd_data ocf:linbit:drbd \ >> params drbd_resource="data" \ >> op monitor interval="30s" >> primitive p_fs_data ocf:heartbeat:Filesystem \ >> params device="/dev/drbd/by-res/data" directory="/mnt/data" >> fstype="ext4" >> primitive p_ip ocf:heartbeat:IPaddr2 \ >> params ip="172.20.50.3" cidr_netmask="255.255.0.0" nic="eth0" \ >> op monitor interval="30s" >> primitive p_rp lsb:rp \ >> op monitor interval="30s" \ >> meta target-role="Started" >> group g_cluster_services p_ip p_fs_data p_rp >> ms ms_drbd p_drbd_data \ >> meta master-max="1" master-node-max="1" clone-max="2" >> clone-node-max="1" notify="true" >> location loc_preferred_master g_cluster_services inf: rpalpha >> colocation colo_mnt_on_master inf: g_cluster_services ms_drbd:Master >> order ord_mount_after_drbd inf: ms_drbd:promote g_cluster_services:start >> property $id="cib-bootstrap-options" \ >> dc-version="1.0.8-042548a451fce8400660f6031f4da6f0223dd5dd" \ >> cluster-infrastructure="Heartbeat" \ >> no-quorum-policy="ignore" \ >> stonith-enabled="false" \ >> expected-quorum-votes="2" \ >> >> >> r...@rpalpha:~$ sudo cat /etc/ha.d/ha.cf >> node rpalpha >> node rpbravo >> >> keepalive 2 >> warntime 5 >> deadtime 15 >> initdead 60 >> >> mcast eth0 239.0.0.43 694 1 0 >> bcast eth1 >> >> use_logd yes >> autojoin none >> crm respawn >> >> >> r...@rpalpha:~$ sudo cat /etc/drbd.conf >> global { >> usage-count no; >> } >> common { >> protocol C; >> >> handlers {} >> >> startup {} >> >> disk {} >> >> net { >> cram-hmac-alg sha1; >> shared-secret "foobar"; >> } >> >> syncer { >> verify-alg sha1; >> rate 100M; >> } >> } >> resource data { >> device /dev/drbd0; >> meta-disk internal; >> on rpalpha { >> disk /dev/mapper/rpalpha-data; >> address 192.168.1.1:7789; >> } >> on rpbravo { >> disk /dev/mapper/rpbravo-data; >> address 192.168.1.2:7789; >> } >> } >> >> _______________________________________________ >> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >> >> Project Home: http://www.clusterlabs.org >> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >> Bugs: >> http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker >> >> > > > _______________________________________________ > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: > http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker >
_______________________________________________ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker