On Wed, Oct 6, 2010 at 10:21 PM, Craig Hurley <li...@thehurley.com> wrote: > I tried using ping instead of pingd and I added "number" to the > evaluation, I get the same results :/ > > primitive p_ping ocf:pacemaker:ping params host_list=172.20.0.254 > clone c_ping p_ping meta globally-unique=false > location loc_ping g_cluster_services rule -inf: not_defined p_ping or > p_ping number:lte 0
The attribute you check for here has to match the one set by your resource. Don't use the name of the resource - try Vadym's suggestion. > > Regards, > Craig. > > > On 6 October 2010 20:43, Jayakrishnan <jayakrishnan...@gmail.com> wrote: >> >> Hello, >> >> Guess the change:-- >> location loc_pingd g_cluster_services rule -inf: not_defined pingd or pingd >> number:lte 0 >> >> should work >> >> >> -- >> Regards, >> >> Jayakrishnan. L >> >> Visit: >> www.foralllinux.blogspot.com >> www.jayakrishnan.bravehost.com >> >> >> On Wed, Oct 6, 2010 at 11:56 AM, Claus Denk <d...@us.es> wrote: >>> >>> I am having a similar problem, so let's wait for the experts, But in the >>> meanwhile, try changing >>> >>> >>> location loc_pingd g_cluster_services rule -inf: not_defined p_pingd >>> or p_pingd lte 0 >>> >>> to >>> >>> location loc_pingd g_cluster_services rule -inf: not_defined pingd >>> or pingd number:lte 0 >>> >>> and see what happens. As far as I have read, it is also more recommended >>> to use the "ping" >>> resource instead of "pingd"... >>> >>> kind regards, Claus >>> >>> >>> >>> >>> >>> >>> On 10/06/2010 05:45 AM, Craig Hurley wrote: >>>> >>>> Hello, >>>> >>>> I have a 2 node cluster, running DRBD, heartbeat and pacemaker in >>>> active/passive mode. On both nodes, eth0 is connected to the main >>>> network, eth1 is used to connect the nodes directly to each other. >>>> The nodes share a virtual IP address on eth0. Pacemaker is also >>>> controlling a custom service with an LSB compliant script in >>>> /etc/init.d/. All of this is working fine and I'm happy with it. >>>> >>>> I'd like to configure the nodes so that they fail over if eth0 goes >>>> down (or if they cannot access a particular gateway), so I tried >>>> adding the following (as per >>>> http://www.clusterlabs.org/wiki/Example_configurations#Set_up_pingd) >>>> >>>> primitive p_pingd ocf:pacemaker:pingd params host_list=172.20.0.254 op >>>> monitor interval=15s timeout=5s >>>> clone c_pingd p_pingd meta globally-unique=false >>>> location loc_pingd g_cluster_services rule -inf: not_defined p_pingd >>>> or p_pingd lte 0 >>>> >>>> ... but when I do add that, all resource are stopped and they don't >>>> come back up on either node. Am I making a basic mistake or do you >>>> need more info from me? >>>> >>>> All help is appreciated, >>>> Craig. >>>> >>>> >>>> pacemaker >>>> Version: 1.0.8+hg15494-2ubuntu2 >>>> >>>> heartbeat >>>> Version: 1:3.0.3-1ubuntu1 >>>> >>>> drbd8-utils >>>> Version: 2:8.3.7-1ubuntu2.1 >>>> >>>> >>>> r...@rpalpha:~$ sudo crm configure show >>>> node $id="32482293-7b0f-466e-b405-c64bcfa2747d" rpalpha >>>> node $id="3f2aac12-05aa-4ac7-b91f-c47fa28efb44" rpbravo >>>> primitive p_drbd_data ocf:linbit:drbd \ >>>> params drbd_resource="data" \ >>>> op monitor interval="30s" >>>> primitive p_fs_data ocf:heartbeat:Filesystem \ >>>> params device="/dev/drbd/by-res/data" directory="/mnt/data" >>>> fstype="ext4" >>>> primitive p_ip ocf:heartbeat:IPaddr2 \ >>>> params ip="172.20.50.3" cidr_netmask="255.255.0.0" nic="eth0" \ >>>> op monitor interval="30s" >>>> primitive p_rp lsb:rp \ >>>> op monitor interval="30s" \ >>>> meta target-role="Started" >>>> group g_cluster_services p_ip p_fs_data p_rp >>>> ms ms_drbd p_drbd_data \ >>>> meta master-max="1" master-node-max="1" clone-max="2" >>>> clone-node-max="1" notify="true" >>>> location loc_preferred_master g_cluster_services inf: rpalpha >>>> colocation colo_mnt_on_master inf: g_cluster_services ms_drbd:Master >>>> order ord_mount_after_drbd inf: ms_drbd:promote g_cluster_services:start >>>> property $id="cib-bootstrap-options" \ >>>> dc-version="1.0.8-042548a451fce8400660f6031f4da6f0223dd5dd" \ >>>> cluster-infrastructure="Heartbeat" \ >>>> no-quorum-policy="ignore" \ >>>> stonith-enabled="false" \ >>>> expected-quorum-votes="2" \ >>>> >>>> >>>> r...@rpalpha:~$ sudo cat /etc/ha.d/ha.cf >>>> node rpalpha >>>> node rpbravo >>>> >>>> keepalive 2 >>>> warntime 5 >>>> deadtime 15 >>>> initdead 60 >>>> >>>> mcast eth0 239.0.0.43 694 1 0 >>>> bcast eth1 >>>> >>>> use_logd yes >>>> autojoin none >>>> crm respawn >>>> >>>> >>>> r...@rpalpha:~$ sudo cat /etc/drbd.conf >>>> global { >>>> usage-count no; >>>> } >>>> common { >>>> protocol C; >>>> >>>> handlers {} >>>> >>>> startup {} >>>> >>>> disk {} >>>> >>>> net { >>>> cram-hmac-alg sha1; >>>> shared-secret "foobar"; >>>> } >>>> >>>> syncer { >>>> verify-alg sha1; >>>> rate 100M; >>>> } >>>> } >>>> resource data { >>>> device /dev/drbd0; >>>> meta-disk internal; >>>> on rpalpha { >>>> disk /dev/mapper/rpalpha-data; >>>> address 192.168.1.1:7789; >>>> } >>>> on rpbravo { >>>> disk /dev/mapper/rpbravo-data; >>>> address 192.168.1.2:7789; >>>> } >>>> } >>>> > > _______________________________________________ > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: > http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker > _______________________________________________ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker