Hi,

The Pacemaker_Explained.pdf document says that

" setting of migration-threshold=2 and failure-timeout=60s would cause the 
resource to move to a new node after 2 failures, and allow it to move back 
(depending on the stickiness and constraint scores) after one minute."

Can you please help me understand what will happen on the following scenarios 
in 2 node active passive configuration?

1 - If one resource failed twice within 60s, it will move to the other node.
This is clear to understand.

2 - If one resource failed once and there is no failure within 60s, will the 
pacemaker reset the failcounts of that resource, so that the failcounts are 
tracked freshly?

The failcounts are not reset if the migration-threshold didn't occur within the 
failure-timeout period. Is that a bug in pacemaker-1.0.5-4.1?

Thanks,
Raffi





> -----Original Message-----
> From: [email protected] [mailto:linux-ha-
> [email protected]] On Behalf Of Andreas Kurz
> Sent: Friday, April 13, 2012 2:49 PM
> To: [email protected]
> Subject: Re: [Linux-HA] problem with pind
> 
> On 04/12/2012 02:59 PM, Trujillo Carmona, Antonio wrote:
> >
> > I'm try to configure a cluster and I have problem with pingd.
> > my config is
> > crm(live)configure# show
> > node proxy-00
> > node proxy-01
> > primitive ip-segura ocf:heartbeat:IPaddr2 \
> >     params ip="10.104.16.123" nic="lan" cidr_netmask="19" \
> >     op monitor interval="10" \
> >     meta target-role="Started"
> > primitive pingd ocf:pacemaker:pingd \
> 
> use ocf:pacemaker:ping
> 
> >     params host_list="10.104.16.157" \
> 
> and you have to define a monitor operation.
> 
> Without any constraints to let the cluster react on connectivity changes
> ping resource is useless ... this may help:
> 
> http://www.hastexo.com/resources/hints-and-kinks/network-connectivity-
> check-pacemaker
> 
> Regards,
> Andreas
> 
> --
> Need help with Pacemaker?
> http://www.hastexo.com/now
> 
> >     meta target-role="Started"
> > property $id="cib-bootstrap-options" \
> >     dc-version="1.1.6-9971ebba4494012a93c03b40a2c58ec0eb60f50c" \
> >     cluster-infrastructure="openais" \
> >     stonith-enabled="false" \
> >     no-quorum-policy="ignore" \
> >     expected-quorum-votes="2"
> >
> > crm(live)# status
> > ============
> > Last updated: Thu Apr 12 14:54:21 2012
> > Last change: Thu Apr 12 14:40:00 2012
> > Stack: openais
> > Current DC: proxy-00 - partition WITHOUT quorum
> > Version: 1.1.6-9971ebba4494012a93c03b40a2c58ec0eb60f50c
> > 2 Nodes configured, 2 expected votes
> > 2 Resources configured.
> > ============
> >
> > Online: [ proxy-00 ]
> > OFFLINE: [ proxy-01 ]
> >
> >  ip-segura  (ocf::heartbeat:IPaddr2):       Started proxy-00
> >
> > Failed actions:
> >     pingd:0_monitor_0 (node=proxy-00, call=5, rc=2, status=complete):
> > invalid parameter
> >     pingd_monitor_0 (node=proxy-00, call=8, rc=2, status=complete):
> > invalid parameter
> >
> > crm(live)resource# start pingd
> > crm(live)resource# status
> >  ip-segura  (ocf::heartbeat:IPaddr2) Started
> >  pingd      (ocf::pacemaker:pingd) Stopped
> >
> > and in the system log I got:
> >
> > Apr 12 14:55:18 proxy-00 crm_resource: [27941]: ERROR: unpack_rsc_op:
> > Hard error - pingd:0_last_failure_0 failed with rc=2: Preventing pingd:0
> > from re-starting on proxy-00
> > Apr 12 14:55:18 proxy-00 crm_resource: [27941]: ERROR: unpack_rsc_op:
> > Hard error - pingd_last_failure_0 failed with rc=2: Preventing pingd
> > from re-starting on proxy-00
> >
> > I have stoped node 2 in order to less problem
> >
> > ¿I can't found any reference to this error?
> > ¿Can you help me? please.
> >
> >
> >
> >
> 
> 

_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to