Hi Andrew, > > About 1.0, I ask Mr. Mori for backporting. > > Will you revise 1.1? > > Yes, I have it modified locally. I'll push it out soon.
I confirmed a revision for your PM1.1. * http://hg.clusterlabs.org/pacemaker/1.1/rev/862936c5bca3 I think to have PM1.0 reflect this revision. I ask Mr. Mori for revision, but is not there the problem even if patch revise PM1.0 in the same way? Best Regards, Hideo Yamauchi. --- Andrew Beekhof <and...@beekhof.net> wrote: > On Thu, Dec 2, 2010 at 1:44 AM, <renayama19661...@ybb.ne.jp> wrote: > > Hi Andrew, > > > >> > Can 1.0 reflect this revision? > >> > Because there is influence else, is it impossible? > >> > >> I have no objection to it being added to 1.0, it should be safe. > > > > Thanks. > > > > About 1.0, I ask Mr. Mori for backporting. > > Will you revise 1.1? > > Yes, I have it modified locally. I'll push it out soon. > > > > > Best Regards, > > Hideo Yamauchi. > > > > > > --- Andrew Beekhof <and...@beekhof.net> wrote: > > > >> On Mon, Nov 29, 2010 at 5:11 AM, �<renayama19661...@ybb.ne.jp> > >> wrote: > >> > Hi Andrew, > >> > > >> > Sorry.... > >> > My response was late. > >> > > >> >> I think the smartest thing to do here is drop the cib_scope_local flag > >> >> from -f > >> > > >> > � � � �if(do_force) { > >> > � � � � � � � > �crm_debug("Forcing..."); > >> > /* � � � � � � > >> > �cib_options |= > > cib_scope_local|cib_quorum_override; */ > >> > � � � � � � � > >> > �cib_options |= > > cib_quorum_override; > >> > � � � �} > >> > > >> > > >> > I confirmed movement with you according to a revision. > >> > The resource moves well. > >> > > >> > Can 1.0 reflect this revision? > >> > Because there is influence else, is it impossible? > >> > >> I have no objection to it being added to 1.0, it should be safe. > >> > >> > > >> > Best Regards, > >> > Hideo Yamauchi. > >> > > >> > --- Andrew Beekhof <and...@beekhof.net> wrote: > >> > > >> >> 2010/11/8 �<renayama19661...@ybb.ne.jp>: > >> >> > Hi, > >> >> > > >> >> > By two simple node constitution, it caused trouble(monitor error) in > >> >> > a resource. > >> >> > > >> >> > ============ > >> >> > Last updated: Mon Nov �8 10:16:50 2010 > >> >> > Stack: Heartbeat > >> >> > Current DC: srv02 (f80f87fd-cc09-43c7-80bc-8d9e96de376b) - partition > >> >> > WITHOUT quorum > >> >> > Version: 1.0.9-0a40fd0cb9f2fcedef9d1967115c912314c57438 > >> >> > 2 Nodes configured, unknown expected votes > >> >> > 1 Resources configured. > >> >> > ============ > >> >> > > >> >> > Online: [ srv01 srv02 ] > >> >> > > >> >> > �Resource Group: grpDummy > >> >> > � � prmDummy1-1 � � � > >> >> > �(ocf::heartbeat:Dummy): > >> Started > >> > srv02 > >> >> > � � prmDummy1-2 � � � > >> >> > �(ocf::heartbeat:Dummy): > >> Started > >> > srv02 > >> >> > � � prmDummy1-3 � � � > >> >> > �(ocf::heartbeat:Dummy): > >> Started > >> > srv02 > >> >> > � � prmDummy1-4 � � � > >> >> > �(ocf::heartbeat:Dummy): > >> Started > >> > srv02 > >> >> > > >> >> > Migration summary: > >> >> > * Node srv02: > >> >> > * Node srv01: > >> >> > � prmDummy1-1: migration-threshold=1 fail-count=1 > >> >> > > >> >> > Failed actions: > >> >> > � �prmDummy1-1_monitor_30000 (node=srv01, call=7, rc=7, > >> >> > status=complete): > not > >> > running > >> >> > > >> >> > > >> >> > I carried out the next command consecutively after a resource > >> >> > exceeded a fail-over. > >> >> > > >> >> > [r...@srv01 ~]# crm_resource -C -r prmDummy1-1 -N srv01;crm_resource > >> >> > -M -r grpDummy -N > >> srv01 > >> >> -f -Q > >> >> > > >> >> > ============ > >> >> > Last updated: Mon Nov �8 10:17:33 2010 > >> >> > Stack: Heartbeat > >> >> > Current DC: srv02 (f80f87fd-cc09-43c7-80bc-8d9e96de376b) - partition > >> >> > WITHOUT quorum > >> >> > Version: 1.0.9-0a40fd0cb9f2fcedef9d1967115c912314c57438 > >> >> > 2 Nodes configured, unknown expected votes > >> >> > 1 Resources configured. > >> >> > ============ > >> >> > > >> >> > Online: [ srv01 srv02 ] > >> >> > > >> >> > �Resource Group: grpDummy > >> >> > � � prmDummy1-1 � � � > >> >> > �(ocf::heartbeat:Dummy): > >> Started > >> > srv02 > >> >> > � � prmDummy1-2 � � � > >> >> > �(ocf::heartbeat:Dummy): > >> Started > >> > srv02 > >> >> > � � prmDummy1-3 � � � > >> >> > �(ocf::heartbeat:Dummy): > >> Started > >> > srv02 > >> >> > � � prmDummy1-4 � � � > >> >> > �(ocf::heartbeat:Dummy): > >> Started > >> > srv02 > >> >> > > >> >> > Migration summary: > >> >> > * Node srv02: > >> >> > * Node srv01: > >> >> > > >> >> > But, the resource does not move to a srv01 node. > >> >> > > >> >> > Does the "crm_resource -M" command have to carry it out after waiting > >> >> > for a S_IDLE > state? > >> >> > > >> >> > Or is this phenomenon a bug? > >> >> > > >> >> > �* I attach a collection of hb_report file > >> >> > >> >> So the problem here is that not only does -f �enable logic in > >> >> move_resource(), but also > >> >> > >> >> � � � � � � � > >> >> cib_options |= > > cib_scope_local|cib_quorum_override; > >> >> > >> >> Combined with the fact that crm_resource -C is not synchronous in 1.0, > >> >> if you run -M on a non-DC node, the updates hit the local cib while > >> >> the cluster is re-probing the resource(s). > >> >> This results in the two CIBs getting out of sync: > >> >> Nov �8 10:17:15 srv01 crmd: [5367]: WARN: cib_native_callback: > >> >> CIB > >> >> command failed: Application of an update diff failed > >> >> Nov �8 10:17:15 srv01 crmd: [5367]: WARN: cib_native_callback: > >> >> CIB > >> >> command failed: Application of an update diff failed > >> >> > >> >> and the process of re-syncing them results in the behavior you saw. > >> >> > >> >> I think the smartest thing to do here is drop the cib_scope_local flag > >> >> from -f > >> >> > >> >> _______________________________________________ > >> >> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > >> >> http://oss.clusterlabs.org/mailman/listinfo/pacemaker > >> >> > >> >> Project Home: http://www.clusterlabs.org > >> >> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > >> >> Bugs: > >> >> http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker > >> >> > >> > > >> > > >> > _______________________________________________ > >> > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > >> > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > >> > > >> > Project Home: http://www.clusterlabs.org > >> > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > >> > Bugs: > >> > http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker > >> > > >> > >> _______________________________________________ > >> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > >> http://oss.clusterlabs.org/mailman/listinfo/pacemaker > >> > >> Project Home: http://www.clusterlabs.org > >> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > >> Bugs: > >> http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker > >> > > > > > > _______________________________________________ > > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > > > Project Home: http://www.clusterlabs.org > > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > > Bugs: > > http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker > > > > _______________________________________________ > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: > http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker > _______________________________________________ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://developerbugs.linux-foundation.org/enter_bug.cgi?product=Pacemaker