On Mon, Sep 10, 2012 at 4:42 PM,  <renayama19661...@ybb.ne.jp> wrote:
> Hi All,
>
> We confirmed movement of the trouble of the clone resource that we combined 
> with Master/Slave resource.
>
> The master / slave resources are replaced under the influence of the trouble 
> of the clonal resource.
>
> We confirmed it in the next procedure.
>
>
> Step1) We start a cluster and send cib.
>
> ============
> Last updated: Mon Sep 10 15:26:25 2012
> Stack: Heartbeat
> Current DC: drbd2 (08607c71-da7b-4abf-b6d5-39ee39552e89) - partition with 
> quorum
> Version: 1.0.12-c6770b8
> 2 Nodes configured, unknown expected votes
> 6 Resources configured.
> ============
>
> Online: [ drbd1 drbd2 ]
>
>  Resource Group: grpPostgreSQLDB
>      prmApPostgreSQLDB  (ocf::pacemaker:Dummy): Started drbd1
>  Resource Group: grpStonith1
>      prmStonith1-2      (stonith:external/ssh): Started drbd2
>      prmStonith1-3      (stonith:meatware):     Started drbd2
>  Resource Group: grpStonith2
>      prmStonith2-2      (stonith:external/ssh): Started drbd1
>      prmStonith2-3      (stonith:meatware):     Started drbd1
>  Master/Slave Set: msDrPostgreSQLDB
>      Masters: [ drbd1 ]
>      Slaves: [ drbd2 ]
>  Clone Set: clnDiskd1
>      Started: [ drbd1 drbd2 ]
>  Clone Set: clnPingd
>      Started: [ drbd1 drbd2 ]
>
> Step2) We cause a monitor error in pingd.
>
> [root@drbd1 ~]# rm -rf /var/run/pingd-default_ping_set
>
> Step3) FailOver is finished.
>
> ============
> Last updated: Mon Sep 10 15:27:08 2012
> Stack: Heartbeat
> Current DC: drbd2 (08607c71-da7b-4abf-b6d5-39ee39552e89) - partition with 
> quorum
> Version: 1.0.12-c6770b8
> 2 Nodes configured, unknown expected votes
> 6 Resources configured.
> ============
>
> Online: [ drbd1 drbd2 ]
>
>  Resource Group: grpPostgreSQLDB
>      prmApPostgreSQLDB  (ocf::pacemaker:Dummy): Started drbd2
>  Resource Group: grpStonith1
>      prmStonith1-2      (stonith:external/ssh): Started drbd2
>      prmStonith1-3      (stonith:meatware):     Started drbd2
>  Resource Group: grpStonith2
>      prmStonith2-2      (stonith:external/ssh): Started drbd1
>      prmStonith2-3      (stonith:meatware):     Started drbd1
>  Master/Slave Set: msDrPostgreSQLDB
>      Masters: [ drbd2 ]
>      Stopped: [ prmDrPostgreSQLDB:1 ]
>  Clone Set: clnDiskd1
>      Started: [ drbd1 drbd2 ]
>  Clone Set: clnPingd
>      Started: [ drbd2 ]
>      Stopped: [ prmPingd:0 ]
>
> Failed actions:
>     prmPingd:0_monitor_10000 (node=drbd1, call=14, rc=7, status=complete): 
> not running
>
>
>
> However, Master/Slave resources seemed to be replaced when we watched log.
>
> Sep 10 15:26:53 drbd2 pengine: [2668]: notice: LogActions: Move    resource 
> prmApPostgreSQLDB#011(Started drbd1 -> drbd2)
> Sep 10 15:26:53 drbd2 pengine: [2668]: notice: LogActions: Leave   resource 
> prmStonith1-2#011(Started drbd2)
> Sep 10 15:26:53 drbd2 pengine: [2668]: notice: LogActions: Leave   resource 
> prmStonith1-3#011(Started drbd2)
> Sep 10 15:26:53 drbd2 pengine: [2668]: notice: LogActions: Leave   resource 
> prmStonith2-2#011(Started drbd1)
> Sep 10 15:26:53 drbd2 pengine: [2668]: notice: LogActions: Leave   resource 
> prmStonith2-3#011(Started drbd1)
> Sep 10 15:26:53 drbd2 pengine: [2668]: notice: LogActions: Move    resource 
> prmDrPostgreSQLDB:0#011(Master drbd1 -> drbd2)
> Sep 10 15:26:53 drbd2 pengine: [2668]: notice: LogActions: Stop    resource 
> prmDrPostgreSQLDB:1#011(drbd2)
> Sep 10 15:26:53 drbd2 pengine: [2668]: notice: LogActions: Leave   resource 
> prmDiskd1:0#011(Started drbd1)
> Sep 10 15:26:53 drbd2 pengine: [2668]: notice: LogActions: Leave   resource 
> prmDiskd1:1#011(Started drbd2)
> Sep 10 15:26:53 drbd2 pengine: [2668]: notice: LogActions: Stop    resource 
> prmPingd:0#011(drbd1)
> Sep 10 15:26:53 drbd2 pengine: [2668]: notice: LogActions: Leave   resource 
> prmPingd:1#011(Started drbd2)
>
> The replacement is unnecessary, and Slave becomes Master, and inoperative 
> Master should have only to originally stop.
>
> However, this problem seems to be solved in Pacemaker1.1.
>
> Will the correction be possible for Pacemaker1.0?
> Because I have a big difference in placement processing with Pacemaker1.1, I 
> think that the correction to Pacemaker1.0 is difficult.

You're probably right.  I will have a look soon.

>
>  * This problem may have been reported as a known problem.
>  * I registered this problem with Bugzilla.
>   * http://bugs.clusterlabs.org/show_bug.cgi?id=5103

great :)

>
> Best Regards,
> Hideo Yamauchi.
>
>
> _______________________________________________
> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
> http://oss.clusterlabs.org/mailman/listinfo/pacemaker
>
> Project Home: http://www.clusterlabs.org
> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
> Bugs: http://bugs.clusterlabs.org

_______________________________________________
Pacemaker mailing list: Pacemaker@oss.clusterlabs.org
http://oss.clusterlabs.org/mailman/listinfo/pacemaker

Project Home: http://www.clusterlabs.org
Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf
Bugs: http://bugs.clusterlabs.org

Reply via email to