Hi Jürgen, I am beginner with Pacemaker and I don't know if I can help. But my configuration is working well. Here is:
# crm configure property no-quorum-policy=ignore # crm configure property stonith-enabled=false primitive net_conn ocf:pacemaker:ping \ params pidfile="/var/run/ping.pid" \ host_list="192.168.188.1" \ op start interval="0" timeout="60s" \ op stop interval="0" timeout="20s" \ op monitor interval="10s" timeout="60s" primitive cluster_ip ocf:heartbeat:IPaddr2 \ params ip="192.168.188.20" cidr_netmask="32"\ op monitor interval="10s" primitive cluster_mon ocf:pacemaker:ClusterMon \ params pidfile="/var/run/crm_mon.pid" htmlfile="/var/tmp/crm_mon.html"\ op start interval="0" timeout="20s"\ op stop interval="0" timeout="20s"\ op monitor interval="10s" timeout="20s" primitive drbd_r8 ocf:linbit:drbd \ params drbd_resource="r8"\ op monitor interval="60s" role="Master"\ op monitor interval="59s" role="Slave" primitive drbd_r8_fs ocf:heartbeat:Filesystem \ params device="/dev/drbd8" directory="/mnt/drbd8" fstype="ext3" clone clone_net_conn net_conn \ meta clone-node-max="1" clone-max="2" ms drbd_r8_ms drbd_r8 \ meta master-max="1" master-node-max="1" clone-max="2" clone-node-max="1" notify="true" location ms_drbd_r8-no-conn drbd_r8_ms \ rule $id="ms_drbd_r8-no-conn-rule" $role="Master" -inf: not_defined pingd or pingd number:lte 0 colocation fs_on_drbd inf: drbd_r8_fs drbd_r8_ms:Master order fs_after_drbd inf: drbd_r8_ms:promote drbd_r8_fs:start Felipe On Tue, Feb 5, 2013 at 10:04 AM, Jürgen Herrmann <juergen.herrm...@xlhost.de > wrote: > Hi there! > > I have the following problem: > > I have a 2 node cluster with a dual primary drbd resource. On top > of it sits an OCFS2 file system. nodes: app1a, app1b > > Now today I had the following scenario (occurred several times now): > - crm node standby app1a > - poweroff app1a for hdd replacement (hw raid controller) > - poweron app1a > - crm node online app1a > > all the other resources come back up as expecte, expect the master > slave set for the dual primary drbd. > > here's the relevant portion of my cluster config: > > node app1a.xlhost.de \ > attributes standby="off" > node app1b.xlhost.de \ > attributes standby="off" > primitive resDLM ocf:pacemaker:controld \ > op start interval="0" timeout="90s" \ > op stop interval="0" timeout="100s" \ > op monitor interval="120s" > primitive resDRBD0 ocf:linbit:drbd \ > op monitor interval="23" role="Slave" timeout="30" \ > op monitor interval="13" role="Master" timeout="20" \ > op start interval="0" timeout="240s" \ > op promote interval="0" timeout="240s" \ > op demote interval="0" timeout="100s" \ > op stop interval="0" timeout="100s" \ > params drbd_resource="drbd0" > primitive resFSDRBD0 ocf:heartbeat:Filesystem \ > params device="/dev/drbd0" directory="/mnt/drbd0" fstype="ocfs2" > options="noatime,intr,**nodiratime,heartbeat=none" \ > op monitor interval="120s" timeout="50s" \ > op start interval="0" timeout="70s" \ > op stop interval="0" timeout="70s" > primitive resO2CB ocf:pacemaker:o2cb \ > op start interval="0" timeout="90s" \ > op stop interval="0" timeout="100s" \ > op monitor interval="120s" > ms msDRBD0 resDRBD0 \ > meta master-max="2" master-node-max="1" clone-max="2" > clone-node-max="1" notify="true" target-role="Master" > clone cloneDLM resDLM \ > meta globally-unique="false" interleave="true" > target-role="Started" > clone cloneFSDRBD0 resFSDRBD0 \ > meta interleave="true" globally-unique="false" > target-role="Started" > clone cloneO2CB resO2CB \ > meta globally-unique="false" interleave="true" > target-role="Started" > colocation colFSDRBD0_DRBD0 inf: cloneFSDRBD0 msDRBD0:Master > colocation colFSDRBD0_O2CB inf: cloneFSDRBD0 cloneO2CB > colocation colO2CB_DLM inf: cloneO2CB cloneDLM > order ordDLM_FSDRBD0 inf: cloneDLM cloneFSDRBD0 > order ordDLM_O2CB inf: cloneDLM cloneO2CB > order ordDRBD0_FSDRBD0 inf: msDRBD0:promote cloneFSDRBD0 > order ordO2CB_FSDRBD0 inf: cloneO2CB cloneFSDRBD0 > > if i take down both nodes and fire them up again, everything goes back > to normal and msDRBD0 is promoted to master on both nodes. > > I suspect this has something to do with ordering or colocation constraints > but i'm not sure though. i've been staring at this problem for dozens of > times now and a vast amount of googling did not turn up my specific > problem either. > > anybody have a clue? :) any hint in the right direction as where too look > etc. would really be appreciated. > > Thanks in advance for your help and best regards, > Jürgen Herrmann > -- > >> XLhost.de ® - Webhosting von supersmall bis eXtra Large << >>> >> > XLhost.de GmbH > Jürgen Herrmann, Geschäftsführer > Boelckestrasse 21, 93051 Regensburg, Germany > > Geschäftsführer: Jürgen Herrmann > Registriert unter: HRB9918 > Umsatzsteuer-**Identifikationsnummer: DE245931218 > > Fon: +49 (0)800 XLHOSTDE [0800 95467833] > Fax: +49 (0)800 95467830 > Web: http://www.XLhost.de > > ______________________________**_________________ > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > http://oss.clusterlabs.org/**mailman/listinfo/pacemaker<http://oss.clusterlabs.org/mailman/listinfo/pacemaker> > > Project Home: http://www.clusterlabs.org > Getting started: > http://www.clusterlabs.org/**doc/Cluster_from_Scratch.pdf<http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf> > Bugs: http://bugs.clusterlabs.org > -- *-- -- Felipe Oliveira Gutierrez -- felipe.o.gutier...@gmail.com -- https://sites.google.com/site/lipe82/Home/diaadia*
_______________________________________________ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org