my first post on this list; will be quite a long one :) environment: OS: RHEL 5.4 x86_64
drbd83.x86_64 8.3.2-6.el5_3 kmod-drbd83.x86_64 8.3.2-6.el5_3 openais.x86_64 0.80.6-8.el5_4.1 heartbeat.x86_64 3.0.0-33.2 resource-agents.x86_64 1.0-31.4 pacemaker.x86_64 1.0.5-4.1 pacemaker-libs.x86_64 1.0.5-4.1 i would like to mention that i first successfully implemented a 2 node active-passive setup with pacemaker with help from the drbd-users-guide. my next try was the three node setup (one backup-server outside of the HA cluster) it is up-and-running when i doing it 'by hand' ! my problems start with the configuration of pacemaker, here i' m lost. i defined the primitive resources, master-slave (without complains of crm/configure/ptest) after colocation + ordering, ptest complains: ------------------------------------------------------------------------------------------------------------------------- crm(live)configure# ptest ptest[15059]: 2009/10/16_11:27:17 WARN: native_color: Resource res.ip.mysql cannot run anywhere ptest[15059]: 2009/10/16_11:27:17 WARN: native_color: Resource res.drbd.mysqlstack:0 cannot run anywhere ------------------------------------------------------------------------------------------------------------------------- ------------------------------------------------------------------------------------------------------------------------- crm(live)configure# show node mysqlHA1.cologne.tecdoc.local node mysqlHA2.cologne.tecdoc.local primitive res.drbd.mysqldb ocf:linbit:drbd \ params drbd_resource="mysqldb" primitive res.drbd.mysqlstack ocf:linbit:drbd \ params drbd_resource="mysqlstack" primitive res.ip.mysql ocf:heartbeat:IPaddr2 \ params ip="172.30.2.10" nic="eth0" \ op monitor interval="2s" timeout="0.5s" ms ms.drbd.mysqldb res.drbd.mysqldb \ meta master-max="1" master-node-max="1" clone-max="2" clone-node-max="1" notify="true" globally_unique="false" ms ms.drbd.mysqlstack res.drbd.mysqlstack \ meta master-max="1" master-node-max="1" clone-max="1" clone-node-max="1" notify="true" globally_unique="false" colocation co.ms.drbd.mysqlstack_on_ms.drbd.mysqldb inf: ms.drbd.mysqlstack ms.drbd.mysqldb:Master colocation co.ms.drbd.mysqlstack_on_res.ip.mysql inf: ms.drbd.mysqlstack res.ip.mysql colocation co.res.ip.mysql_on_ms.drbd.mysqldb_master inf: res.ip.mysql ms.drbd.mysqldb:Master order o.ip.mysql_before_ms.drbd.mysqlstack inf: res.ip.mysql ms.drbd.mysqlstack:start order o.ms.drbd.mysqldb_before_ms.drbd.mysqlstack inf: ms.drbd.mysqldb:promote ms.drbd.mysqlstack:start property $id="cib-bootstrap-options" \ stonith-enabled="false" ------------------------------------------------------------------------------------------------------------------------- i have no idea why the cluster-ip cannot be run anywhere (and of course prevent res.drbd.mysqlstack from starting) if you're interested, my /etc/drbd.conf: ------------------------------------------------------------------------------------------------------------------------- global { usage-count no; } common { syncer { rate 10M; verify-alg sha1; } protocol C; } resource mysqldb { handlers { fence-peer "/usr/lib/drbd/crm-fence-peer.sh"; after-resync-target "/usr/lib/drbd/crm-unfence-peer.sh"; split-brain "/usr/lib/drbd/notify-split-brain.sh root"; } startup { degr-wfc-timeout 60; # 1 minute outdated-wfc-timeout 2; # 2 seconds. } disk { on-io-error detach; } net { cram-hmac-alg "sha1"; shared-secret "secret"; after-sb-0pri disconnect; after-sb-1pri disconnect; after-sb-2pri disconnect; rr-conflict disconnect; } syncer { rate 20M; al-extents 257; } device /dev/drbd0; disk /dev/sdb; meta-disk internal; on mysqlHA1.cologne.tecdoc.local { address 10.6.0.127:7788; } on mysqlHA2.cologne.tecdoc.local { address 10.6.0.128:7788; } } resource mysqlstack { protocol A; device /dev/drbd10; stacked-on-top-of mysqldb { address 172.30.2.10:7789; # Cluster IP } on mysqlHAoffsite.cologne.tecdoc.local { disk /dev/sdb; address 172.30.2.78:7789; # public IP of backup-node meta-disk internal; } } ------------------------------------------------------------------------------------------------------------------------- Mit freundlichen Grüßen / with kind regards Torsten Schmidt System Manager
_______________________________________________ Pacemaker mailing list Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker