Hi list, i configured pacemaker to take care of a three node drbd stetup but the monitor on the stacked drbd resouce failed and the resource never gets started. Doing the necessary steps by hand everything works fine (on-node-with-ip: drbdadm --stacked adjust mysqlstack ; drbdadm --stacked primary mysqlstack)
is it possible that the drbd-monitor can't deal properly the a stacked-resource or do i miss something in my configuration? ptest ------------------------------------------------------------------------------- ptest[2395]: 2009/10/19_12:27:47 info: determine_online_status: Node mysqlHA2.cologne.tecdoc.local is online ptest[2395]: 2009/10/19_12:27:47 info: unpack_rsc_op: res.drbd.mysqlstack:0_monitor_0 on mysqlHA2.cologne.tecdoc.local returned 6 (not configured) instead of the expected value: 7 (not running) ptest[2395]: 2009/10/19_12:27:47 ERROR: unpack_rsc_op: Hard error - res.drbd.mysqlstack:0_monitor_0 failed with rc=6: Preventing ms.drbd.mysqlstack from re-starting anywhere in the cluster ptest[2395]: 2009/10/19_12:27:47 info: unpack_rsc_op: res.ip.mysql_monitor_0 on mysqlHA2.cologne.tecdoc.local returned 0 (ok) instead of the expected value: 7 (not running) ptest[2395]: 2009/10/19_12:27:47 notice: unpack_rsc_op: Operation res.ip.mysql_monitor_0 found resource res.ip.mysql active on mysqlHA2.cologne.tecdoc.local ptest[2395]: 2009/10/19_12:27:47 info: determine_online_status: Node mysqlHA1.cologne.tecdoc.local is online ptest[2395]: 2009/10/19_12:27:47 info: unpack_rsc_op: res.drbd.mysqlstack:0_monitor_0 on mysqlHA1.cologne.tecdoc.local returned 6 (not configured) instead of the expected value: 7 (not running) ptest[2395]: 2009/10/19_12:27:47 ERROR: unpack_rsc_op: Hard error - res.drbd.mysqlstack:0_monitor_0 failed with rc=6: Preventing ms.drbd.mysqlstack from re-starting anywhere in the cluster ptest[2395]: 2009/10/19_12:27:48 debug: native_assign_node: All nodes for resource res.drbd.mysqlstack:0 are unavailable, unclean or shutting down (mysqlHA2.cologne.tecdoc.local: 1, -1000000) ptest[2395]: 2009/10/19_12:27:48 WARN: native_color: Resource res.drbd.mysqlstack:0 cannot run anywhere ------------------------------------------------------------------------------- crm_mon ------------------------------------------------------------------------------- ============ Last updated: Mon Oct 19 12:10:48 2009 Stack: openais Current DC: mysqlHA1.cologne.tecdoc.local - partition with quorum Version: 1.0.5-462f1569a43740667daf7b0f6b521742e9eb8fa7 2 Nodes configured, 2 expected votes 3 Resources configured. ============ Online: [ mysqlHA2.cologne.tecdoc.local mysqlHA1.cologne.tecdoc.local ] Master/Slave Set: ms.drbd.mysqldb Masters: [ mysqlHA2.cologne.tecdoc.local ] Slaves: [ mysqlHA1.cologne.tecdoc.local ] res.ip.mysql (ocf::heartbeat:IPaddr2): Started mysqlHA2.cologne.tecdoc.local Failed actions: res.drbd.mysqlstack:0_monitor_0 (node=(null), call=3, rc=6, status=complete): not configured ------------------------------------------------------------------------------- cat /proc/drbd on node mysqlHA2: ------------------------------------------------------------------------------- version: 8.3.2 (api:88/proto:86-90) GIT-hash: dd7985327f146f33b86d4bff5ca8c94234ce840e build by mockbu...@v20z-x86-64.home.local, 2009-08-29 14:07:55 0: cs:Connected ro:Primary/Secondary ds:UpToDate/UpToDate C r---- ns:4 nr:0 dw:0 dr:4 al:0 bm:1 lo:0 pe:0 ua:0 ap:0 ep:1 wo:b oos:0 ------------------------------------------------------------------------------- her my pacemaker config: ------------------------------------------------------------------------------- node mysqlHA1.cologne.tecdoc.local node mysqlHA2.cologne.tecdoc.local primitive res.drbd.mysqldb ocf:linbit:drbd \ params drbd_resource="mysqldb" primitive res.drbd.mysqlstack ocf:linbit:drbd \ params drbd_resource="mysqlstack" \ meta target-role="Started" primitive res.ip.mysql ocf:heartbeat:IPaddr2 \ params ip="172.30.2.10" nic="eth0" netmask="255.255.255.0" ms ms.drbd.mysqldb res.drbd.mysqldb \ meta master-max="1" master-node-max="1" clone-max="2" clone-node-max="1" notify="true" globally_unique="false" ms ms.drbd.mysqlstack res.drbd.mysqlstack \ meta master-max="1" master-node-max="1" clone-max="1" clone-node-max="1" notify="true" globally_unique="false" t colocation co.ms.drbd.mysqlstack_on_ms.drbd.mysqldb inf: ms.drbd.mysqlstack ms.drbd.mysqldb:Master colocation co.res.ip.mysql_on_ms.drbd.mysqldb_master inf: res.ip.mysql ms.drbd.mysqldb:Master order o.ip.mysql_before_ms.drbd.mysqlstack inf: res.ip.mysql ms.drbd.mysqlstack:start order o.ms.drbd.mysqldb_before_ms.drbd.mysqlstack inf: ms.drbd.mysqldb:promote ms.drbd.mysqlstack:start property $id="cib-bootstrap-options" \ dc-version="1.0.5-462f1569a43740667daf7b0f6b521742e9eb8fa7" \ cluster-infrastructure="openais" \ expected-quorum-votes="2" \ stonith-enabled="false" \ last-lrm-refresh="1255946421" ------------------------------------------------------------------------------- drbd.conf ------------------------------------------------------------------------------- global { usage-count no; } common { syncer { rate 10M; verify-alg sha1; } protocol C; } resource mysqldb { handlers { fence-peer "/usr/lib/drbd/crm-fence-peer.sh"; after-resync-target "/usr/lib/drbd/crm-unfence-peer.sh"; split-brain "/usr/lib/drbd/notify-split-brain.sh root"; } startup { degr-wfc-timeout 60; # 1 minute outdated-wfc-timeout 2; # 2 seconds. } disk { on-io-error detach; } net { cram-hmac-alg "sha1"; shared-secret "supersecret"; after-sb-0pri disconnect; after-sb-1pri disconnect; after-sb-2pri disconnect; rr-conflict disconnect; } syncer { rate 20M; al-extents 257; } device /dev/drbd0; disk /dev/sdb; meta-disk internal; on mysqlHA1.cologne.tecdoc.local { address 10.6.0.127:7788; } on mysqlHA2.cologne.tecdoc.local { address 10.6.0.128:7788; } } resource mysqlstack { protocol A; device /dev/drbd10; stacked-on-top-of mysqldb { address 172.30.2.10:7789; # Cluster IP } on mysqlHAoffsite.cologne.tecdoc.local { disk /dev/sdb; address 172.30.2.78:7789; # public IP of backup-node meta-disk internal; } } ------------------------------------------------------------------------------- Mit freundlichen Grüßen / with kind regards Torsten Schmidt System Manager
_______________________________________________ Pacemaker mailing list Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker