dear all, i configure pacemaker and corosync on 2 Centos 6.3 servers by following instruction on 'Cluster from Scratch'. on the beginning, i follow 'Cluster from Scratch' edition 5. but, since i use centos, i change to 'Cluster from Scratch' edition 3 to configure active/active servers. Now on 1st server (cluster1), the Filesystem resource cannot start. the gfs2 filesystem can't be mounted.
this is the crm configuration [root@cluster2 ~]# crm configure show node cluster1 \ attributes standby="off" node cluster2 \ attributes standby="off" primitive ClusterIP ocf:heartbeat:IPaddr2 \ params ip="xxx.xxx.xxx.229" cidr_netmask="32" clusterip_hash="sourceip" \ op monitor interval="30s" primitive WebData ocf:linbit:drbd \ params drbd_resource="wwwdata" \ op monitor interval="60s" primitive WebFS ocf:heartbeat:Filesystem \ params device="/dev/drbd/by-res/wwwdata" directory="/var/www/html" fstype="gfs2" primitive WebSite ocf:heartbeat:apache \ params configfile="/etc/httpd/conf/httpd.conf" statusurl=" http://localhost/server-status" \ op monitor interval="1min" ms WebDataClone WebData \ meta master-max="2" master-node-max="1" clone-max="2" clone-node-max="1" notify="true" clone WebFSClone WebFS clone WebIP ClusterIP \ meta globally-unique="true" clone-max="2" clone-node-max="1" interleave="false" clone WebSiteClone WebSite \ meta interleave="false" colocation WebSite-with-WebFS inf: WebSiteClone WebFSClone colocation colocation-WebSite-ClusterIP-INFINITY inf: WebSiteClone WebIP colocation fs_on_drbd inf: WebFSClone WebDataClone:Master order WebFS-after-WebData inf: WebDataClone:promote WebFSClone:start order WebSite-after-WebFS inf: WebFSClone WebSiteClone order order-ClusterIP-WebSite-mandatory : WebIP:start WebSiteClone:start property $id="cib-bootstrap-options" \ dc-version="1.1.7-6.el6-148fccfd5985c5590cc601123c6c16e966b85d14" \ cluster-infrastructure="cman" \ expected-quorum-votes="2" \ stonith-enabled="false" \ no-quorum-policy="ignore" rsc_defaults $id="rsc-options" \ resource-stickiness="100" when i want to mount the filesystem manually, this message appear : [root@cluster1 ~]# mount /dev/drbd1 /mnt/ mount point already used or other mount in progress error mounting lockproto lock_dlm but when i check the mount, there is no mount from drbd there is another strange thing, the 1st server (cluster1) cannot reboot. it hangs with message 'please standby while rebooting the system'. in the reboot process, there are 2 failed action which is related to fencing. i didn't configure any fencing yet. one of the failed action is : 'stopping cluster leaving fence domain .... found dlm lockspace /sys/kernel/dlm/web fence_tool : cannot leave due to active system [FAILED]' please help me with this problem -- Best Regards, Soni Maula Harriz Database Administrator PT. Data Aksara Sangkuriang
_______________________________________________ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org