Hi. You don't need cman in your configuration.The best way is use gfs-pcmk and dlm-pcmk.
16 февраля 2012 г. 12:30 пользователь emmanuel segura <emi2f...@gmail.com>написал: > I don't think so > > I remember gfs2 && ocfs2 on pacemaker & redhat cluster use clvm > > In redhat cluster dlm it's started with cman > > On pacemaker you should do a clone resource > > > 2012/2/16 diego fanesi <diego.fan...@gmail.com> > >> Try to set expected-quorum-votes to 1. I'm not sure but in my opinion >> that is the problem. >> Il giorno 14/feb/2012 06:43, "Богомолов Дмитрий Викторович" < >> beats...@mail.ru> ha scritto: >> >> Hi, >>> I have a trouble with my test configuration. >>> I build an Actice/Active cluster >>> Ubuntu(11.10)+DRBD+Cman+Pacemaker+gfs2+Xen for test purpose. >>> Now i am doing some tests with availability. I am try to start cluster >>> on one node. >>> >>> Trouble is - the Filesystem primitive ClusterFS (fs type=gfs2) does not >>> start when one of two nodes is switched off. >>> >>> Here my configuration: >>> >>> node blaster \ >>> attributes standby="off" >>> node turrel \ >>> attributes standby="off" >>> primitive ClusterData ocf:linbit:drbd \ >>> params drbd_resource="clusterdata" \ >>> op monitor interval="60s" >>> primitive ClusterFS ocf:heartbeat:Filesystem \ >>> params device="/dev/drbd/by-res/clusterdata" >>> directory="/mnt/cluster" fstype="gfs2" \ >>> op start interval="0" timeout="60s" \ >>> op stop interval="0" timeout="60s" \ >>> op monitor interval="60s" timeout="60s" >>> primitive ClusterIP ocf:heartbeat:IPaddr2 \ >>> params ip="192.168.122.252" cidr_netmask="32" >>> clusterip_hash="sourceip" \ >>> op monitor interval="30s" >>> primitive SSH-stonith stonith:ssh \ >>> params hostlist="turrel blaster" \ >>> op monitor interval="60s" >>> primitive XenDom ocf:heartbeat:Xen \ >>> params xmfile="/etc/xen/xen1.example.com.cfg" \ >>> meta allow-migrate="true" is-managed="true" target-role="Stopped" >>> \ >>> utilization cores="1" mem="512" \ >>> op monitor interval="30s" timeout="30s" \ >>> op start interval="0" timeout="90s" \ >>> op stop interval="0" timeout="300s" >>> ms ClusterDataClone ClusterData \ >>> meta master-max="2" master-node-max="1" clone-max="2" >>> clone-node-max="1" notify="true" >>> clone ClusterFSClone ClusterFS \ >>> meta target-role="Started" is-managed="true" >>> clone IP ClusterIP \ >>> meta globally-unique="true" clone-max="2" clone-node-max="2" >>> clone SSH-stonithClone SSH-stonith >>> location prefere-blaster XenDom 50: blaster >>> colocation XenDom-with-ClusterFS inf: XenDom ClusterFSClone >>> colocation fs_on_drbd inf: ClusterFSClone ClusterDataClone:Master >>> order ClusterFS-after-ClusterData inf: ClusterDataClone:promote >>> ClusterFSClone:start >>> order XenDom-after-ClusterFS inf: ClusterFSClone XenDom >>> property $id="cib-bootstrap-options" \ >>> dc-version="1.1.5-01e86afaaa6d4a8c4836f68df80ababd6ca3902f" \ >>> cluster-infrastructure="cman" \ >>> expected-quorum-votes="2" \ >>> stonith-enabled="true" \ >>> no-quorum-policy="ignore" \ >>> last-lrm-refresh="1329194925" >>> rsc_defaults $id="rsc-options" \ >>> resource-stickiness="100" >>> >>> Here is an $crm resource show: >>> >>> Master/Slave Set: ClusterDataClone [ClusterData] >>> Masters: [ turrel ] >>> Stopped: [ ClusterData:1 ] >>> Clone Set: IP [ClusterIP] (unique) >>> ClusterIP:0 (ocf::heartbeat:IPaddr2) Started >>> ClusterIP:1 (ocf::heartbeat:IPaddr2) Started >>> Clone Set: ClusterFSClone [ClusterFS] >>> Stopped: [ ClusterFS:0 ClusterFS:1 ] >>> Clone Set: SSH-stonithClone [SSH-stonith] >>> Started: [ turrel ] >>> Stopped: [ SSH-stonith:1 ] >>> XenDom (ocf::heartbeat:Xen) Stopped >>> >>> I tryed: >>> crm(live)resource# cleanup ClusterFSClone >>> Cleaning up ClusterFS:0 on turrel >>> Cleaning up ClusterFS:1 on turrel >>> Waiting for 3 replies from the CRMd... OK >>> >>> I can see only warn message in /var/log/cluster/corosync.log >>> Feb 14 16:25:56 turrel pengine: [1640]: WARN: unpack_rsc_op: Processing >>> failed op ClusterFS:0_start_0 on turrel: unknown exec error (-2) >>> and >>> Feb 14 16:25:56 turrel pengine: [1640]: WARN: common_apply_stickiness: >>> Forcing ClusterFSClone away from turrel after 1000000 failures (max=1000000) >>> Feb 14 16:25:56 turrel pengine: [1640]: WARN: common_apply_stickiness: >>> Forcing ClusterFSClone away from turrel after 1000000 failures (max=1000000) >>> >>> Direct me, please, what i need to check or else? >>> >>> Best regards, >>> Dmitriy Bogomolov >>> >>> _______________________________________________ >>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>> >>> Project Home: http://www.clusterlabs.org >>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>> Bugs: http://bugs.clusterlabs.org >>> >> >> _______________________________________________ >> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >> >> Project Home: http://www.clusterlabs.org >> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >> Bugs: http://bugs.clusterlabs.org >> >> > > > -- > esta es mi vida e me la vivo hasta que dios quiera > > _______________________________________________ > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org > > -- Viacheslav Biriukov BR http://biriukov.com
_______________________________________________ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org