On 13/06/2013, at 5:15 PM, Xiaomin Zhang <zhangxiao...@gmail.com> wrote:
> Thanks Andrew. > Yes, the fs_ssn service (ocf:FileSystem) is still running when the machine > loses network. I configure it as primitive: > primitive fs_ssn ocf:heartbeat:Filesystem \ > op monitor interval="15s" \ > params device="/dev/drbd0" directory="/drbd" fstype="ext3" \ > meta target-role="Started" > As I assume this resource can only be started on 1 node, I think it should be > stopped automatically when pacemaker detects it's not in a HA cluster. > Is this incorrect assumption? No. But I'd need to see logs from all the nodes (please use attachments) to be able to comment further. > Thanks. > > > > On Thu, Jun 13, 2013 at 1:50 PM, Andrew Beekhof <and...@beekhof.net> wrote: > > On 13/06/2013, at 2:43 PM, Xiaomin Zhang <zhangxiao...@gmail.com> wrote: > > > Andrew Beekhof <andrew@...> writes: > > > >> > >> Try increasing your stickiness as it is being exceeded by the location > > constraints. > >> For the biggest stick, try 'infinity' which means - never move unless the > > node dies. > >> > > > > Thanks, Andrew, I applied infinity resource stickiness. However, the sst > > resource is still switched to the node which is online back from failure. > > And I found sth in the log: > > > > Jun 13 11:46:29 node3 pengine[27813]: warning: unpack_rsc_op: Processing > > failed op monitor for ip_ssn on node2: not running (7) > > Jun 13 11:46:29 node3 pengine[27813]: error: native_create_actions: > > Resource fs_ssn (ocf::Filesystem) is active on 2 nodes attempting recovery > > Jun 13 11:46:29 node3 pengine[27813]: warning: native_create_actions: See > > http://clusterlabs.org/wiki/FAQ#Resource_is_Too_Active for more information. > > > > Is this log showing that pacemaker tries to restart all the resource when > > the failed node is back again? > > No, thats a log showing the services were already running there when > pacemaker started. > > > > > > >>> Thanks. > >>> > >>> Below is my configure: > >>> ------------------CONFIG START-------------------------------------- > >>> node node3 \ > >>> attributes standby="on" > >>> node node1 > >>> node node2 > >>> primitive drbd_ssn ocf:linbit:drbd \ > >>> params drbd_resource="r0" \ > >>> op monitor interval="15s" > >>> primitive fs_ssn ocf:heartbeat:Filesystem \ > >>> op monitor interval="15s" \ > >>> params device="/dev/drbd0" directory="/drbd" fstype="ext3" \ > >>> meta target-role="Started" > >>> primitive ip_ssn ocf:heartbeat:IPaddr2 \ > >>> params ip="192.168.241.1" cidr_netmask="32" \ > >>> op monitor interval="15s" \ > >>> meta target-role="Started" > >>> primitive ip_sst ocf:heartbeat:IPaddr2 \ > >>> params ip="192.168.241.2" cidr_netmask="32" \ > >>> op monitor interval="15s" \ > >>> meta target-role="Started" > >>> primitive sst lsb:sst \ > >>> op monitor interval="15s" \ > >>> meta target-role="stopped" > >>> primitive ssn lsb:ssn \ > >>> op monitor interval="15s" \ > >>> meta target-role="stopped" > >>> ms ms_drbd_ssn drbd_ssn \ > >>> meta master-max="1" master-node-max="1" clone-max="2" > >>> clone-node-max="1" > >>> notify="true" target-role="Started" > >>> location sst_ip_prefer ip_sst 50: node1 > >>> location drbd_ssn_prefer ms_drbd_ssn 50: node1 > >>> colocation fs_ssn_coloc inf: ip_ssn fs_ssn > >>> colocation fs_on_drbd_coloc inf: fs_ssn ms_drbd_ssn:Master > >>> colocation sst_ip_coloc inf: sst ip_sst > >>> colocation ssn_ip_coloc inf: ssn ip_ssn > >>> order ssn_after_drbd inf: ms_drbd_ssn:promote fs_ssn:start > >>> order ip_after_fs inf: fs_ssn:start ip_ssn:start > >>> order sst_after_ip inf: ip_sst:start sst:start > >>> order sst_after_ssn inf: ssn:start sst:start > >>> order ssn_after_ip inf: ip_ssn:start ssn:start > >>> property $id="cib-bootstrap-options" \ > >>> dc-version="1.1.8-7.el6-394e906" \ > >>> cluster-infrastructure="classic openais (with plugin)" \ > >>> expected-quorum-votes="3" \ > >>> stonith-enabled="false" > >>> rsc_defaults $id="rsc-options" \ > >>> resource-stickiness="100" > >>> > >>> -------------------CONFIG END---------------------------------------- > >>> > > Best Regards. > > Xiaomin > > > > > > > > > > > > > > > > _______________________________________________ > > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > > > Project Home: http://www.clusterlabs.org > > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > > Bugs: http://bugs.clusterlabs.org > > > _______________________________________________ > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org > > _______________________________________________ > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org _______________________________________________ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org