26.03.2013 04:23, Dennis Jacobfeuerborn wrote: > I have now reduced the configuration further and removed LVM from the > picture. Still the cluster fails when I set the master node to standby. > What's interesting is that things get fixed when I issue a simple > "cleanup" for the filesystem resource. > > This is what my current config looks like: > > node nfs1 \ > attributes standby="off" > node nfs2 > primitive p_drbd_web1 ocf:linbit:drbd \ > params drbd_resource="web1" \ > op monitor interval="15" role="Master" \ > op monitor interval="30" role="Slave" > primitive p_fs_web1 ocf:heartbeat:Filesystem \ > params device="/dev/drbd0" \ > directory="/srv/nfs/web1" fstype="ext4" \ > op monitor interval="10s" > ms ms_drbd_web1 p_drbd_web1 \ > meta master-max="1" master-node-max="1" \ > clone-max="2" clone-node-max="1" notify="true" > colocation c_web1_on_drbd inf: ms_drbd_web1:Master p_fs_web1
Above means: "colocate ms_drbd_web1:Master with p_fs_web1", or "promote ms_drbd_web1 where p_fs_web1 is (or "is about to be")". Probably that is not exactly what you want (although that is also valid, but uses different logic internally). I usually place resources in a different order in colocation and order constraints, and that works. > order o_drbd_before_web1 inf: ms_drbd_web1:promote p_fs_web1 > property $id="cib-bootstrap-options" \ > dc-version="1.1.8-7.el6-394e906" \ > cluster-infrastructure="classic openais (with plugin)" \ > expected-quorum-votes="2" \ > stonith-enabled="false" \ > no-quorum-policy="ignore" \ > last-lrm-refresh="1364259713" \ > maintenance-mode="false" > rsc_defaults $id="rsc-options" \ > resource-stickiness="100" > > I cannot figure out what is wrong with this configuration. > > Regards, > Dennis > > On 25.03.2013 13:09, Dennis Jacobfeuerborn wrote: >> I just found the following in the dmesg output which might or might not >> add to understanding the problem: >> >> device-mapper: table: 253:2: linear: dm-linear: Device lookup failed >> device-mapper: ioctl: error adding target to table >> >> Regards, >> Dennis >> >> On 25.03.2013 13:04, Dennis Jacobfeuerborn wrote: >>> Hi, >>> I'm currently trying create a two node redundant NFS setup on CentOS 6.4 >>> using pacemaker and crmsh. >>> >>> I use this Document as a starting poing: >>> https://www.suse.com/documentation/sle_ha/singlehtml/book_sleha_techguides/book_sleha_techguides.html >>> >>> >>> >>> >>> The first issue is that using these instructions I get the cluster up >>> and running but the moment I try to stop the pacemaker service on the >>> current master node several resources just fail and everything goes >>> pear-shaped. >>> >>> Since the problem seems to relate to the nfs bits in the configuration I >>> removed these in order to get to a minimal working setup and then add >>> things piece by piece in order to find the source of the problem. >>> >>> Now I am at a point where I basically have only >>> DRBD+LVM+Filesystems+IPAddr2 configured and now LVM seems to act up. >>> >>> I can start the cluster and everything is fine but the moment I stop >>> pacemaker on the master i end up with this as a status: >>> >>> === >>> Node nfs2: standby >>> Online: [ nfs1 ] >>> >>> Master/Slave Set: ms_drbd_nfs [p_drbd_nfs] >>> Masters: [ nfs1 ] >>> Stopped: [ p_drbd_nfs:1 ] >>> >>> Failed actions: >>> p_lvm_nfs_start_0 (node=nfs1, call=505, rc=1, status=complete): >>> unknown error >>> === >>> >>> and in the log on nfs1 I see: >>> LVM(p_lvm_nfs)[7515]: 2013/03/25_12:34:21 ERROR: device-mapper: >>> reload ioctl on failed: Invalid argument device-mapper: reload ioctl on >>> failed: Invalid argument 2 logical volume(s) in volume group "nfs" now >>> active >>> >>> However a lvs in this state shows: >>> [root@nfs1 ~]# lvs >>> LV VG Attr LSize Pool Origin Data% Move Log >>> web1 nfs -wi------ 2,00g >>> web2 nfs -wi------ 2,00g >>> lv_root vg_nfs1.local -wi-ao--- 2,45g >>> lv_swap vg_nfs1.local -wi-ao--- 256,00m >>> >>> So the volume group is present. >>> >>> My current configuration looks like this: >>> >>> node nfs1 \ >>> attributes standby="off" >>> node nfs2 \ >>> attributes standby="on" >>> primitive p_drbd_nfs ocf:linbit:drbd \ >>> params drbd_resource="nfs" \ >>> op monitor interval="15" role="Master" \ >>> op monitor interval="30" role="Slave" >>> primitive p_fs_web1 ocf:heartbeat:Filesystem \ >>> params device="/dev/nfs/web1" \ >>> directory="/srv/nfs/web1" \ >>> fstype="ext4" \ >>> op monitor interval="10s" >>> primitive p_fs_web2 ocf:heartbeat:Filesystem \ >>> params device="/dev/nfs/web2" \ >>> directory="/srv/nfs/web2" \ >>> fstype="ext4" \ >>> op monitor interval="10s" >>> primitive p_ip_nfs ocf:heartbeat:IPaddr2 \ >>> params ip="10.99.0.142" cidr_netmask="24" \ >>> op monitor interval="30s" >>> primitive p_lvm_nfs ocf:heartbeat:LVM \ >>> params volgrpname="nfs" \ >>> op monitor interval="30s" >>> group g_nfs p_lvm_nfs p_fs_web1 p_fs_web2 p_ip_nfs >>> ms ms_drbd_nfs p_drbd_nfs \ >>> meta master-max="1" \ >>> master-node-max="1" \ >>> clone-max="2" \ >>> clone-node-max="1" \ >>> notify="true" >>> colocation c_nfs_on_drbd inf: g_nfs ms_drbd_nfs:Master >>> property $id="cib-bootstrap-options" \ >>> dc-version="1.1.8-7.el6-394e906" \ >>> cluster-infrastructure="classic openais (with plugin)" \ >>> expected-quorum-votes="2" \ >>> stonith-enabled="false" \ >>> no-quorum-policy="ignore" \ >>> last-lrm-refresh="1364212090" \ >>> maintenance-mode="false" >>> rsc_defaults $id="rsc_defaults-options" \ >>> resource-stickiness="100" >>> >>> Any ideas why this isn't working? >>> >>> Regards, >>> Dennis >>> >>> _______________________________________________ >>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>> >>> Project Home: http://www.clusterlabs.org >>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>> Bugs: http://bugs.clusterlabs.org >> >> >> _______________________________________________ >> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >> >> Project Home: http://www.clusterlabs.org >> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >> Bugs: http://bugs.clusterlabs.org > > > _______________________________________________ > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org _______________________________________________ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org