by the way.. just to note that.. for a normal testing (manual failover, rebooting the active node)... the cluster is working fine. I only encounter this error if I try to poweroff/shutoff the active node.
On Mon, Dec 29, 2014 at 4:05 PM, Marlon Guao <[email protected]> wrote: > Hi. > > > Dec 29 13:47:16 s1 LVM(vg1)[1601]: WARNING: LVM Volume cluvg1 is not > available (stopped) > Dec 29 13:47:16 s1 crmd[1515]: notice: process_lrm_event: Operation > vg1_monitor_0: not running (node= > s1, call=23, rc=7, cib-update=40, confirmed=true) > Dec 29 13:47:16 s1 crmd[1515]: notice: te_rsc_command: Initiating action > 9: monitor fs1_monitor_0 on > s1 (local) > Dec 29 13:47:16 s1 crmd[1515]: notice: te_rsc_command: Initiating action > 16: monitor vg1_monitor_0 on > s2 > Dec 29 13:47:16 s1 Filesystem(fs1)[1618]: WARNING: Couldn't find device > [/dev/mapper/cluvg1-clulv1]. Ex > pected /dev/??? to exist > > > from the LVM agent, it checked if the volume is already available.. and > will raise the above error if not. But, I don't see that it tries to > activate it before raising the VG. Perhaps, it assumes that the VG is > already activated... so, I'm not sure who should be activating it (should > it be LVM?). > > > if [ $rc -ne 0 ]; then > ocf_log $loglevel "LVM Volume $1 is not available > (stopped)" > rc=$OCF_NOT_RUNNING > else > case $(get_vg_mode) in > 1) # exclusive with tagging. > # If vg is running, make sure the correct tag is > present. Otherwise we > # can not guarantee exclusive activation. > if ! check_tags; then > ocf_exit_reason "WARNING: > $OCF_RESKEY_volgrpname is active without the cluster tag, \"$OUR_TAG\"" > > On Mon, Dec 29, 2014 at 3:36 PM, emmanuel segura <[email protected]> > wrote: > >> logs? >> >> 2014-12-29 6:54 GMT+01:00 Marlon Guao <[email protected]>: >> > Hi, >> > >> > just want to ask regarding the LVM resource agent on pacemaker/corosync. >> > >> > I setup 2 nodes cluster (opensuse13.2 -- my config below). The cluster >> > works as expected, like doing a manual failover (via crm resource move), >> > and automatic failover (by rebooting the active node for instance). >> But, if >> > i try to just "shutoff" the active node (it's a VM, so I can do a >> > poweroff). The resources won't be able to failover to the passive node. >> > when I did an investigation, it's due to an LVM resource not starting >> > (specifically, the VG). I found out that the LVM resource won't try to >> > activate the volume group in the passive node. Is this an expected >> > behaviour? >> > >> > what I really expect is that, in the event that the active node be >> shutoff >> > (by a power outage for instance), all resources should be failover >> > automatically to the passive. LVM should re-activate the VG. >> > >> > >> > here's my config. >> > >> > node 1: s1 >> > node 2: s2 >> > primitive cluIP IPaddr2 \ >> > params ip=192.168.13.200 cidr_netmask=32 \ >> > op monitor interval=30s >> > primitive clvm ocf:lvm2:clvmd \ >> > params daemon_timeout=30 \ >> > op monitor timeout=90 interval=30 >> > primitive dlm ocf:pacemaker:controld \ >> > op monitor interval=60s timeout=90s on-fail=ignore \ >> > op start interval=0 timeout=90 >> > primitive fs1 Filesystem \ >> > params device="/dev/mapper/cluvg1-clulv1" directory="/data" fstype=btrfs >> > primitive mariadb mysql \ >> > params config="/etc/my.cnf" >> > primitive sbd stonith:external/sbd \ >> > op monitor interval=15s timeout=60s >> > primitive vg1 LVM \ >> > params volgrpname=cluvg1 exclusive=yes \ >> > op start timeout=10s interval=0 \ >> > op stop interval=0 timeout=10 \ >> > op monitor interval=10 timeout=30 on-fail=restart depth=0 >> > group base-group dlm clvm >> > group rgroup cluIP vg1 fs1 mariadb \ >> > meta target-role=Started >> > clone base-clone base-group \ >> > meta interleave=true target-role=Started >> > property cib-bootstrap-options: \ >> > dc-version=1.1.12-1.1.12.git20140904.266d5c2 \ >> > cluster-infrastructure=corosync \ >> > no-quorum-policy=ignore \ >> > last-lrm-refresh=1419514875 \ >> > cluster-name=xxx \ >> > stonith-enabled=true >> > rsc_defaults rsc-options: \ >> > resource-stickiness=100 >> > >> > -- >> >>>> import this >> > _______________________________________________ >> > Linux-HA mailing list >> > [email protected] >> > http://lists.linux-ha.org/mailman/listinfo/linux-ha >> > See also: http://linux-ha.org/ReportingProblems >> >> >> >> -- >> esta es mi vida e me la vivo hasta que dios quiera >> _______________________________________________ >> Linux-HA mailing list >> [email protected] >> http://lists.linux-ha.org/mailman/listinfo/linux-ha >> See also: http://linux-ha.org/ReportingProblems >> > > > > -- > >>> import this > -- >>> import this _______________________________________________ Linux-HA mailing list [email protected] http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
