Environment

fedora17+corosync-2.0.1-1.fc17.x86_64+pacemaker-1.1.7-2.fc17.x86_64

two node cluster:

#corosync-quorumtool -l
Membership information
----------------------
    Nodeid      Votes Name
       200          1 xen02
       100          1 xen01

# crm status
============
Last updated: Mon Jun 18 11:51:42 2012
Last change: Mon Jun 18 11:51:39 2012 via cibadmin on xen01
Stack: corosync
Current DC: xen02 (200) - partition with quorum
Version: 1.1.7-2.fc17-ee0730e13d124c3d58f00016c3376a1de5323cff
2 Nodes configured, unknown expected votes
3 Resources configured.
============

Online: [ xen01 xen02 ]

BUT

# crm node list
xen01(100): normal
xen02(200): normal

with one resource ClusterIP ocf:heartbeat:IPaddr2 cluster is working and 
migration gone with no problems
but for ms_drbd resource migration is working only once

if i do crm resource migrate (first migration finished correctry), then delete 
location constraint and then do back-migration  

1. using crm resource migrate, i'v got error

Error performing operation: ms_drbd_iscsi_0 is already active on xen02

2. if i create location constraint using crm configure, 

#crm configure
# location cli-standby-ms_drbd_iscsi_0 ms_drbd_iscsi_0 rule 
$id="cli-standby-rule-ms_drbd_iscsi_0" inf: #uname eq xen02
WARNING: cli-standby-ms_drbd_iscsi_0: referenced node xen02 does not exist
#commit
WARNING: cli-standby-ms_drbd_iscsi_0: referenced node xen02 does not exist

migration DONE with warning 

and

# crm_verify -L
#
# crm configure show
node $id="100" xen01
node $id="200" xen02
primitive ClusterIP ocf:heartbeat:IPaddr2 \
        params ip="172.16.0.4" cidr_netmask="32" \
        op monitor interval="30s" \
        meta is-managed="true" target-role="Stopped"
primitive p_drbd_iscsi_0 ocf:linbit:drbd \
        params drbd_resource="iscsi0" \
        op start interval="0" timeout="240s" \
        op stop interval="0" timeout="100s" \
        op monitor interval="10s" role="Master" \
        op monitor interval="60s" role="Slave"
ms ms_drbd_iscsi_0 p_drbd_iscsi_0 \
        meta master-max="1" master-node-max="1" clone-max="2" 
clone-node-max="1" notify="true" target-role="Started" is-managed="true"
location cli-standby-ms_drbd_iscsi_0 ms_drbd_iscsi_0 \
        rule $id="cli-standby-rule-ms_drbd_iscsi_0" inf: #uname eq xen02
property $id="cib-bootstrap-options" \
        dc-version="1.1.7-2.fc17-ee0730e13d124c3d58f00016c3376a1de5323cff" \
        cluster-infrastructure="corosync" \
        stonith-enabled="false" \
        no-quorum-policy="ignore"
rsc_defaults $id="rsc-options" \
        resource-stickiness="100"

I think that the problem is in nodes names

# crm node list

xen01(100): normal

xen02(200): normal

because in fedora16  with  pacemaker-1.1.7-2.fc16.x86_64 + 
corosync-1.4.3-1.fc16.x86_64

# crm node list
xen01: normal
        standby: off
xen02: normal
        standby: off

--
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to