The thing is, even though it's a test system, pacemaker and DRBD will
still operate as if it is critical. Turning off stonith won't properly
emulate production because when a node enters an unknown state, the
system will no longer behave predictably.
digimer
On 2019-04-15 12:13 p.m., Graham Smith wrote:
Thanks for the reply
I'm just doing some basic testing and not using stonith, but I
understand why this is important.
I rebuilt my setup from scratch and it worked the 2nd time. Not sure
what I got wrong, but I now have the basics working.
--
best regards
Graham
-----Original Message-----
*From*: digimer <[email protected]
<mailto:digimer%20%[email protected]%3e>>
*To*: Graham Smith <[email protected]
<mailto:graham%20smith%20%[email protected]%3e>>,
[email protected] <mailto:[email protected]>
*Subject*: Re: [DRBD-user] I need a little help with drbd and
pacemaker - node never promoted
*Date*: Mon, 15 Apr 2019 11:24:36 -0400
For one; Enable and test stonith in Pacemaker. When a node can be
failed and fenced, then configure drbd to use fencing:
resource-and-stonith; and set the {un,}fence-handler to
crm-{un,}fence-peer.sh.
digimer
On 2019-04-12 11:51 p.m., Graham Smith wrote:
Hi
1st time user, trying to set up drbd with pacemaker for a potental
NFS server project. I've tried following 3 guides and I can not get
drbd and pacemaker to work together. And guidance is gratefully received.
Centos 7.5 on 2 nodes
uname -a
Linux node-1 3.10.0-957.10.1.el7.x86_64 #1 SMP Mon Mar 18 15:06:45
UTC 2019 x86_64 x86_64 x86_64 GNU/Linux
I have drbd installed and initially it is syncing a disk between 2
nodes (prior to pcs cluster)
drbdadm --version
DRBDADM_BUILDTAG=GIT-hash:\ d458166f5f4740625e5ff215f62366aca60ca37b\
build\ by\ mockbuild@\,\ 2018-11-03\ 14:14:44
DRBDADM_API_VERSION=2
DRBD_KERNEL_VERSION_CODE=0x090010
DRBD_KERNEL_VERSION=9.0.16
DRBDADM_VERSION_CODE=0x090600
DRBDADM_VERSION=9.6.0
drbdadm status
drbd0 role:Primary
disk:UpToDate
node-2 role:Secondary
peer-disk:UpToDate
So far so good
cat /proc/drbd
version: 9.0.16-1 (api:2/proto:86-114)
GIT-hash: ab9777dfeaf9d619acc9a5201bfcae8103e9529c build by
mockbuild@, 2018-11-03 13:54:24
Transports (api:16): tcp (9.0.16-1)
# the output is different than I see in online guides (maybe
different in v9?).
https://clusterlabs.org/pacemaker/doc/en-US/Pacemaker/2.0/html/Clusters_from_Scratch/_initialize_drbd.html
---------
[
root@pcmk-
<mailto:root@pcmk->
1 ~]# cat /proc/drbd
version: 8.4.11-1 (api:1/proto:86-101)
GIT-hash: 66145a308421e9c124ec391a7848ac20203bb03c build by mockbuild@,
2018-04-26 12:10:42
1: cs:WFConnection ro:Secondary/Unknown ds:Inconsistent/DUnknown C r----s
ns:0 nr:0 dw:0 dr:0 al:8 bm:0 lo:0 pe:0 ua:0 ap:0 ep:1 wo:f oos:524236
--------
When I create a PCS cluster and add drbd as a resource, I do not get
a node to become a master
pcs -f drbd_cfg resource create NFSData ocf:linbit:drbd
drbd_resource=drbd0 op monitor interval=60
pcs -f drbd_cfg resource master NFSDataClone NFSData master-max=1
master-node-max=1 clone-max=2 clone-node-max=1 notify=true
pcs status
Cluster name: mycluster
Stack: corosync
Current DC: node-1 (version 1.1.19-8.el7_6.4-c3c624ea3d) - partition
with quorum
Last updated: Fri Apr 12 23:43:36 2019
Last change: Fri Apr 12 23:43:32 2019 by root via cibadmin on node-1
2 nodes configured
2 resources configured
Online: [ node-1 node-2 ]
Full list of resources:
Master/Slave Set: NFSDataClone [NFSData]
Slaves: [ node-1 node-2 ]
Daemon Status:
corosync: active/disabled
pacemaker: active/disabled
pcsd: active/enabled
Here is the output of pcs config
pcs config
Cluster Name: mycluster
Corosync Nodes:
node-1 node-2
Pacemaker Nodes:
node-1 node-2
Resources:
Master: NFSDataClone
Meta Attrs: master-node-max=1 clone-max=2 notify=true master-max=1
clone-node-max=1
Resource: NFSData (class=ocf provider=linbit type=drbd)
Attributes: drbd_resource=drbd0
Operations: demote interval=0s timeout=90 (NFSData-demote-interval-0s)
monitor interval=60s (NFSData-monitor-interval-60s)
notify interval=0s timeout=90 (NFSData-notify-interval-0s)
promote interval=0s timeout=90
(NFSData-promote-interval-0s)
reload interval=0s timeout=30 (NFSData-reload-interval-0s)
start interval=0s timeout=240 (NFSData-start-interval-0s)
stop interval=0s timeout=100 (NFSData-stop-interval-0s)
Stonith Devices:
Fencing Levels:
Location Constraints:
Ordering Constraints:
Colocation Constraints:
Ticket Constraints:
Alerts:
No alerts defined
Resources Defaults:
resource-stickiness: 100
Operations Defaults:
No defaults set
Cluster Properties:
cluster-infrastructure: corosync
cluster-name: mycluster
dc-version: 1.1.19-8.el7_6.4-c3c624ea3d
default-resource-stickiness: INFINITY
have-watchdog: false
no-quorum-policy: ignore
stonith-enabled: false
Quorum:
Options:
------
After creating the PCS cluster drbd now has no primary
drbdadm status
drbd0 role:Secondary
disk:UpToDate
node-2 role:Secondary
peer-disk:UpToDate
--------
I'm stumped on how to get this to work, are there better guides, is
this a know bug. Thanks in advance for any guidance.
--
best regards
Graham
_______________________________________________
drbd-user mailing list
[email protected]
http://lists.linbit.com/mailman/listinfo/drbd-user
_______________________________________________
drbd-user mailing list
[email protected]
http://lists.linbit.com/mailman/listinfo/drbd-user