Re: [Linux-HA] Resource Route+IPaddr2 not really working
Hello. Try grep lrmd.* /var/log/messages Darren -Original Message- From: linux-ha-boun...@lists.linux-ha.org [mailto:linux-ha-boun...@lists.linux-ha.org] On Behalf Of Zausel Sent: 12 January 2010 18:02 To: linux-ha@lists.linux-ha.org Subject: [Linux-HA] Resource Route+IPaddr2 not really working Hi everybody! I want to make a "simple" cluster resource. I have 2 nodes, every node have a ipaddress and a default gw. If one node is going down, the oter node is getting the ipaddress and gateway. So far so good. Know I restarting node1, the server is going down and the resource switched to the other node. When node1 is going online the rescources are out of order in my crm_mon. I can't switch the resource anymore. Only a hard kill of the openais will help. Does anybody have a solution? SoftwareInfos: SLES 11 64bit pacemaker-1.0.5-0.2.3 heartbeat-common-3.0.0-0.3.2 heartbeat-ldirectord-3.0.0-0.1.5 heartbeat-resources-3.0.0-0.1.5 openais-0.80.3-26.6.3 config: node server1 \ attributes standby="off" node server2 \ attributes standby="off" primitive myroute ocf:heartbeat:Route \ params destination="0.0.0.0/0" gateway="192.168.2.254" device="eth1" \ meta failure-timeout="3s" \ meta target-role="Started" primitive myroute2 ocf:heartbeat:Route \ params destination="0.0.0.0/0" gateway="192.168.2.254" device="eth1" \ meta failure-timeout="3s" \ meta target-role="Started" primitive myip1 ocf:heartbeat:IPaddr2 \ params ip="192.168.2.11" broadcast="192.168.2.255" cidr_netmask="24" nic="eth1" \ meta target-role="Started" \ op monitor interval="2s" primitive myip2 ocf:heartbeat:IPaddr2 \ params ip="192.168.2.12" broadcast="192.168.2.255" cidr_netmask="24" nic="eth1" \ meta target-role="Started" \ op monitor interval="2s" group server1s myip1 myroute group server2s myip2 myroute2 location MyHostIP1 server1s 100: server1 location MyHostIP2 server2s 100: server2 property $id="cib-bootstrap-options" \ dc-version="1.0.5-05c8b63cbca7ce95182bb41881b3c5677f20bd5c" \ cluster-infrastructure="openais" \ expected-quorum-votes="2" \ no-quorum-policy="ignore" \ stonith-enabled="false" \ last-lrm-refresh="1263306062" ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
[Linux-HA] Resource Route+IPaddr2 not really working
Hi Darren, I got a lot of the same entries in the logfile: Jan 12 17:34:32 server2 lrmd: [15375]: info: RA output: (myroute:monitor:stderr) logd is not running Jan 12 17:34:32 server2 lrmd: [15375]: info: RA output: (myroute2:monitor:stderr) logd is not running Jan 12 17:34:32 server2 lrmd: [15375]: info: RA output: (myroute:monitor:stderr) 2010/01/12_17:34:32 ERROR: Gateway address 192.168.2.254 is unreachable. Jan 12 17:34:32 server2 lrmd: [15375]: info: RA output: (myroute2:monitor:stderr) 2010/01/12_17:34:32 ERROR: Gateway address 192.168.2.254 is unreachable. bye > Hello. > > Try grep lrmd.* /var/log/messages > > Darren > > > -Original Message- > From: linux-ha-boun...@lists.linux-ha.org > [mailto:linux-ha-boun...@lists.linux-ha.org] On Behalf Of Zausel > Sent: 12 January 2010 18:02 > To: linux-ha@lists.linux-ha.org > Subject: [Linux-HA] Resource Route+IPaddr2 not really working > > Hi everybody! > > I want to make a "simple" cluster resource. I have 2 nodes, every node > have a ipaddress and a default gw. If one node is going down, the oter > node is getting the ipaddress and gateway. So far so good. Know I > restarting node1, the server is going down and the resource switched > to the other node. When node1 is going online the rescources are out > of order in my crm_mon. I can't switch the resource anymore. Only a > hard kill of the openais will help. Does anybody have a solution? > > > > SoftwareInfos: > > SLES 11 64bit > pacemaker-1.0.5-0.2.3 > heartbeat-common-3.0.0-0.3.2 > heartbeat-ldirectord-3.0.0-0.1.5 > heartbeat-resources-3.0.0-0.1.5 > openais-0.80.3-26.6.3 > > config: > > node server1 \ > attributes standby="off" > node server2 \ > attributes standby="off" > primitive myroute ocf:heartbeat:Route \ > params destination="0.0.0.0/0" gateway="192.168.2.254" > device="eth1" \ > meta failure-timeout="3s" \ > meta target-role="Started" > primitive myroute2 ocf:heartbeat:Route \ > params destination="0.0.0.0/0" gateway="192.168.2.254" > device="eth1" \ > meta failure-timeout="3s" \ > meta target-role="Started" > primitive myip1 ocf:heartbeat:IPaddr2 \ > params ip="192.168.2.11" broadcast="192.168.2.255" > cidr_netmask="24" nic="eth1" \ > meta target-role="Started" \ > op monitor interval="2s" > primitive myip2 ocf:heartbeat:IPaddr2 \ > params ip="192.168.2.12" broadcast="192.168.2.255" > cidr_netmask="24" nic="eth1" \ > meta target-role="Started" \ > op monitor interval="2s" > group server1s myip1 myroute > group server2s myip2 myroute2 > location MyHostIP1 server1s 100: server1 > location MyHostIP2 server2s 100: server2 > property $id="cib-bootstrap-options" \ > dc-version="1.0.5-05c8b63cbca7ce95182bb41881b3c5677f20bd5c" \ > cluster-infrastructure="openais" \ > expected-quorum-votes="2" \ > no-quorum-policy="ignore" \ > stonith-enabled="false" \ > last-lrm-refresh="1263306062" > > > ___ > Linux-HA mailing list > Linux-HA@lists.linux-ha.org > http://lists.linux-ha.org/mailman/listinfo/linux-ha > See also: http://linux-ha.org/ReportingProblems > ___ > Linux-HA mailing list > Linux-HA@lists.linux-ha.org > http://lists.linux-ha.org/mailman/listinfo/linux-ha > See also: http://linux-ha.org/ReportingProblems > ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
Re: [Linux-HA] Resource Route+IPaddr2 not really working
It does seem to be fairly straightforward what the problem is indicating - it can't reach the gateway to add it. Are you able to add it manually when that happens? -Original Message- From: linux-ha-boun...@lists.linux-ha.org [mailto:linux-ha-boun...@lists.linux-ha.org] On Behalf Of Zausel Sent: 13 January 2010 10:23 To: General Linux-HA mailing list Subject: [Linux-HA] Resource Route+IPaddr2 not really working Hi Darren, I got a lot of the same entries in the logfile: Jan 12 17:34:32 server2 lrmd: [15375]: info: RA output: (myroute:monitor:stderr) logd is not running Jan 12 17:34:32 server2 lrmd: [15375]: info: RA output: (myroute2:monitor:stderr) logd is not running Jan 12 17:34:32 server2 lrmd: [15375]: info: RA output: (myroute:monitor:stderr) 2010/01/12_17:34:32 ERROR: Gateway address 192.168.2.254 is unreachable. Jan 12 17:34:32 server2 lrmd: [15375]: info: RA output: (myroute2:monitor:stderr) 2010/01/12_17:34:32 ERROR: Gateway address 192.168.2.254 is unreachable. bye > Hello. > > Try grep lrmd.* /var/log/messages > > Darren > > > -Original Message- > From: linux-ha-boun...@lists.linux-ha.org > [mailto:linux-ha-boun...@lists.linux-ha.org] On Behalf Of Zausel > Sent: 12 January 2010 18:02 > To: linux-ha@lists.linux-ha.org > Subject: [Linux-HA] Resource Route+IPaddr2 not really working > > Hi everybody! > > I want to make a "simple" cluster resource. I have 2 nodes, every node > have a ipaddress and a default gw. If one node is going down, the oter > node is getting the ipaddress and gateway. So far so good. Know I > restarting node1, the server is going down and the resource switched > to the other node. When node1 is going online the rescources are out > of order in my crm_mon. I can't switch the resource anymore. Only a > hard kill of the openais will help. Does anybody have a solution? > > > > SoftwareInfos: > > SLES 11 64bit > pacemaker-1.0.5-0.2.3 > heartbeat-common-3.0.0-0.3.2 > heartbeat-ldirectord-3.0.0-0.1.5 > heartbeat-resources-3.0.0-0.1.5 > openais-0.80.3-26.6.3 > > config: > > node server1 \ > attributes standby="off" > node server2 \ > attributes standby="off" > primitive myroute ocf:heartbeat:Route \ > params destination="0.0.0.0/0" gateway="192.168.2.254" > device="eth1" \ > meta failure-timeout="3s" \ > meta target-role="Started" > primitive myroute2 ocf:heartbeat:Route \ > params destination="0.0.0.0/0" gateway="192.168.2.254" > device="eth1" \ > meta failure-timeout="3s" \ > meta target-role="Started" > primitive myip1 ocf:heartbeat:IPaddr2 \ > params ip="192.168.2.11" broadcast="192.168.2.255" > cidr_netmask="24" nic="eth1" \ > meta target-role="Started" \ > op monitor interval="2s" > primitive myip2 ocf:heartbeat:IPaddr2 \ > params ip="192.168.2.12" broadcast="192.168.2.255" > cidr_netmask="24" nic="eth1" \ > meta target-role="Started" \ > op monitor interval="2s" > group server1s myip1 myroute > group server2s myip2 myroute2 > location MyHostIP1 server1s 100: server1 > location MyHostIP2 server2s 100: server2 > property $id="cib-bootstrap-options" \ > dc-version="1.0.5-05c8b63cbca7ce95182bb41881b3c5677f20bd5c" \ > cluster-infrastructure="openais" \ > expected-quorum-votes="2" \ > no-quorum-policy="ignore" \ > stonith-enabled="false" \ > last-lrm-refresh="1263306062" > > > ___ > Linux-HA mailing list > Linux-HA@lists.linux-ha.org > http://lists.linux-ha.org/mailman/listinfo/linux-ha > See also: http://linux-ha.org/ReportingProblems > ___ > Linux-HA mailing list > Linux-HA@lists.linux-ha.org > http://lists.linux-ha.org/mailman/listinfo/linux-ha > See also: http://linux-ha.org/ReportingProblems > ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
[Linux-HA] Multiple Choice test for cluster knowledge
Hi, together with the "Clusterbau" book I developed a quick & dirty multiple choice test for your knowledge about pacemaker. Download the test under: http://www.oreilly.de/catalog/linuxhacluster2ger/Clusterbau_Exam4Fun.zip Sorry, the test is at the moment only in German. The book is also in German. If there is a demand of an English translation, please let me know. Greetings, -- Dr. Michael Schwartzkopff MultiNET Services GmbH Addresse: Bretonischer Ring 7; 85630 Grasbrunn; Germany Tel: +49 - 89 - 45 69 11 0 Fax: +49 - 89 - 45 69 11 21 mob: +49 - 174 - 343 28 75 mail: mi...@multinet.de web: www.multinet.de Sitz der Gesellschaft: 85630 Grasbrunn Registergericht: Amtsgericht München HRB 114375 Geschäftsführer: Günter Jurgeneit, Hubert Martens --- PGP Fingerprint: F919 3919 FF12 ED5A 2801 DEA6 AA77 57A4 EDD8 979B Skype: misch42 ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
Re: [Linux-HA] Multiple Choice test for cluster knowledge
Yes please in English for both! Have you (or anyone else) thought of doing a Linux-HA certification? Darren -Original Message- From: linux-ha-boun...@lists.linux-ha.org [mailto:linux-ha-boun...@lists.linux-ha.org] On Behalf Of Michael Schwartzkopff Sent: 13 January 2010 13:17 To: The Pacemaker cluster resource manager; General Linux-HA mailing list Subject: [Linux-HA] Multiple Choice test for cluster knowledge Hi, together with the "Clusterbau" book I developed a quick & dirty multiple choice test for your knowledge about pacemaker. Download the test under: http://www.oreilly.de/catalog/linuxhacluster2ger/Clusterbau_Exam4Fun.zip Sorry, the test is at the moment only in German. The book is also in German. If there is a demand of an English translation, please let me know. Greetings, -- Dr. Michael Schwartzkopff MultiNET Services GmbH Addresse: Bretonischer Ring 7; 85630 Grasbrunn; Germany Tel: +49 - 89 - 45 69 11 0 Fax: +49 - 89 - 45 69 11 21 mob: +49 - 174 - 343 28 75 mail: mi...@multinet.de web: www.multinet.de Sitz der Gesellschaft: 85630 Grasbrunn Registergericht: Amtsgericht München HRB 114375 Geschäftsführer: Günter Jurgeneit, Hubert Martens --- PGP Fingerprint: F919 3919 FF12 ED5A 2801 DEA6 AA77 57A4 EDD8 979B Skype: misch42 ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
Re: [Linux-HA] Multiple Choice test for cluster knowledge
Am Mittwoch, 13. Januar 2010 14:19:47 schrieb darren.mans...@opengi.co.uk: > Yes please in English for both! > > Have you (or anyone else) thought of doing a Linux-HA certification? > > Darren > > -Original Message- > From: linux-ha-boun...@lists.linux-ha.org > [mailto:linux-ha-boun...@lists.linux-ha.org] On Behalf Of Michael > Schwartzkopff Sent: 13 January 2010 13:17 > To: The Pacemaker cluster resource manager; General Linux-HA mailing list > Subject: [Linux-HA] Multiple Choice test for cluster knowledge > > Hi, > > together with the "Clusterbau" book I developed a quick & dirty multiple > choice test for your knowledge about pacemaker. Download the test under: > > http://www.oreilly.de/catalog/linuxhacluster2ger/Clusterbau_Exam4Fun.zip > > Sorry, the test is at the moment only in German. The book is also in > German. If there is a demand of an English translation, please let me know. > > Greetings, If you visit my course you get a certificate ;-) See: http://www.gurulabs.com/linux-training/courses/GL640/ Greetings, -- Dr. Michael Schwartzkopff MultiNET Services GmbH Addresse: Bretonischer Ring 7; 85630 Grasbrunn; Germany Tel: +49 - 89 - 45 69 11 0 Fax: +49 - 89 - 45 69 11 21 mob: +49 - 174 - 343 28 75 mail: mi...@multinet.de web: www.multinet.de Sitz der Gesellschaft: 85630 Grasbrunn Registergericht: Amtsgericht München HRB 114375 Geschäftsführer: Günter Jurgeneit, Hubert Martens --- PGP Fingerprint: F919 3919 FF12 ED5A 2801 DEA6 AA77 57A4 EDD8 979B Skype: misch42 ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
Re: [Linux-HA] [Pacemaker] Multiple Choice test for cluster knowledge
On 2010-01-13 14:19, darren.mans...@opengi.co.uk wrote: > Yes please in English for both! > > Have you (or anyone else) thought of doing a Linux-HA certification? Yes. We have. It's called DRBD Certified Engineer but actually covers not only DRBD, but Heartbeat, Corosync and Pacemaker also and will thus be renamed LINBIT Certified Cluster Engineer very soon. Shameless self-plug: we have on-line training coming up shortly; see http://www.linbit.com/en/education/training-schedule/drbd-total-feb-1-mar-5-2010-online/ (I just noticed the web site still says Heartbeat only, rather than Heartbeat/Corosync/Pacemaker; that's an oversight. Will be changed promptly.) Cheers, Florian signature.asc Description: OpenPGP digital signature ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
[Linux-HA] Colocation of 2 resources so that it can't run together
Hello All, Thanks for suggesting me to upgrade from heartbeat-2.1.3 to pacemaker + heartbeat.. Now the things seems to be so fine..but still I have some doubts. It will be very grateful to get those doubts cleared.. I've got a 3 node cluster of OpenVZ (running VM's managed by pacemaker with Heartbeat on iSCSI) that I'm trying to get working in a stable fashion, but I'm having a hard time avoiding what I've dubbed the "startled herd" problem. Basically, the whole cluster sits there quite happily unless I defined colocations for the resources, which I am running on the 2 machines with 1 stand by machine. Might be the way I am defining colocations is wrong..please suggest me how I can define the colocations for my these 2 resources. Basically the scenario is: I have configured 2 resources groups containing all necessary file system, ip address and vz-script. The 3 node are as: 1) node_master ---> Active node 2) node_slave ---> Passive node 3) node3 -> Active node And the 2 resources are as: 1) group_vz_1 2) group_vz_2 My need is I don't want to have these 2 groups run on same machine(In case of both active machine get crashes and resource failover to standby machine(node_slave)). Please let me know how I can define the colocation fro the 2 resources so that it won't run together. Output of crm configure show are as: [r...@node_slave ~]# crm configure show node $id="075961d1-4492-4ba9-b4ad-e8c27b9e3f4b" node_master \ attributes $id="nodes-075961d1-4492-4ba9-b4ad-e8c27b9e3f4b" node $id="7e5fdac9-80dc-41a7-bd8f-a5591a1b69a0" node3 \ attributes $id="nodes-7e5fdac9-80dc-41a7-bd8f-a5591a1b69a0" node $id="d90b1ed2--44ac-9a4d-b435a6befd36" node_slave \ attributes $id="nodes-d90b1ed2--44ac-9a4d-b435a6befd36" primitive filesystem1_Filesystem ocf:heartbeat:Filesystem \ operations $id="filesystem1_Filesystem-operations" \ op start interval="0" timeout="60" \ op stop interval="0" timeout="60" \ op notify interval="0" timeout="60" \ op monitor interval="20" timeout="40" start-delay="0" \ op validate-all interval="0" timeout="5" \ op meta-data interval="0" timeout="5" \ params device="/dev/disk/by-uuid/f5feb406-685a-41f8-a4f7-170ae0925901" directory="/vz" fstype="ext3" options="_netdev,noatime" \ meta resource-stickiness="2000" primitive filesystem2_Filesystem ocf:heartbeat:Filesystem \ operations $id="filesystem2_Filesystem-operations" \ op start interval="0" timeout="60" \ op stop interval="0" timeout="60" \ op notify interval="0" timeout="60" \ op monitor interval="20" timeout="40" start-delay="0" \ op validate-all interval="0" timeout="5" \ op meta-data interval="0" timeout="5" \ params device="/dev/disk/by-uuid/81c3845e-c2f6-4cb0-a0cd-e00c074942fb" directory="/vz" fstype="ext3" options="_netdev,noatime" \ meta resource-stickiness="2000" primitive vip2_ipaddr2 ocf:heartbeat:IPaddr2 \ operations $id="vip2_ipaddr2-operations" \ op start interval="0" timeout="90" \ op stop interval="0" timeout="100" \ op status interval="10s" timeout="20s" start-delay="0" \ op monitor interval="10s" timeout="20s" start-delay="0" \ op meta-data interval="0" timeout="5s" \ op validate-all interval="0" timeout="20s" \ params ip="66.199.245.207" nic="eth0" cidr_netmask="255.255.255.224" \ meta resource-stickiness="2000" primitive vip_ipaddr2 ocf:heartbeat:IPaddr2 \ operations $id="vip_ipaddr2-operations" \ op start interval="0" timeout="90" \ op stop interval="0" timeout="100" \ op status interval="10s" timeout="20s" start-delay="0" \ op monitor interval="10s" timeout="20s" start-delay="0" \ op meta-data interval="0" timeout="5s" \ op validate-all interval="0" timeout="20s" \ params ip="66.199.245.204" nic="eth0" cidr_netmask="255.255.255.224" \ meta resource-stickiness="2000" primitive vz1_script lsb:vz \ operations $id="vz1_script-operations" \ op start interval="0" timeout="250" \ op stop interval="0" timeout="250" \ op status interval="10" timeout="10" \ op monitor interval="15" timeout="15" start-delay="15" \ op meta-data interval="0" timeout="5" \ meta is-managed="true" resource-stickiness="2000" primitive vz2_script lsb:vz \ operations $id="vz2_script-operations" \ op start interval="0" timeout="250" \ op stop interval="0" timeout="200" \ op status interval="10" timeout="10" \ op monitor interval="10" timeout="10" start-delay="15" \ op meta-data interval="0" timeout="5" \ meta resource-stickiness="2000" \ params $id="vz2_script-instance_attributes" group group_vz_1 vip_ipaddr2 filesystem1_Filesystem vz1_script \ meta target-role="started" group group_vz_2 vip2_ipaddr2 files
Re: [Linux-HA] [Pacemaker] Multiple Choice test for cluster knowledge
On Jan 13, 2010, at 7:39 AM, Florian Haas wrote: Shameless self-plug: we have on-line training coming up shortly; see http://www.linbit.com/en/education/training-schedule/drbd-total-feb-1-mar-5-2010-online/ I would like to sign up, but Feb 1 - Mar 5 is bad timing for me. I would definitely appreciate an announcement about future online sessions. I'm primarily interested in heartbeat+pacemaker, but drbd is useful to know about as well. regards, alex PGP.sig Description: This is a digitally signed message part ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
[Linux-HA] CRM strange errors
Hello folkz, I have compiled Corosync/OpenAIS, Pacemaker, DRBD and OCFS2 cluster environment, generated cib.xml file with haresources2cib.py script, and put it into /var/lib/heartbeat/crm directory When I try to run either 'crm configure show' or 'crm configure' I'm getting these errors: INFO: no curses support: you won't see colors Signon to CIB failed: connection failed Init failed, could not perform requested operations ERROR: cannot parse xml: no element found: line 1, column 0 ERROR: No CIB! Does it look like a problem with openSUSE 11.2 implementations of ncurses, libxml2 or libxslt? Could somebody please give me a hint? Thanks in advance, Alex ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
[Linux-HA] Virtual mac address
Hello everyone. I've been trying to use heartbeat version 2 with the script IPaddr2 and I've managed to get it running with cibadmin, crm_mon, etc. The problem is that it doesn't get the virtual mac address that I establish in the xml configuration. "crm_mon" tells me this: Last updated: Wed Jan 13 17:55:01 2010 Stack: Heartbeat Current DC: heart (883e1da9-c47e-4eeb-850c-e118a31a2d55) - partition with quorum Version: 1.0.5-3840e6b5a305ccb803d29b468556739e75532d56 2 Nodes configured, unknown expected votes 1 Resources configured. Online: [ heart bacula ] Resource Group: group_1 rscIPaddr2 (ocf::heartbeat:IPaddr2): Started bacula So IPaddr2 it's running, but when I look at the ip configuration and make tests with ping and look at ARP cache, I don't see the virtual mac anywhere. And before someone warns me about this topic, yes, I've read this old thread: http://lists.linux-ha.org/pipermail/linux-ha/2008-July/033776.html Both servers are running ubuntu-server 9.10 Karmic, the master node with kernel 2.6.31-16-386 and the slave with 2.6.31.14-generic-pae. I only want to put in high-availability a virtual ip and a virtual mac, to solve ARP problems on switchs. This is my cib.xml: Thanks in advance. ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
[Linux-HA] Question
Hello All, I'm looking for advice. I'm using drbd-8.3.5. I have created a drbd cluster active/active with gfs2 on the two nodes. Can I access this cluster from other nodes besides the two I've created drbd on and maintain sync. In other words, can I add four nodes to the gfs2 cluster (only two are the drbd) and access it from the four nodes using gfs2 and maintain sync? Note: I am not increasing my drbd cluster size (still two nodes), I am just adding gfs2 nodes to make a total of four. I want to say that drbd is great :-) I have been using it for ten years. Philip, you won't remember me but I still owe you a drink :-) Mark ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
Re: [Linux-HA] Colocation of 2 resources so that it can't run together
On 2010-01-13T20:38:00, jaspal singla wrote: > group group_vz_1 vip_ipaddr2 filesystem1_Filesystem vz1_script \ > meta target-role="started" > group group_vz_2 vip2_ipaddr2 filesystem2_Filesystem vz2_script \ > meta target-role="started" > location location_master group_vz_1 700: node_master > location location_node3 group_vz_2 600: node3 > location location_slave_1 group_vz_1 0: node_slave > location location_slave_2 group_vz_2 0: node_slave > colocation colocation_vz_test -inf: group_vz_1 group_vz_2 The anti-collocation rule you have is correct, and this should result in the resources not being placed on the same node. > property $id="cib-bootstrap-options" \ > dc-version="1.0.5-462f1569a43740667daf7b0f6b521742e9eb8fa7" \ > cluster-infrastructure="Heartbeat" \ > stonith-enabled="false" \ > symmetric-cluster="false" \ > last-lrm-refresh="1263311374" Disabling stonith is not a good idea if you're running shared storage. > rsc_defaults $id="rsc_defaults-options" You may want to enable resource-stickiness to avoid resources shuffling around needlessly. Regards, Lars -- Architect Storage/HA, OPS Engineering, Novell, Inc. SUSE LINUX Products GmbH, GF: Markus Rex, HRB 16746 (AG Nürnberg) "Experience is the name everyone gives to their mistakes." -- Oscar Wilde ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
Re: [Linux-HA] Question
Hello Mark, Am Mittwoch, den 13.01.2010, 15:57 -0500 schrieb Yeary, Mark A.: > Hello All, > > I'm looking for advice. I'm using drbd-8.3.5. I have created a > drbd cluster active/active with gfs2 on the two nodes. Can I access > this cluster from other nodes besides the two I've created drbd on and > maintain sync. In other words, can I add four nodes to the gfs2 cluster > (only two are the drbd) and access it from the four nodes using gfs2 and > maintain sync? Note: I am not increasing my drbd cluster size (still > two nodes), I am just adding gfs2 nodes to make a total of four. But how should ne these "just-gfs-nodes" be able to see or attach these storage, since your drbd-device is _not_ shared to this hosts. Your Nodes are not able to attch I would suggest, you want to export your drbd-device through aoe, iscsi, fcoe or whatever. A high-available storage-server through linux-ha could be the name to the next topic you open ;-) > I want to say that drbd is great :-) I have been using it for ten > years. Philip, you won't remember me but I still owe you a drink :-) > > Mark hth, Thomas > ___ > Linux-HA mailing list > Linux-HA@lists.linux-ha.org > http://lists.linux-ha.org/mailman/listinfo/linux-ha > See also: http://linux-ha.org/ReportingProblems ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
[Linux-HA] high-available storage-server through linux-ha
Hi Thomas, I should have been more clear in my explanation. We have a two physical node drbd cluster with GFS2. We have tried to add two more nodes by increasing the gfs2 journal to four, the two in the drbd cluster and two separate nodes; all four with gfs2 formatting. All four nodes are connected via virtual machine (KVM) virtio disks. We can access them but they do not maintain synchronization. Becoming frustrated (only a little :-), we went back to the two with drbd and gfs2. We then tried to connect with two more nodes (a total of four) via iscsi. That didn't seem to work either. Anything you can tell us would be helpful :-) Truly, Mark -Original Message- From: linux-ha-boun...@lists.linux-ha.org [mailto:linux-ha-boun...@lists.linux-ha.org] On Behalf Of Thomas Halinka Sent: Wednesday, January 13, 2010 4:40 PM To: General Linux-HA mailing list Subject: Re: [Linux-HA] Question Hello Mark, Am Mittwoch, den 13.01.2010, 15:57 -0500 schrieb Yeary, Mark A.: > Hello All, > > I'm looking for advice. I'm using drbd-8.3.5. I have created a > drbd cluster active/active with gfs2 on the two nodes. Can I access > this cluster from other nodes besides the two I've created drbd on and > maintain sync. In other words, can I add four nodes to the gfs2 cluster > (only two are the drbd) and access it from the four nodes using gfs2 and > maintain sync? Note: I am not increasing my drbd cluster size (still > two nodes), I am just adding gfs2 nodes to make a total of four. But how should ne these "just-gfs-nodes" be able to see or attach these storage, since your drbd-device is _not_ shared to this hosts. Your Nodes are not able to attch I would suggest, you want to export your drbd-device through aoe, iscsi, fcoe or whatever. A high-available storage-server through linux-ha could be the name to the next topic you open ;-) > I want to say that drbd is great :-) I have been using it for ten > years. Philip, you won't remember me but I still owe you a drink :-) > > Mark hth, Thomas > ___ > Linux-HA mailing list > Linux-HA@lists.linux-ha.org > http://lists.linux-ha.org/mailman/listinfo/linux-ha > See also: http://linux-ha.org/ReportingProblems ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems No virus found in this incoming message. Checked by AVG - www.avg.com Version: 8.5.432 / Virus Database: 270.14.138/2618 - Release Date: 01/13/10 07:35:00 ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
Re: [Linux-HA] high-available storage-server through linux-ha
Hi Mark, Am Mittwoch, den 13.01.2010, 17:38 -0500 schrieb Yeary, Mark A.: > Hi Thomas, > > I should have been more clear in my explanation. We have a two > physical node drbd cluster with GFS2. We have tried to add two more > nodes by increasing the gfs2 journal to four, the two in the drbd > cluster and two separate nodes; all four with gfs2 formatting. All four > nodes are connected via virtual machine (KVM) virtio disks. So your hosts are vms? > We can > access them but they do not maintain synchronization. .. because they all access _one common_ blockdevice passed through by kvm? Two of them play drbd with this device while the others do not? > Becoming frustrated (only a little :-), we went back to the two > with drbd and gfs2. We then tried to connect with two more nodes (a > total of four) via iscsi. That didn't seem to work either. Why? Just dd if=/dev/zero of=/MOUNTPOINT_GFS/10GB bs=1M count=1 And add this as a Lun to e.g. IET > Anything > you can tell us would be helpful :-) We should get more clearness about what you need - what is your goal? Do you want to access a Filesystem from all this hosts? Well known NFS with or GFS/OCFS2 on DRBD with iSCSI, AoE or FCoE could be a Solution... Or do you want to Share your _Raw Storage_? iSCSI, AoE, FCoE on LVM on top of DRBD would be nice... > > Truly, > > Mark > cu, thomas ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
Re: [Linux-HA] heartbeat waits for initdead even after all nodes have joined
I don't have autojoin in my ha.cf, and I believe it defaults to "autojoin none", so that wouldn't explain why heartbeat keeps waiting after all nodes have joined. I can see in /var/log/messages where crmd is doing the waiting for my 900-second initdead: 2010-01-11T13:51:15.428916-05:00 crmd: [4273]: info: do_started: The local CRM is operational 2010-01-11T13:51:15.428924-05:00 crmd: [4273]: info: do_state_transition: State transition S_STARTING -> S_PENDING [ input=I_PENDING cause=C_FSA_INTERNAL origin=do_started ] 2010-01-11T14:06:15.964307-05:00 crmd: [4273]: info: crm_timer_popped: Election Trigger (I_DC_TIMEOUT) just popped! 2010-01-11T14:06:15.964337-05:00 crmd: [4273]: WARN: do_log: [[FSA]] Input I_DC_TIMEOUT from crm_timer_popped() received in state (S_PENDING) 2010-01-11T14:06:15.964348-05:00 crmd: [4273]: info: do_state_transition: State transition S_PENDING -> S_ELECTION [ input=I_DC_TIMEOUT cause=C_TIMER_POPPED origin=crm_timer_popped ] I am using "Version 2 Resource Manager". I didn't previously realize this was the last version before the split. I am also using DRBD, and yesterday I discovered that its wait-for-connection timeout (wfc-timeout) works as I had hoped initdead would, and by putting it before heartbeat in the startup sequence, it turns out I don't really need initdead after all. Thanks, David -Original Message- From: linux-ha-boun...@lists.linux-ha.org [mailto:linux-ha-boun...@lists.linux-ha.org] On Behalf Of Dejan Muhamedagic Sent: Tuesday, January 12, 2010 3:51 AM To: General Linux-HA mailing list Subject: Re: [Linux-HA] heartbeat waits for initdead even after all nodes have joined Hi, On Mon, Jan 11, 2010 at 03:21:05PM -0500, David Sickmiller wrote: > Hi, > > > > I was hoping to configure my 2-node cluster to start as soon as both > nodes were present but wait up to 15 minutes if the other node was > missing upon system startup. In my case, a delay of several minutes is > better than a split-brain scenario. The Linux-HA documentation says > "The initdead parameter is used to set the time that it takes to declare > a cluster node dead when Heartbeat is first started.", so I figured I > could just set "initdead 900" in ha.cf. Unfortunately, heartbeat seems > to be waiting for the entire initdead time interval regardless of > whether all the nodes are present. > > > > Does this match others' experiences? Is there a different setting that > could accomplish my objective? > > > > It seems like the documentation would be more accurate if it said "The > initdead parameter is used to set the time that heartbeat waits before > starting any resources, which allows time for additional nodes to join." If you have autojoin set to "any". > However, I would much prefer that Linux-HA behaved according to the > original documentation. > > > > I'm using Heartbeat 2.1.4 on RHEL 5.4. Please switch to Pacemaker/heartbeat or Pacemaker/corosync. Or are you using v1/haresources? Thanks, Dejan ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems ___ Linux-HA mailing list Linux-HA@lists.linux-ha.org http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems