Re: [Linux-HA] Resource Route+IPaddr2 not really working

2010-01-13 Thread Darren.Mansell
Hello.

Try grep lrmd.* /var/log/messages

Darren


-Original Message-
From: linux-ha-boun...@lists.linux-ha.org
[mailto:linux-ha-boun...@lists.linux-ha.org] On Behalf Of Zausel
Sent: 12 January 2010 18:02
To: linux-ha@lists.linux-ha.org
Subject: [Linux-HA] Resource Route+IPaddr2 not really working

Hi everybody!

I want to make a "simple" cluster resource. I have 2 nodes, every node  
have a ipaddress and a default gw. If one node is going down, the oter  
node is getting the ipaddress and gateway. So far so good. Know I  
restarting node1, the server is going down and the resource switched  
to the other node. When node1 is going online the rescources are out  
of order in my crm_mon. I can't switch the resource anymore. Only a  
hard kill of the openais will help. Does anybody have a solution?



SoftwareInfos:

SLES 11 64bit
pacemaker-1.0.5-0.2.3
heartbeat-common-3.0.0-0.3.2
heartbeat-ldirectord-3.0.0-0.1.5
heartbeat-resources-3.0.0-0.1.5
openais-0.80.3-26.6.3

config:

node server1 \
attributes standby="off"
node server2 \
attributes standby="off"
primitive myroute ocf:heartbeat:Route \
params destination="0.0.0.0/0" gateway="192.168.2.254"
device="eth1" \
meta failure-timeout="3s" \
meta target-role="Started"
primitive myroute2 ocf:heartbeat:Route \
params destination="0.0.0.0/0" gateway="192.168.2.254"
device="eth1" \
meta failure-timeout="3s" \
meta target-role="Started"
primitive myip1 ocf:heartbeat:IPaddr2 \
params ip="192.168.2.11" broadcast="192.168.2.255"
cidr_netmask="24" nic="eth1" \
meta target-role="Started" \
op monitor interval="2s"
primitive myip2 ocf:heartbeat:IPaddr2 \
params ip="192.168.2.12" broadcast="192.168.2.255"
cidr_netmask="24" nic="eth1" \
meta target-role="Started" \
op monitor interval="2s"
group server1s myip1 myroute
group server2s myip2 myroute2
location MyHostIP1 server1s 100: server1
location MyHostIP2 server2s 100: server2
property $id="cib-bootstrap-options" \
dc-version="1.0.5-05c8b63cbca7ce95182bb41881b3c5677f20bd5c" \
cluster-infrastructure="openais" \
expected-quorum-votes="2" \
no-quorum-policy="ignore" \
stonith-enabled="false" \
last-lrm-refresh="1263306062"


___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


[Linux-HA] Resource Route+IPaddr2 not really working

2010-01-13 Thread Zausel
Hi Darren,

I got a lot of the same entries in the logfile:

Jan 12 17:34:32 server2 lrmd: [15375]: info: RA output:
(myroute:monitor:stderr) logd is not running
Jan 12 17:34:32 server2 lrmd: [15375]: info: RA output:
(myroute2:monitor:stderr) logd is not running
Jan 12 17:34:32 server2 lrmd: [15375]: info: RA output:
(myroute:monitor:stderr) 2010/01/12_17:34:32 ERROR: Gateway address
192.168.2.254 is unreachable.
Jan 12 17:34:32 server2 lrmd: [15375]: info: RA output:
(myroute2:monitor:stderr) 2010/01/12_17:34:32 ERROR: Gateway address
192.168.2.254 is unreachable.

bye



> Hello.
>
> Try grep lrmd.* /var/log/messages
>
> Darren
>
>
> -Original Message-
> From: linux-ha-boun...@lists.linux-ha.org
> [mailto:linux-ha-boun...@lists.linux-ha.org] On Behalf Of Zausel
> Sent: 12 January 2010 18:02
> To: linux-ha@lists.linux-ha.org
> Subject: [Linux-HA] Resource Route+IPaddr2 not really working
>
> Hi everybody!
>
> I want to make a "simple" cluster resource. I have 2 nodes, every node
> have a ipaddress and a default gw. If one node is going down, the oter
> node is getting the ipaddress and gateway. So far so good. Know I
> restarting node1, the server is going down and the resource switched
> to the other node. When node1 is going online the rescources are out
> of order in my crm_mon. I can't switch the resource anymore. Only a
> hard kill of the openais will help. Does anybody have a solution?
>
>
>
> SoftwareInfos:
>
> SLES 11 64bit
> pacemaker-1.0.5-0.2.3
> heartbeat-common-3.0.0-0.3.2
> heartbeat-ldirectord-3.0.0-0.1.5
> heartbeat-resources-3.0.0-0.1.5
> openais-0.80.3-26.6.3
>
> config:
>
> node server1 \
> attributes standby="off"
> node server2 \
> attributes standby="off"
> primitive myroute ocf:heartbeat:Route \
> params destination="0.0.0.0/0" gateway="192.168.2.254"
> device="eth1" \
> meta failure-timeout="3s" \
> meta target-role="Started"
> primitive myroute2 ocf:heartbeat:Route \
> params destination="0.0.0.0/0" gateway="192.168.2.254"
> device="eth1" \
> meta failure-timeout="3s" \
> meta target-role="Started"
> primitive myip1 ocf:heartbeat:IPaddr2 \
> params ip="192.168.2.11" broadcast="192.168.2.255"
> cidr_netmask="24" nic="eth1" \
> meta target-role="Started" \
> op monitor interval="2s"
> primitive myip2 ocf:heartbeat:IPaddr2 \
> params ip="192.168.2.12" broadcast="192.168.2.255"
> cidr_netmask="24" nic="eth1" \
> meta target-role="Started" \
> op monitor interval="2s"
> group server1s myip1 myroute
> group server2s myip2 myroute2
> location MyHostIP1 server1s 100: server1
> location MyHostIP2 server2s 100: server2
> property $id="cib-bootstrap-options" \
> dc-version="1.0.5-05c8b63cbca7ce95182bb41881b3c5677f20bd5c" \
> cluster-infrastructure="openais" \
> expected-quorum-votes="2" \
> no-quorum-policy="ignore" \
> stonith-enabled="false" \
> last-lrm-refresh="1263306062"
>
>
> ___
> Linux-HA mailing list
> Linux-HA@lists.linux-ha.org
> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> See also: http://linux-ha.org/ReportingProblems
> ___
> Linux-HA mailing list
> Linux-HA@lists.linux-ha.org
> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> See also: http://linux-ha.org/ReportingProblems
>


___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


Re: [Linux-HA] Resource Route+IPaddr2 not really working

2010-01-13 Thread Darren.Mansell
It does seem to be fairly straightforward what the problem is indicating
- it can't reach the gateway to add it. Are you able to add it manually
when that happens?

-Original Message-
From: linux-ha-boun...@lists.linux-ha.org
[mailto:linux-ha-boun...@lists.linux-ha.org] On Behalf Of Zausel
Sent: 13 January 2010 10:23
To: General Linux-HA mailing list
Subject: [Linux-HA] Resource Route+IPaddr2 not really working

Hi Darren,

I got a lot of the same entries in the logfile:

Jan 12 17:34:32 server2 lrmd: [15375]: info: RA output:
(myroute:monitor:stderr) logd is not running
Jan 12 17:34:32 server2 lrmd: [15375]: info: RA output:
(myroute2:monitor:stderr) logd is not running
Jan 12 17:34:32 server2 lrmd: [15375]: info: RA output:
(myroute:monitor:stderr) 2010/01/12_17:34:32 ERROR: Gateway address
192.168.2.254 is unreachable.
Jan 12 17:34:32 server2 lrmd: [15375]: info: RA output:
(myroute2:monitor:stderr) 2010/01/12_17:34:32 ERROR: Gateway address
192.168.2.254 is unreachable.

bye



> Hello.
>
> Try grep lrmd.* /var/log/messages
>
> Darren
>
>
> -Original Message-
> From: linux-ha-boun...@lists.linux-ha.org
> [mailto:linux-ha-boun...@lists.linux-ha.org] On Behalf Of Zausel
> Sent: 12 January 2010 18:02
> To: linux-ha@lists.linux-ha.org
> Subject: [Linux-HA] Resource Route+IPaddr2 not really working
>
> Hi everybody!
>
> I want to make a "simple" cluster resource. I have 2 nodes, every node
> have a ipaddress and a default gw. If one node is going down, the oter
> node is getting the ipaddress and gateway. So far so good. Know I
> restarting node1, the server is going down and the resource switched
> to the other node. When node1 is going online the rescources are out
> of order in my crm_mon. I can't switch the resource anymore. Only a
> hard kill of the openais will help. Does anybody have a solution?
>
>
>
> SoftwareInfos:
>
> SLES 11 64bit
> pacemaker-1.0.5-0.2.3
> heartbeat-common-3.0.0-0.3.2
> heartbeat-ldirectord-3.0.0-0.1.5
> heartbeat-resources-3.0.0-0.1.5
> openais-0.80.3-26.6.3
>
> config:
>
> node server1 \
> attributes standby="off"
> node server2 \
> attributes standby="off"
> primitive myroute ocf:heartbeat:Route \
> params destination="0.0.0.0/0" gateway="192.168.2.254"
> device="eth1" \
> meta failure-timeout="3s" \
> meta target-role="Started"
> primitive myroute2 ocf:heartbeat:Route \
> params destination="0.0.0.0/0" gateway="192.168.2.254"
> device="eth1" \
> meta failure-timeout="3s" \
> meta target-role="Started"
> primitive myip1 ocf:heartbeat:IPaddr2 \
> params ip="192.168.2.11" broadcast="192.168.2.255"
> cidr_netmask="24" nic="eth1" \
> meta target-role="Started" \
> op monitor interval="2s"
> primitive myip2 ocf:heartbeat:IPaddr2 \
> params ip="192.168.2.12" broadcast="192.168.2.255"
> cidr_netmask="24" nic="eth1" \
> meta target-role="Started" \
> op monitor interval="2s"
> group server1s myip1 myroute
> group server2s myip2 myroute2
> location MyHostIP1 server1s 100: server1
> location MyHostIP2 server2s 100: server2
> property $id="cib-bootstrap-options" \
> dc-version="1.0.5-05c8b63cbca7ce95182bb41881b3c5677f20bd5c" \
> cluster-infrastructure="openais" \
> expected-quorum-votes="2" \
> no-quorum-policy="ignore" \
> stonith-enabled="false" \
> last-lrm-refresh="1263306062"
>
>
> ___
> Linux-HA mailing list
> Linux-HA@lists.linux-ha.org
> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> See also: http://linux-ha.org/ReportingProblems
> ___
> Linux-HA mailing list
> Linux-HA@lists.linux-ha.org
> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> See also: http://linux-ha.org/ReportingProblems
>


___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


[Linux-HA] Multiple Choice test for cluster knowledge

2010-01-13 Thread Michael Schwartzkopff
Hi,

together with the "Clusterbau" book I developed a quick & dirty multiple 
choice test for your knowledge about pacemaker. Download the test under:

http://www.oreilly.de/catalog/linuxhacluster2ger/Clusterbau_Exam4Fun.zip

Sorry, the test is at the moment only in German. The book is also in German. 
If there is a demand of an English translation, please let me know.

Greetings,

-- 
Dr. Michael Schwartzkopff
MultiNET Services GmbH
Addresse: Bretonischer Ring 7; 85630 Grasbrunn; Germany
Tel: +49 - 89 - 45 69 11 0
Fax: +49 - 89 - 45 69 11 21
mob: +49 - 174 - 343 28 75

mail: mi...@multinet.de
web: www.multinet.de

Sitz der Gesellschaft: 85630 Grasbrunn
Registergericht: Amtsgericht München HRB 114375
Geschäftsführer: Günter Jurgeneit, Hubert Martens

---

PGP Fingerprint: F919 3919 FF12 ED5A 2801 DEA6 AA77 57A4 EDD8 979B
Skype: misch42
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


Re: [Linux-HA] Multiple Choice test for cluster knowledge

2010-01-13 Thread Darren.Mansell
Yes please in English for both!

Have you (or anyone else) thought of doing a Linux-HA certification?

Darren

-Original Message-
From: linux-ha-boun...@lists.linux-ha.org 
[mailto:linux-ha-boun...@lists.linux-ha.org] On Behalf Of Michael Schwartzkopff
Sent: 13 January 2010 13:17
To: The Pacemaker cluster resource manager; General Linux-HA mailing list
Subject: [Linux-HA] Multiple Choice test for cluster knowledge

Hi,

together with the "Clusterbau" book I developed a quick & dirty multiple 
choice test for your knowledge about pacemaker. Download the test under:

http://www.oreilly.de/catalog/linuxhacluster2ger/Clusterbau_Exam4Fun.zip

Sorry, the test is at the moment only in German. The book is also in German. 
If there is a demand of an English translation, please let me know.

Greetings,

-- 
Dr. Michael Schwartzkopff
MultiNET Services GmbH
Addresse: Bretonischer Ring 7; 85630 Grasbrunn; Germany
Tel: +49 - 89 - 45 69 11 0
Fax: +49 - 89 - 45 69 11 21
mob: +49 - 174 - 343 28 75

mail: mi...@multinet.de
web: www.multinet.de

Sitz der Gesellschaft: 85630 Grasbrunn
Registergericht: Amtsgericht München HRB 114375
Geschäftsführer: Günter Jurgeneit, Hubert Martens

---

PGP Fingerprint: F919 3919 FF12 ED5A 2801 DEA6 AA77 57A4 EDD8 979B
Skype: misch42
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


Re: [Linux-HA] Multiple Choice test for cluster knowledge

2010-01-13 Thread Michael Schwartzkopff
Am Mittwoch, 13. Januar 2010 14:19:47 schrieb darren.mans...@opengi.co.uk:
> Yes please in English for both!
>
> Have you (or anyone else) thought of doing a Linux-HA certification?
>
> Darren
>
> -Original Message-
> From: linux-ha-boun...@lists.linux-ha.org
> [mailto:linux-ha-boun...@lists.linux-ha.org] On Behalf Of Michael
> Schwartzkopff Sent: 13 January 2010 13:17
> To: The Pacemaker cluster resource manager; General Linux-HA mailing list
> Subject: [Linux-HA] Multiple Choice test for cluster knowledge
>
> Hi,
>
> together with the "Clusterbau" book I developed a quick & dirty multiple
> choice test for your knowledge about pacemaker. Download the test under:
>
> http://www.oreilly.de/catalog/linuxhacluster2ger/Clusterbau_Exam4Fun.zip
>
> Sorry, the test is at the moment only in German. The book is also in
> German. If there is a demand of an English translation, please let me know.
>
> Greetings,

If you visit my course you get a certificate ;-)

See: http://www.gurulabs.com/linux-training/courses/GL640/

Greetings,

-- 
Dr. Michael Schwartzkopff
MultiNET Services GmbH
Addresse: Bretonischer Ring 7; 85630 Grasbrunn; Germany
Tel: +49 - 89 - 45 69 11 0
Fax: +49 - 89 - 45 69 11 21
mob: +49 - 174 - 343 28 75

mail: mi...@multinet.de
web: www.multinet.de

Sitz der Gesellschaft: 85630 Grasbrunn
Registergericht: Amtsgericht München HRB 114375
Geschäftsführer: Günter Jurgeneit, Hubert Martens

---

PGP Fingerprint: F919 3919 FF12 ED5A 2801 DEA6 AA77 57A4 EDD8 979B
Skype: misch42
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


Re: [Linux-HA] [Pacemaker] Multiple Choice test for cluster knowledge

2010-01-13 Thread Florian Haas
On 2010-01-13 14:19, darren.mans...@opengi.co.uk wrote:
> Yes please in English for both!
> 
> Have you (or anyone else) thought of doing a Linux-HA certification?

Yes. We have. It's called DRBD Certified Engineer but actually covers
not only DRBD, but Heartbeat, Corosync and Pacemaker also and will thus
be renamed LINBIT Certified Cluster Engineer very soon.

Shameless self-plug: we have on-line training coming up shortly; see
http://www.linbit.com/en/education/training-schedule/drbd-total-feb-1-mar-5-2010-online/

(I just noticed the web site still says Heartbeat only, rather than
Heartbeat/Corosync/Pacemaker; that's an oversight. Will be changed
promptly.)

Cheers,
Florian





signature.asc
Description: OpenPGP digital signature
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

[Linux-HA] Colocation of 2 resources so that it can't run together

2010-01-13 Thread jaspal singla
Hello All,

Thanks for suggesting me to upgrade from heartbeat-2.1.3 to pacemaker +
heartbeat..

Now the things seems to be so fine..but still I have some doubts. It will be
very grateful to get those doubts cleared..

I've got a 3 node cluster of OpenVZ (running VM's managed by pacemaker with
Heartbeat on iSCSI) that I'm trying to get working in a stable fashion, but
I'm having a hard time avoiding what I've dubbed the "startled herd"
problem.

Basically, the whole cluster sits there quite happily unless I defined
colocations for the resources, which I am running on the 2 machines with 1
stand by machine. Might be the way I am defining colocations is
wrong..please suggest me how I can define the colocations for my these 2
resources.

Basically the scenario is:

I have configured 2 resources groups containing all necessary file system,
ip address and vz-script.

The 3 node are as:

1) node_master ---> Active node
2) node_slave  ---> Passive node
3) node3 -> Active node

And the 2 resources are as:

1) group_vz_1
2) group_vz_2

My need is I don't want to have these 2 groups run on same machine(In case
of both active machine get crashes and resource failover to standby
machine(node_slave)).

Please let me know how I can define the colocation fro the 2 resources so
that it won't run together.

Output of crm configure show are as:

[r...@node_slave ~]# crm  configure show
node $id="075961d1-4492-4ba9-b4ad-e8c27b9e3f4b" node_master \
attributes $id="nodes-075961d1-4492-4ba9-b4ad-e8c27b9e3f4b"
node $id="7e5fdac9-80dc-41a7-bd8f-a5591a1b69a0" node3 \
attributes $id="nodes-7e5fdac9-80dc-41a7-bd8f-a5591a1b69a0"
node $id="d90b1ed2--44ac-9a4d-b435a6befd36" node_slave \
attributes $id="nodes-d90b1ed2--44ac-9a4d-b435a6befd36"
primitive filesystem1_Filesystem ocf:heartbeat:Filesystem \
operations $id="filesystem1_Filesystem-operations" \
op start interval="0" timeout="60" \
op stop interval="0" timeout="60" \
op notify interval="0" timeout="60" \
op monitor interval="20" timeout="40" start-delay="0" \
op validate-all interval="0" timeout="5" \
op meta-data interval="0" timeout="5" \
params
device="/dev/disk/by-uuid/f5feb406-685a-41f8-a4f7-170ae0925901"
directory="/vz" fstype="ext3" options="_netdev,noatime" \
meta resource-stickiness="2000"
primitive filesystem2_Filesystem ocf:heartbeat:Filesystem \
operations $id="filesystem2_Filesystem-operations" \
op start interval="0" timeout="60" \
op stop interval="0" timeout="60" \
op notify interval="0" timeout="60" \
op monitor interval="20" timeout="40" start-delay="0" \
op validate-all interval="0" timeout="5" \
op meta-data interval="0" timeout="5" \
params
device="/dev/disk/by-uuid/81c3845e-c2f6-4cb0-a0cd-e00c074942fb"
directory="/vz" fstype="ext3" options="_netdev,noatime" \
meta resource-stickiness="2000"
primitive vip2_ipaddr2 ocf:heartbeat:IPaddr2 \
operations $id="vip2_ipaddr2-operations" \
op start interval="0" timeout="90" \
op stop interval="0" timeout="100" \
op status interval="10s" timeout="20s" start-delay="0" \
op monitor interval="10s" timeout="20s" start-delay="0" \
op meta-data interval="0" timeout="5s" \
op validate-all interval="0" timeout="20s" \
params ip="66.199.245.207" nic="eth0" cidr_netmask="255.255.255.224"
\
meta resource-stickiness="2000"
primitive vip_ipaddr2 ocf:heartbeat:IPaddr2 \
operations $id="vip_ipaddr2-operations" \
op start interval="0" timeout="90" \
op stop interval="0" timeout="100" \
op status interval="10s" timeout="20s" start-delay="0" \
op monitor interval="10s" timeout="20s" start-delay="0" \
op meta-data interval="0" timeout="5s" \
op validate-all interval="0" timeout="20s" \
params ip="66.199.245.204" nic="eth0" cidr_netmask="255.255.255.224"
\
meta resource-stickiness="2000"
primitive vz1_script lsb:vz \
operations $id="vz1_script-operations" \
op start interval="0" timeout="250" \
op stop interval="0" timeout="250" \
op status interval="10" timeout="10" \
op monitor interval="15" timeout="15" start-delay="15" \
op meta-data interval="0" timeout="5" \
meta is-managed="true" resource-stickiness="2000"
primitive vz2_script lsb:vz \
operations $id="vz2_script-operations" \
op start interval="0" timeout="250" \
op stop interval="0" timeout="200" \
op status interval="10" timeout="10" \
op monitor interval="10" timeout="10" start-delay="15" \
op meta-data interval="0" timeout="5" \
meta resource-stickiness="2000" \
params $id="vz2_script-instance_attributes"
group group_vz_1 vip_ipaddr2 filesystem1_Filesystem vz1_script \
meta target-role="started"
group group_vz_2 vip2_ipaddr2 files

Re: [Linux-HA] [Pacemaker] Multiple Choice test for cluster knowledge

2010-01-13 Thread Alex Dean


On Jan 13, 2010, at 7:39 AM, Florian Haas wrote:



Shameless self-plug: we have on-line training coming up shortly; see
http://www.linbit.com/en/education/training-schedule/drbd-total-feb-1-mar-5-2010-online/


I would like to sign up, but Feb 1 - Mar 5 is bad timing for me.  I  
would definitely appreciate an announcement about future online  
sessions.  I'm primarily interested in heartbeat+pacemaker, but drbd  
is useful to know about as well.


regards,
alex


PGP.sig
Description: This is a digitally signed message part
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

[Linux-HA] CRM strange errors

2010-01-13 Thread Hunny Bunny
Hello folkz,
I have compiled Corosync/OpenAIS, Pacemaker, DRBD and OCFS2 cluster
environment,  generated cib.xml file with haresources2cib.py script, and put it 
into
/var/lib/heartbeat/crm directory

When I try to run either 'crm configure show' or 'crm configure'
I'm getting these errors:

INFO: no curses support: you won't see colors
Signon to CIB failed: connection failed
Init failed, could not perform requested operations
ERROR: cannot parse xml: no element found: line 1, column 0
ERROR: No CIB!

Does it look like a problem with openSUSE 11.2 implementations of ncurses, 
libxml2 or libxslt?

Could somebody please give me a hint?

Thanks in advance,

Alex

 








  
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


[Linux-HA] Virtual mac address

2010-01-13 Thread Benon Fuertes
Hello everyone.

I've been trying to use heartbeat version 2 with the script IPaddr2 and I've 
managed to get it running with cibadmin, crm_mon, etc.
The problem is that it doesn't get the virtual mac address that I establish in 
the xml configuration. 

"crm_mon" tells me this:


Last updated: Wed Jan 13 17:55:01 2010
Stack: Heartbeat
Current DC: heart (883e1da9-c47e-4eeb-850c-e118a31a2d55) - partition with quorum
Version: 1.0.5-3840e6b5a305ccb803d29b468556739e75532d56
2 Nodes configured, unknown expected votes
1 Resources configured.


Online: [ heart bacula ]

Resource Group: group_1
rscIPaddr2  (ocf::heartbeat:IPaddr2):   Started bacula





So IPaddr2 it's running, but when I look at the ip configuration and make tests 
with ping and look at ARP cache, I don't see the virtual mac anywhere.

And before someone warns me about this topic, yes, I've read this old thread:  
http://lists.linux-ha.org/pipermail/linux-ha/2008-July/033776.html



Both servers are running ubuntu-server 9.10 Karmic, the master node with kernel 
2.6.31-16-386 and the slave with 2.6.31.14-generic-pae.
I only want to put in high-availability a virtual ip and a virtual mac, to 
solve ARP problems on switchs.

This is my cib.xml:


  

  

  
  
  
  
  
  
  
  
  
  
  
  
  
  
  
  
  
  
  
  

  


  
  


  

  

  
  

  
  
  
  
  

  

  


  

  

  

  



Thanks in advance.


___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


[Linux-HA] Question

2010-01-13 Thread Yeary, Mark A.
Hello All,

 I'm looking for advice.  I'm using drbd-8.3.5.  I have created a
drbd cluster active/active with gfs2 on the two nodes.  Can I access
this cluster from other nodes besides the two I've created drbd on and
maintain sync.  In other words, can I add four nodes to the gfs2 cluster
(only two are the drbd) and access it from the four nodes using gfs2 and
maintain sync?  Note: I am not increasing my drbd cluster size (still
two nodes), I am just adding gfs2 nodes to make a total of four.

 I want to say that drbd is great :-)  I have been using it for ten
years.  Philip, you won't remember me but I still owe you a drink :-)  

Mark  
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


Re: [Linux-HA] Colocation of 2 resources so that it can't run together

2010-01-13 Thread Lars Marowsky-Bree
On 2010-01-13T20:38:00, jaspal singla  wrote:

> group group_vz_1 vip_ipaddr2 filesystem1_Filesystem vz1_script \
> meta target-role="started"
> group group_vz_2 vip2_ipaddr2 filesystem2_Filesystem vz2_script \
> meta target-role="started"
> location location_master group_vz_1 700: node_master
> location location_node3 group_vz_2 600: node3
> location location_slave_1 group_vz_1 0: node_slave
> location location_slave_2 group_vz_2 0: node_slave
> colocation colocation_vz_test -inf: group_vz_1 group_vz_2

The anti-collocation rule you have is correct, and this should result in
the resources not being placed on the same node.

> property $id="cib-bootstrap-options" \
> dc-version="1.0.5-462f1569a43740667daf7b0f6b521742e9eb8fa7" \
> cluster-infrastructure="Heartbeat" \
> stonith-enabled="false" \
> symmetric-cluster="false" \
> last-lrm-refresh="1263311374"

Disabling stonith is not a good idea if you're running shared storage.

> rsc_defaults $id="rsc_defaults-options"

You may want to enable resource-stickiness to avoid resources shuffling
around needlessly.


Regards,
Lars

-- 
Architect Storage/HA, OPS Engineering, Novell, Inc.
SUSE LINUX Products GmbH, GF: Markus Rex, HRB 16746 (AG Nürnberg)
"Experience is the name everyone gives to their mistakes." -- Oscar Wilde

___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


Re: [Linux-HA] Question

2010-01-13 Thread Thomas Halinka
Hello Mark,

Am Mittwoch, den 13.01.2010, 15:57 -0500 schrieb Yeary, Mark A.:
> Hello All,
> 
>  I'm looking for advice.  I'm using drbd-8.3.5.  I have created a
> drbd cluster active/active with gfs2 on the two nodes.  Can I access
> this cluster from other nodes besides the two I've created drbd on and
> maintain sync.  In other words, can I add four nodes to the gfs2 cluster
> (only two are the drbd) and access it from the four nodes using gfs2 and
> maintain sync?  Note: I am not increasing my drbd cluster size (still
> two nodes), I am just adding gfs2 nodes to make a total of four.

But how should ne these "just-gfs-nodes" be able to see or attach these
storage, since your drbd-device is _not_ shared to this hosts. Your
Nodes are not able to attch 

I would suggest, you want to export your drbd-device through aoe, iscsi,
fcoe or whatever.

A high-available storage-server through linux-ha could be the name to
the next topic you open ;-)

>  I want to say that drbd is great :-)  I have been using it for ten
> years.  Philip, you won't remember me but I still owe you a drink :-)  
> 
> Mark  

hth,

Thomas

> ___
> Linux-HA mailing list
> Linux-HA@lists.linux-ha.org
> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> See also: http://linux-ha.org/ReportingProblems

___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


[Linux-HA] high-available storage-server through linux-ha

2010-01-13 Thread Yeary, Mark A.
Hi Thomas,

 I should have been more clear in my explanation.  We have a two
physical node drbd cluster with GFS2.  We have tried to add two more
nodes by increasing the gfs2 journal to four, the two in the drbd
cluster and two separate nodes; all four with gfs2 formatting.  All four
nodes are connected via virtual machine (KVM) virtio disks.  We can
access them but they do not maintain synchronization.  

 Becoming frustrated (only a little :-), we went back to the two
with drbd and gfs2.  We then tried to connect with two more nodes (a
total of four) via iscsi.  That didn't seem to work either.  Anything
you can tell us would be helpful :-)

Truly,

Mark


 

-Original Message-
From: linux-ha-boun...@lists.linux-ha.org
[mailto:linux-ha-boun...@lists.linux-ha.org] On Behalf Of Thomas Halinka
Sent: Wednesday, January 13, 2010 4:40 PM
To: General Linux-HA mailing list
Subject: Re: [Linux-HA] Question

Hello Mark,

Am Mittwoch, den 13.01.2010, 15:57 -0500 schrieb Yeary, Mark A.:
> Hello All,
> 
>  I'm looking for advice.  I'm using drbd-8.3.5.  I have created a
> drbd cluster active/active with gfs2 on the two nodes.  Can I access
> this cluster from other nodes besides the two I've created drbd on and
> maintain sync.  In other words, can I add four nodes to the gfs2
cluster
> (only two are the drbd) and access it from the four nodes using gfs2
and
> maintain sync?  Note: I am not increasing my drbd cluster size (still
> two nodes), I am just adding gfs2 nodes to make a total of four.

But how should ne these "just-gfs-nodes" be able to see or attach these
storage, since your drbd-device is _not_ shared to this hosts. Your
Nodes are not able to attch 

I would suggest, you want to export your drbd-device through aoe, iscsi,
fcoe or whatever.

A high-available storage-server through linux-ha could be the name to
the next topic you open ;-)

>  I want to say that drbd is great :-)  I have been using it for
ten
> years.  Philip, you won't remember me but I still owe you a drink :-)

> 
> Mark  

hth,

Thomas

> ___
> Linux-HA mailing list
> Linux-HA@lists.linux-ha.org
> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> See also: http://linux-ha.org/ReportingProblems

___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

No virus found in this incoming message.
Checked by AVG - www.avg.com 
Version: 8.5.432 / Virus Database: 270.14.138/2618 - Release Date:
01/13/10 07:35:00
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


Re: [Linux-HA] high-available storage-server through linux-ha

2010-01-13 Thread Thomas Halinka
Hi Mark,

Am Mittwoch, den 13.01.2010, 17:38 -0500 schrieb Yeary, Mark A.:
> Hi Thomas,
> 
>  I should have been more clear in my explanation.  We have a two
> physical node drbd cluster with GFS2.  We have tried to add two more
> nodes by increasing the gfs2 journal to four, the two in the drbd
> cluster and two separate nodes; all four with gfs2 formatting.  All four
> nodes are connected via virtual machine (KVM) virtio disks.

So your hosts are vms?

>   We can
> access them but they do not maintain synchronization.  

.. because they all access _one common_ blockdevice passed through by
kvm? Two of them play drbd with this device while the others do not?

>  Becoming frustrated (only a little :-), we went back to the two
> with drbd and gfs2.  We then tried to connect with two more nodes (a
> total of four) via iscsi.  That didn't seem to work either.

Why?

Just dd if=/dev/zero of=/MOUNTPOINT_GFS/10GB bs=1M count=1

And add this as a Lun to e.g. IET

>  Anything
> you can tell us would be helpful :-)

We should get more clearness about what you need - what is your goal?

Do you want to access a Filesystem from all this hosts?
Well known NFS with or GFS/OCFS2 on DRBD with iSCSI, AoE or FCoE
could be a Solution...

Or do you want to Share your _Raw Storage_?
iSCSI, AoE, FCoE on LVM on top of DRBD would be nice...


> 
> Truly,
> 
> Mark
> 
cu,

thomas



___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems


Re: [Linux-HA] heartbeat waits for initdead even after all nodes have joined

2010-01-13 Thread David Sickmiller
I don't have autojoin in my ha.cf, and I believe it defaults to
"autojoin none", so that wouldn't explain why heartbeat keeps waiting
after all nodes have joined.

I can see in /var/log/messages where crmd is doing the waiting for my
900-second initdead:

2010-01-11T13:51:15.428916-05:00 crmd: [4273]: info: do_started: The
local CRM is operational
2010-01-11T13:51:15.428924-05:00 crmd: [4273]: info:
do_state_transition: State transition S_STARTING -> S_PENDING [
input=I_PENDING cause=C_FSA_INTERNAL origin=do_started ]
2010-01-11T14:06:15.964307-05:00 crmd: [4273]: info: crm_timer_popped:
Election Trigger (I_DC_TIMEOUT) just popped!
2010-01-11T14:06:15.964337-05:00 crmd: [4273]: WARN: do_log: [[FSA]]
Input I_DC_TIMEOUT from crm_timer_popped() received in state (S_PENDING)
2010-01-11T14:06:15.964348-05:00 crmd: [4273]: info:
do_state_transition: State transition S_PENDING -> S_ELECTION [
input=I_DC_TIMEOUT cause=C_TIMER_POPPED origin=crm_timer_popped ]

I am using "Version 2 Resource Manager".  I didn't previously realize
this was the last version before the split.

I am also using DRBD, and yesterday I discovered that its
wait-for-connection timeout (wfc-timeout) works as I had hoped initdead
would, and by putting it before heartbeat in the startup sequence, it
turns out I don't really need initdead after all.

Thanks,
David


-Original Message-
From: linux-ha-boun...@lists.linux-ha.org
[mailto:linux-ha-boun...@lists.linux-ha.org] On Behalf Of Dejan
Muhamedagic
Sent: Tuesday, January 12, 2010 3:51 AM
To: General Linux-HA mailing list
Subject: Re: [Linux-HA] heartbeat waits for initdead even after all
nodes have joined

Hi,

On Mon, Jan 11, 2010 at 03:21:05PM -0500, David Sickmiller wrote:
> Hi,
> 
>  
> 
> I was hoping to configure my 2-node cluster to start as soon as both
> nodes were present but wait up to 15 minutes if the other node was
> missing upon system startup.  In my case, a delay of several minutes
is
> better than a split-brain scenario.  The Linux-HA documentation says
> "The initdead parameter is used to set the time that it takes to
declare
> a cluster node dead when Heartbeat is first started.", so I figured I
> could just set "initdead 900" in ha.cf.  Unfortunately, heartbeat
seems
> to be waiting for the entire initdead time interval regardless of
> whether all the nodes are present.
> 
>  
> 
> Does this match others' experiences?  Is there a different setting
that
> could accomplish my objective?
> 
>  
> 
> It seems like the documentation would be more accurate if it said "The
> initdead parameter is used to set the time that heartbeat waits before
> starting any resources, which allows time for additional nodes to
join."

If you have autojoin set to "any".

> However, I would much prefer that Linux-HA behaved according to the
> original documentation.
> 
>  
> 
> I'm using Heartbeat 2.1.4 on RHEL 5.4.

Please switch to Pacemaker/heartbeat or Pacemaker/corosync. Or
are you using v1/haresources?

Thanks,

Dejan
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems
___
Linux-HA mailing list
Linux-HA@lists.linux-ha.org
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems