On Thu, Feb 14, 2008 at 1:52 PM, Schmidt, Florian <[EMAIL PROTECTED]> wrote: > Hey there, > > I think I do have a valid v2-Configuration now. I got the python-script > running, with adding the variables manually. I don't understand > everything in the CIB, but I think it works. ;) > > Now I want to specify ONE ping node. (the switch, where both nodes are > plugged in) to check if the line between the switch an the nodes are OK > or not. > > I think I have to configure this using a pingd-clone and a constraint?!? > > The clone will run on both nodes and when one node will lose the > connection to the switch, all resources will be switched to the other > node. > If both nodes don't have connection to the switch, the resources don't > need to switch. > > Is this possible with this snippets, I copied from > http://linux-ha.org/v2/faq/pingd?highlight=%28v2/faq/%29 > > <clone id="pingd-clone"> > <meta_attributes id="pingd-clone-ma"> > <attributes> > <nvpair id="pingd-clone-1" name="globally_unique" value="false"/> > </attributes> > </meta_attributes> > <primitive id="pingd-child" provider="heartbeat" class="ocf" > type="pingd"> > <operations> > <op id="pingd-child-monitor" name="monitor" interval="20s" > timeout="60s" prereq="nothing"/> > <op id="pingd-child-start" name="start" prereq="nothing"/> > </operations> > <instance_attributes id="pingd_inst_attr"> > <attributes> > <nvpair id="pingd-1" name="dampen" value="5s"/> > <nvpair id="pingd-2" name="multiplier" value="100"/> > </attributes> > </instance_attributes> > </primitive> > </clone> > > and > > > <rsc_location id="my_resource:connected" rsc="my_resource"> > <rule id="my_resource:connected:rule" > score_attribute="default_ping_set"> > <expression id="my_resource:connected:expr:defined" > attribute="default_ping_set" operation="defined"/> > </rule> > </rsc_location> > > Are this the right snippets for this purpose?
looks like it > > > Next question: > The haresources2cib.py-script created the following lines at each of my > resources > > <op id="Filesystem_3_mon" interval="120s" name="monitor" timeout="60s"/> > > Does this mean, that the status of the Filesystem gets monitored? yes > If > somebody will unmount the filesystem by hand, will in the worst case be > discovered after the interval of 120 seconds and then Heartbeat will > try to mount it again. yes > If this failes it will switch the resources to > the second node after the timeout of 60 seconds?!? no. that is determined by the failure stickiness settings > One more question: > How to update the CIB? I read somewhere not to manipulate the cib.xml > with an editor. But I didn't find a way to add the lines ((e.g for the > clone above) with one of the tools. Is there a tool or do I have to edit > it manually? http://clusterlabs.org/mw/Image:Configuration_Explained.pdf > > So far....further questions will follow soon :P > > Tanks for answering > > Florian > > _______________________________________________ > Linux-HA mailing list > [email protected] > http://lists.linux-ha.org/mailman/listinfo/linux-ha > See also: http://linux-ha.org/ReportingProblems > _______________________________________________ Linux-HA mailing list [email protected] http://lists.linux-ha.org/mailman/listinfo/linux-ha See also: http://linux-ha.org/ReportingProblems
