On Thu, Jun 21, 2012 at 12:11 AM, emmanuel segura <emi2f...@gmail.com> wrote: > I don't know but see the fail it's in the operation lx0_monitor_0, so i ask > to someone with more experience then me, if pacemaker does a monitor > operation before start?
Always. We never start a resource unless we know for sure its not already running somewhere. Thats what we use non-recurring monitor operations for. > > maybe when you restart the resource something goes wrong and the resource > fail and after that it's blocked > > ================ > on-fail="block" > ================ > > > > 2012/6/20 Kadlecsik József <kadlecsik.joz...@wigner.mta.hu> >> >> On Wed, 20 Jun 2012, emmanuel segura wrote: >> >> > Why you say there is not error in the message >> > ========================================================= >> > Jun 20 11:57:25 atlas4 lrmd: [17568]: info: operation monitor[35] on lx0 >> > for client 17571: pid 30179 exited with return code 7 >> > Jun 20 11:57:25 atlas4 crmd: [17571]: debug: create_operation_update: >> > do_update_resource: Updating resouce lx0 after complete monitor op >> > (interval=0) >> > Jun 20 11:57:25 atlas4 crmd: [17571]: info: process_lrm_event: LRM >> > operation lx0_monitor_0 (call=35, rc=7, cib-update=61, confirmed=true) >> > not >> > running >> >> I interpreted those lines as a checking that the resource hasn't been >> started yet (confirmed=true). And indeed, it's not running so the return >> code is OCF_NOT_RUNNING. >> >> There's no log message about an attempt to start the resource. >> >> Best regards, >> Jozsef >> >> > 2012/6/20 Kadlecsik József <kadlecsik.joz...@wigner.mta.hu> >> > Hello, >> > >> > Somehow a VirtualDomain resource after a "crm resource restart", >> > which did >> > *not* start the resource but just stop, the resource cannot be >> > started >> > anymore. The most baffling is that I do not see an error >> > message. The >> > resource in question, named 'lx0', can be started directly via >> > virsh/libvirt and libvirtd is running on all cluster nodes. >> > >> > We run corosync 1.4.2-1~bpo60+1, pacemaker 1.1.6-2~bpo60+1 >> > (debian). >> > >> > # crm status >> > ============ >> > Last updated: Wed Jun 20 15:14:44 2012 >> > Last change: Wed Jun 20 14:07:40 2012 via cibadmin on atlas0 >> > Stack: openais >> > Current DC: atlas0 - partition with quorum >> > Version: 1.1.6-9971ebba4494012a93c03b40a2c58ec0eb60f50c >> > 7 Nodes configured, 7 expected votes >> > 18 Resources configured. >> > ============ >> > >> > Online: [ atlas0 atlas1 atlas2 atlas3 atlas4 atlas5 atlas6 ] >> > >> > kerberos (ocf::heartbeat:VirtualDomain): Started atlas0 >> > stonith-atlas3 (stonith:ipmilan): Started atlas4 >> > stonith-atlas1 (stonith:ipmilan): Started atlas4 >> > stonith-atlas2 (stonith:ipmilan): Started atlas4 >> > stonith-atlas0 (stonith:ipmilan): Started atlas4 >> > stonith-atlas4 (stonith:ipmilan): Started atlas3 >> > mailman (ocf::heartbeat:VirtualDomain): Started atlas6 >> > indico (ocf::heartbeat:VirtualDomain): Started atlas0 >> > papi (ocf::heartbeat:VirtualDomain): Started atlas1 >> > wwwd (ocf::heartbeat:VirtualDomain): Started atlas2 >> > webauth (ocf::heartbeat:VirtualDomain): Started atlas3 >> > caladan (ocf::heartbeat:VirtualDomain): Started atlas4 >> > radius (ocf::heartbeat:VirtualDomain): Started atlas5 >> > mail0 (ocf::heartbeat:VirtualDomain): Started atlas6 >> > stonith-atlas5 (stonith:apcmastersnmp): Started atlas4 >> > stonith-atlas6 (stonith:apcmastersnmp): Started atlas4 >> > w0 (ocf::heartbeat:VirtualDomain): Started atlas2 >> > >> > # crm resource show >> > kerberos (ocf::heartbeat:VirtualDomain) Started >> > stonith-atlas3 (stonith:ipmilan) Started >> > stonith-atlas1 (stonith:ipmilan) Started >> > stonith-atlas2 (stonith:ipmilan) Started >> > stonith-atlas0 (stonith:ipmilan) Started >> > stonith-atlas4 (stonith:ipmilan) Started >> > mailman (ocf::heartbeat:VirtualDomain) Started >> > indico (ocf::heartbeat:VirtualDomain) Started >> > papi (ocf::heartbeat:VirtualDomain) Started >> > wwwd (ocf::heartbeat:VirtualDomain) Started >> > webauth (ocf::heartbeat:VirtualDomain) Started >> > caladan (ocf::heartbeat:VirtualDomain) Started >> > radius (ocf::heartbeat:VirtualDomain) Started >> > mail0 (ocf::heartbeat:VirtualDomain) Started >> > stonith-atlas5 (stonith:apcmastersnmp) Started >> > stonith-atlas6 (stonith:apcmastersnmp) Started >> > w0 (ocf::heartbeat:VirtualDomain) Started >> > lx0 (ocf::heartbeat:VirtualDomain) Stopped >> > >> > # crm configure show >> > node atlas0 \ >> > attributes standby="false" \ >> > utilization memory="24576" >> > node atlas1 \ >> > attributes standby="false" \ >> > utilization memory="24576" >> > node atlas2 \ >> > attributes standby="false" \ >> > utilization memory="24576" >> > node atlas3 \ >> > attributes standby="false" \ >> > utilization memory="24576" >> > node atlas4 \ >> > attributes standby="false" \ >> > utilization memory="24576" >> > node atlas5 \ >> > attributes standby="off" \ >> > utilization memory="20480" >> > node atlas6 \ >> > attributes standby="off" \ >> > utilization memory="20480" >> > primitive caladan ocf:heartbeat:VirtualDomain \ >> > params config="/etc/libvirt/crm/caladan.xml" >> > hypervisor="qemu:///system" \ >> > meta allow-migrate="true" target-role="Started" >> > is-managed="true" \ >> > op start interval="0" timeout="120s" \ >> > op stop interval="0" timeout="120s" \ >> > op monitor interval="10s" timeout="40s" depth="0" \ >> > op migrate_to interval="0" timeout="240s" on-fail="block" >> > \ >> > op migrate_from interval="0" timeout="240s" >> > on-fail="block" \ >> > utilization memory="4608" >> > primitive indico ocf:heartbeat:VirtualDomain \ >> > params config="/etc/libvirt/crm/indico.xml" >> > hypervisor="qemu:///system" \ >> > meta allow-migrate="true" target-role="Started" >> > is-managed="true" \ >> > op start interval="0" timeout="120s" \ >> > op stop interval="0" timeout="120s" \ >> > op monitor interval="10s" timeout="40s" depth="0" \ >> > op migrate_to interval="0" timeout="240s" on-fail="block" >> > \ >> > op migrate_from interval="0" timeout="240s" >> > on-fail="block" \ >> > utilization memory="5120" >> > primitive kerberos ocf:heartbeat:VirtualDomain \ >> > params config="/etc/libvirt/qemu/kerberos.xml" >> > hypervisor="qemu:///system" \ >> > meta allow-migrate="true" target-role="Started" >> > is-managed="true" \ >> > op start interval="0" timeout="120s" \ >> > op stop interval="0" timeout="120s" \ >> > op monitor interval="10s" timeout="40s" depth="0" \ >> > op migrate_to interval="0" timeout="240s" on-fail="block" >> > \ >> > op migrate_from interval="0" timeout="240s" >> > on-fail="block" \ >> > utilization memory="4608" >> > primitive lx0 ocf:heartbeat:VirtualDomain \ >> > params config="/etc/libvirt/crm/lx0.xml" >> > hypervisor="qemu:///system" \ >> > meta allow-migrate="true" target-role="Started" >> > is-managed="true" \ >> > op start interval="0" timeout="120s" \ >> > op stop interval="0" timeout="120s" \ >> > op monitor interval="10s" timeout="40s" depth="0" \ >> > op migrate_to interval="0" timeout="240s" on-fail="block" >> > \ >> > op migrate_from interval="0" timeout="240s" >> > on-fail="block" \ >> > utilization memory="4608" >> > primitive mail0 ocf:heartbeat:VirtualDomain \ >> > params config="/etc/libvirt/crm/mail0.xml" >> > hypervisor="qemu:///system" \ >> > meta allow-migrate="true" target-role="Started" >> > is-managed="true" \ >> > op start interval="0" timeout="120s" \ >> > op stop interval="0" timeout="120s" \ >> > op monitor interval="10s" timeout="40s" depth="0" \ >> > op migrate_to interval="0" timeout="240s" on-fail="block" >> > \ >> > op migrate_from interval="0" timeout="240s" >> > on-fail="block" \ >> > utilization memory="4608" >> > primitive mailman ocf:heartbeat:VirtualDomain \ >> > params config="/etc/libvirt/crm/mailman.xml" >> > hypervisor="qemu:///system" \ >> > meta allow-migrate="true" target-role="Started" >> > is-managed="true" \ >> > op start interval="0" timeout="120s" \ >> > op stop interval="0" timeout="120s" \ >> > op monitor interval="10s" timeout="40s" depth="0" \ >> > op migrate_to interval="0" timeout="240s" on-fail="block" >> > \ >> > op migrate_from interval="0" timeout="240s" >> > on-fail="block" \ >> > utilization memory="5120" >> > primitive papi ocf:heartbeat:VirtualDomain \ >> > params config="/etc/libvirt/crm/papi.xml" >> > hypervisor="qemu:///system" \ >> > meta allow-migrate="true" target-role="Started" >> > is-managed="true" \ >> > op start interval="0" timeout="120s" \ >> > op stop interval="0" timeout="120s" \ >> > op monitor interval="10s" timeout="40s" depth="0" \ >> > op migrate_to interval="0" timeout="240s" on-fail="block" >> > \ >> > op migrate_from interval="0" timeout="240s" >> > on-fail="block" \ >> > utilization memory="6144" >> > primitive radius ocf:heartbeat:VirtualDomain \ >> > params config="/etc/libvirt/crm/radius.xml" >> > hypervisor="qemu:///system" \ >> > meta allow-migrate="true" target-role="Started" >> > is-managed="true" \ >> > op start interval="0" timeout="120s" \ >> > op stop interval="0" timeout="120s" \ >> > op monitor interval="10s" timeout="40s" depth="0" \ >> > op migrate_to interval="0" timeout="240s" on-fail="block" >> > \ >> > op migrate_from interval="0" timeout="240s" >> > on-fail="block" \ >> > utilization memory="4608" >> > primitive stonith-atlas0 stonith:ipmilan \ >> > params hostname="atlas0" ipaddr="192.168.40.20" >> > port="623" >> > auth="md5" priv="admin" login="root" password="XXXXX" \ >> > op start interval="0" timeout="120s" \ >> > meta target-role="Started" >> > primitive stonith-atlas1 stonith:ipmilan \ >> > params hostname="atlas1" ipaddr="192.168.40.21" >> > port="623" >> > auth="md5" priv="admin" login="root" password="XXXX" \ >> > op start interval="0" timeout="120s" \ >> > meta target-role="Started" >> > primitive stonith-atlas2 stonith:ipmilan \ >> > params hostname="atlas2" ipaddr="192.168.40.22" >> > port="623" >> > auth="md5" priv="admin" login="root" password="XXXX" \ >> > op start interval="0" timeout="120s" \ >> > meta target-role="Started" >> > primitive stonith-atlas3 stonith:ipmilan \ >> > params hostname="atlas3" ipaddr="192.168.40.23" >> > port="623" >> > auth="md5" priv="admin" login="root" password="XXXX" \ >> > op start interval="0" timeout="120s" \ >> > meta target-role="Started" >> > primitive stonith-atlas4 stonith:ipmilan \ >> > params hostname="atlas4" ipaddr="192.168.40.24" >> > port="623" >> > auth="md5" priv="admin" login="root" password="XXXX" \ >> > op start interval="0" timeout="120s" \ >> > meta target-role="Started" >> > primitive stonith-atlas5 stonith:apcmastersnmp \ >> > params ipaddr="192.168.40.252" port="161" >> > community="XXXX" >> > pcmk_host_list="atlas5" pcmk_host_check="static-list" >> > primitive stonith-atlas6 stonith:apcmastersnmp \ >> > params ipaddr="192.168.40.252" port="161" >> > community="XXXX" >> > pcmk_host_list="atlas6" pcmk_host_check="static-list" >> > primitive w0 ocf:heartbeat:VirtualDomain \ >> > params config="/etc/libvirt/crm/w0.xml" >> > hypervisor="qemu:///system" \ >> > meta allow-migrate="true" target-role="Started" \ >> > op start interval="0" timeout="120s" \ >> > op stop interval="0" timeout="120s" \ >> > op monitor interval="10s" timeout="40s" depth="0" \ >> > op migrate_to interval="0" timeout="240s" on-fail="block" >> > \ >> > op migrate_from interval="0" timeout="240s" >> > on-fail="block" \ >> > utilization memory="4608" >> > primitive webauth ocf:heartbeat:VirtualDomain \ >> > params config="/etc/libvirt/crm/webauth.xml" >> > hypervisor="qemu:///system" \ >> > meta allow-migrate="true" target-role="Started" >> > is-managed="true" \ >> > op start interval="0" timeout="120s" \ >> > op stop interval="0" timeout="120s" \ >> > op monitor interval="10s" timeout="40s" depth="0" \ >> > op migrate_to interval="0" timeout="240s" on-fail="block" >> > \ >> > op migrate_from interval="0" timeout="240s" >> > on-fail="block" \ >> > utilization memory="4608" >> > primitive wwwd ocf:heartbeat:VirtualDomain \ >> > params config="/etc/libvirt/crm/wwwd.xml" >> > hypervisor="qemu:///system" \ >> > meta allow-migrate="true" target-role="Started" >> > is-managed="true" \ >> > op start interval="0" timeout="120s" \ >> > op stop interval="0" timeout="120s" \ >> > op monitor interval="10s" timeout="40s" depth="0" \ >> > op migrate_to interval="0" timeout="240s" on-fail="block" >> > \ >> > op migrate_from interval="0" timeout="240s" >> > on-fail="block" \ >> > utilization memory="5120" >> > location location-stonith-atlas0 stonith-atlas0 -inf: atlas0 >> > location location-stonith-atlas1 stonith-atlas1 -inf: atlas1 >> > location location-stonith-atlas2 stonith-atlas2 -inf: atlas2 >> > location location-stonith-atlas3 stonith-atlas3 -inf: atlas3 >> > location location-stonith-atlas4 stonith-atlas4 -inf: atlas4 >> > location location-stonith-atlas5 stonith-atlas5 -inf: atlas5 >> > location location-stonith-atlas6 stonith-atlas6 -inf: atlas6 >> > property $id="cib-bootstrap-options" \ >> > >> > dc-version="1.1.6-9971ebba4494012a93c03b40a2c58ec0eb60f50c" \ >> > cluster-infrastructure="openais" \ >> > expected-quorum-votes="7" \ >> > stonith-enabled="true" \ >> > no-quorum-policy="stop" \ >> > last-lrm-refresh="1340193431" \ >> > symmetric-cluster="true" \ >> > maintenance-mode="false" \ >> > stop-all-resources="false" \ >> > is-managed-default="true" \ >> > placement-strategy="balanced" >> > >> > # crm_verify -L -VV >> > [...] >> > crm_verify[19320]: 2012/06/20_15:25:50 notice: LogActions: Leave >> > w0 >> > (Started atlas2) >> > crm_verify[19320]: 2012/06/20_15:25:50 notice: LogActions: Leave >> > stonith-atlas6 (Started atlas4) >> > crm_verify[19320]: 2012/06/20_15:25:50 notice: LogActions: Leave >> > stonith-atlas5 (Started atlas4) >> > crm_verify[19320]: 2012/06/20_15:25:50 notice: LogActions: Leave >> > stonith-atlas4 (Started atlas3) >> > crm_verify[19320]: 2012/06/20_15:25:50 notice: LogActions: Leave >> > stonith-atlas3 (Started atlas4) >> > crm_verify[19320]: 2012/06/20_15:25:50 notice: LogActions: Leave >> > stonith-atlas2 (Started atlas4) >> > crm_verify[19320]: 2012/06/20_15:25:50 notice: LogActions: Leave >> > stonith-atlas1 (Started atlas4) >> > crm_verify[19320]: 2012/06/20_15:25:50 notice: LogActions: Leave >> > stonith-atlas0 (Started atlas4) >> > crm_verify[19320]: 2012/06/20_15:25:50 notice: LogActions: Start >> > lx0 >> > (atlas4) >> > >> > I have tried to delete the resource and add again, did not help. >> > The corresponding log entries: >> > >> > Jun 20 11:57:25 atlas4 crmd: [17571]: info: delete_resource: >> > Removing >> > resource lx0 for 28654_crm_resource (internal) on atlas0 >> > Jun 20 11:57:25 atlas4 lrmd: [17568]: debug: lrmd_rsc_destroy: >> > removing >> > resource lx0 >> > Jun 20 11:57:25 atlas4 crmd: [17571]: debug: delete_rsc_entry: >> > sync: >> > Sending delete op for lx0 >> > Jun 20 11:57:25 atlas4 crmd: [17571]: info: notify_deleted: >> > Notifying >> > 28654_crm_resource on atlas0 that lx0 was deleted >> > Jun 20 11:57:25 atlas4 crmd: [17571]: WARN: >> > decode_transition_key: Bad >> > UUID (crm-resource-28654) in sscanf result (3) for >> > 0:0:crm-resource-28654 >> > Jun 20 11:57:25 atlas4 crmd: [17571]: debug: >> > create_operation_update: >> > send_direct_ack: Updating resouce lx0 after complete delete op >> > (interval=60000) >> > Jun 20 11:57:25 atlas4 crmd: [17571]: info: send_direct_ack: >> > ACK'ing >> > resource op lx0_delete_60000 from 0:0:crm-resource-28654: >> > lrm_invoke-lrmd-1340186245-16 >> > Jun 20 11:57:25 atlas4 corosync[17530]: [TOTEM ] mcasted >> > message added >> > to pending queue >> > Jun 20 11:57:25 atlas4 corosync[17530]: [TOTEM ] mcasted >> > message added >> > to pending queue >> > Jun 20 11:57:25 atlas4 corosync[17530]: [TOTEM ] Delivering >> > 10d5 to 10d7 >> > Jun 20 11:57:25 atlas4 corosync[17530]: [TOTEM ] Delivering >> > MCAST >> > message with seq 10d6 to pending delivery queue >> > Jun 20 11:57:25 atlas4 corosync[17530]: [TOTEM ] Delivering >> > MCAST >> > message with seq 10d7 to pending delivery queue >> > Jun 20 11:57:25 atlas4 corosync[17530]: [TOTEM ] Received >> > ringid(192.168.40.60:22264) seq 10d6 >> > Jun 20 11:57:25 atlas4 corosync[17530]: [TOTEM ] Received >> > ringid(192.168.40.60:22264) seq 10d7 >> > Jun 20 11:57:25 atlas4 crmd: [17571]: debug: notify_deleted: >> > Triggering a >> > refresh after 28654_crm_resource deleted lx0 from the LRM >> > Jun 20 11:57:25 atlas4 cib: [17567]: debug: cib_process_xpath: >> > Processing >> > cib_query op for >> > >> > //cib/configuration/crm_config//cluster_property_set//nvpair[@name='last-lr >> > m-refresh'] >> > (/cib/configuration/crm_config/cluster_property_set/nvpair[6]) >> > >> > >> > Jun 20 11:57:25 atlas4 lrmd: [17568]: debug: >> > on_msg_add_rsc:client [17571] >> > adds resource lx0 >> > Jun 20 11:57:25 atlas4 corosync[17530]: [TOTEM ] Delivering >> > 149e to 149f >> > Jun 20 11:57:25 atlas4 corosync[17530]: [TOTEM ] Delivering >> > MCAST >> > message with seq 149f to pending delivery queue >> > Jun 20 11:57:25 atlas4 corosync[17530]: [TOTEM ] Received >> > ringid(192.168.40.60:22264) seq 14a0 >> > Jun 20 11:57:25 atlas4 corosync[17530]: [TOTEM ] Delivering >> > 149f to 14a0 >> > Jun 20 11:57:25 atlas4 corosync[17530]: [TOTEM ] Delivering >> > MCAST >> > message with seq 14a0 to pending delivery queue >> > Jun 20 11:57:25 atlas4 corosync[17530]: [TOTEM ] releasing >> > messages up >> > to and including 149e >> > Jun 20 11:57:25 atlas4 crmd: [17571]: info: do_lrm_rsc_op: >> > Performing >> > key=26:10266:7:e7426ec7-3bae-4a4b-a4ae-c3f80f17e058 >> > op=lx0_monitor_0 ) >> > Jun 20 11:57:25 atlas4 lrmd: [17568]: debug: >> > on_msg_perform_op:2396: >> > copying parameters for rsc lx0 >> > Jun 20 11:57:25 atlas4 lrmd: [17568]: debug: on_msg_perform_op: >> > add an >> > operation operation monitor[35] on lx0 for client 17571, its >> > parameters: >> > crm_feature_set=[3.0.5] config=[/etc/libvirt/crm/lx0.xml] >> > CRM_meta_timeout=[20000] hypervisor=[qemu:///system] to the >> > operation >> > list. >> > Jun 20 11:57:25 atlas4 corosync[17530]: [TOTEM ] releasing >> > messages up >> > to and including 149f >> > Jun 20 11:57:25 atlas4 lrmd: [17568]: info: rsc:lx0 probe[35] >> > (pid 30179) >> > Jun 20 11:57:25 atlas4 VirtualDomain[30179]: INFO: Domain name >> > "lx0" saved >> > to /var/run/resource-agents/VirtualDomain-lx0.state. >> > Jun 20 11:57:25 atlas4 corosync[17530]: [TOTEM ] releasing >> > messages up >> > to and including 14bc >> > Jun 20 11:57:25 atlas4 VirtualDomain[30179]: DEBUG: Virtual >> > domain lx0 is >> > currently shut off. >> > Jun 20 11:57:25 atlas4 lrmd: [17568]: WARN: Managed lx0:monitor >> > process >> > 30179 exited with return code 7. >> > Jun 20 11:57:25 atlas4 lrmd: [17568]: info: operation >> > monitor[35] on lx0 >> > for client 17571: pid 30179 exited with return code 7 >> > Jun 20 11:57:25 atlas4 crmd: [17571]: debug: >> > create_operation_update: >> > do_update_resource: Updating resouce lx0 after complete monitor >> > op >> > (interval=0) >> > Jun 20 11:57:25 atlas4 crmd: [17571]: info: process_lrm_event: >> > LRM >> > operation lx0_monitor_0 (call=35, rc=7, cib-update=61, >> > confirmed=true) not >> > running >> > Jun 20 11:57:25 atlas4 crmd: [17571]: debug: >> > update_history_cache: >> > Appending monitor op to history for 'lx0' >> > Jun 20 11:57:25 atlas4 crmd: [17571]: debug: get_xpath_object: >> > No match >> > for //cib_update_result//diff-added//crm_config in >> > /notify/cib_update_result/diff >> > >> > What can be wrong in the setup/configuration? And what on the >> > earth >> > happened? >> > >> > Best regards, >> > Jozsef >> > -- >> > E-mail : kadlecsik.joz...@wigner.mta.hu >> > PGP key: http://www.kfki.hu/~kadlec/pgp_public_key.txt >> > Address: Wigner Research Centre for Physics, Hungarian Academy >> > of Sciences >> > H-1525 Budapest 114, POB. 49, Hungary >> > >> > _______________________________________________ >> > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >> > http://oss.clusterlabs.org/mailman/listinfo/pacemaker >> > >> > Project Home: http://www.clusterlabs.org >> > Getting started: >> > http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >> > Bugs: http://bugs.clusterlabs.org >> > >> > >> > >> > >> > -- >> > esta es mi vida e me la vivo hasta que dios quiera >> > >> > >> >> -- >> E-mail : kadlecsik.joz...@wigner.mta.hu >> PGP key: http://www.kfki.hu/~kadlec/pgp_public_key.txt >> Address: Wigner Research Centre for Physics, Hungarian Academy of Sciences >> H-1525 Budapest 114, POB. 49, Hungary >> >> _______________________________________________ >> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >> >> Project Home: http://www.clusterlabs.org >> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >> Bugs: http://bugs.clusterlabs.org >> > > > > -- > esta es mi vida e me la vivo hasta que dios quiera > > _______________________________________________ > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org > _______________________________________________ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org