Oh, my VMs are in productive enviroment, so I shutdown the cluster and start Xen on one node whithout cluster.
Mit freundlichen Grüßen Frank Meier UNIX-Basis Hamm Reno Group GmbH Industriegebiet West | D-66987 Thaleischweiler-Fröschen T.+49(0)6334 444-8322 | F.+49(0)6334 444-8190 frank.me...@hr-group.de | www.reno.de ___________________________________________________________________ Sitz: Am Tie 7 | D-49086 Osnabrück Handelsregister Osnabrück HRB 19587 Geschäftsführer: Hans-Jürgen de Fries, Jens Gransee, Manfred Klumpp, Robert Reisch Am 19.04.2012 10:07, schrieb pacemaker-requ...@oss.clusterlabs.org: > Send Pacemaker mailing list submissions to > pacemaker@oss.clusterlabs.org > > To subscribe or unsubscribe via the World Wide Web, visit > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > or, via email, send a message with subject or body 'help' to > pacemaker-requ...@oss.clusterlabs.org > > You can reach the person managing the list at > pacemaker-ow...@oss.clusterlabs.org > > When replying, please edit your Subject line so it is more specific > than "Re: Contents of Pacemaker digest..." > > > Today's Topics: > > 1. Re: Pacemaker Digest, Vol 53, Issue 40 (emmanuel segura) > > > ---------------------------------------------------------------------- > > Message: 1 > Date: Thu, 19 Apr 2012 10:12:37 +0200 > From: emmanuel segura <emi2f...@gmail.com> > To: The Pacemaker cluster resource manager > <pacemaker@oss.clusterlabs.org> > Subject: Re: [Pacemaker] Pacemaker Digest, Vol 53, Issue 40 > Message-ID: > <CAE7pJ3B5L1m2s=OQ=cUDd9=mqy50orhp+eidtecrp5_6jlr...@mail.gmail.com> > Content-Type: text/plain; charset="iso-8859-1" > > vgchange -cy XenVG > > your vg it's not clustered and check the clvmd process > > ps -ef | grep clvmd > > Il giorno 19 aprile 2012 09:45, Frank Meier <frank.me...@hr-group.de> ha > scritto: > >> xencluster1:~ # vgs >> VG #PV #LV #SN Attr VSize VFree >> XenVG 8 32 0 wz--n- 938.16g 530.98g >> >> >> Mit freundlichen Gr??en >> >> Frank Meier >> UNIX-Basis >> >> Hamm Reno Group GmbH >> Industriegebiet West | D-66987 Thaleischweiler-Fr?schen >> T.+49(0)6334 444-8322 | F.+49(0)6334 444-8190 >> frank.me...@hr-group.de | www.reno.de >> ___________________________________________________________________ >> >> Sitz: Am Tie 7 | D-49086 Osnabr?ck >> Handelsregister Osnabr?ck HRB 19587 >> Gesch?ftsf?hrer: Hans-J?rgen de Fries, >> Jens Gransee, Manfred Klumpp, >> Robert Reisch >> >> Am 19.04.2012 09:32, schrieb pacemaker-requ...@oss.clusterlabs.org: >>> Send Pacemaker mailing list submissions to >>> pacemaker@oss.clusterlabs.org >>> >>> To subscribe or unsubscribe via the World Wide Web, visit >>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>> or, via email, send a message with subject or body 'help' to >>> pacemaker-requ...@oss.clusterlabs.org >>> >>> You can reach the person managing the list at >>> pacemaker-ow...@oss.clusterlabs.org >>> >>> When replying, please edit your Subject line so it is more specific >>> than "Re: Contents of Pacemaker digest..." >>> >>> >>> Today's Topics: >>> >>> 1. Re: Pacemaker Digest, Vol 53, Issue 38 (emmanuel segura) >>> >>> >>> ---------------------------------------------------------------------- >>> >>> Message: 1 >>> Date: Thu, 19 Apr 2012 09:37:31 +0200 >>> From: emmanuel segura <emi2f...@gmail.com> >>> To: The Pacemaker cluster resource manager >>> <pacemaker@oss.clusterlabs.org> >>> Subject: Re: [Pacemaker] Pacemaker Digest, Vol 53, Issue 38 >>> Message-ID: >>> <CAE7pJ3CovBAsjf+vbbJGEKuqTTvFiWWpn0S=UvOTSoyQpeTa= >> w...@mail.gmail.com> >>> Content-Type: text/plain; charset="iso-8859-1" >>> >>> give this command >>> >>> vgs >>> >>> Il giorno 19 aprile 2012 09:32, Frank Meier <frank.me...@hr-group.de> ha >>> scritto: >>> >>>> Hi, >>>> >>>> here my lvm.conf: >>>> >>>> devices { >>>> dir = "/dev" >>>> scan = [ "/dev" ] >>>> filter = [ "a|/dev/mapper|", "r/.*/" ] >>>> cache = "/etc/lvm/.cache" >>>> write_cache_state = 1 >>>> types = [ "device-mapper", 16 ] >>>> sysfs_scan = 1 >>>> md_component_detection = 1 >>>> } >>>> log { >>>> verbose = 0 >>>> syslog = 1 >>>> overwrite = 0 >>>> level = 0 >>>> indent = 1 >>>> command_names = 0 >>>> prefix = " " >>>> } >>>> backup { >>>> backup = 1 >>>> backup_dir = "/etc/lvm/backup" >>>> archive = 1 >>>> archive_dir = "/etc/lvm/archive" >>>> retain_min = 10 >>>> retain_days = 30 >>>> } >>>> shell { >>>> history_size = 100 >>>> } >>>> global { >>>> umask = 077 >>>> test = 0 >>>> activation = 1 >>>> proc = "/proc" >>>> locking_type = 3 >>>> locking_dir = "/var/lock/lvm" >>>> } >>>> activation { >>>> missing_stripe_filler = "error" >>>> mirror_region_size = 512 >>>> reserved_stack = 256 >>>> reserved_memory = 8192 >>>> process_priority = -18 >>>> } >>>> dmeventd { >>>> mirror_library = "libdevmapper-event-lvm2mirror.so.2.02" >>>> snapshot_library = "libdevmapper-event-lvm2snapshot.so.2.02" >>>> } >>>> >>>> >>>> Mit freundlichen Gr??en >>>> >>>> Frank Meier >>>> UNIX-Basis >>>> >>>> Hamm Reno Group GmbH >>>> Industriegebiet West | D-66987 Thaleischweiler-Fr?schen >>>> T.+49(0)6334 444-8322 | F.+49(0)6334 444-8190 >>>> frank.me...@hr-group.de | www.reno.de >>>> ___________________________________________________________________ >>>> >>>> Sitz: Am Tie 7 | D-49086 Osnabr?ck >>>> Handelsregister Osnabr?ck HRB 19587 >>>> Gesch?ftsf?hrer: Hans-J?rgen de Fries, >>>> Jens Gransee, Manfred Klumpp, >>>> Robert Reisch >>>> >>>> Am 19.04.2012 09:13, schrieb pacemaker-requ...@oss.clusterlabs.org: >>>>> Send Pacemaker mailing list submissions to >>>>> pacemaker@oss.clusterlabs.org >>>>> >>>>> To subscribe or unsubscribe via the World Wide Web, visit >>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>>> or, via email, send a message with subject or body 'help' to >>>>> pacemaker-requ...@oss.clusterlabs.org >>>>> >>>>> You can reach the person managing the list at >>>>> pacemaker-ow...@oss.clusterlabs.org >>>>> >>>>> When replying, please edit your Subject line so it is more specific >>>>> than "Re: Contents of Pacemaker digest..." >>>>> >>>>> >>>>> Today's Topics: >>>>> >>>>> 1. Re: LVM restarts after SLES upgrade (emmanuel segura) >>>>> >>>>> >>>>> ---------------------------------------------------------------------- >>>>> >>>>> Message: 1 >>>>> Date: Thu, 19 Apr 2012 09:18:37 +0200 >>>>> From: emmanuel segura <emi2f...@gmail.com> >>>>> To: The Pacemaker cluster resource manager >>>>> <pacemaker@oss.clusterlabs.org> >>>>> Subject: Re: [Pacemaker] LVM restarts after SLES upgrade >>>>> Message-ID: >>>>> <CAE7pJ3Ab1ZzSpou0=vvrO= >>>> mb3ur2+k1xhalljepj-soouef...@mail.gmail.com> >>>>> Content-Type: text/plain; charset="iso-8859-1" >>>>> >>>>> maybe you have problem with clvm >>>>> >>>>> can you show me your vm configuration? >>>>> >>>>> Il giorno 19 aprile 2012 08:29, Frank Meier <frank.me...@hr-group.de> >> ha >>>>> scritto: >>>>> >>>>>> Hi, >>>>>> >>>>>> I've installed a 2-Node Xen-Cluster with SLES 11 SP1. >>>>>> >>>>>> After an upgrade to SLES11 SP2 the cluster won't work as the old one. >>>>>> >>>>>> After some time, all LVs reorganize and so the VM where restarted. >>>>>> >>>>>> Any clue? >>>>>> >>>>>> CIB: >>>>>> >>>>>> node xencluster1 \ >>>>>> attributes standby="off" >>>>>> node xencluster2 \ >>>>>> attributes standby="off" >>>>>> primitive STONITH--Node1 stonith:meatware \ >>>>>> meta is-managed="true" target-role="Started" \ >>>>>> operations $id="STONITH--Node1-operations" \ >>>>>> op monitor interval="15" timeout="15" start-delay="15" \ >>>>>> params hostlist="xencluster1" >>>>>> primitive STONITH-Node2 stonith:meatware \ >>>>>> meta target-role="started" is-managed="true" \ >>>>>> operations $id="STONITH-Node2-operations" \ >>>>>> op monitor interval="15" timeout="15" start-delay="15" \ >>>>>> params hostlist="xencluster2" >>>>>> primitive clvm ocf:lvm2:clvmd \ >>>>>> operations $id="clvm-operations" \ >>>>>> op monitor interval="10" timeout="20" \ >>>>>> op start interval="0" timeout="90" \ >>>>>> op stop interval="0" timeout="100" \ >>>>>> params daemon_timeout="30" >>>>>> primitive clvm-xenvg ocf:heartbeat:LVM \ >>>>>> operations $id="clvm-xenvg-operations" \ >>>>>> op monitor interval="10" timeout="30" \ >>>>>> params volgrpname="XenVG" >>>>>> primitive cmirror ocf:lvm2:cmirrord \ >>>>>> operations $id="cmirror-operations" \ >>>>>> op monitor interval="10" timeout="20" \ >>>>>> op start interval="0" timeout="90" \ >>>>>> op stop interval="0" timeout="100" \ >>>>>> params daemon_timeout="30" >>>>>> primitive dlm ocf:pacemaker:controld \ >>>>>> operations $id="dlm-operations" \ >>>>>> op monitor interval="10" timeout="20" start-delay="0" \ >>>>>> op start interval="0" timeout="90" \ >>>>>> op stop interval="0" timeout="100" >>>>>> primitive fs-config-xen ocf:heartbeat:Filesystem \ >>>>>> meta is-managed="true" target-role="Started" \ >>>>>> operations $id="fs-config-xen-operations" \ >>>>>> op monitor interval="20" timeout="40" \ >>>>>> params device="/dev/mapper/XenVG-xenconfig" >>>> directory="/etc/xen/vm" >>>>>> fstype="ocfs2" >>>>>> primitive o2cb ocf:ocfs2:o2cb \ >>>>>> operations $id="o2cb-operations" \ >>>>>> op monitor interval="10" timeout="20" \ >>>>>> op start interval="0" timeout="90" \ >>>>>> op stop interval="0" timeout="100" >>>>>> primitive vm-cim ocf:heartbeat:Xen \ >>>>>> meta target-role="Started" is-managed="true" >>>> allow-migrate="true" \ >>>>>> operations $id="vm-cim-operations" \ >>>>>> op monitor interval="10" timeout="30" \ >>>>>> op start interval="0" timeout="60" \ >>>>>> op stop interval="0" timeout="60" \ >>>>>> op migrate_to interval="0" timeout="360" \ >>>>>> params xmfile="/etc/xen/vm/cim" >>>>>> primitive vm-deprepo ocf:heartbeat:Xen \ >>>>>> meta target-role="started" is-managed="true" >>>> allow-migrate="true" \ >>>>>> operations $id="vm-deprepo-operations" \ >>>>>> op monitor interval="10" timeout="30" \ >>>>>> op start interval="0" timeout="60" \ >>>>>> op stop interval="0" timeout="60" \ >>>>>> op migrate_to interval="0" timeout="360" \ >>>>>> params xmfile="/etc/xen/vm/debrepo" >>>>>> >>>>>> >>>>>> group clvm-glue dlm clvm o2cb cmirror \ >>>>>> meta target-role="Started" >>>>>> group xen-vg-fs clvm-xenvg fs-config-xen >>>>>> clone c-clvm-glue clvm-glue \ >>>>>> meta interleave="true" ordered="true" target-role="Started" >>>>>> clone c-xen-vg-fs xen-vg-fs \ >>>>>> meta interleave="true" ordered="true" is-managed="true" >>>>>> target-role="Started" >>>>>> location STONITHnode1Constraint STONITH--Node1 -inf: xencluster1 >>>>>> colocation colo-clvmglue-xenvgsfs inf: c-xen-vg-fs c-clvm-glue >>>>>> order o-cim inf: c-xen-vg-fs vm-cim >>>>>> order o-clvmglue-xenvgfs inf: c-clvm-glue c-xen-vg-fs >>>>>> order o-deprepo inf: c-xen-vg-fs vm-deprepo >>>>>> order o-ilmt inf: c-xen-vg-fs vm-ilmt >>>>>> order o-intranet inf: c-xen-vg-fs vm-intranet >>>>>> order o-nagois inf: c-xen-vg-fs vm-nagios >>>>>> order o-oneiroi inf: c-xen-vg-fs vm-oneiroi >>>>>> order o-otbridge inf: c-xen-vg-fs vm-otbridge >>>>>> order o-soa inf: c-xen-vg-fs vm-soa >>>>>> order o-topdesk-entw inf: c-xen-vg-fs vm-topdesk-entw >>>>>> order o-traumschiff inf: c-xen-vg-fs vm-traumschiff >>>>>> order o-virenscanner inf: c-xen-vg-fs vm-virenscanner >>>>>> property $id="cib-bootstrap-options" \ >>>>>> dc-version="1.1.6-b988976485d15cb702c9307df55512d323831a5e" \ >>>>>> no-quorum-policy="ignore" \ >>>>>> default-resource-stickiness="100000" \ >>>>>> last-lrm-refresh="1334518148" \ >>>>>> cluster-infrastructure="openais" \ >>>>>> expected-quorum-votes="2" \ >>>>>> default-action-timeout="120s" \ >>>>>> maintenance-mode="true" >>>>>> op_defaults $id="op_defaults-options" \ >>>>>> record-pending="false" >>>>>> >>>>>> Node 1Logs: >>>>>> >>>>>> Apr 15 22:01:10 xencluster1 clvmd[8763]: Sending message to all >> cluster >>>>>> nodes >>>>>> Apr 15 22:01:10 xencluster1 clvmd[8763]: process_work_item: local >>>>>> Apr 15 22:01:10 xencluster1 clvmd[8763]: process_local_command: >> unknown >>>>>> (0x2d) msg=0x7f5c40010e80, msglen =32, client=0x7f5c400578d0 >>>>>> Apr 15 22:01:10 xencluster1 clvmd[8763]: Syncing device names >>>>>> Apr 15 22:01:10 xencluster1 clvmd[8763]: Reply from node 4914420a: 0 >>>> bytes >>>>>> Apr 15 22:01:10 xencluster1 clvmd[8763]: Got 1 replies, expecting: 2 >>>>>> Apr 15 22:01:10 xencluster1 clvmd[8763]: LVM thread waiting for work >>>>>> Apr 15 22:01:10 xencluster1 clvmd[8763]: 1226064394 got message from >>>>>> nodeid 1226064394 for 0. len 32 >>>>>> Apr 15 22:01:40 xencluster1 lrmd: [8093]: WARN: clvm-xenvg:0:monitor >>>>>> process (PID 31785) timed out (try 1). Killing with signal SIGTERM >>>> (15). >>>>>> Apr 15 22:01:40 xencluster1 lrmd: [8093]: WARN: operation monitor[92] >> on >>>>>> clvm-xenvg:0 for client 8096: pid 31785 timed out >>>>>> Apr 15 22:01:40 xencluster1 crmd: [8096]: ERROR: process_lrm_event: >> LRM >>>>>> operation clvm-xenvg:0_monitor_10000 (92) Timed Out (timeout=30000ms) >>>>>> Apr 15 22:01:42 xencluster1 corosync[8071]: [TOTEM ] Retransmit >> List: >>>>>> 21ac >>>>>> Apr 15 22:01:42 xencluster1 corosync[8071]: [TOTEM ] Retransmit >> List: >>>>>> 21ab >>>>>> Apr 15 22:01:42 xencluster1 corosync[8071]: [TOTEM ] Retransmit >> List: >>>>>> 21ac >>>>>> Apr 15 22:01:42 xencluster1 corosync[8071]: [TOTEM ] Retransmit >> List: >>>>>> 21ab >>>>>> >>>>>> >>>>>> pr 15 22:01:50 xencluster1 clvmd[8763]: Send local reply >>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: Read on local socket 14, len >> = >>>> 32 >>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: check_all_clvmds_running >>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: down_callback. node >> 1226064394, >>>>>> state = 3 >>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: down_callback. node >> 1309950474, >>>>>> state = 3 >>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: Got pre command condition... >>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: Writing status 0 down pipe 16 >>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: Waiting to do post command - >>>>>> state = 0 >>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: read on PIPE 15: 4 bytes: >>>> status: >>>>>> 0 >>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: background routine status was >>>>>> 0, sock_client=0x7f5c40057c10 >>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: distribute command: XID = >> 15473 >>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: num_nodes = 2 >>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: add_to_lvmqueue: >>>>>> cmd=0x7f5c4000dbd0. client=0x7f5c40057c10, msg=0x7f5c40057880, len=32, >>>>>> csid=(nil), xid=15473 >>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: Sending message to all >> cluster >>>>>> nodes >>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: process_work_item: local >>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: process_local_command: >> unknown >>>>>> (0x2d) msg=0x7f5c40010e80, msglen =32, client=0x7f5c40057c10 >>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: Syncing device names >>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: Reply from node 4914420a: 0 >>>> bytes >>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: Got 1 replies, expecting: 2 >>>>>> Apr 15 22:01:50 xencluster1 clvmd[8763]: LVM thread waiting for work >>>>>> Apr 15 22:01:50 xencluster1 corosync[8071]: [TOTEM ] Retransmit >> List: >>>>>> 21ab >>>>>> Apr 15 22:01:50 xencluster1 corosync[8071]: [TOTEM ] Retransmit >> List: >>>>>> 21ac >>>>>> Apr 15 22:01:50 xencluster1 corosync[8071]: [TOTEM ] Retransmit >> List: >>>>>> 21ab >>>>>> ... >>>>>> Apr 15 22:02:19 xencluster1 corosync[8071]: [TOTEM ] Retransmit >> List: >>>>>> 21ac >>>>>> Apr 15 22:02:20 xencluster1 corosync[8071]: [TOTEM ] Retransmit >> List: >>>>>> 21ab >>>>>> Apr 15 22:02:20 xencluster1 lrmd: [8093]: WARN: clvm-xenvg:0:monitor >>>>>> process (PID 771) timed out (try 1). Killing with signal SIGTERM >> (15). >>>>>> Apr 15 22:02:20 xencluster1 lrmd: [8093]: WARN: operation monitor[92] >> on >>>>>> clvm-xenvg:0 for client 8096: pid 771 timed out >>>>>> Apr 15 22:02:20 xencluster1 crmd: [8096]: ERROR: process_lrm_event: >> LRM >>>>>> operation clvm-xenvg:0_monitor_10000 (92) Timed Out (timeout=30000ms) >>>>>> Apr 15 22:02:20 xencluster1 corosync[8071]: [TOTEM ] Retransmit >> List: >>>>>> 21ac >>>>>> Apr 15 22:02:20 xencluster1 corosync[8071]: [TOTEM ] Retransmit >> List: >>>>>> 21ab >>>>>> ... >>>>>> Apr 15 22:02:30 xencluster1 clvmd[8763]: Got new connection on fd 17 >>>>>> Apr 15 22:02:30 xencluster1 clvmd[8763]: Read on local socket 17, len >> = >>>> 28 >>>>>> Apr 15 22:02:30 xencluster1 clvmd[8763]: creating pipe, [18, 19] >>>>>> Apr 15 22:02:30 xencluster1 clvmd[8763]: Creating pre&post thread >>>>>> Apr 15 22:02:30 xencluster1 clvmd[8763]: Created pre&post thread, >> state >>>> = 0 >>>>>> Apr 15 22:02:30 xencluster1 clvmd[8763]: in sub thread: client = >>>>>> 0x7f5c40058080 >>>>>> Apr 15 22:02:30 xencluster1 clvmd[8763]: Sub thread ready for work. >>>>>> Apr 15 22:02:30 xencluster1 clvmd[8763]: doing PRE command LOCK_VG >>>>>> 'V_XenVG' at 1 (client=0x7f5c40058080) >>>>>> Apr 15 22:02:30 xencluster1 clvmd[8763]: lock_resource 'V_XenVG', >>>>>> flags=0, mode=3 >>>>>> Apr 15 22:02:30 xencluster1 clvmd[8763]: lock_resource returning 0, >>>>>> lock_id=3a0001 >>>>>> >>>>>> >>>>>> >>>>>> Apr 15 22:02:48 xencluster1 corosync[8071]: [TOTEM ] Retransmit >> List: >>>>>> 21ab >>>>>> Apr 15 22:02:48 xencluster1 cib: [8092]: info: cib_stats: Processed 4 >>>>>> operations (2500.00us average, 0% utilization) in the last 10min >>>>>> Apr 15 22:02:48 xencluster1 corosync[8071]: [TOTEM ] Retransmit >> List: >>>>>> 21ac >>>>>> >>>>>> >>>>>> >>>>>> Apr 15 22:03:00 xencluster1 corosync[8071]: [TOTEM ] Retransmit >> List: >>>>>> 21ac >>>>>> Apr 15 22:03:00 xencluster1 lrmd: [8093]: WARN: clvm-xenvg:0:monitor >>>>>> process (PID 2256) timed out (try 1). Killing with signal SIGTERM >> (15). >>>>>> Apr 15 22:03:00 xencluster1 lrmd: [8093]: WARN: operation monitor[92] >> on >>>>>> clvm-xenvg:0 for client 8096: pid 2256 timed out >>>>>> Apr 15 22:03:00 xencluster1 crmd: [8096]: ERROR: process_lrm_event: >> LRM >>>>>> operation clvm-xenvg:0_monitor_10000 (92) Timed Out (timeout=30000ms) >>>>>> Apr 15 22:03:00 xencluster1 corosync[8071]: [TOTEM ] Retransmit >> List: >>>>>> 21ab >>>>>> >>>>>> >>>>>> Node2 Log >>>>>> >>>>>> Apr 15 22:00:51 xencluster2 clvmd[12666]: Joined child thread >>>>>> Apr 15 22:00:51 xencluster2 clvmd[12666]: ret == 0, errno = 0. >> removing >>>>>> client >>>>>> Apr 15 22:00:51 xencluster2 clvmd[12666]: add_to_lvmqueue: >>>>>> cmd=0x7fa3d0028710. client=0x7fa3d0028760, msg=(nil), len=0, >> csid=(nil), >>>>>> xid=14778 >>>>>> Apr 15 22:00:51 xencluster2 clvmd[12666]: process_work_item: free fd >> -1 >>>>>> Apr 15 22:00:51 xencluster2 clvmd[12666]: LVM thread waiting for work >>>>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: crm_timer_popped: >>>>>> PEngine Recheck Timer (I_PE_CALC) just popped (900000ms) >>>>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: do_state_transition: >>>>>> State transition S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC >>>>>> cause=C_TIMER_POPPED origin=crm_timer_popped ] >>>>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: do_state_transition: >>>>>> Progressed to state S_POLICY_ENGINE after C_TIMER_POPPED >>>>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: do_state_transition: >> All >>>>>> 2 cluster nodes are eligible to run resources. >>>>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: do_pe_invoke: Query >> 983: >>>>>> Requesting the current CIB: S_POLICY_ENGINE >>>>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: do_pe_invoke_callback: >>>>>> Invoking the PE: query=983, ref=pe_calc-dc-1334520052-604, seq=2212, >>>>>> quorate=1 >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_config: On >>>>>> loss of CCM Quorum: Ignore >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_rsc_op: >>>>>> Operation clvm:1_last_failure_0 found resource clvm:1 active on >>>> xencluster2 >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_rsc_op: >>>>>> Operation STONITH-Node2_last_failure_0 found resource STONITH-Node2 >>>>>> active on xencluster2 >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_rsc_op: >>>>>> Operation STONITH--Node1_last_failure_0 found resource STONITH--Node1 >>>>>> active on xencluster2 >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_rsc_op: >>>>>> Operation cmirror:1_last_failure_0 found resource cmirror:1 active on >>>>>> xencluster2 >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_rsc_op: >>>>>> Operation dlm:1_last_failure_0 found resource dlm:1 active on >>>> xencluster2 >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_rsc_op: >>>>>> Operation o2cb:1_last_failure_0 found resource o2cb:1 active on >>>> xencluster2 >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_rsc_op: >>>>>> Operation dlm:0_last_failure_0 found resource dlm:0 active on >>>> xencluster1 >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_rsc_op: >>>>>> Operation clvm:0_last_failure_0 found resource clvm:0 active on >>>> xencluster1 >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_rsc_op: >>>>>> Operation o2cb:0_last_failure_0 found resource o2cb:0 active on >>>> xencluster1 >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: unpack_rsc_op: >>>>>> Operation clvm-xenvg:0_last_failure_0 found resource clvm-xenvg:0 >> active >>>>>> on xencluster1 >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: WARN: unpack_rsc_op: >>>>>> Processing failed op vm-deprepo_last_failure_0 on xencluster1: unknown >>>>>> error (1) >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> STONITH-Node2 (Started xencluster2) >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> STONITH--Node1 (Started xencluster2) >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> dlm:0 (Started xencluster1) >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> clvm:0 (Started xencluster1) >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> o2cb:0 (Started xencluster1) >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> cmirror:0 (Started xencluster1) >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> dlm:1 (Started xencluster2) >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> clvm:1 (Started xencluster2) >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> o2cb:1 (Started xencluster2) >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> cmirror:1 (Started xencluster2) >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> clvm-xenvg:0 (Started xencluster1) >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> fs-config-xen:0 (Started xencluster1) >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> clvm-xenvg:1 (Started xencluster2) >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> fs-config-xen:1 (Started xencluster2) >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> vm-ilmt (Started xencluster1) >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> vm-topdesk-entw (Started xencluster1) >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> vm-otbridge (Started xencluster2) >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> vm-virenscanner (Started xencluster1) >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> vm-deprepo (Started xencluster1) >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> vm-traumschiff (Started xencluster2) >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> vm-cim (Started xencluster2) >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> vm-oneiroi (Started xencluster1) >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> vm-intranet (Started xencluster1) >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> vm-soa (Started xencluster1) >>>>>> Apr 15 22:00:52 xencluster2 pengine: [7677]: notice: >> process_pe_message: >>>>>> Transition 80: PEngine Input stored in: >>>> /var/lib/pengine/pe-input-271.bz2 >>>>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: do_state_transition: >>>>>> State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ >>>>>> input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ] >>>>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: unpack_graph: Unpacked >>>>>> transition 80: 0 actions in 0 synapses >>>>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: do_te_invoke: >> Processing >>>>>> graph 80 (ref=pe_calc-dc-1334520052-604) derived from >>>>>> /var/lib/pengine/pe-input-271.bz2 >>>>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: run_graph: >>>>>> ==================================================== >>>>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: notice: run_graph: >> Transition >>>>>> 80 (Complete=0, Pending=0, Fired=0, Skipped=0, Incomplete=0, >>>>>> Source=/var/lib/pengine/pe-input-271.bz2): Complete >>>>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: te_graph_trigger: >>>>>> Transition 80 is now complete >>>>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: notify_crmd: >> Transition >>>>>> 80 status: done - <null> >>>>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: do_state_transition: >>>>>> State transition S_TRANSITION_ENGINE -> S_IDLE [ input=I_TE_SUCCESS >>>>>> cause=C_FSA_INTERNAL origin=notify_crmd ] >>>>>> Apr 15 22:00:52 xencluster2 crmd: [7678]: info: do_state_transition: >>>>>> Starting PEngine Recheck Timer >>>>>> Apr 15 22:01:00 xencluster2 clvmd[12666]: 1309950474 got message from >>>>>> nodeid 1226064394 for 0. len 32 >>>>>> Apr 15 22:01:00 xencluster2 clvmd[12666]: add_to_lvmqueue: >>>>>> cmd=0x7fa3d0028780. client=0x6934a0, msg=0x7fa3d910063c, len=32, >>>>>> csid=0x7fffb4e5d944, xid=0 >>>>>> Apr 15 22:01:00 xencluster2 clvmd[12666]: process_work_item: remote >>>>>> Apr 15 22:01:00 xencluster2 clvmd[12666]: process_remote_command >> unknown >>>>>> (0x2d) for clientid 0x5000000 XID 15337 on node 4914420a >>>>>> Apr 15 22:01:00 xencluster2 clvmd[12666]: Syncing device names >>>>>> Apr 15 22:01:00 xencluster2 clvmd[12666]: LVM thread waiting for work >>>>>> Apr 15 22:01:00 xencluster2 clvmd[12666]: 1309950474 got message from >>>>>> nodeid 1309950474 for 1226064394. len 18 >>>>>> Apr 15 22:01:00 xencluster2 clvmd[12666]: 1309950474 got message from >>>>>> nodeid 1226064394 for 0. len 32 >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> Apr 15 22:01:02 xencluster2 clvmd[12666]: Joined child thread >>>>>> Apr 15 22:01:02 xencluster2 clvmd[12666]: ret == 0, errno = 0. >> removing >>>>>> client >>>>>> Apr 15 22:01:02 xencluster2 clvmd[12666]: add_to_lvmqueue: >>>>>> cmd=0x7fa3d004d7a0. client=0x7fa3d004d7f0, msg=(nil), len=0, >> csid=(nil), >>>>>> xid=14848 >>>>>> Apr 15 22:01:02 xencluster2 clvmd[12666]: process_work_item: free fd >> -1 >>>>>> Apr 15 22:01:02 xencluster2 clvmd[12666]: LVM thread waiting for work >>>>>> Apr 15 22:01:06 xencluster2 stonith-ng: [8283]: info: stonith_command: >>>>>> Processed st_execute from lrmd: rc=-1 >>>>>> Apr 15 22:01:06 xencluster2 stonith: parse config info >> info=xencluster1 >>>>>> Apr 15 22:01:06 xencluster2 stonith: meatware device OK. >>>>>> Apr 15 22:01:06 xencluster2 stonith-ng: [8283]: info: log_operation: >>>>>> STONITH--Node1: Performing: stonith -t meatware -S >>>>>> Apr 15 22:01:06 xencluster2 stonith-ng: [8283]: info: log_operation: >>>>>> STONITH--Node1: success: 0 >>>>>> Apr 15 22:01:06 xencluster2 stonith-ng: [8283]: info: >>>>>> stonith_device_execute: Nothing to do for STONITH--Node1 >>>>>> Apr 15 22:01:06 xencluster2 stonith-ng: [8283]: info: stonith_command: >>>>>> Processed st_execute from lrmd: rc=-1 >>>>>> Apr 15 22:01:06 xencluster2 stonith: parse config info >> info=xencluster2 >>>>>> Apr 15 22:01:06 xencluster2 stonith: meatware device OK. >>>>>> Apr 15 22:01:06 xencluster2 stonith-ng: [8283]: info: log_operation: >>>>>> STONITH-Node2: Performing: stonith -t meatware -S >>>>>> Apr 15 22:01:06 xencluster2 stonith-ng: [8283]: info: log_operation: >>>>>> STONITH-Node2: success: 0 >>>>>> Apr 15 22:01:06 xencluster2 stonith-ng: [8283]: info: >>>>>> stonith_device_execute: Nothing to do for STONITH-Node2 >>>>>> Apr 15 22:01:10 xencluster2 clvmd[12666]: 1309950474 got message from >>>>>> nodeid 1226064394 for 0. len 32 >>>>>> Apr 15 22:01:10 xencluster2 clvmd[12666]: add_to_lvmqueue: >>>>>> cmd=0x7fa3d004d810. client=0x6934a0, msg=0x7fa3d9100cfc, len=32, >>>>>> csid=0x7fffb4e5d944, xid=0 >>>>>> Apr 15 22:01:10 xencluster2 clvmd[12666]: process_work_item: remote >>>>>> Apr 15 22:01:10 xencluster2 clvmd[12666]: process_remote_command >> unknown >>>>>> (0x2d) for clientid 0x5000000 XID 15407 on node 4914420a >>>>>> Apr 15 22:01:10 xencluster2 clvmd[12666]: Syncing device names >>>>>> Apr 15 22:01:10 xencluster2 clvmd[12666]: LVM thread waiting for work >>>>>> Apr 15 22:01:10 xencluster2 clvmd[12666]: 1309950474 got message from >>>>>> nodeid 1309950474 for 1226064394. len 18 >>>>>> Apr 15 22:01:10 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 >>>>>> Apr 15 22:01:10 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 >>>>>> Apr 15 22:01:10 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 >>>>>> .... >>>>>> Apr 15 22:01:11 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 >>>>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: Got new connection on fd 5 >>>>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: Read on local socket 5, len >> = >>>> 28 >>>>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: creating pipe, [12, 13] >>>>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: Creating pre&post thread >>>>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: Created pre&post thread, >>>> state = >>>>>> 0 >>>>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: in sub thread: client = >>>>>> 0x7fa3d004d810 >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: Syncing device names >>>>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: Reply from node 4e14420a: 0 >>>> bytes >>>>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: Got 1 replies, expecting: 2 >>>>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: LVM thread waiting for work >>>>>> Apr 15 22:01:12 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 >>>>>> Apr 15 22:01:12 xencluster2 clvmd[12666]: 1309950474 got message from >>>>>> nodeid 1309950474 for 0. len 32 >>>>>> Apr 15 22:01:12 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 >>>>>> Apr 15 22:01:12 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 >>>>>> .... >>>>>> Apr 15 22:01:21 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 2197 >>>>>> Apr 15 22:01:21 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 2197 >>>>>> Apr 15 22:01:21 xencluster2 stonith-ng: [8283]: info: stonith_command: >>>>>> Processed st_execute from lrmd: rc=-1 >>>>>> Apr 15 22:01:21 xencluster2 stonith: parse config info >> info=xencluster1 >>>>>> Apr 15 22:01:21 xencluster2 stonith: meatware device OK. >>>>>> Apr 15 22:01:21 xencluster2 stonith-ng: [8283]: info: log_operation: >>>>>> STONITH--Node1: Performing: stonith -t meatware -S >>>>>> Apr 15 22:01:21 xencluster2 stonith-ng: [8283]: info: log_operation: >>>>>> STONITH--Node1: success: 0 >>>>>> Apr 15 22:01:21 xencluster2 stonith-ng: [8283]: info: >>>>>> stonith_device_execute: Nothing to do for STONITH--Node1 >>>>>> Apr 15 22:01:21 xencluster2 stonith-ng: [8283]: info: stonith_command: >>>>>> Processed st_execute from lrmd: rc=-1 >>>>>> Apr 15 22:01:21 xencluster2 stonith: parse config info >> info=xencluster2 >>>>>> Apr 15 22:01:21 xencluster2 stonith: meatware device OK. >>>>>> Apr 15 22:01:21 xencluster2 stonith-ng: [8283]: info: log_operation: >>>>>> STONITH-Node2: Performing: stonith -t meatware -S >>>>>> Apr 15 22:01:21 xencluster2 stonith-ng: [8283]: info: log_operation: >>>>>> STONITH-Node2: success: 0 >>>>>> Apr 15 22:01:21 xencluster2 stonith-ng: [8283]: info: >>>>>> stonith_device_execute: Nothing to do for STONITH-Node2 >>>>>> Apr 15 22:01:21 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 2197 >>>>>> Apr 15 22:01:22 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 2197 >>>>>> Apr 15 22:01:22 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 2197 >>>>>> .... >>>>>> Apr 15 22:01:36 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 2197 >>>>>> Apr 15 22:01:36 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 2197 >>>>>> Apr 15 22:01:36 xencluster2 stonith-ng: [8283]: info: stonith_command: >>>>>> Processed st_execute from lrmd: rc=-1 >>>>>> Apr 15 22:01:36 xencluster2 stonith: parse config info >> info=xencluster1 >>>>>> Apr 15 22:01:36 xencluster2 stonith: meatware device OK. >>>>>> Apr 15 22:01:36 xencluster2 stonith-ng: [8283]: info: log_operation: >>>>>> STONITH--Node1: Performing: stonith -t meatware -S >>>>>> Apr 15 22:01:36 xencluster2 stonith-ng: [8283]: info: log_operation: >>>>>> STONITH--Node1: success: 0 >>>>>> Apr 15 22:01:36 xencluster2 stonith-ng: [8283]: info: >>>>>> stonith_device_execute: Nothing to do for STONITH--Node1 >>>>>> Apr 15 22:01:36 xencluster2 stonith-ng: [8283]: info: stonith_command: >>>>>> Processed st_execute from lrmd: rc=-1 >>>>>> Apr 15 22:01:36 xencluster2 stonith: parse config info >> info=xencluster2 >>>>>> Apr 15 22:01:36 xencluster2 stonith: meatware device OK. >>>>>> Apr 15 22:01:36 xencluster2 stonith-ng: [8283]: info: log_operation: >>>>>> STONITH-Node2: Performing: stonith -t meatware -S >>>>>> Apr 15 22:01:36 xencluster2 stonith-ng: [8283]: info: log_operation: >>>>>> STONITH-Node2: success: 0 >>>>>> Apr 15 22:01:36 xencluster2 stonith-ng: [8283]: info: >>>>>> stonith_device_execute: Nothing to do for STONITH-Node2 >>>>>> Apr 15 22:01:36 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 2197 >>>>>> Apr 15 22:01:37 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 2197 >>>>>> .... >>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 2197 >>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 2197 >>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: process_graph_event: >>>>>> Detected action clvm-xenvg:0_monitor_10000 from a different >> transition: >>>>>> 62 vs. 80 >>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: >> abort_transition_graph: >>>>>> process_graph_event:476 - Triggered transition abort (complete=1, >>>>>> tag=lrm_rsc_op, id=clvm-xenvg:0_monitor_10000, >>>>>> magic=2:-2;61:62:0:72497771-fdb2-4ca7-a723-2cc360861bb0, >> cib=0.5816.4) : >>>>>> Old event >>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: WARN: update_failcount: >>>>>> Updating failcount for clvm-xenvg:0 on xencluster1 after failed >> monitor: >>>>>> rc=-2 (update=value++, time=1334520100) >>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: do_state_transition: >>>>>> State transition S_IDLE -> S_POLICY_ENGINE [ input=I_PE_CALC >>>>>> cause=C_FSA_INTERNAL origin=abort_transition_graph ] >>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: do_state_transition: >> All >>>>>> 2 cluster nodes are eligible to run resources. >>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: do_pe_invoke: Query >> 984: >>>>>> Requesting the current CIB: S_POLICY_ENGINE >>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 2197 >>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 2197 >>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 2197 >>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 2197 >>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 2197 >>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 2197 >>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 2197 219a 219b 219c 219d >>>>>> ... >>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 2197 219a 219b 219c 219d >>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: do_pe_invoke_callback: >>>>>> Invoking the PE: query=984, ref=pe_calc-dc-1334520100-605, seq=2212, >>>>>> quorate=1 >>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 2197 219a 219b 219c 219d >>>>>> ... >>>>>> >>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 2197 219a 219b 219c 219d >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_config: On >>>>>> loss of CCM Quorum: Ignore >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op: >>>>>> Operation clvm:1_last_failure_0 found resource clvm:1 active on >>>> xencluster2 >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op: >>>>>> Operation STONITH-Node2_last_failure_0 found resource STONITH-Node2 >>>>>> active on xencluster2 >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op: >>>>>> Operation STONITH--Node1_last_failure_0 found resource STONITH--Node1 >>>>>> active on xencluster2 >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op: >>>>>> Operation cmirror:1_last_failure_0 found resource cmirror:1 active on >>>>>> xencluster2 >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op: >>>>>> Operation dlm:1_last_failure_0 found resource dlm:1 active on >>>> xencluster2 >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op: >>>>>> Operation o2cb:1_last_failure_0 found resource o2cb:1 active on >>>> xencluster2 >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op: >>>>>> Operation cmirror:0_last_failure_0 found resource cmirror:0 active on >>>>>> xencluster1 >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op: >>>>>> Operation dlm:0_last_failure_0 found resource dlm:0 active on >>>> xencluster1 >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op: >>>>>> Operation clvm:0_last_failure_0 found resource clvm:0 active on >>>> xencluster1 >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op: >>>>>> Operation o2cb:0_last_failure_0 found resource o2cb:0 active on >>>> xencluster1 >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: unpack_rsc_op: >>>>>> Operation clvm-xenvg:0_last_failure_0 found resource clvm-xenvg:0 >> active >>>>>> on xencluster1 >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: WARN: unpack_rsc_op: >>>>>> Processing failed op clvm-xenvg:0_monitor_10000 on xencluster1: >> unknown >>>>>> exec error (-2) >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: WARN: unpack_rsc_op: >>>>>> Processing failed op vm-deprepo_last_failure_0 on xencluster1: unknown >>>>>> error (1) >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: RecurringOp: >> Start >>>>>> recurring monitor (10s) for clvm-xenvg:0 on xencluster1 >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> STONITH-Node2 (Started xencluster2) >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> STONITH--Node1 (Started xencluster2) >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> dlm:0 (Started xencluster1) >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> clvm:0 (Started xencluster1) >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> o2cb:0 (Started xencluster1) >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> cmirror:0 (Started xencluster1) >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> dlm:1 (Started xencluster2) >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> clvm:1 (Started xencluster2) >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> o2cb:1 (Started xencluster2) >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> cmirror:1 (Started xencluster2) >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: >> Recover >>>>>> clvm-xenvg:0 (Started xencluster1) >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: >> Restart >>>>>> fs-config-xen:0 (Started xencluster1) >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> clvm-xenvg:1 (Started xencluster2) >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> fs-config-xen:1 (Started xencluster2) >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: >> Restart >>>>>> vm-ilmt (Started xencluster1) >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: >> Restart >>>>>> vm-topdesk-entw (Started xencluster1) >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> vm-otbridge (Started xencluster2) >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: >> Restart >>>>>> vm-virenscanner (Started xencluster1) >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: >> Restart >>>>>> vm-deprepo (Started xencluster1) >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> vm-traumschiff (Started xencluster2) >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: Leave >>>>>> vm-cim (Started xencluster2) >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: >> Restart >>>>>> vm-oneiroi (Started xencluster1) >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: >> Restart >>>>>> vm-intranet (Started xencluster1) >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: LogActions: >> Restart >>>>>> vm-soa (Started xencluster1) >>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: do_state_transition: >>>>>> State transition S_POLICY_ENGINE -> S_TRANSITION_ENGINE [ >>>>>> input=I_PE_SUCCESS cause=C_IPC_MESSAGE origin=handle_response ] >>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: unpack_graph: Unpacked >>>>>> transition 81: 39 actions in 39 synapses >>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: do_te_invoke: >> Processing >>>>>> graph 81 (ref=pe_calc-dc-1334520100-605) derived from >>>>>> /var/lib/pengine/pe-input-272.bz2 >>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: te_rsc_command: >>>>>> Initiating action 82: stop vm-ilmt_stop_0 on xencluster1 >>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: te_rsc_command: >>>>>> Initiating action 86: stop vm-topdesk-entw_stop_0 on xencluster1 >>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: te_rsc_command: >>>>>> Initiating action 90: stop vm-virenscanner_stop_0 on xencluster1 >>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: te_rsc_command: >>>>>> Initiating action 92: stop vm-deprepo_stop_0 on xencluster1 >>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: te_rsc_command: >>>>>> Initiating action 98: stop vm-oneiroi_stop_0 on xencluster1 >>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: te_rsc_command: >>>>>> Initiating action 100: stop vm-intranet_stop_0 on xencluster1 >>>>>> Apr 15 22:01:40 xencluster2 crmd: [7678]: info: te_rsc_command: >>>>>> Initiating action 102: stop vm-soa_stop_0 on xencluster1 >>>>>> Apr 15 22:01:40 xencluster2 pengine: [7677]: notice: >> process_pe_message: >>>>>> Transition 81: PEngine Input stored in: >>>> /var/lib/pengine/pe-input-272.bz2 >>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 2197 219a 219b 219c 219d >>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 2197 219a 219b 219c 219d >>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 2197 219a 219b 219c 219d >>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 2197 219a 219b 219c 219d >>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 2197 219a 219b 219c 219d >>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 2197 219a 219b 219c 219d 219e 219f 21a0 21a1 21a2 21a3 21a4 21a5 >>>>>> Apr 15 22:01:40 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 2197 219a 219b 219c 219d 219e 219f 21a0 21a1 21a2 21a3 21a4 21a5 >>>>>> ... >>>>>> Apr 15 22:01:41 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 2197 219a 219b 219c 219d 219e 219f 21a0 21a1 21a2 21a3 21a4 21a5 >>>>>> Apr 15 22:01:42 xencluster2 lrmd: [7675]: WARN: clvm-xenvg:1:monitor >>>>>> process (PID 29785) timed out (try 1). Killing with signal SIGTERM >>>> (15). >>>>>> Apr 15 22:01:42 xencluster2 lrmd: [7675]: WARN: operation monitor[125] >>>>>> on clvm-xenvg:1 for client 7678: pid 29785 timed out >>>>>> Apr 15 22:01:42 xencluster2 crmd: [7678]: ERROR: process_lrm_event: >> LRM >>>>>> operation clvm-xenvg:1_monitor_10000 (125) Timed Out (timeout=30000ms) >>>>>> Apr 15 22:01:42 xencluster2 crmd: [7678]: info: process_graph_event: >>>>>> Detected action clvm-xenvg:1_monitor_10000 from a different >> transition: >>>>>> 68 vs. 81 >>>>>> Apr 15 22:01:42 xencluster2 crmd: [7678]: info: >> abort_transition_graph: >>>>>> process_graph_event:476 - Triggered transition abort (complete=0, >>>>>> tag=lrm_rsc_op, id=clvm-xenvg:1_monitor_10000, >>>>>> magic=2:-2;61:68:0:72497771-fdb2-4ca7-a723-2cc360861bb0, >> cib=0.5816.5) : >>>>>> Old event >>>>>> Apr 15 22:01:42 xencluster2 crmd: [7678]: info: update_abort_priority: >>>>>> Abort priority upgraded from 0 to 1000000 >>>>>> Apr 15 22:01:42 xencluster2 crmd: [7678]: info: update_abort_priority: >>>>>> Abort action done superceeded by restart >>>>>> Apr 15 22:01:42 xencluster2 crmd: [7678]: WARN: update_failcount: >>>>>> Updating failcount for clvm-xenvg:1 on xencluster2 after failed >> monitor: >>>>>> rc=-2 (update=value++, time=1334520102) >>>>>> Apr 15 22:01:42 xencluster2 attrd: [7676]: notice: >> attrd_trigger_update: >>>>>> Sending flush op to all hosts for: fail-count-clvm-xenvg:1 (1) >>>>>> Apr 15 22:01:42 xencluster2 attrd: [7676]: notice: >> attrd_perform_update: >>>>>> Sent update 448: fail-count-clvm-xenvg:1=1 >>>>>> Apr 15 22:01:42 xencluster2 attrd: [7676]: notice: >> attrd_trigger_update: >>>>>> Sending flush op to all hosts for: last-failure-clvm-xenvg:1 >>>> (1334520102) >>>>>> Apr 15 22:01:42 xencluster2 crmd: [7678]: info: >> abort_transition_graph: >>>>>> te_update_diff:176 - Triggered transition abort (complete=0, >> tag=nvpair, >>>>>> id=status-xencluster2-fail-count-clvm-xenvg.1, >>>>>> name=fail-count-clvm-xenvg:1, value=1, magic=NA, cib=0.5816.6) : >>>>>> Transient attribute: update >>>>>> Apr 15 22:01:42 xencluster2 attrd: [7676]: notice: >> attrd_perform_update: >>>>>> Sent update 451: last-failure-clvm-xenvg:1=1334520102 >>>>>> Apr 15 22:01:42 xencluster2 crmd: [7678]: info: >> abort_transition_graph: >>>>>> te_update_diff:176 - Triggered transition abort (complete=0, >> tag=nvpair, >>>>>> id=status-xencluster2-last-failure-clvm-xenvg.1, >>>>>> name=last-failure-clvm-xenvg:1, value=1334520102, magic=NA, >>>>>> cib=0.5816.7) : Transient attribute: update >>>>>> Apr 15 22:01:42 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 2197 219a 219b 219c 219d 219e 219f 21a0 21a1 21a2 21a3 21a4 21a5 >>>>>> Apr 15 22:01:42 xencluster2 corosync[7666]: [TOTEM ] Retransmit >> List: >>>>>> 2196 2197 219a 219b 219c 219d 219e 219f 21a0 21a1 21a2 21a3 21a4 21a5 >>>>>> >>>>>> -- >>>>>> Mit freundlichen Gr??en >>>>>> >>>>>> Frank Meier >>>>>> UNIX-Basis >>>>>> >>>>>> Hamm Reno Group GmbH >>>>>> Industriegebiet West | D-66987 Thaleischweiler-Fr?schen >>>>>> T.+49(0)6334 444-8322 | F.+49(0)6334 444-8190 >>>>>> frank.me...@hr-group.de | www.reno.de >>>>>> ___________________________________________________________________ >>>>>> >>>>>> Sitz: Am Tie 7 | D-49086 Osnabr?ck >>>>>> Handelsregister Osnabr?ck HRB 19587 >>>>>> Gesch?ftsf?hrer: Hans-J?rgen de Fries, >>>>>> Jens Gransee, Manfred Klumpp, >>>>>> Robert Reisch >>>>>> >>>>>> _______________________________________________ >>>>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>>>> >>>>>> Project Home: http://www.clusterlabs.org >>>>>> Getting started: >>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>>> Bugs: http://bugs.clusterlabs.org >>>>>> >>>>> >>>>> >>>>> >>>>> -- >>>>> esta es mi vida e me la vivo hasta que dios quiera >>>>> -------------- next part -------------- >>>>> An HTML attachment was scrubbed... >>>>> URL: < >>>> >> http://oss.clusterlabs.org/pipermail/pacemaker/attachments/20120419/00a36dbe/attachment.html >>>>> >>>>> >>>>> ------------------------------ >>>>> >>>>> _______________________________________________ >>>>> Pacemaker mailing list >>>>> Pacemaker@oss.clusterlabs.org >>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>>> >>>>> >>>>> End of Pacemaker Digest, Vol 53, Issue 38 >>>>> ***************************************** >>>> >>>> _______________________________________________ >>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>> >>>> Project Home: http://www.clusterlabs.org >>>> Getting started: >> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>> Bugs: http://bugs.clusterlabs.org >>>> >>> >>> >>> >>> -- >>> esta es mi vida e me la vivo hasta que dios quiera >>> -------------- next part -------------- >>> An HTML attachment was scrubbed... >>> URL: < >> http://oss.clusterlabs.org/pipermail/pacemaker/attachments/20120419/63b888a3/attachment.html >>> >>> >>> ------------------------------ >>> >>> _______________________________________________ >>> Pacemaker mailing list >>> Pacemaker@oss.clusterlabs.org >>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>> >>> >>> End of Pacemaker Digest, Vol 53, Issue 40 >>> ***************************************** >> >> _______________________________________________ >> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >> >> Project Home: http://www.clusterlabs.org >> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >> Bugs: http://bugs.clusterlabs.org >> > > > > -- > esta es mi vida e me la vivo hasta que dios quiera > -------------- next part -------------- > An HTML attachment was scrubbed... > URL: > <http://oss.clusterlabs.org/pipermail/pacemaker/attachments/20120419/f21bf7c7/attachment.html> > > ------------------------------ > > _______________________________________________ > Pacemaker mailing list > Pacemaker@oss.clusterlabs.org > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > > End of Pacemaker Digest, Vol 53, Issue 42 > ***************************************** _______________________________________________ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org