On 20 Feb 2014, at 4:18 am, Andrey Groshev <gre...@yandex.ru> wrote:
> > > 19.02.2014, 06:47, "Andrew Beekhof" <and...@beekhof.net>: >> On 18 Feb 2014, at 9:29 pm, Andrey Groshev <gre...@yandex.ru> wrote: >> >>> Hi, ALL and Andrew! >>> >>> Today is a good day - I killed a lot, and a lot of shooting at me. >>> In general - I am happy (almost like an elephant) :) >>> Except resources on the node are important to me eight processes: >>> corosync,pacemakerd,cib,stonithd,lrmd,attrd,pengine,crmd. >>> I killed them with different signals (4,6,11 and even 9). >>> Behavior does not depend of number signal - it's good. >>> If STONITH send reboot to the node - it rebooted and rejoined the cluster >>> - too it's good. >>> But the behavior is different from killing various demons. >>> >>> Turned four groups: >>> 1. corosync,cib - STONITH work 100%. >>> Kill via any signals - call STONITH and reboot. >> >> excellent >> >>> 3. stonithd,attrd,pengine - not need STONITH >>> This daemons simple restart, resources - stay running. >> >> right >> >>> 2. lrmd,crmd - strange behavior STONITH. >>> Sometimes called STONITH - and the corresponding reaction. >>> Sometimes restart daemon >> >> The daemon will always try to restart, the only variable is how long it >> takes the peer to notice and initiate fencing. >> If the failure happens just before a they're due to receive totem token, the >> failure will be very quickly detected and the node fenced. >> If the failure happens just after, then detection will take longer - giving >> the node longer to recover and not be fenced. >> >> So fence/not fence is normal and to be expected. >> >>> and restart resources with large delay MS:pgsql. >>> One time after restart crmd - pgsql don't restart. >> >> I would not expect pgsql to ever restart - if the RA does its job properly >> anyway. >> In the case the node is not fenced, the crmd will respawn and the the PE >> will request that it re-detect the state of all resources. >> >> If the agent reports "all good", then there is nothing more to do. >> If the agent is not reporting "all good", you should really be asking why. >> >>> 4. pacemakerd - nothing happens. >> >> On non-systemd based machines, correct. >> >> On a systemd based machine pacemakerd is respawned and reattaches to the >> existing daemons. >> Any subsequent daemon failure will be detected and the daemon respawned. > > And! I almost forgot about IT! > Exist another (NORMAL) the variants, the methods, the ideas? > Without this ... @$%#$%&$%^&$%^&##@#$$^$%& !!!!! > Otherwise - it's a full epic fail ;) -ENOPARSE > >>> And then I can kill any process of the third group. They do not restart. >> >> Until they become needed. >> Eg. if the DC goes to invoke the policy engine, that will fail causing the >> crmd to fail and the node to be fenced. >> >>> Generaly don't touch corosync,cib and maybe lrmd,crmd. >>> >>> What do you think about this? >>> The main question of this topic - we decided. >>> But this varied behavior - another big problem. >>> >>> 17.02.2014, 08:52, "Andrey Groshev" <gre...@yandex.ru>: >>>> 17.02.2014, 02:27, "Andrew Beekhof" <and...@beekhof.net>: >>>>> With no quick follow-up, dare one hope that means the patch worked? :-) >>>> Hi, >>>> No, unfortunately the chief changed my plans on Friday and all day I was >>>> engaged in a parallel project. >>>> I hope that today have time to carry out the necessary tests. >>>>> On 14 Feb 2014, at 3:37 pm, Andrey Groshev <gre...@yandex.ru> wrote: >>>>>> Yes, of course. Now beginning build world and test ) >>>>>> >>>>>> 14.02.2014, 04:41, "Andrew Beekhof" <and...@beekhof.net>: >>>>>>> The previous patch wasn't quite right. >>>>>>> Could you try this new one? >>>>>>> >>>>>>> http://paste.fedoraproject.org/77123/13923376/ >>>>>>> >>>>>>> [11:23 AM] beekhof@f19 ~/Development/sources/pacemaker/devel ☺ # git >>>>>>> diff >>>>>>> diff --git a/crmd/callbacks.c b/crmd/callbacks.c >>>>>>> index ac4b905..d49525b 100644 >>>>>>> --- a/crmd/callbacks.c >>>>>>> +++ b/crmd/callbacks.c >>>>>>> @@ -199,8 +199,7 @@ peer_update_callback(enum crm_status_type type, >>>>>>> crm_node_t * node, const void *d >>>>>>> stop_te_timer(down->timer); >>>>>>> >>>>>>> flags |= node_update_join | node_update_expected; >>>>>>> - crm_update_peer_join(__FUNCTION__, node, >>>>>>> crm_join_none); >>>>>>> - crm_update_peer_expected(__FUNCTION__, node, >>>>>>> CRMD_JOINSTATE_DOWN); >>>>>>> + crmd_peer_down(node, FALSE); >>>>>>> check_join_state(fsa_state, __FUNCTION__); >>>>>>> >>>>>>> update_graph(transition_graph, down); >>>>>>> diff --git a/crmd/crmd_utils.h b/crmd/crmd_utils.h >>>>>>> index bc472c2..1a2577a 100644 >>>>>>> --- a/crmd/crmd_utils.h >>>>>>> +++ b/crmd/crmd_utils.h >>>>>>> @@ -100,6 +100,7 @@ void crmd_join_phase_log(int level); >>>>>>> const char *get_timer_desc(fsa_timer_t * timer); >>>>>>> gboolean too_many_st_failures(void); >>>>>>> void st_fail_count_reset(const char * target); >>>>>>> +void crmd_peer_down(crm_node_t *peer, bool full); >>>>>>> >>>>>>> # define fsa_register_cib_callback(id, flag, data, fn) do { >>>>>>> \ >>>>>>> fsa_cib_conn->cmds->register_callback( >>>>>>> \ >>>>>>> diff --git a/crmd/te_actions.c b/crmd/te_actions.c >>>>>>> index f31d4ec..3bfce59 100644 >>>>>>> --- a/crmd/te_actions.c >>>>>>> +++ b/crmd/te_actions.c >>>>>>> @@ -80,11 +80,8 @@ send_stonith_update(crm_action_t * action, const >>>>>>> char *target, const char *uuid) >>>>>>> crm_info("Recording uuid '%s' for node '%s'", uuid, target); >>>>>>> peer->uuid = strdup(uuid); >>>>>>> } >>>>>>> - crm_update_peer_proc(__FUNCTION__, peer, crm_proc_none, NULL); >>>>>>> - crm_update_peer_state(__FUNCTION__, peer, CRM_NODE_LOST, 0); >>>>>>> - crm_update_peer_expected(__FUNCTION__, peer, >>>>>>> CRMD_JOINSTATE_DOWN); >>>>>>> - crm_update_peer_join(__FUNCTION__, peer, crm_join_none); >>>>>>> >>>>>>> + crmd_peer_down(peer, TRUE); >>>>>>> node_state = >>>>>>> do_update_node_cib(peer, >>>>>>> node_update_cluster | node_update_peer | >>>>>>> node_update_join | >>>>>>> diff --git a/crmd/te_utils.c b/crmd/te_utils.c >>>>>>> index ad7e573..0c92e95 100644 >>>>>>> --- a/crmd/te_utils.c >>>>>>> +++ b/crmd/te_utils.c >>>>>>> @@ -247,10 +247,7 @@ tengine_stonith_notify(stonith_t * st, >>>>>>> stonith_event_t * st_event) >>>>>>> >>>>>>> } >>>>>>> >>>>>>> - crm_update_peer_proc(__FUNCTION__, peer, crm_proc_none, >>>>>>> NULL); >>>>>>> - crm_update_peer_state(__FUNCTION__, peer, CRM_NODE_LOST, 0); >>>>>>> - crm_update_peer_expected(__FUNCTION__, peer, >>>>>>> CRMD_JOINSTATE_DOWN); >>>>>>> - crm_update_peer_join(__FUNCTION__, peer, crm_join_none); >>>>>>> + crmd_peer_down(peer, TRUE); >>>>>>> } >>>>>>> } >>>>>>> >>>>>>> diff --git a/crmd/utils.c b/crmd/utils.c >>>>>>> index 3988cfe..2df53ab 100644 >>>>>>> --- a/crmd/utils.c >>>>>>> +++ b/crmd/utils.c >>>>>>> @@ -1077,3 +1077,13 @@ update_attrd_remote_node_removed(const char >>>>>>> *host, const char *user_name) >>>>>>> crm_trace("telling attrd to clear attributes for remote host >>>>>>> %s", host); >>>>>>> update_attrd_helper(host, NULL, NULL, user_name, TRUE, 'C'); >>>>>>> } >>>>>>> + >>>>>>> +void crmd_peer_down(crm_node_t *peer, bool full) >>>>>>> +{ >>>>>>> + if(full && peer->state == NULL) { >>>>>>> + crm_update_peer_state(__FUNCTION__, peer, CRM_NODE_LOST, 0); >>>>>>> + crm_update_peer_proc(__FUNCTION__, peer, crm_proc_none, >>>>>>> NULL); >>>>>>> + } >>>>>>> + crm_update_peer_join(__FUNCTION__, peer, crm_join_none); >>>>>>> + crm_update_peer_expected(__FUNCTION__, peer, >>>>>>> CRMD_JOINSTATE_DOWN); >>>>>>> +} >>>>>>> >>>>>>> On 16 Jan 2014, at 7:24 pm, Andrey Groshev <gre...@yandex.ru> wrote: >>>>>>>> 16.01.2014, 01:30, "Andrew Beekhof" <and...@beekhof.net>: >>>>>>>>> On 16 Jan 2014, at 12:41 am, Andrey Groshev <gre...@yandex.ru> >>>>>>>>> wrote: >>>>>>>>>> 15.01.2014, 02:53, "Andrew Beekhof" <and...@beekhof.net>: >>>>>>>>>>> On 15 Jan 2014, at 12:15 am, Andrey Groshev <gre...@yandex.ru> >>>>>>>>>>> wrote: >>>>>>>>>>>> 14.01.2014, 10:00, "Andrey Groshev" <gre...@yandex.ru>: >>>>>>>>>>>>> 14.01.2014, 07:47, "Andrew Beekhof" <and...@beekhof.net>: >>>>>>>>>>>>>> Ok, here's what happens: >>>>>>>>>>>>>> >>>>>>>>>>>>>> 1. node2 is lost >>>>>>>>>>>>>> 2. fencing of node2 starts >>>>>>>>>>>>>> 3. node2 reboots (and cluster starts) >>>>>>>>>>>>>> 4. node2 returns to the membership >>>>>>>>>>>>>> 5. node2 is marked as a cluster member >>>>>>>>>>>>>> 6. DC tries to bring it into the cluster, but needs to >>>>>>>>>>>>>> cancel the active transition first. >>>>>>>>>>>>>> Which is a problem since the node2 fencing operation >>>>>>>>>>>>>> is part of that >>>>>>>>>>>>>> 7. node2 is in a transition (pending) state until fencing >>>>>>>>>>>>>> passes or fails >>>>>>>>>>>>>> 8a. fencing fails: transition completes and the node >>>>>>>>>>>>>> joins the cluster >>>>>>>>>>>>>> >>>>>>>>>>>>>> Thats in theory, except we automatically try again. Which >>>>>>>>>>>>>> isn't appropriate. >>>>>>>>>>>>>> This should be relatively easy to fix. >>>>>>>>>>>>>> >>>>>>>>>>>>>> 8b. fencing passes: the node is incorrectly marked as >>>>>>>>>>>>>> offline >>>>>>>>>>>>>> >>>>>>>>>>>>>> This I have no idea how to fix yet. >>>>>>>>>>>>>> >>>>>>>>>>>>>> On another note, it doesn't look like this agent works at >>>>>>>>>>>>>> all. >>>>>>>>>>>>>> The node has been back online for a long time and the >>>>>>>>>>>>>> agent is still timing out after 10 minutes. >>>>>>>>>>>>>> So "Once the script makes sure that the victim will >>>>>>>>>>>>>> rebooted and again available via ssh - it exit with 0." does not >>>>>>>>>>>>>> seem true. >>>>>>>>>>>>> Damn. Looks like you're right. At some time I broke my >>>>>>>>>>>>> agent and had not noticed it. Who will understand. >>>>>>>>>>>> I repaired my agent - after send reboot he is wait STDIN. >>>>>>>>>>>> Returned "normally" a behavior - hangs "pending", until >>>>>>>>>>>> manually send reboot. :) >>>>>>>>>>> Right. Now you're in case 8b. >>>>>>>>>>> >>>>>>>>>>> Can you try this patch: >>>>>>>>>>> http://paste.fedoraproject.org/68450/38973966 >>>>>>>>>> Killed all day experiences. >>>>>>>>>> It turns out here that: >>>>>>>>>> 1. Did cluster. >>>>>>>>>> 2. On the node-2 send signal (-4) - killed corosink >>>>>>>>>> 3. From node-1 (there DC) - stonith sent reboot >>>>>>>>>> 4. Noda rebooted and resources start. >>>>>>>>>> 5. Again. On the node-2 send signal (-4) - killed corosink >>>>>>>>>> 6. Again. From node-1 (there DC) - stonith sent reboot >>>>>>>>>> 7. Noda-2 rebooted and hangs in "pending" >>>>>>>>>> 8. Waiting, waiting..... manually reboot. >>>>>>>>>> 9. Noda-2 reboot and raised resources start. >>>>>>>>>> 10. GOTO p.2 >>>>>>>>> Logs? >>>>>>>> Yesterday I wrote an additional letter why not put the logs. >>>>>>>> Read it please, it contains a few more questions. >>>>>>>> Today again began to hang and continue along the same cycle. >>>>>>>> Logs here http://send2me.ru/crmrep2.tar.bz2 >>>>>>>>>>>> New logs: http://send2me.ru/crmrep1.tar.bz2 >>>>>>>>>>>>>> On 14 Jan 2014, at 1:19 pm, Andrew Beekhof >>>>>>>>>>>>>> <and...@beekhof.net> wrote: >>>>>>>>>>>>>>> Apart from anything else, your timeout needs to be >>>>>>>>>>>>>>> bigger: >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> Jan 13 12:21:36 [17223] >>>>>>>>>>>>>>> dev-cluster2-node1.unix.tensor.ru stonith-ng: ( >>>>>>>>>>>>>>> commands.c:1321 ) error: log_operation: Operation 'reboot' >>>>>>>>>>>>>>> [11331] (call 2 from crmd.17227) for host >>>>>>>>>>>>>>> 'dev-cluster2-node2.unix.tensor.ru' with device 'st1' returned: >>>>>>>>>>>>>>> -62 (Timer expired) >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> On 14 Jan 2014, at 7:18 am, Andrew Beekhof >>>>>>>>>>>>>>> <and...@beekhof.net> wrote: >>>>>>>>>>>>>>>> On 13 Jan 2014, at 8:31 pm, Andrey Groshev >>>>>>>>>>>>>>>> <gre...@yandex.ru> wrote: >>>>>>>>>>>>>>>>> 13.01.2014, 02:51, "Andrew Beekhof" >>>>>>>>>>>>>>>>> <and...@beekhof.net>: >>>>>>>>>>>>>>>>>> On 10 Jan 2014, at 9:55 pm, Andrey Groshev >>>>>>>>>>>>>>>>>> <gre...@yandex.ru> wrote: >>>>>>>>>>>>>>>>>>> 10.01.2014, 14:31, "Andrey Groshev" >>>>>>>>>>>>>>>>>>> <gre...@yandex.ru>: >>>>>>>>>>>>>>>>>>>> 10.01.2014, 14:01, "Andrew Beekhof" >>>>>>>>>>>>>>>>>>>> <and...@beekhof.net>: >>>>>>>>>>>>>>>>>>>>> On 10 Jan 2014, at 5:03 pm, Andrey Groshev >>>>>>>>>>>>>>>>>>>>> <gre...@yandex.ru> wrote: >>>>>>>>>>>>>>>>>>>>>> 10.01.2014, 05:29, "Andrew Beekhof" >>>>>>>>>>>>>>>>>>>>>> <and...@beekhof.net>: >>>>>>>>>>>>>>>>>>>>>>> On 9 Jan 2014, at 11:11 pm, Andrey Groshev >>>>>>>>>>>>>>>>>>>>>>> <gre...@yandex.ru> wrote: >>>>>>>>>>>>>>>>>>>>>>>> 08.01.2014, 06:22, "Andrew Beekhof" >>>>>>>>>>>>>>>>>>>>>>>> <and...@beekhof.net>: >>>>>>>>>>>>>>>>>>>>>>>>> On 29 Nov 2013, at 7:17 pm, Andrey Groshev >>>>>>>>>>>>>>>>>>>>>>>>> <gre...@yandex.ru> wrote: >>>>>>>>>>>>>>>>>>>>>>>>>> Hi, ALL. >>>>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>>>> I'm still trying to cope with the fact >>>>>>>>>>>>>>>>>>>>>>>>>> that after the fence - node hangs in "pending". >>>>>>>>>>>>>>>>>>>>>>>>> Please define "pending". Where did you >>>>>>>>>>>>>>>>>>>>>>>>> see this? >>>>>>>>>>>>>>>>>>>>>>>> In crm_mon: >>>>>>>>>>>>>>>>>>>>>>>> ...... >>>>>>>>>>>>>>>>>>>>>>>> Node dev-cluster2-node2 (172793105): pending >>>>>>>>>>>>>>>>>>>>>>>> ...... >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> The experiment was like this: >>>>>>>>>>>>>>>>>>>>>>>> Four nodes in cluster. >>>>>>>>>>>>>>>>>>>>>>>> On one of them kill corosync or pacemakerd >>>>>>>>>>>>>>>>>>>>>>>> (signal 4 or 6 oк 11). >>>>>>>>>>>>>>>>>>>>>>>> Thereafter, the remaining start it >>>>>>>>>>>>>>>>>>>>>>>> constantly reboot, under various pretexts, "softly >>>>>>>>>>>>>>>>>>>>>>>> whistling", "fly low", "not a cluster member!" ... >>>>>>>>>>>>>>>>>>>>>>>> Then in the log fell out "Too many failures >>>>>>>>>>>>>>>>>>>>>>>> ...." >>>>>>>>>>>>>>>>>>>>>>>> All this time in the status in crm_mon is >>>>>>>>>>>>>>>>>>>>>>>> "pending". >>>>>>>>>>>>>>>>>>>>>>>> Depending on the wind direction changed to >>>>>>>>>>>>>>>>>>>>>>>> "UNCLEAN" >>>>>>>>>>>>>>>>>>>>>>>> Much time has passed and I can not >>>>>>>>>>>>>>>>>>>>>>>> accurately describe the behavior... >>>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>>> Now I am in the following state: >>>>>>>>>>>>>>>>>>>>>>>> I tried locate the problem. Came here with >>>>>>>>>>>>>>>>>>>>>>>> this. >>>>>>>>>>>>>>>>>>>>>>>> I set big value in property >>>>>>>>>>>>>>>>>>>>>>>> stonith-timeout="600s". >>>>>>>>>>>>>>>>>>>>>>>> And got the following behavior: >>>>>>>>>>>>>>>>>>>>>>>> 1. pkill -4 corosync >>>>>>>>>>>>>>>>>>>>>>>> 2. from node with DC call my fence agent >>>>>>>>>>>>>>>>>>>>>>>> "sshbykey" >>>>>>>>>>>>>>>>>>>>>>>> 3. It sends reboot victim and waits until >>>>>>>>>>>>>>>>>>>>>>>> she comes to life again. >>>>>>>>>>>>>>>>>>>>>>> Hmmm.... what version of pacemaker? >>>>>>>>>>>>>>>>>>>>>>> This sounds like a timing issue that we fixed >>>>>>>>>>>>>>>>>>>>>>> a while back >>>>>>>>>>>>>>>>>>>>>> Was a version 1.1.11 from December 3. >>>>>>>>>>>>>>>>>>>>>> Now try full update and retest. >>>>>>>>>>>>>>>>>>>>> That should be recent enough. Can you create a >>>>>>>>>>>>>>>>>>>>> crm_report the next time you reproduce? >>>>>>>>>>>>>>>>>>>> Of course yes. Little delay.... :) >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> ...... >>>>>>>>>>>>>>>>>>>> cc1: warnings being treated as errors >>>>>>>>>>>>>>>>>>>> upstart.c: In function ‘upstart_job_property’: >>>>>>>>>>>>>>>>>>>> upstart.c:264: error: implicit declaration of >>>>>>>>>>>>>>>>>>>> function ‘g_variant_lookup_value’ >>>>>>>>>>>>>>>>>>>> upstart.c:264: error: nested extern declaration of >>>>>>>>>>>>>>>>>>>> ‘g_variant_lookup_value’ >>>>>>>>>>>>>>>>>>>> upstart.c:264: error: assignment makes pointer >>>>>>>>>>>>>>>>>>>> from integer without a cast >>>>>>>>>>>>>>>>>>>> gmake[2]: *** [libcrmservice_la-upstart.lo] Error 1 >>>>>>>>>>>>>>>>>>>> gmake[2]: Leaving directory >>>>>>>>>>>>>>>>>>>> `/root/ha/pacemaker/lib/services' >>>>>>>>>>>>>>>>>>>> make[1]: *** [all-recursive] Error 1 >>>>>>>>>>>>>>>>>>>> make[1]: Leaving directory `/root/ha/pacemaker/lib' >>>>>>>>>>>>>>>>>>>> make: *** [core] Error 1 >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> I'm trying to solve this a problem. >>>>>>>>>>>>>>>>>>> Do not get solved quickly... >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> https://developer.gnome.org/glib/2.28/glib-GVariant.html#g-variant-lookup-value >>>>>>>>>>>>>>>>>>> g_variant_lookup_value () Since 2.28 >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> # yum list installed glib2 >>>>>>>>>>>>>>>>>>> Loaded plugins: fastestmirror, rhnplugin, security >>>>>>>>>>>>>>>>>>> This system is receiving updates from RHN Classic >>>>>>>>>>>>>>>>>>> or Red Hat Satellite. >>>>>>>>>>>>>>>>>>> Loading mirror speeds from cached hostfile >>>>>>>>>>>>>>>>>>> Installed Packages >>>>>>>>>>>>>>>>>>> glib2.x86_64 >>>>>>>>>>>>>>>>>>> 2.26.1-3.el6 >>>>>>>>>>>>>>>>>>> installed >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> # cat /etc/issue >>>>>>>>>>>>>>>>>>> CentOS release 6.5 (Final) >>>>>>>>>>>>>>>>>>> Kernel \r on an \m >>>>>>>>>>>>>>>>>> Can you try this patch? >>>>>>>>>>>>>>>>>> Upstart jobs wont work, but the code will compile >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> diff --git a/lib/services/upstart.c >>>>>>>>>>>>>>>>>> b/lib/services/upstart.c >>>>>>>>>>>>>>>>>> index 831e7cf..195c3a4 100644 >>>>>>>>>>>>>>>>>> --- a/lib/services/upstart.c >>>>>>>>>>>>>>>>>> +++ b/lib/services/upstart.c >>>>>>>>>>>>>>>>>> @@ -231,12 +231,21 @@ upstart_job_exists(const char >>>>>>>>>>>>>>>>>> *name) >>>>>>>>>>>>>>>>>> static char * >>>>>>>>>>>>>>>>>> upstart_job_property(const char *obj, const gchar * >>>>>>>>>>>>>>>>>> iface, const char *name) >>>>>>>>>>>>>>>>>> { >>>>>>>>>>>>>>>>>> + char *output = NULL; >>>>>>>>>>>>>>>>>> + >>>>>>>>>>>>>>>>>> +#if !GLIB_CHECK_VERSION(2,28,0) >>>>>>>>>>>>>>>>>> + static bool err = TRUE; >>>>>>>>>>>>>>>>>> + >>>>>>>>>>>>>>>>>> + if(err) { >>>>>>>>>>>>>>>>>> + crm_err("This version of glib is too old to >>>>>>>>>>>>>>>>>> support upstart jobs"); >>>>>>>>>>>>>>>>>> + err = FALSE; >>>>>>>>>>>>>>>>>> + } >>>>>>>>>>>>>>>>>> +#else >>>>>>>>>>>>>>>>>> GError *error = NULL; >>>>>>>>>>>>>>>>>> GDBusProxy *proxy; >>>>>>>>>>>>>>>>>> GVariant *asv = NULL; >>>>>>>>>>>>>>>>>> GVariant *value = NULL; >>>>>>>>>>>>>>>>>> GVariant *_ret = NULL; >>>>>>>>>>>>>>>>>> - char *output = NULL; >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> crm_info("Calling GetAll on %s", obj); >>>>>>>>>>>>>>>>>> proxy = get_proxy(obj, BUS_PROPERTY_IFACE); >>>>>>>>>>>>>>>>>> @@ -272,6 +281,7 @@ upstart_job_property(const char >>>>>>>>>>>>>>>>>> *obj, const gchar * iface, const char *name) >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> g_object_unref(proxy); >>>>>>>>>>>>>>>>>> g_variant_unref(_ret); >>>>>>>>>>>>>>>>>> +#endif >>>>>>>>>>>>>>>>>> return output; >>>>>>>>>>>>>>>>>> } >>>>>>>>>>>>>>>>> Ok :) I patch source. >>>>>>>>>>>>>>>>> Type "make rc" - the same error. >>>>>>>>>>>>>>>> Because its not building your local changes >>>>>>>>>>>>>>>>> Make new copy via "fetch" - the same error. >>>>>>>>>>>>>>>>> It seems that if not exist >>>>>>>>>>>>>>>>> ClusterLabs-pacemaker-Pacemaker-1.1.11-rc3.tar.gz, then >>>>>>>>>>>>>>>>> download it. >>>>>>>>>>>>>>>>> Otherwise use exist archive. >>>>>>>>>>>>>>>>> Cutted log ....... >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> # make rc >>>>>>>>>>>>>>>>> make TAG=Pacemaker-1.1.11-rc3 rpm >>>>>>>>>>>>>>>>> make[1]: Entering directory `/root/ha/pacemaker' >>>>>>>>>>>>>>>>> rm -f pacemaker-dirty.tar.* pacemaker-tip.tar.* >>>>>>>>>>>>>>>>> pacemaker-HEAD.tar.* >>>>>>>>>>>>>>>>> if [ ! -f >>>>>>>>>>>>>>>>> ClusterLabs-pacemaker-Pacemaker-1.1.11-rc3.tar.gz ]; then >>>>>>>>>>>>>>>>> \ >>>>>>>>>>>>>>>>> rm -f pacemaker.tar.*; >>>>>>>>>>>>>>>>> \ >>>>>>>>>>>>>>>>> if [ Pacemaker-1.1.11-rc3 = dirty ]; then >>>>>>>>>>>>>>>>> \ >>>>>>>>>>>>>>>>> git commit -m "DO-NOT-PUSH" -a; >>>>>>>>>>>>>>>>> \ >>>>>>>>>>>>>>>>> git archive >>>>>>>>>>>>>>>>> --prefix=ClusterLabs-pacemaker-Pacemaker-1.1.11-rc3/ HEAD | >>>>>>>>>>>>>>>>> gzip > ClusterLabs-pacemaker-Pacemaker-1.1.11-rc3.tar.gz; >>>>>>>>>>>>>>>>> \ >>>>>>>>>>>>>>>>> git reset --mixed HEAD^; >>>>>>>>>>>>>>>>> \ >>>>>>>>>>>>>>>>> else >>>>>>>>>>>>>>>>> \ >>>>>>>>>>>>>>>>> git archive >>>>>>>>>>>>>>>>> --prefix=ClusterLabs-pacemaker-Pacemaker-1.1.11-rc3/ >>>>>>>>>>>>>>>>> Pacemaker-1.1.11-rc3 | gzip > >>>>>>>>>>>>>>>>> ClusterLabs-pacemaker-Pacemaker-1.1.11-rc3.tar.gz; \ >>>>>>>>>>>>>>>>> fi; >>>>>>>>>>>>>>>>> \ >>>>>>>>>>>>>>>>> echo `date`: Rebuilt >>>>>>>>>>>>>>>>> ClusterLabs-pacemaker-Pacemaker-1.1.11-rc3.tar.gz; >>>>>>>>>>>>>>>>> \ >>>>>>>>>>>>>>>>> else >>>>>>>>>>>>>>>>> \ >>>>>>>>>>>>>>>>> echo `date`: Using existing tarball: >>>>>>>>>>>>>>>>> ClusterLabs-pacemaker-Pacemaker-1.1.11-rc3.tar.gz; >>>>>>>>>>>>>>>>> \ >>>>>>>>>>>>>>>>> fi >>>>>>>>>>>>>>>>> Mon Jan 13 13:23:21 MSK 2014: Using existing tarball: >>>>>>>>>>>>>>>>> ClusterLabs-pacemaker-Pacemaker-1.1.11-rc3.tar.gz >>>>>>>>>>>>>>>>> ....... >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Well, "make rpm" - build rpms and I create cluster. >>>>>>>>>>>>>>>>> I spent the same tests and confirmed the behavior. >>>>>>>>>>>>>>>>> crm_reoprt log here - http://send2me.ru/crmrep.tar.bz2 >>>>>>>>>>>>>>>> Thanks! >>>>>>>>>>>>>> , >>>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>>>>>>>>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>>>>>>>>>>>> >>>>>>>>>>>>>> Project Home: http://www.clusterlabs.org >>>>>>>>>>>>>> Getting started: >>>>>>>>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>>>>>>>>>>> Bugs: http://bugs.clusterlabs.org >>>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>>>>>>>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>>>>>>>>>>> >>>>>>>>>>>>> Project Home: http://www.clusterlabs.org >>>>>>>>>>>>> Getting started: >>>>>>>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>>>>>>>>>> Bugs: http://bugs.clusterlabs.org >>>>>>>>>>>> _______________________________________________ >>>>>>>>>>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>>>>>>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>>>>>>>>>> >>>>>>>>>>>> Project Home: http://www.clusterlabs.org >>>>>>>>>>>> Getting started: >>>>>>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>>>>>>>>> Bugs: http://bugs.clusterlabs.org >>>>>>>>>>> , >>>>>>>>>>> _______________________________________________ >>>>>>>>>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>>>>>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>>>>>>>>> >>>>>>>>>>> Project Home: http://www.clusterlabs.org >>>>>>>>>>> Getting started: >>>>>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>>>>>>>> Bugs: http://bugs.clusterlabs.org >>>>>>>>>> _______________________________________________ >>>>>>>>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>>>>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>>>>>>>> >>>>>>>>>> Project Home: http://www.clusterlabs.org >>>>>>>>>> Getting started: >>>>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>>>>>>> Bugs: http://bugs.clusterlabs.org >>>>>>>>> , >>>>>>>>> _______________________________________________ >>>>>>>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>>>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>>>>>>> >>>>>>>>> Project Home: http://www.clusterlabs.org >>>>>>>>> Getting started: >>>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>>>>>> Bugs: http://bugs.clusterlabs.org >>>>>>>> _______________________________________________ >>>>>>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>>>>>> >>>>>>>> Project Home: http://www.clusterlabs.org >>>>>>>> Getting started: >>>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>>>>> Bugs: http://bugs.clusterlabs.org >>>>>>> , >>>>>>> _______________________________________________ >>>>>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>>>>> >>>>>>> Project Home: http://www.clusterlabs.org >>>>>>> Getting started: >>>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>>>> Bugs: http://bugs.clusterlabs.org >>>>>> _______________________________________________ >>>>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>>>> >>>>>> Project Home: http://www.clusterlabs.org >>>>>> Getting started: >>>>>> http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>>> Bugs: http://bugs.clusterlabs.org >>>>> , >>>>> _______________________________________________ >>>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>>> >>>>> Project Home: http://www.clusterlabs.org >>>>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>>> Bugs: http://bugs.clusterlabs.org >>>> _______________________________________________ >>>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>>> >>>> Project Home: http://www.clusterlabs.org >>>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>>> Bugs: http://bugs.clusterlabs.org >>> _______________________________________________ >>> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >>> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >>> >>> Project Home: http://www.clusterlabs.org >>> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >>> Bugs: http://bugs.clusterlabs.org >> >> , >> _______________________________________________ >> Pacemaker mailing list: Pacemaker@oss.clusterlabs.org >> http://oss.clusterlabs.org/mailman/listinfo/pacemaker >> >> Project Home: http://www.clusterlabs.org >> Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf >> Bugs: http://bugs.clusterlabs.org > > _______________________________________________ > Pacemaker mailing list: Pacemaker@oss.clusterlabs.org > http://oss.clusterlabs.org/mailman/listinfo/pacemaker > > Project Home: http://www.clusterlabs.org > Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf > Bugs: http://bugs.clusterlabs.org
signature.asc
Description: Message signed with OpenPGP using GPGMail
_______________________________________________ Pacemaker mailing list: Pacemaker@oss.clusterlabs.org http://oss.clusterlabs.org/mailman/listinfo/pacemaker Project Home: http://www.clusterlabs.org Getting started: http://www.clusterlabs.org/doc/Cluster_from_Scratch.pdf Bugs: http://bugs.clusterlabs.org