Well, the clone runs on all nodes (see crm_mon output). That's fine and I guess 
it isn't globally unique then. It is just not verified correctly by crm_verify 
(I suppose).

Here's the clone-config:


       <clone id="clone_ldirectord">

         <instance_attributes id="clone_ldirectord_inst_attr">
           <attributes>
             <nvpair id="clone_ldirector_conf_meta_attr_clone_max" 
name="clone_max" value="3"/>
             <nvpair id="clone_ldirector_conf_meta_attr_clone_node_max" 
name="clone_node_max" value="1"/>
           </attributes>
         </instance_attributes>

         <group id="group_clone_ldirectord_lvs-monitor">

           <primitive id="resource_ldirectord" class="ocf" type="ldirectord" 
provider="heartbeat">
             <operations>
               <op id="resource_ldirectord_operation_op" name="monitor" 
description="ldirectord-monitor" interval="10" timeout="3" start_delay="10s" 
disabled="false" role="Started" prereq="nothing" on_
fail="restart"/>
             </operations>
             <instance_attributes id="ldirectord_inst_attributes">
               <attributes>
                 <nvpair id="ldirector_attr_configfile" name="configfile" 
value="/etc/ha.d/ldirectord.cf"/>
                 <nvpair id="ldirector_attr_binary" name="ldirectord" 
value="/usr/sbin/ldirectord"/>
               </attributes>
             </instance_attributes>
           </primitive>
           <meta_attributes id="clone_ldirectord-meta-options">
             <attributes>
               <nvpair id="clone_ldirectord-meta-options-target-role" 
name="target-role" value="Started"/>
               <nvpair id="clone_ldirectord-meta-options-is-managed" 
name="is-managed" value="true"/>
               <nvpair id="clone_ldirectord-meta-options-notify" name="notify" 
value="true"/>
             </attributes>
           </meta_attributes>

           <primitive class="lsb" id="lvs-monitor" type="lvs-monitor" 
restart_type="restart">
             <operations>
               <op id="lvs-monitor_op_start" name="start" timeout="2s"/>
               <op id="lvs-monitor_op_stop" name="stop" timeout="2s"/>
               <op id="lvs-monitor_op_status" name="monitor" interval="5s" 
timeout="2s"/>
             </operations>
           </primitive>

         </group>

       </clone> 

-----Ursprüngliche Nachricht-----
Von: [email protected] 
[mailto:[email protected]] Im Auftrag von Dejan Muhamedagic
Gesendet: Freitag, 9. Januar 2009 13:04
An: General Linux-HA mailing list
Betreff: Re: [Linux-HA] crm_verify bug?

Hi,

On Fri, Jan 09, 2009 at 12:39:34PM +0100, Arndt Roth wrote:
> Hi *,
> 
>  
> 
> I have a group with one OCF and one LSB resource defined into a clone
> (OCF-Clone + LSB-primitive in a group didn't work). 
> 
>  
> 
> Clone Set: clone_ldirectord
> 
>     Resource Group: group_clone_ldirectord_lvs-monitor:0
> 
>         resource_ldirectord:0   (ocf::heartbeat:ldirectord):    Started
> server2
> 
>         lvs-monitor:0   (lsb:lvs-monitor):      Started server2
> 
>     Resource Group: group_clone_ldirectord_lvs-monitor:1
> 
>         resource_ldirectord:1   (ocf::heartbeat:ldirectord):    Started
> server1
> 
>         lvs-monitor:1   (lsb:lvs-monitor):      Started server1
> 
>     Resource Group: group_clone_ldirectord_lvs-monitor:2
> 
>         resource_ldirectord:2   (ocf::heartbeat:ldirectord):    Started
> server3
> 
>         lvs-monitor:2   (lsb:lvs-monitor):      Started server3
> 
>  
> 
> Now I get this confusing error-message on all 3 nodes when verifying
> with crm_verify:
> 
>  
> 
> [r...@server2:~]$ crm_verify -LV
> 
> crm_verify[20534]: 2009/01/09_12:28:27 WARN: unpack_rsc_op:
> resource_ldirectord:0_monitor_0 found active resource_ldirectord:0 on
> server2
> 
> crm_verify[20534]: 2009/01/09_12:28:27 WARN: unpack_rsc_op:
> resource_ldirectord:1_monitor_0 found active resource_ldirectord:1 on
> server2
> 
> crm_verify[20534]: 2009/01/09_12:28:27 WARN: unpack_rsc_op:
> resource_ldirectord:2_monitor_0 found active resource_ldirectord:2 on
> server2
> 
> crm_verify[20534]: 2009/01/09_12:28:27 WARN: unpack_rsc_op:
> lvs-monitor:2_monitor_0 found active lvs-monitor:2 on server2
> 
> crm_verify[20534]: 2009/01/09_12:28:27 WARN: unpack_rsc_op:
> lvs-monitor:0_monitor_0 found active lvs-monitor:0 on server2
> 
> crm_verify[20534]: 2009/01/09_12:28:27 WARN: unpack_rsc_op:
> lvs-monitor:1_monitor_0 found active lvs-monitor:1 on server2
> 
>  
> 
> Why does crm_verify display on all nodes that all resources are active
> on server2?
> 
> Apart from the error-message everything looks ok, but how can I confirm
> that it is only a crm_verify problem or bug?

Perhaps you can show the configuration too. Maybe this is due to
globally-unique set to true (the default) for the clone?

Thanks,

Dejan

>  
> 
> Thanks for advice,
> 
>  
> 
> Arndt
> 
>  
> 
> 
> 
> 
> _______________________________________________
> Linux-HA mailing list
> [email protected]
> http://lists.linux-ha.org/mailman/listinfo/linux-ha
> See also: http://linux-ha.org/ReportingProblems
_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems



_______________________________________________
Linux-HA mailing list
[email protected]
http://lists.linux-ha.org/mailman/listinfo/linux-ha
See also: http://linux-ha.org/ReportingProblems

Reply via email to