Hi, Alena,

Im trying to follow your steps:

* disable cluster
Succeed.

* enable maintenance for the primary storage in the cluster
Maintenance on VMware cluster failed for the first two trys, with error message 
like: 
Unable to create a deployment for VM[ConsoleProxy|v-38-VM]

WARN  [cloud.consoleproxy.ConsoleProxyManagerImpl] (consoleproxy-1:) Exception 
while trying to start console proxy
com.cloud.exception.InsufficientServerCapacityException: Unable to create a 
deployment for VM[ConsoleProxy|v-47-VM]Scope=interface com.cloud.dc.DataCenter; 
id=1

seems each time a new system VM was created, but still on VMware cluster, which 
leads to failure
The maintenance succeed in the third try.

* put hosts in cluster into maintenance mode
Succeed

* destroy system vms
Destroying them does not stop them re-create

* delete hosts and primary storage
Failed to delete primary storage, with message: there are still volumes 
associated with this pool

* delete the cluster


Putting hosts/storage into maintenance mode does not stop system VMs re-create
>From codes I can see management server get supported hypervisorTypes and 
>always fetch the first one, and the first one in my environment happens to be 
>vmware.

I have changed expunge.interval = expunge.delay = 120
Should I set consoleproxy.restart = false and update db to set 
secondary.storage.vm=false ?

Regards
Mice

-----邮件原件-----
发件人: Alena Prokharchyk [mailto:alena.prokharc...@citrix.com] 
发送时间: 2012年7月12日 10:03
收件人: cloudstack-dev@incubator.apache.org
主题: Re: System VMs restarted on a disabled cluster

On 7/11/12 6:29 PM, "Mice Xia" <mice_...@tcloudcomputing.com> wrote:

>Hi, All
>
> 
>
>I've set up an environment with two clusters (in the same pod), one
>Xenserver and the other is VMware, based on 3.0.x ASF branch.
>
>Now I'm trying to remove the VMware cluster begin with disabling it and
>destroying the system VMS running on it, but the systemVMs restarted
>immediately on VMware cluster, which blocks cluster removal.
>
> 
>
>I wonder if this is the expected result by design, or should it be
>better that the system VMs get allocated on an enabled cluster?
>
> 
>
> 
>
>Regards
>
>Mice 
>
>



It's by design. Disabled cluster just can't be used for creating new /
starting existing user vms / routers; but it still can be used by system
resources (SSVM and Console proxy).

To delete the cluster, you need to:

* disable cluster
* enable maintenance for the primary storage in the cluster
* put hosts in cluster into maintenance mode

* destroy system vms
* delete hosts and primary storage
* delete the cluster

-Alena.

Reply via email to