Hi, Alena, Im trying to follow your steps:
* disable cluster Succeed. * enable maintenance for the primary storage in the cluster Maintenance on VMware cluster failed for the first two trys, with error message like: Unable to create a deployment for VM[ConsoleProxy|v-38-VM] WARN [cloud.consoleproxy.ConsoleProxyManagerImpl] (consoleproxy-1:) Exception while trying to start console proxy com.cloud.exception.InsufficientServerCapacityException: Unable to create a deployment for VM[ConsoleProxy|v-47-VM]Scope=interface com.cloud.dc.DataCenter; id=1 seems each time a new system VM was created, but still on VMware cluster, which leads to failure The maintenance succeed in the third try. * put hosts in cluster into maintenance mode Succeed * destroy system vms Destroying them does not stop them re-create * delete hosts and primary storage Failed to delete primary storage, with message: there are still volumes associated with this pool * delete the cluster Putting hosts/storage into maintenance mode does not stop system VMs re-create >From codes I can see management server get supported hypervisorTypes and >always fetch the first one, and the first one in my environment happens to be >vmware. I have changed expunge.interval = expunge.delay = 120 Should I set consoleproxy.restart = false and update db to set secondary.storage.vm=false ? Regards Mice -----邮件原件----- 发件人: Alena Prokharchyk [mailto:alena.prokharc...@citrix.com] 发送时间: 2012年7月12日 10:03 收件人: cloudstack-dev@incubator.apache.org 主题: Re: System VMs restarted on a disabled cluster On 7/11/12 6:29 PM, "Mice Xia" <mice_...@tcloudcomputing.com> wrote: >Hi, All > > > >I've set up an environment with two clusters (in the same pod), one >Xenserver and the other is VMware, based on 3.0.x ASF branch. > >Now I'm trying to remove the VMware cluster begin with disabling it and >destroying the system VMS running on it, but the systemVMs restarted >immediately on VMware cluster, which blocks cluster removal. > > > >I wonder if this is the expected result by design, or should it be >better that the system VMs get allocated on an enabled cluster? > > > > > >Regards > >Mice > > It's by design. Disabled cluster just can't be used for creating new / starting existing user vms / routers; but it still can be used by system resources (SSVM and Console proxy). To delete the cluster, you need to: * disable cluster * enable maintenance for the primary storage in the cluster * put hosts in cluster into maintenance mode * destroy system vms * delete hosts and primary storage * delete the cluster -Alena.