Actually, minor correction. When adding to VM/templates the name of the detail is rootDiskController for Root controller and dataDiskController for additional disks. Also, if you want to make changes on a global scale the changes need to go to vm_template_details and user_vm_details tables respectively.
On 2/21/17, 8:03 PM, "Sergey Levitskiy" <sergey.levits...@autodesk.com> wrote: Here it is the logic. 1. Default value is taken from a global configuration vmware.root.disk.controller 2. To override add the same config to template or VM (starting from 4.10 UI allows adding advanced settings to templates and/or VMs). If added to a template all VMs deployed from it will inherit this value. If added to VM and then template is created it will also inherits all advanced settings. On 2/21/17, 7:06 PM, "Nathan Johnson" <njohn...@ena.com> wrote: Sergey Levitskiy <sergey.levits...@autodesk.com> wrote: > On vmware rootdiskcontroller is passed over to the hypervisor in VM start > command. I know for the fact that the following rootdiskcontroller option > specified in template/vm details work fine: > ide > scsi > lsilogic > lsilogic1068 > > In general, any scsi controller option that vmware recognizes should work. > > Thanks, > Sergey Thanks Sergey! So do you happen to know where on the management server side the determination is made as to which rootDiskController parameter to pass? > > > On 2/21/17, 6:13 PM, "Nathan Johnson" <njohn...@ena.com> wrote: > > Wido den Hollander <w...@widodh.nl> wrote: > >>> Op 25 januari 2017 om 4:44 schreef Simon Weller <swel...@ena.com>: >>> >>> >>> Maybe this is a good opportunity to discuss modernizing the OS >>> selections so that drivers (and other features) could be selectable per >>> OS. >> >> That seems like a good idea. If you select Ubuntu 16.04 or CentOS 7.3 >> then for example it will give you a VirtIO SCSI disk on KVM, anything >> previous to that will get VirtIO-blk. > > So one thing I noticed, there is a possibility of a rootDiskController > parameter passed to the Start Command. So this means that the Management > server could control whether to use scsi or virtio, assuming I’m reading > this correctly, and we shouldn’t necessarily have to rely on the os type > name inside the agent code. From a quick glance at the vmware code, it > looks like maybe they already use this parameter? It would be great if > someone familiar with the vmware code could chime in here. > > Thanks, > > Nathan > > > >> Wido >> >>> Thoughts? >>> >>> >>> ________________________________ >>> From: Syed Ahmed <sah...@cloudops.com> >>> Sent: Tuesday, January 24, 2017 10:46 AM >>> To: dev@cloudstack.apache.org >>> Cc: Simon Weller >>> Subject: Re: Adding VirtIO SCSI to KVM hypervisors >>> >>> To maintain backward compatibility we would have to add a config option >>> here unfortunately. I do like the idea however. We can make the default >>> VirtIO ISCSI and keep the VirtIO-blk as an alternative for existing >>> installations. >>> >>> On Mon, Jan 23, 2017 at 8:05 AM, Wido den Hollander >>> <w...@widodh.nl<mailto:w...@widodh.nl>> wrote: >>> >>>> Op 21 januari 2017 om 23:50 schreef Wido den Hollander >>>> <w...@widodh.nl<mailto:w...@widodh.nl>>: >>>> >>>> >>>> >>>> >>>>> Op 21 jan. 2017 om 22:59 heeft Syed Ahmed >>>>> <sah...@cloudops.com<mailto:sah...@cloudops.com>> het volgende >>>>> geschreven: >>>>> >>>>> Exposing this via an API would be tricky but it can definitely be >>>>> added as >>>>> a cluster-wide or a global setting in my opinion. By enabling that, >>>>> all the >>>>> instances would be using VirtIO SCSI. Is there a reason you'd want some >>>>> instances to use VirtIIO and others to use VirtIO SCSI? >>>> >>>> Even a global setting would be a bit of work and hacky as well. >>>> >>>> I do not see any reason to keep VirtIO, it os just that devices will be >>>> named sdX instead of vdX in the guest. >>> >>> To add, the Qemu wiki [0] says: >>> >>> "A virtio storage interface for efficient I/O that overcomes virtio-blk >>> limitations and supports advanced SCSI hardware." >>> >>> At OpenStack [1] they also say: >>> >>> "It has been designed to replace virtio-blk, increase it's performance >>> and improve scalability." >>> >>> So it seems that VirtIO is there to be removed. I'd say switch to VirtIO >>> SCSI at version 5.X? :) >>> >>> Wido >>> >>> [0]: http://wiki.qemu.org/Features/VirtioSCSI >>> [1]: https://wiki.openstack.org/wiki/LibvirtVirtioScsi >>> >>>> That might break existing Instances when not using labels or UUIDs in >>>> the Instance when mounting. >>>> >>>> Wido >>>> >>>>>> On Sat, Jan 21, 2017 at 4:22 PM, Simon Weller >>>>>> <swel...@ena.com<mailto:swel...@ena.com>> wrote: >>>>>> >>>>>> For the record, we've been looking into this as well. >>>>>> Has anyone tried it with Windows VMs before? The standard virtio >>>>>> driver >>>>>> doesn't support spanned disks and that's something we'd really like to >>>>>> enable for our customers. >>>>>> >>>>>> >>>>>> >>>>>> Simon Weller/615-312-6068<tel:615-312-6068> <(615)%20312-6068> >>>>>> >>>>>> >>>>>> -----Original Message----- >>>>>> *From:* Wido den Hollander [w...@widodh.nl<mailto:w...@widodh.nl>] >>>>>> *Received:* Saturday, 21 Jan 2017, 2:56PM >>>>>> *To:* Syed Ahmed [sah...@cloudops.com<mailto:sah...@cloudops.com>]; >>>>>> dev@cloudstack.apache.org<mailto:dev@cloudstack.apache.org> [ >>>>>> dev@cloudstack.apache.org<mailto:dev@cloudstack.apache.org>] >>>>>> *Subject:* Re: Adding VirtIO SCSI to KVM hypervisors >>>>>> >>>>>> >>>>>>> Op 21 januari 2017 om 16:15 schreef Syed Ahmed >>>>>>> <sah...@cloudops.com<mailto:sah...@cloudops.com>>: >>>>>>> >>>>>>> >>>>>>> Wido, >>>>>>> >>>>>>> Were you thinking of adding this as a global setting? I can see why >>>>>>> it >>>>>> will >>>>>>> be useful. I'm happy to review any ideas you might have around this. >>>>>> >>>>>> Well, not really. We don't have any structure for this in place right >>>>>> now >>>>>> to define what type of driver/disk we present to a guest. >>>>>> >>>>>> See my answer below. >>>>>> >>>>>>> Thanks, >>>>>>> -Syed >>>>>>> On Sat, Jan 21, 2017 at 04:46 Laszlo Hornyak >>>>>>> <laszlo.horn...@gmail.com<mailto:laszlo.horn...@gmail.com>> >>>>>>> wrote: >>>>>>> >>>>>>>> Hi Wido, >>>>>>>> >>>>>>>> If I understand correctly from the documentation and your examples, >>>>>> virtio >>>>>>>> provides virtio interface to the guest while virtio-scsi provides >>>>>>>> scsi >>>>>>>> interface, therefore an IaaS service should not replace it without >>>>>>>> user >>>>>>>> request / approval. It would be probably better to let the user set >>>>>> what >>>>>>>> kind of IO interface the VM needs. >>>>>> >>>>>> You'd say, but we already do those. Some Operating Systems get a IDE >>>>>> disk, >>>>>> others a SCSI disk and when Linux guest support it according to our >>>>>> database we use VirtIO. >>>>>> >>>>>> CloudStack has no way of telling how to present a volume to a guest. I >>>>>> think it would be a bit to much to just make that configurable. That >>>>>> would >>>>>> mean extra database entries, API calls. A bit overkill imho in this >>>>>> case. >>>>>> >>>>>> VirtIO SCSI is supported by all Linux distributions for a very long >>>>>> time. >>>>>> >>>>>> Wido >>>>>> >>>>>>>> Best regards, >>>>>>>> Laszlo >>>>>>>> >>>>>>>> On Fri, Jan 20, 2017 at 10:21 PM, Wido den Hollander >>>>>>>> <w...@widodh.nl<mailto:w...@widodh.nl>> >>>>>>>> wrote: >>>>>>>> >>>>>>>>> Hi, >>>>>>>>> >>>>>>>>> VirtIO SCSI [0] has been supported a while now by Linux and all >>>>>> kernels, >>>>>>>>> but inside CloudStack we are not using it. There is a issue for >>>>>>>>> this >>>>>> [1]. >>>>>>>>> It would bring more (theoretical) performance to VMs, but one of >>>>>>>>> the >>>>>>>>> motivators (for me) is that we can support TRIM/DISCARD [2]. >>>>>>>>> >>>>>>>>> This would allow for RBD images on Ceph to shrink, but it can also >>>>>> give >>>>>>>>> back free space on QCOW2 images if quests run fstrim. Something all >>>>>>>> modern >>>>>>>>> distributions all do weekly in a CRON. >>>>>>>>> >>>>>>>>> Now, it is simple to swap VirtIO for VirtIO SCSI. This would >>>>>>>>> however >>>>>> mean >>>>>>>>> that disks inside VMs are then called /dev/sdX instead of /dev/vdX. >>>>>>>>> >>>>>>>>> For GRUB and such this is no problems. This usually work on UUIDs >>>>>> and/or >>>>>>>>> labels, but for static mounts on /dev/vdb1 for example things >>>>>>>>> break. >>>>>>>>> >>>>>>>>> We currently don't have any configuration method on how we want to >>>>>>>> present >>>>>>>>> a disk to a guest, so when attaching a volume we can't say that we >>>>>> want >>>>>>>> to >>>>>>>>> use a different driver. If we think that a Operating System >>>>>>>>> supports >>>>>>>> VirtIO >>>>>>>>> we use that driver in KVM. >>>>>>>>> >>>>>>>>> Any suggestion on how to add VirtIO SCSI support? >>>>>>>>> >>>>>>>>> Wido >>>>>>>>> >>>>>>>>> >>>>>>>>> [0]: http://wiki.qemu.org/Features/VirtioSCSI >>>>>>>>> [1]: https://issues.apache.org/jira/browse/CLOUDSTACK-8239 >>>>>>>>> [2]: https://issues.apache.org/jira/browse/CLOUDSTACK-8104 >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> -- >>>>>>>> >>>>>>>> EOF > > > Nathan Johnson > R&D Engineer > > > > 618 Grassmere Park Drive, Suite 12 > Nashville, TN 37211 > General Office: 615-312-6000 > > website | blog | support