Sergey Levitskiy <sergey.levits...@autodesk.com> wrote: > Actually, minor correction. When adding to VM/templates the name of the > detail is rootDiskController for Root controller and dataDiskController > for additional disks. > Also, if you want to make changes on a global scale the changes need to > go to vm_template_details and user_vm_details tables respectively.
Thanks! Very helpful > > On 2/21/17, 8:03 PM, "Sergey Levitskiy" <sergey.levits...@autodesk.com> > wrote: > > Here it is the logic. > 1. Default value is taken from a global configuration > vmware.root.disk.controller > 2. To override add the same config to template or VM (starting from 4.10 > UI allows adding advanced settings to templates and/or VMs). If added to a > template all VMs deployed from it will inherit this value. If added to VM and > then template is created it will also inherits all advanced settings. > > > > > On 2/21/17, 7:06 PM, "Nathan Johnson" <njohn...@ena.com> wrote: > > Sergey Levitskiy <sergey.levits...@autodesk.com> wrote: > >> On vmware rootdiskcontroller is passed over to the hypervisor in VM start >> command. I know for the fact that the following rootdiskcontroller option >> specified in template/vm details work fine: >> ide >> scsi >> lsilogic >> lsilogic1068 >> >> In general, any scsi controller option that vmware recognizes should work. >> >> Thanks, >> Sergey > > Thanks Sergey! So do you happen to know where on the management > server > side the determination is made as to which rootDiskController > parameter to > pass? > > > > >> On 2/21/17, 6:13 PM, "Nathan Johnson" <njohn...@ena.com> wrote: >> >> Wido den Hollander <w...@widodh.nl> wrote: >> >>>> Op 25 januari 2017 om 4:44 schreef Simon Weller <swel...@ena.com>: >>>> >>>> >>>> Maybe this is a good opportunity to discuss modernizing the OS >>>> selections so that drivers (and other features) could be selectable per >>>> OS. >>> >>> That seems like a good idea. If you select Ubuntu 16.04 or CentOS 7.3 >>> then for example it will give you a VirtIO SCSI disk on KVM, anything >>> previous to that will get VirtIO-blk. >> >> So one thing I noticed, there is a possibility of a rootDiskController >> parameter passed to the Start Command. So this means that the Management >> server could control whether to use scsi or virtio, assuming I’m reading >> this correctly, and we shouldn’t necessarily have to rely on the os type >> name inside the agent code. From a quick glance at the vmware code, it >> looks like maybe they already use this parameter? It would be great if >> someone familiar with the vmware code could chime in here. >> >> Thanks, >> >> Nathan >> >> >> >>> Wido >>> >>>> Thoughts? >>>> >>>> >>>> ________________________________ >>>> From: Syed Ahmed <sah...@cloudops.com> >>>> Sent: Tuesday, January 24, 2017 10:46 AM >>>> To: dev@cloudstack.apache.org >>>> Cc: Simon Weller >>>> Subject: Re: Adding VirtIO SCSI to KVM hypervisors >>>> >>>> To maintain backward compatibility we would have to add a config option >>>> here unfortunately. I do like the idea however. We can make the default >>>> VirtIO ISCSI and keep the VirtIO-blk as an alternative for existing >>>> installations. >>>> >>>> On Mon, Jan 23, 2017 at 8:05 AM, Wido den Hollander >>>> <w...@widodh.nl<mailto:w...@widodh.nl>> wrote: >>>> >>>>> Op 21 januari 2017 om 23:50 schreef Wido den Hollander >>>>> <w...@widodh.nl<mailto:w...@widodh.nl>>: >>>>> >>>>> >>>>> >>>>> >>>>>> Op 21 jan. 2017 om 22:59 heeft Syed Ahmed >>>>>> <sah...@cloudops.com<mailto:sah...@cloudops.com>> het volgende >>>>>> geschreven: >>>>>> >>>>>> Exposing this via an API would be tricky but it can definitely be >>>>>> added as >>>>>> a cluster-wide or a global setting in my opinion. By enabling that, >>>>>> all the >>>>>> instances would be using VirtIO SCSI. Is there a reason you'd want >>>>>> some >>>>>> instances to use VirtIIO and others to use VirtIO SCSI? >>>>> >>>>> Even a global setting would be a bit of work and hacky as well. >>>>> >>>>> I do not see any reason to keep VirtIO, it os just that devices will be >>>>> named sdX instead of vdX in the guest. >>>> >>>> To add, the Qemu wiki [0] says: >>>> >>>> "A virtio storage interface for efficient I/O that overcomes virtio-blk >>>> limitations and supports advanced SCSI hardware." >>>> >>>> At OpenStack [1] they also say: >>>> >>>> "It has been designed to replace virtio-blk, increase it's performance >>>> and improve scalability." >>>> >>>> So it seems that VirtIO is there to be removed. I'd say switch to VirtIO >>>> SCSI at version 5.X? :) >>>> >>>> Wido >>>> >>>> [0]: http://wiki.qemu.org/Features/VirtioSCSI >>>> [1]: https://wiki.openstack.org/wiki/LibvirtVirtioScsi >>>> >>>>> That might break existing Instances when not using labels or UUIDs in >>>>> the Instance when mounting. >>>>> >>>>> Wido >>>>> >>>>>>> On Sat, Jan 21, 2017 at 4:22 PM, Simon Weller >>>>>>> <swel...@ena.com<mailto:swel...@ena.com>> wrote: >>>>>>> >>>>>>> For the record, we've been looking into this as well. >>>>>>> Has anyone tried it with Windows VMs before? The standard virtio >>>>>>> driver >>>>>>> doesn't support spanned disks and that's something we'd really like >>>>>>> to >>>>>>> enable for our customers. >>>>>>> >>>>>>> >>>>>>> >>>>>>> Simon Weller/615-312-6068<tel:615-312-6068> <(615)%20312-6068> >>>>>>> >>>>>>> >>>>>>> -----Original Message----- >>>>>>> *From:* Wido den Hollander [w...@widodh.nl<mailto:w...@widodh.nl>] >>>>>>> *Received:* Saturday, 21 Jan 2017, 2:56PM >>>>>>> *To:* Syed Ahmed [sah...@cloudops.com<mailto:sah...@cloudops.com>]; >>>>>>> dev@cloudstack.apache.org<mailto:dev@cloudstack.apache.org> [ >>>>>>> dev@cloudstack.apache.org<mailto:dev@cloudstack.apache.org>] >>>>>>> *Subject:* Re: Adding VirtIO SCSI to KVM hypervisors >>>>>>> >>>>>>> >>>>>>>> Op 21 januari 2017 om 16:15 schreef Syed Ahmed >>>>>>>> <sah...@cloudops.com<mailto:sah...@cloudops.com>>: >>>>>>>> >>>>>>>> >>>>>>>> Wido, >>>>>>>> >>>>>>>> Were you thinking of adding this as a global setting? I can see why >>>>>>>> it >>>>>>> will >>>>>>>> be useful. I'm happy to review any ideas you might have around this. >>>>>>> >>>>>>> Well, not really. We don't have any structure for this in place right >>>>>>> now >>>>>>> to define what type of driver/disk we present to a guest. >>>>>>> >>>>>>> See my answer below. >>>>>>> >>>>>>>> Thanks, >>>>>>>> -Syed >>>>>>>> On Sat, Jan 21, 2017 at 04:46 Laszlo Hornyak >>>>>>>> <laszlo.horn...@gmail.com<mailto:laszlo.horn...@gmail.com>> >>>>>>>> wrote: >>>>>>>> >>>>>>>>> Hi Wido, >>>>>>>>> >>>>>>>>> If I understand correctly from the documentation and your examples, >>>>>>> virtio >>>>>>>>> provides virtio interface to the guest while virtio-scsi provides >>>>>>>>> scsi >>>>>>>>> interface, therefore an IaaS service should not replace it without >>>>>>>>> user >>>>>>>>> request / approval. It would be probably better to let the user set >>>>>>> what >>>>>>>>> kind of IO interface the VM needs. >>>>>>> >>>>>>> You'd say, but we already do those. Some Operating Systems get a IDE >>>>>>> disk, >>>>>>> others a SCSI disk and when Linux guest support it according to our >>>>>>> database we use VirtIO. >>>>>>> >>>>>>> CloudStack has no way of telling how to present a volume to a >>>>>>> guest. I >>>>>>> think it would be a bit to much to just make that configurable. That >>>>>>> would >>>>>>> mean extra database entries, API calls. A bit overkill imho in this >>>>>>> case. >>>>>>> >>>>>>> VirtIO SCSI is supported by all Linux distributions for a very long >>>>>>> time. >>>>>>> >>>>>>> Wido >>>>>>> >>>>>>>>> Best regards, >>>>>>>>> Laszlo >>>>>>>>> >>>>>>>>> On Fri, Jan 20, 2017 at 10:21 PM, Wido den Hollander >>>>>>>>> <w...@widodh.nl<mailto:w...@widodh.nl>> >>>>>>>>> wrote: >>>>>>>>> >>>>>>>>>> Hi, >>>>>>>>>> >>>>>>>>>> VirtIO SCSI [0] has been supported a while now by Linux and all >>>>>>> kernels, >>>>>>>>>> but inside CloudStack we are not using it. There is a issue for >>>>>>>>>> this >>>>>>> [1]. >>>>>>>>>> It would bring more (theoretical) performance to VMs, but one of >>>>>>>>>> the >>>>>>>>>> motivators (for me) is that we can support TRIM/DISCARD [2]. >>>>>>>>>> >>>>>>>>>> This would allow for RBD images on Ceph to shrink, but it can also >>>>>>> give >>>>>>>>>> back free space on QCOW2 images if quests run fstrim. Something >>>>>>>>>> all >>>>>>>>> modern >>>>>>>>>> distributions all do weekly in a CRON. >>>>>>>>>> >>>>>>>>>> Now, it is simple to swap VirtIO for VirtIO SCSI. This would >>>>>>>>>> however >>>>>>> mean >>>>>>>>>> that disks inside VMs are then called /dev/sdX instead of >>>>>>>>>> /dev/vdX. >>>>>>>>>> >>>>>>>>>> For GRUB and such this is no problems. This usually work on UUIDs >>>>>>> and/or >>>>>>>>>> labels, but for static mounts on /dev/vdb1 for example things >>>>>>>>>> break. >>>>>>>>>> >>>>>>>>>> We currently don't have any configuration method on how we want to >>>>>>>>> present >>>>>>>>>> a disk to a guest, so when attaching a volume we can't say that we >>>>>>> want >>>>>>>>> to >>>>>>>>>> use a different driver. If we think that a Operating System >>>>>>>>>> supports >>>>>>>>> VirtIO >>>>>>>>>> we use that driver in KVM. >>>>>>>>>> >>>>>>>>>> Any suggestion on how to add VirtIO SCSI support? >>>>>>>>>> >>>>>>>>>> Wido >>>>>>>>>> >>>>>>>>>> >>>>>>>>>> [0]: http://wiki.qemu.org/Features/VirtioSCSI >>>>>>>>>> [1]: https://issues.apache.org/jira/browse/CLOUDSTACK-8239 >>>>>>>>>> [2]: https://issues.apache.org/jira/browse/CLOUDSTACK-8104 >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> -- >>>>>>>>> >>>>>>>>> EOF >> >> >> Nathan Johnson >> R&D Engineer >> >> >> >> 618 Grassmere Park Drive, Suite 12 >> Nashville, TN 37211 >> General Office: 615-312-6000 >> >> website | blog | support