No problem, John.

I still want to re-review it by myself before coming up with a new patch
file.

Also, maybe I should first wait for Wei's changes to be checked in and
merge those into mine before generating a new patch file?


On Wed, Jun 12, 2013 at 10:40 AM, John Burwell <jburw...@basho.com> wrote:

> Mike,
>
> I just realized that I forgot to publish my review.  I am offline ATM,
> but I will publish it in the next couple of hours.
>
> Do you plan to update your the patch in Review Board?
>
> Sorry for the oversight,
> -John
>
>
>
>
> On Jun 12, 2013, at 2:26 AM, Mike Tutkowski
> <mike.tutkow...@solidfire.com> wrote:
>
> > Hi Edison, John, and Wei (and whoever else is reading this :) ),
> >
> > Just an FYI that I believe I have implemented all the areas we wanted
> > addressed.
> >
> > I plan to review the code again tomorrow morning or afternoon, then send
> in
> > another patch.
> >
> > Thanks for all the work on this everyone!
> >
> >
> > On Tue, Jun 11, 2013 at 12:29 PM, Mike Tutkowski <
> > mike.tutkow...@solidfire.com> wrote:
> >
> >> Sure, that sounds good.
> >>
> >>
> >> On Tue, Jun 11, 2013 at 12:11 PM, Wei ZHOU <ustcweiz...@gmail.com>
> wrote:
> >>
> >>> Hi Mike,
> >>>
> >>> It looks the two feature do not have many conflicts in Java code,
> except
> >>> the cloudstack UI.
> >>> If you do not mind, I will merge disk_io_throttling branch into master
> >>> this
> >>> week, so that you can develop based on it.
> >>>
> >>> -Wei
> >>>
> >>>
> >>> 2013/6/11 Mike Tutkowski <mike.tutkow...@solidfire.com>
> >>>
> >>>> Hey John,
> >>>>
> >>>> The SolidFire patch does not depend on the object_store branch, but -
> as
> >>>> Edison mentioned - it might be easier if we merge the SolidFire branch
> >>> into
> >>>> the object_store branch before object_store goes into master.
> >>>>
> >>>> I'm not sure how the disk_io_throttling fits into this merge strategy.
> >>>> Perhaps Wei can chime in on that.
> >>>>
> >>>>
> >>>> On Tue, Jun 11, 2013 at 11:07 AM, John Burwell <jburw...@basho.com>
> >>> wrote:
> >>>>
> >>>>> Mike,
> >>>>>
> >>>>> We have a delicate merge dance to perform.  The disk_io_throttling,
> >>>>> solidfire, and object_store appear to have a number of overlapping
> >>>>> elements.  I understand the dependencies between the patches to be as
> >>>>> follows:
> >>>>>
> >>>>>        object_store <- solidfire -> disk_io_throttling
> >>>>>
> >>>>> Am I correct that the device management aspects of SolidFire are
> >>> additive
> >>>>> to the object_store branch or there are circular dependency between
> >>> the
> >>>>> branches?  Once we understand the dependency graph, we can determine
> >>> the
> >>>>> best approach to land the changes in master.
> >>>>>
> >>>>> Thanks,
> >>>>> -John
> >>>>>
> >>>>>
> >>>>> On Jun 10, 2013, at 11:10 PM, Mike Tutkowski <
> >>>> mike.tutkow...@solidfire.com>
> >>>>> wrote:
> >>>>>
> >>>>>> Also, if we are good with Edison merging my code into his branch
> >>> before
> >>>>>> going into master, I am good with that.
> >>>>>>
> >>>>>> We can remove the StoragePoolType.Dynamic code after his merge and
> >>> we
> >>>> can
> >>>>>> deal with Burst IOPS then, as well.
> >>>>>>
> >>>>>>
> >>>>>> On Mon, Jun 10, 2013 at 9:08 PM, Mike Tutkowski <
> >>>>>> mike.tutkow...@solidfire.com> wrote:
> >>>>>>
> >>>>>>> Let me make sure I follow where we're going here:
> >>>>>>>
> >>>>>>> 1) There should be NO references to hypervisor code in the storage
> >>>>>>> plug-ins code (this includes the default storage plug-in, which
> >>>>> currently
> >>>>>>> sends several commands to the hypervisor in use (although it does
> >>> not
> >>>>> know
> >>>>>>> which hypervisor (XenServer, ESX(i), etc.) is actually in use))
> >>>>>>>
> >>>>>>> 2) managed=true or managed=false can be placed in the url field (if
> >>>> not
> >>>>>>> present, we default to false). This info is stored in the
> >>>>>>> storage_pool_details table.
> >>>>>>>
> >>>>>>> 3) When the "attach" command is sent to the hypervisor in
> >>> question, we
> >>>>>>> pass the managed property along (this takes the place of the
> >>>>>>> StoragePoolType.Dynamic check).
> >>>>>>>
> >>>>>>> 4) execute(AttachVolumeCommand) in the hypervisor checks for the
> >>>> managed
> >>>>>>> property. If true for an attach, the necessary hypervisor data
> >>>>> structure is
> >>>>>>> created and the rest of the attach command executes to attach the
> >>>>> volume.
> >>>>>>>
> >>>>>>> 5) When execute(AttachVolumeCommand) is invoked to detach a volume,
> >>>> the
> >>>>>>> same check is made. If managed, the hypervisor data structure is
> >>>>> removed.
> >>>>>>>
> >>>>>>> 6) I do not see an clear way to support Burst IOPS in 4.2 unless
> >>> it is
> >>>>>>> stored in the volumes and disk_offerings table. If we have some
> >>> idea,
> >>>>>>> that'd be cool.
> >>>>>>>
> >>>>>>> Thanks!
> >>>>>>>
> >>>>>>>
> >>>>>>> On Mon, Jun 10, 2013 at 8:58 PM, Mike Tutkowski <
> >>>>>>> mike.tutkow...@solidfire.com> wrote:
> >>>>>>>
> >>>>>>>> "+1 -- Burst IOPS can be implemented while avoiding implementation
> >>>>>>>> attributes.  I always wondered about the details field.  I think
> >>> we
> >>>>> should
> >>>>>>>> beef up the description in the documentation regarding the
> >>> expected
> >>>>> format
> >>>>>>>> of the field.  In 4.1, I noticed that the details are not
> >>> returned on
> >>>>> the
> >>>>>>>> createStoratePool updateStoragePool, or listStoragePool response.
> >>>> Why
> >>>>>>>> don't we return it?  It seems like it would be useful for clients
> >>> to
> >>>> be
> >>>>>>>> able to inspect the contents of the details field."
> >>>>>>>>
> >>>>>>>> Not sure how this would work storing Burst IOPS here.
> >>>>>>>>
> >>>>>>>> Burst IOPS need to be variable on a Disk Offering-by-Disk Offering
> >>>>>>>> basis. For each Disk Offering created, you have to be able to
> >>>> associate
> >>>>>>>> unique Burst IOPS. There is a disk_offering_details table. Maybe
> >>> it
> >>>>> could
> >>>>>>>> go there?
> >>>>>>>>
> >>>>>>>> I'm also not sure how you would accept the Burst IOPS in the GUI
> >>> if
> >>>>> it's
> >>>>>>>> not stored like the Min and Max fields are in the DB.
> >>>>>>>
> >>>>>>>
> >>>>>>>
> >>>>>>> --
> >>>>>>> *Mike Tutkowski*
> >>>>>>> *Senior CloudStack Developer, SolidFire Inc.*
> >>>>>>> e: mike.tutkow...@solidfire.com
> >>>>>>> o: 303.746.7302
> >>>>>>> Advancing the way the world uses the cloud<
> >>>>> http://solidfire.com/solution/overview/?video=play>
> >>>>>>> *™*
> >>>>>>
> >>>>>>
> >>>>>>
> >>>>>> --
> >>>>>> *Mike Tutkowski*
> >>>>>> *Senior CloudStack Developer, SolidFire Inc.*
> >>>>>> e: mike.tutkow...@solidfire.com
> >>>>>> o: 303.746.7302
> >>>>>> Advancing the way the world uses the
> >>>>>> cloud<http://solidfire.com/solution/overview/?video=play>
> >>>>>> *™*
> >>>>
> >>>>
> >>>> --
> >>>> *Mike Tutkowski*
> >>>> *Senior CloudStack Developer, SolidFire Inc.*
> >>>> e: mike.tutkow...@solidfire.com
> >>>> o: 303.746.7302
> >>>> Advancing the way the world uses the
> >>>> cloud<http://solidfire.com/solution/overview/?video=play>
> >>>> *™*
> >>
> >>
> >>
> >> --
> >> *Mike Tutkowski*
> >> *Senior CloudStack Developer, SolidFire Inc.*
> >> e: mike.tutkow...@solidfire.com
> >> o: 303.746.7302
> >> Advancing the way the world uses the cloud<
> http://solidfire.com/solution/overview/?video=play>
> >> *™*
> >
> >
> >
> > --
> > *Mike Tutkowski*
> > *Senior CloudStack Developer, SolidFire Inc.*
> > e: mike.tutkow...@solidfire.com
> > o: 303.746.7302
> > Advancing the way the world uses the
> > cloud<http://solidfire.com/solution/overview/?video=play>
> > *™*
>



-- 
*Mike Tutkowski*
*Senior CloudStack Developer, SolidFire Inc.*
e: mike.tutkow...@solidfire.com
o: 303.746.7302
Advancing the way the world uses the
cloud<http://solidfire.com/solution/overview/?video=play>
*™*

Reply via email to