Thank you for the clarification. I'm excited to give it a shot.

On Jul 25, 2017 3:24 AM, "David" <dclistsli...@gmail.com> wrote:

> My understanding was Cinder is needed to create/delete/manage etc. on
> volumes but I/O to the volumes is direct from the hypervisors. In theory
> you could lose your Cinder service and VMs would stay up.
>
> On 25 Jul 2017 4:18 a.m., "Brady Deetz" <bde...@gmail.com> wrote:
>
> Thanks for pointing to some documentation. I'd seen that and it is
> certainly an option. From my understanding, with a Cinder deployment, you'd
> have the same failure domains and similar performance characteristics to an
> oVirt + NFS + RBD deployment. This is acceptable. But, the dream I have in
> my head is where the RBD images are mounted and controlled on each
> hypervisor instead of a central storage authority like Cinder. Does that
> exist for anything or is this a fundamentally flawed idea?
>
> On Mon, Jul 24, 2017 at 9:41 PM, Jason Dillaman <jdill...@redhat.com>
> wrote:
>
>> oVirt 3.6 added Cinder/RBD integration [1] and it looks like they are
>> currently working on integrating Cinder within a container to simplify
>> the integration [2].
>>
>> [1] http://www.ovirt.org/develop/release-management/features/sto
>> rage/cinder-integration/
>> [2] http://www.ovirt.org/develop/release-management/features/cin
>> derglance-docker-integration/
>>
>> On Mon, Jul 24, 2017 at 10:27 PM, Brady Deetz <bde...@gmail.com> wrote:
>> > Funny enough, I just had a call with Redhat where the OpenStack
>> engineer was
>> > voicing his frustration that there wasn't any movement on RBD for oVirt.
>> > This is important to me because I'm building out a user-facing private
>> cloud
>> > that just isn't going to be big enough to justify OpenStack and its
>> > administrative overhead. But, I already have 1.75PB (soon to be 2PB) of
>> > CephFS in production. So, it puts me in a really difficult design
>> position.
>> >
>> > On Mon, Jul 24, 2017 at 9:09 PM, Dino Yancey <dino2...@gmail.com>
>> wrote:
>> >>
>> >> I was as much as told by Redhat in a sales call that they push Gluster
>> >> for oVirt/RHEV and Ceph for OpenStack, and don't have any plans to
>> >> change that in the short term. (note this was about a year ago, i
>> >> think - so this isn't super current information).
>> >>
>> >> I seem to recall the hangup was that oVirt had no orchestration
>> >> capability for RBD comparable to OpenStack, and that CephFS wasn't
>> >> (yet?) viable for use as a "POSIX filesystem" oVirt storage domain.
>> >> Personally, I feel like Redhat is worried about competing with
>> >> themselves with GlusterFS versus CephFS and is choosing to focus on
>> >> Gluster as a filesystem, and Ceph as everything minus the filesystem.
>> >>
>> >> Which is a shame, as I'm a fan of both Ceph and oVirt and would love
>> >> to use my existing RHEV infrastructure to bring Ceph into my
>> >> environment.
>> >>
>> >>
>> >> On Mon, Jul 24, 2017 at 8:39 PM, Brady Deetz <bde...@gmail.com> wrote:
>> >> > I haven't seen much talk about direct integration with oVirt.
>> Obviously
>> >> > it
>> >> > kind of comes down to oVirt being interested in participating. But,
>> is
>> >> > the
>> >> > only hold-up getting development time toward an integration or is
>> there
>> >> > some
>> >> > kind of friction between the dev teams?
>> >> >
>> >> > _______________________________________________
>> >> > ceph-users mailing list
>> >> > ceph-users@lists.ceph.com
>> >> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> >> >
>> >>
>> >>
>> >>
>> >> --
>> >> ______________________________
>> >> Dino Yancey
>> >> 2GNT.com Admin
>> >
>> >
>> >
>> > _______________________________________________
>> > ceph-users mailing list
>> > ceph-users@lists.ceph.com
>> > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>> >
>>
>>
>>
>> --
>> Jason
>>
>
>
> _______________________________________________
> ceph-users mailing list
> ceph-users@lists.ceph.com
> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com
>
>
>
_______________________________________________
ceph-users mailing list
ceph-users@lists.ceph.com
http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com

Reply via email to