Hello, On Tue, 9 Aug 2016 14:15:59 -0400 Jeff Bailey wrote:
> > > On 8/9/2016 10:43 AM, Wido den Hollander wrote: > > > >> Op 9 augustus 2016 om 16:36 schreef Александр Пивушков <p...@mail.ru>: > >> > >> > >> > >> Hello dear community! > >>>>>> I'm new to the Ceph and not long ago took up the theme of building > >>>>>> clusters. > >>>>>> Therefore it is very important to your opinion. > >>>>>> It is necessary to create a cluster from 1.2 PB storage and very rapid > >>>>>> access to data. Earlier disks of "Intel® SSD DC P3608 Series 1.6TB > >>>>>> NVMe PCIe 3.0 x4 Solid State Drive" were used, their speed of all > >>>>>> satisfies, but with increase of volume of storage, the price of such > >>>>>> cluster very strongly grows and therefore there was an idea to use > >>>>>> Ceph. > >>>>> > >>>>> You may want to tell us more about your environment, use case and in > >>>>> particular what your clients are. > >>>>> Large amounts of data usually means graphical or scientific data, > >>>>> extremely high speed (IOPS) requirements usually mean database > >>>>> like applications, which one is it, or is it a mix? > >>>> > >>>> This is a mixed project, with combined graphics and science. Project > >>>> linking the vast array of image data. Like google MAP :) > >>>> Previously, customers were Windows that are connected to powerful > >>>> servers directly. > >>>> Ceph cluster connected on FC to servers of the virtual machines is now > >>>> planned. Virtualization - oVirt. > >>> > >>> Stop right there. oVirt, despite being from RedHat, doesn't really support > >>> Ceph directly all that well, last I checked. > >>> That is probably where you get the idea/need for FC from. > >>> > >>> If anyhow possible, you do NOT want another layer and protocol conversion > >>> between Ceph and the VMs, like a FC gateway or iSCSI or NFS. > >>> > >>> So if you're free to choose your Virtualization platform, use KVM/qemu at > >>> the bottom and something like Openstack, OpenNebula, ganeti, Pacemake with > >>> KVM resource agents on top. > >> oh, that's too bad ... > >> I do not understand something... > >> > >> oVirt built on kvm > >> https://www.ovirt.org/documentation/introduction/about-ovirt/ > >> > >> Ceph, such as support kvm > >> http://docs.ceph.com/docs/master/architecture/ > >> > > > > KVM is just the hypervisor. oVirt is a tool which controls KVM and it > > doesn't have support for Ceph. That means that it can't pass down the > > proper arguments to KVM to talk to RBD. > > > >> What could be the overhead costs and how big they are? > >> > >> > >> I do not understand why oVirt bad, and the qemu in the Openstack, it's > >> good. > >> What can be read? > >> > > > > Like I said above. oVirt and OpenStack both control KVM. OpenStack also > > knows how to 'configure' KVM to use RBD, oVirt doesn't. > > > > Maybe Proxmox is a better solution in your case. > > > > oVirt can use ceph through cinder. It doesn't currently provide all the > functionality of > other oVirt storage domains but it does work. > Well, I saw this before I gave my answer: http://www.ovirt.org/develop/release-management/features/storage/cinder-integration/ And based on that I would say oVirt is not a good fit for Ceph at this time. Even less so than OpenNebula, which currently needs an additional shared network FS or hacks to allow live migration with RBD. Christian > > Wido > > > >> > >> -- > >> Александр Пивушков_______________________________________________ > >> ceph-users mailing list > >> ceph-users@lists.ceph.com > >> http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > _______________________________________________ > > ceph-users mailing list > > ceph-users@lists.ceph.com > > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com > > > _______________________________________________ > ceph-users mailing list > ceph-users@lists.ceph.com > http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com -- Christian Balzer Network/Systems Engineer ch...@gol.com Global OnLine Japan/Rakuten Communications http://www.gol.com/ _______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com