Вторник, 9 августа 2016, 17:43 +03:00 от Wido den Hollander <w...@42on.com>: > > >> Op 9 augustus 2016 om 16:36 schreef Александр Пивушков < p...@mail.ru >: >> >> >> > >> Hello dear community! >> >> >> I'm new to the Ceph and not long ago took up the theme of building >> >> >> clusters. >> >> >> Therefore it is very important to your opinion. >> >> >> It is necessary to create a cluster from 1.2 PB storage and very rapid >> >> >> access to data. Earlier disks of "Intel® SSD DC P3608 Series 1.6TB >> >> >> NVMe PCIe 3.0 x4 Solid State Drive" were used, their speed of all >> >> >> satisfies, but with increase of volume of storage, the price of such >> >> >> cluster very strongly grows and therefore there was an idea to use >> >> >> Ceph. >> >> > >> >> >You may want to tell us more about your environment, use case and in >> >> >particular what your clients are. >> >> >Large amounts of data usually means graphical or scientific data, >> >> >extremely high speed (IOPS) requirements usually mean database >> >> >like applications, which one is it, or is it a mix? >> >> >> >>This is a mixed project, with combined graphics and science. Project >> >>linking the vast array of image data. Like google MAP :) >> >> Previously, customers were Windows that are connected to powerful servers >> >> directly. >> >> Ceph cluster connected on FC to servers of the virtual machines is now >> >> planned. Virtualization - oVirt. >> > >> >Stop right there. oVirt, despite being from RedHat, doesn't really support >> >Ceph directly all that well, last I checked. >> >That is probably where you get the idea/need for FC from. >> > >> >If anyhow possible, you do NOT want another layer and protocol conversion >> >between Ceph and the VMs, like a FC gateway or iSCSI or NFS. >> > >> >So if you're free to choose your Virtualization platform, use KVM/qemu at >> >the bottom and something like Openstack, OpenNebula, ganeti, Pacemake with >> >KVM resource agents on top. >> oh, that's too bad ... >> I do not understand something... >> >> oVirt built on kvm >> https://www.ovirt.org/documentation/introduction/about-ovirt/ >> >> Ceph, such as support kvm >> http://docs.ceph.com/docs/master/architecture/ >> > >KVM is just the hypervisor. oVirt is a tool which controls KVM and it doesn't >have support for Ceph. That means that it can't pass down the proper arguments >to KVM to talk to RBD. > >> What could be the overhead costs and how big they are? >> >> >> I do not understand why oVirt bad, and the qemu in the Openstack, it's good. >> What can be read? >> > >Like I said above. oVirt and OpenStack both control KVM. OpenStack also knows >how to 'configure' KVM to use RBD, oVirt doesn't. > >Maybe Proxmox is a better solution in your case. No, I Openstack implement, would be to understand why! :) Why Openstack support for Ceph For example, in an oVirt running Centos 7. I have it mounted directory Ceph. Cluster Ceph installed on 13 physical servers. Do I understand correctly, the overhead is only because of the virtualized network card?
-- Александр Пивушков
_______________________________________________ ceph-users mailing list ceph-users@lists.ceph.com http://lists.ceph.com/listinfo.cgi/ceph-users-ceph.com