On Mon, Feb 21, 2022 at 12:27 PM Thomas Hoberg <[email protected]> wrote:

> That's exactly the direction I originally understood oVirt would go, with
> the ability to run VMs and container side-by-side on the bare metal or
> nested with containers inside VMs for stronger resource or security
> isolation and network virtualization. To me it sounded especially
> attractive with an HCI underpinning so you could deploy it also in the
> field with small 3 node clusters.
>

I think in general a big part of the industry is going down the path of
moving most things behind the k8s API/resource model. This means different
things for different companies. For instance vmware keeps its traditional
virt-stack, adding additional k8s apis in front of it, and crossing the
bridges to k8s clusters behind the scenes to get a unified view, while
other parts are choosing k8s (be it vanilla k8s, openshift, harvester, ...)
and then take for instance KubeVirt to deploy additional k8s clusters on
top of it, unifying the stack this way.

It is definitely true that k8s works significantly different to other
solutions like oVirt or OpenStack, but once you get into it, I think one
would be surprised how simple the architecture of k8s actually is, and also
how little resources core k8s actually takes.

Having said that, as an ex oVirt engineer I would be glad to see oVirt
continue to thrive. The simplicity of oVirt was always appealing to me.

Best regards,
Roman



>
> But combining all those features evidently comes at too high a cost for
> all the integration and the customer base is either too small or too poor:
> the cloud players are all out on making sure you no longer run any hardware
> and then it's really just about pushing your applications there as cloud
> native or "IaaS" compatible as needed.
>
> E.g. I don't see PCI pass-through coming to kubevirt to enable GPU use,
> because it ties the machine to a specific host and goes against the grain
> of K8 as I understand it.
>
> Memory overcommit is quite funny, really, because it's the same issue as
> the original virtual memory: essentially you lie to your consumer about the
> resources available and then swap pages forth and back in an attempt to
> make all your consumers happy. It was processes for virtual memory, it's
> VMs now for the hypervisor and in both cases it's about the consumer and
> the provider not continously negotiating for the resources they need and
> the price they are willing to pay.
>
> That negotiation is always better at the highest level of abstraction, the
> application itself, which why implementing it at the lower levels (e.g.
> VMs) becomes less useful and needed.
>
> And then there is technology like CXL which essentially turns RAM in to a
> fabric and your local CPU will just get RAM from another piece of hardware
> when your application needs more RAM and is willing to pay the premium
> something will charge for it.
>
> With that type of hardware much of what hypervisors used to do goes into
> DPUs/IPUs and CPUs are just running applications making hypercalls. The
> kernel is just there to bootstrap.
>
> Not sure we'll see that type of hardware at home or in the edge, though...
> _______________________________________________
> Users mailing list -- [email protected]
> To unsubscribe send an email to [email protected]
> Privacy Statement: https://www.ovirt.org/privacy-policy.html
> oVirt Code of Conduct:
> https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
> https://lists.ovirt.org/archives/list/[email protected]/message/PC5SDUMCPUEHQCE6SCMITTQWK5QKGMWT/
>
_______________________________________________
Users mailing list -- [email protected]
To unsubscribe send an email to [email protected]
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/[email protected]/message/4IOI2KD7MNKFS3QQ5W4RVYPKBRZIDF5L/

Reply via email to