On 4/1/25 03:50, DERUMIER, Alexandre wrote:
my 2cents, but everybody in the industry is calling this
affinity/antiafifnity (vmware, nutanix, hyperv, openstack, ...).
More precisely, vm affinity rules (vm<->vm) vs node affinity rules
(vm->node , the current HA group)
Personnally I don't care, it's just a name ^_^ .
But I have a lot of customers asking about "does proxmox support
affinity/anti-affinity". and if they are doing their own research, they
will think that it doesnt exist.
(or at minimum, write somewhere in the doc something like "aka vm
affinity" or in commercial presentation ^_^)
I see your point and also called it affinity/anti-affinity before, but
if we go for the HA Rules route here, it'd be really neat to have
"Location Rules" and "Colocation Rules" in the end to coexist and
clearly show the distinction between them, as both are affinity rules at
least for me.
I'd definitely make sure that it is clear from the release notes and
documentation, that this adds the feature to assign affinity between
services, but let's wait for some other comments on this ;).
On 4/1/25 03:50, DERUMIER, Alexandre wrote:
More serious question : Don't have read yet all the code, but how does
it play with the current topsis placement algorithm ?
I currently implemented the colocation rules to put a constraint on
which nodes the manager can select from for the to-be-migrated service.
So if users use the static load scheduler (and the basic / service count
scheduler for that matter too), the colocation rules just make sure that
no recovery node is selected, which contradicts the colocation rules. So
the TOPSIS algorithm isn't changed at all.
There are two things that should/could be changed in the future (besides
the many future ideas that I pointed out already), which are
- (1) the schedulers will still consider all online nodes, i.e. even
though HA groups and/or colocation rules restrict the allowed nodes in
the end, the calculation is done for all nodes which could be
significant for larger clusters, and
- (2) the service (generally) are currently recovered one-by-one in a
best-fit fashion, i.e. there's no order on the service's needed
resources, etc. There could be some edge cases (e.g. think about a
failing node with a bunch of service to be kept together; these should
now be migrated to the same node, if possible, or put them on the
minimum amount of nodes), where the algorithm could find better
solutions if it either orders the to-be-recovered services, and/or the
utilization scheduler has knowledge about the 'keep together'
colocations and considers these (and all subsets) as a single service.
For the latter, the complexity explodes a bit and is harder to test for,
which is why I've gone for the current implementation, as it also
reduces the burden on users to think about what could happen with a
specific set of rules and already allows the notion of MUST/SHOULD. This
gives enough flexibility to improve the decision making of the scheduler
in the future.
On 4/1/25 03:50, DERUMIER, Alexandre wrote:
Small feature request from students && customers: they are a lot
asking to be able to use vm tags in the colocation/affinity
Good idea! We were thinking about this too and I forgot to add it to the
list, thanks for bringing it up again!
Yes, the idea would be to make pools and tags available as selectors for
rules here, so that the changes can be made rather dynamic by just
adding a tag to a service.
The only thing we have to consider here is that HA rules have some
verification phase and invalid rules will be dropped or modified to make
them applicable. Also these external changes must be identified somehow
in the HA stack, as I want to keep the amount of runs through the
verification code to a minimum, i.e. only when the configuration is
changed by the user. But that will be a discussion for another series ;).
_______________________________________________
pve-devel mailing list
pve-devel@lists.proxmox.com
https://lists.proxmox.com/cgi-bin/mailman/listinfo/pve-devel