CA DRA: support priority-based preempting pods using DRA #7683
Labels
area/cluster-autoscaler
area/core-autoscaler
Denotes an issue that is related to the core autoscaler and is not specific to any provider.
kind/feature
Categorizes issue or PR as related to a new feature.
wg/device-management
Categorizes an issue or PR as relevant to WG Device Management.
Which component are you using?:
/area cluster-autoscaler
/area core-autoscaler
/wg device-management
Is your feature request designed to solve a problem? If so describe the problem this feature should solve.:
If CA sees an unschedulable pod waiting for scheduler preemption (with
nominatedNodeName
set), it adds the pod to thenominatedNodeName
in the snapshot without checking predicates, or even removing the preempted pod (so the Node can be effectively "overscheduled").For DRA autoscaling MVP, we still just force-add such a Pod to the snapshot without modifying its ResourceClaims. This means that CA doesn't see the Node's ResourceSlices as used and can just schedule another Pod to use them in the simulations. We need to fix this scenario for production use.
Describe the solution you'd like.:
Describe any alternative solutions you've considered.:
Keep force-adding the pods to the snapshot. This doesn't sound like a good idea for the reasons explained above.
Additional context.:
This is a part of Dynamic Resource Allocation (DRA) support in Cluster Autoscaler. An MVP of the support was implemented in #7530 (with the whole implementation tracked in kubernetes/kubernetes#118612). There are a number of post-MVP follow-ups to be addressed before DRA autoscaling is ready for production use - this is one of them.
The text was updated successfully, but these errors were encountered: