Improvements: Cluster Autoscaling with GPU-sharing Pods & Support for Scheduling Gates #125
EkinKarabulut
announced in
Announcements
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi everyone,
We are excited to share that there are some significant improvements to KAI Scheduler:
Support for Cluster Autoscaling with GPU-Sharing Pods via node-scale-adjuster (see docs)
Introducing
node-scale-adjuster
, which enables cluster autoscalers like Karpenter to work with pods using GPU sharing.Problem
Cluster autoscalers rely on pending pods with resource requests to trigger node provisioning. In KAI, GPU-sharing pods define GPU needs in annotations, not resources.requests, making them invisible to the autoscaler. As a result, no scale-up was triggered (see issue #111)
Solution
node-scale-adjuster
watches for unschedulable GPU-sharing pods and creates temporary utility pods that request full GPUs via standard resources.requests. These utility pods allow the autoscaler to react as expected.Behavior
For more details, refer to the documentation.
Kubernetes Scheduling Gates Support
KAI now respects Kubernetes pod scheduling gates, which allow pending pods to delay scheduling until certain conditions are met.
Behavior with PodGroups
minMember
checks in a PodGroup.minMember = 4
and 4 pending pods, where 1 is gated, will not be scheduled.minMember
, those pods are scheduled.We’d love your feedback on these updates! 🚀
Feel free to:
Beta Was this translation helpful? Give feedback.
All reactions