← Back to News

Kubernetes v1.36: More Drivers, New Features, and the Next Era of DRA

If you’ve struggled with allocating GPUs, TPUs, or other specialized hardware in Kubernetes, you’re not alone. Platform teams have long faced a frustrating reality: the standard resource request model (CPU and memory) doesn’t capture the complexity of modern hardware accelerators. Kubernetes v1.36 tackles this head-on with significant maturity improvements to Dynamic Resource Allocation (DRA), a feature that’s reshaping how teams manage specialized hardware at scale. This release marks a pivotal moment where DRA moves from experimental territory into something platform administrators can seriously consider for production workloads.

Dynamic Resource Allocation fundamentally changes the game by decoupling resource requests from the scheduler’s core logic. Instead of the kubelet trying to figure out GPU assignments on the fly, DRA introduces resource drivers—plugins that understand the nuances of specific hardware and can make intelligent allocation decisions. Here’s how it works: when a pod requests a specialized resource, DRA delegates the allocation to a dedicated driver that understands that hardware’s constraints and capabilities. This might involve checking GPU memory, determining which device offers the best performance characteristics, or even considering topology constraints between multiple accelerators. The scheduler handles pod placement, while the driver handles the specific hardware assignment. It’s a cleaner separation of concerns that scales much better than the previous node selector approach.

The practical implications are significant. Consider a machine learning platform where you’re running both training jobs (which need high-bandwidth GPU memory) and inference services (which prioritize low-latency access). With DRA, you can write drivers that understand these workload patterns and allocate hardware accordingly—something that’s nearly impossible with static kubelet plugins. In v1.36, the feature extends beyond accelerators to native resources like memory and CPU, meaning you can now apply this sophisticated allocation logic to any resource type. Teams deploying inference clusters, high-performance computing workloads, or multi-tenant GPU infrastructure will find new opportunities to optimize hardware utilization without resorting to workarounds or custom controllers.

The maturity improvements in this release—including new driver implementations, better observability, and stabilized APIs—suggest DRA is ready for teams willing to invest in understanding how it fits their infrastructure. If your organization is managing heterogeneous hardware or hitting the limits of node selectors and taints/tolerations, this is worth investigating. The learning curve exists, but the flexibility DRA provides makes it a worthwhile investment for platform teams serious about resource optimization.

Source
↗ Kubernetes Blog