Kubernetes v1.35: The AI Conformance Leap
By Dillip Chowdary • Mar 24, 2026
For years, running **AI** workloads on **Kubernetes** felt like forcing a square peg into a round hole. While K8s was designed for stateless web microservices, AI models require stateful, high-throughput, and hardware-dependent environments. Today, the **CNCF** (Cloud Native Computing Foundation) has officially bridged that gap with the release of **Kubernetes v1.35** and the accompanying **AI Conformance v1.35** standard.
Dynamic Resource Allocation (DRA) goes GA
The headline feature of v1.35 is the General Availability of **Dynamic Resource Allocation** (DRA). Before DRA, managing GPUs in Kubernetes was limited to simple integer requests (e.g., `nvidia.com/gpu: 1`). This led to massive resource fragmentation and underutilization. DRA allows for much more granular control, enabling pods to request specific hardware features like **NVLink** topologies, fractional GPU slicing (MIG/MPS), and even specific memory bandwidth profiles.
Technically, DRA introduces a new resource claim model that decouples device discovery from the core kubelet. This allows vendors to write specialized resource drivers that can handle the complex "gang scheduling" required for large-scale **distributed training** jobs. If a pod needs 8 GPUs connected via a specific high-speed interconnect, DRA ensures that the scheduler only places that pod on a node that meets the exact topological requirements.
The AI Conformance Program
Alongside the software release, the CNCF has launched the **AI Conformance Program**. Similar to the standard Kubernetes conformance tests, this program ensures that cloud providers (AWS, Google, Azure) provide a consistent experience for AI developers. A "CNCF AI Certified" cluster must support specific APIs for **VRAM** isolation, automated driver lifecycle management, and native integration with the **Kueue** job queueing system.
This is a major win for **Sovereign Cloud** providers. By adhering to the AI Conformance standard, smaller providers can offer a "Kubernetes-native" AI experience that is compatible with the same **Helm** charts and **Kueflow** pipelines used in the major public clouds. This reduces vendor lock-in and allows enterprises to move their training workloads to the most cost-effective region without rewriting their infrastructure code.
Technical Insight: Multi-Cluster AI Mesh
Kubernetes v1.35 also introduces alpha support for Multi-Cluster AI Mesh. This allows a single training job to span across multiple physical clusters, utilizing Submariner or Cilium ClusterMesh to handle the high-speed cross-cluster networking required for gradient synchronization.
Optimizing for Inference: Sidecar Containers and WASM
While training gets the headlines, v1.35 also brings significant improvements for **AI Inference**. The new **Inference Sidecar** pattern allows for model weights to be loaded from a shared volume into a sidecar container, reducing pod startup times (Cold Starts) by up to 80%. There is also improved support for **WebAssembly** (Wasm) runtimes, which are increasingly being used to run lightweight models at the edge with minimal overhead.
With v1.35, Kubernetes has officially transitioned from a container orchestrator to an **AI Operating System**. The ecosystem is now focused on "Day 2" operations: observability for GPU metrics via **Prometheus**, automated scaling based on inference latency, and secure multi-tenancy for shared AI clusters. The road to **Artificial General Intelligence** (AGI) is being paved with YAML.