
AI infrastructure has had an awkward teenage phase: fast growth, expensive habits, and a tendency to break in public. Over the past couple of years, organizations have raced from “we have a neat demo” to “this model is in production and it’s now the CEO’s favorite app.” The problem: the infrastructure layer underneath AI workloads—especially on Kubernetes—has often been stitched together from bespoke YAML, vendor-specific GPU magic, and tribal knowledge passed along like an ancient spell.
That’s why a new initiative from the Cloud Native Computing Foundation (CNCF) matters: the Certified Kubernetes AI Conformance Program, announced at KubeCon + CloudNativeCon North America in Atlanta on November 11, 2025. The headline promise is simple: create a community-defined baseline for what a Kubernetes-based platform must provide to run AI/ML workloads reliably, consistently, and portably. citeturn2view0
This article takes as its starting point the Giant Swarm blog post “Infrastructure for AI is finally getting a standard” by Puja Abbassi (published November 11, 2025). citeturn0view0 We’ll go beyond the announcement to unpack what “AI conformance” means in practice, why it’s showing up now, how it relates to the existing Kubernetes conformance model, and what platform teams and security folks should do next.
Why AI on Kubernetes has been… complicated
Kubernetes won the container orchestration wars by doing two things well: standardizing the “how” of running workloads, and building a massive ecosystem around that standard. For classic web apps, “a pod is a pod” is mostly true. For AI, “a pod is a pod” is true in the same way “a car is a car” is true—technically correct, but not helpful when you’re towing a boat uphill in winter.
AI workloads behave differently than typical microservices
AI/ML workloads are often:
- GPU/accelerator dependent and sensitive to device availability and driver versions.
- Batchy and spiky: training runs for hours/days, then stops; inference can be quiet, then suddenly you’re trending on social media.
- Distributed: multi-node training needs coordinated scheduling, networking behavior, and storage throughput.
- Data hungry: storage and I/O patterns matter as much as CPU limits.
All of this produces a predictable outcome: organizations build custom platform extensions, custom GPU node pools, custom schedulers, custom admission policies—and then discover that “custom” has a recurring subscription cost called “your on-call rotation.”
Fragmentation is the tax we’ve been paying
Without a shared baseline, vendors can honestly say their platform is “AI-ready,” while meaning wildly different things. One platform might support GPU device plugins but not handle distributed scheduling well. Another might offer a polished inference experience but require proprietary components. Teams end up comparing stacks by vibes, not verifiable capabilities.
The CNCF’s bet is that conformance—done the Kubernetes way—can reduce that uncertainty.
What the Kubernetes AI Conformance Program is (and isn’t)
The CNCF describes the program as a community-led effort to define and validate standards for running AI workloads reliably and consistently on Kubernetes. citeturn2view0 The key word is validate: this isn’t just a marketing badge; it’s intended to be something platforms can test against and demonstrate compliance with—similar in spirit to the long-running Certified Kubernetes Conformance Program. citeturn1search5
It’s a baseline, not a full AI platform
Conformance programs don’t guarantee you’ll like a product. They don’t guarantee the UI is pleasant or that your vendor’s support is fast. What they do is narrow the universe of surprises. If a platform is AI-conformant, you should be able to assume a certain minimum set of Kubernetes-based capabilities exists and behaves predictably.
Developed in the open, with public artifacts
The conformance work is being developed openly on GitHub in the CNCF repository, inviting vendors and end users to contribute and submit results for certification. citeturn1search2 That matters because AI infrastructure is moving fast; if requirements are locked in a private PDF, they’ll be obsolete before they’re printed.
Tied to a Kubernetes community working group
The work is guided by the AI Conformance Working Group, which is also operating in the open and documents how requirements evolve and how tests should be designed. citeturn1search6 This “requirements plus tests” loop is crucial: standards without tests are aspirations. Tests without standards are chaos with a CI pipeline.
Why now: production AI has forced the issue
CNCF’s announcement aligns with a broader trend: AI is no longer a research curiosity for many organizations; it’s business infrastructure. The CNCF points to Linux Foundation research indicating that 82% of organizations are already building custom AI solutions and 58% use Kubernetes to support those workloads. citeturn2view0turn3search7 If you’re a platform team, you don’t need a research report to know this—your GPU budget already told you.
But the data reinforces an important point: as AI adoption scales, fragmentation becomes expensive. Different teams inside the same enterprise may end up with incompatible “AI platforms” that can’t share governance, observability, or cost controls. A conformance baseline is one way to keep the ecosystem from splintering into incompatible islands.
Giant Swarm’s angle: standards as a platform engineering strategy
In the Giant Swarm post, Puja Abbassi frames the program as a move away from “custom configs, opaque tooling, and vendor-specific lock-in” toward a shared baseline. citeturn0view0 Giant Swarm also states it is among the first platforms certified under the new program. citeturn0view0turn2view0
There’s a subtext here that platform engineers will appreciate: AI/ML platforms are increasingly treated as an extension of internal developer platforms (IDPs). The same patterns apply—self-service, governance, fleet management, security baselines, observability—except now the workloads are heavier, pricier, and more sensitive to “oops.”
What “conformance” tends to standardize in Kubernetes-land
The CNCF announcement says the program defines a minimum set of capabilities and configurations required to run widely used AI/ML frameworks on Kubernetes. citeturn2view0 While the exact test suite will evolve, we can infer the kinds of things that matter by looking at the real-world building blocks teams already depend on.
1) GPUs and accelerators: device plugins, scheduling, and sharing
On Kubernetes, hardware accelerators are typically exposed through the device plugin framework, where vendors advertise resources (for example nvidia.com/gpu) to the kubelet so pods can request them. citeturn5search2 This is foundational—but not sufficient.
In real clusters, GPU utilization is often the difference between “this platform is viable” and “we’re setting money on fire.” That’s why GPU sharing and smarter scheduling are hot topics. NVIDIA’s GPU Operator, for example, supports GPU time-slicing to oversubscribe GPUs so multiple pods can share compute time, with clear trade-offs compared to hardware-partitioning features like MIG. citeturn3search1
Conformance doesn’t need to dictate NVIDIA’s implementation (or AMD’s, or Intel’s). But it can set expectations like:
- How accelerators are advertised and requested.
- Whether the platform supports predictable scheduling behavior for GPU workloads.
- Whether basic GPU observability hooks exist.
2) Job-level workflows: queueing, quotas, and fairness
AI training workloads are often “jobs” rather than “services,” and they compete for scarce GPUs across teams. That’s where queueing and quota management enter the picture. Kubernetes itself has primitives, but the ecosystem has added purpose-built tools.
Kueue is one such Kubernetes-native system that manages quotas and decides when jobs should wait, be admitted, or be preempted. It’s explicitly designed to sit on top of Kubernetes rather than replace core components like the scheduler or cluster autoscaler. citeturn5search1
From a standards perspective, this matters because “AI platform” often means “multi-tenant GPU governance.” If conformance can nudge platforms toward consistent patterns here—queues, quotas, admission checks—it reduces the amount of bespoke glue each organization has to invent.
3) Distributed training: controllers, CRDs, and gang scheduling
Distributed training on Kubernetes is typically orchestrated via controllers and CRDs. Kubeflow’s Training Operator (and its newer Trainer efforts) describes itself as a Kubernetes-native project for scalable distributed training across frameworks like PyTorch and TensorFlow, and notes integration with advanced scheduling systems such as Kueue and Volcano. citeturn4search0
Meanwhile, batch schedulers like Volcano focus on capabilities like gang scheduling and queue resource management to support AI and data workloads. citeturn4search1
Again, the conformance program probably won’t require “thou shalt install Volcano.” But it can standardize what a conformant platform must support so that these tools work consistently across distributions.
4) Inference serving: standard APIs and operational behavior
Inference is where AI meets product. It’s also where “works on my laptop” goes to die under real traffic.
KServe, a CNCF incubating project, positions itself as a standardized inference platform for predictive and generative AI on Kubernetes. citeturn4search2 It’s explicitly trying to make inference behavior more consistent across frameworks, including with production features such as scaling and GPU support.
If you’re a platform team building a standard AI runway internally, conformance plus standard inference tooling is the dream: fewer snowflakes, more repeatability.
So what changes for platform teams?
Standards feel abstract until you’re the one trying to write a runbook at 3:00 a.m. Here are the pragmatic ways an AI conformance program can change daily life.
Procurement gets less subjective
Many organizations are shopping for “AI platforms” without an objective yardstick. Conformance doesn’t eliminate due diligence, but it can remove a whole category of questions: “Does it implement the expected APIs and baseline behaviors?” Instead of arguing about marketing claims, teams can ask: “Show me your conformance results.”
Multi-cloud and hybrid bets become less risky
AI infrastructure decisions tend to calcify quickly because migrating AI stacks is painful. A conformance baseline can improve portability: you can move workloads (or at least your expectations) across certified environments with fewer unpleasant surprises.
Security and governance can anchor on shared assumptions
Security teams love standards when they reduce ambiguity. If the program’s baseline includes repeatable patterns for workload isolation, storage handling, and GPU exposure, it becomes easier to build policy templates and automated checks that apply across vendors.
This is especially relevant in a world where organizations are increasingly concerned with control over data and AI capabilities—one of the drivers behind the broader “sovereign AI” conversation. Linux Foundation commentary on sovereign AI highlights motivations such as data control and national security, and emphasizes open source and open standards as foundational. citeturn3search7
Operational consistency improves (slowly, then suddenly)
The Kubernetes ecosystem has shown that conformance can work. The original Kubernetes conformance program helped ensure consistent core behavior across 100+ distributions and platforms. citeturn2view0 The AI conformance effort aims to do something similar for AI workloads—creating a shared baseline so “AI on Kubernetes” isn’t a completely different sport depending on where you run it.
What doesn’t change (and what you still need to design)
Let’s not oversell it: conformance does not magically solve AI infrastructure. You still have to make real engineering decisions. The certification won’t pick your model serving framework or decide whether you should run training on-prem or in the cloud.
You still need to manage GPU economics
Even with conformance, the GPU question remains: do you want exclusive access per pod, or do you want sharing? Time-slicing can raise utilization but changes the performance isolation story, and documentation is explicit that time-sliced “replicas” don’t provide the memory/fault isolation that MIG does. citeturn3search1
Conformance can ensure the basics work. It can’t eliminate the need for capacity planning and performance testing under your workloads.
You still need an opinionated internal platform experience
Most organizations don’t want “raw Kubernetes” for AI users. They want templates, guardrails, curated images, standardized pipelines, and a sane path from experiment to production. Conformance helps make the underlying infrastructure more predictable, but you still have to build (or buy) the developer experience layer.
Case-study style scenario: the three stages of enterprise AI pain
To make this concrete, here’s a common progression I’ve seen (with the names changed to protect the innocent and the negligent).
Stage 1: “We have GPUs and a dream”
A team spins up a Kubernetes cluster, installs a GPU device plugin, and deploys a notebook or a training job. It works. Everyone is happy. The cluster is mostly idle, but that’s a future problem.
Stage 2: “Why is everything bespoke?”
More teams arrive. Now you need quotas, queueing, and policies. Somebody introduces Kueue to manage fair sharing and admissions. citeturn5search1 Another team needs better batch scheduling and introduces Volcano. citeturn4search1 Suddenly, you’ve built a mini economy where GPUs are currency and YAML is the tax code.
Stage 3: “We need a platform, not a pile”
At this point, leadership wants reliability and predictability. The platform team wants standardization. This is where conformance programs are useful: they provide a baseline for infrastructure capabilities so internal tooling isn’t built on shifting sand.
How AI conformance relates to other Kubernetes standards
Conformance programs are not new in Kubernetes. The most important precedent is the Certified Kubernetes Conformance Program, with established submission processes and tooling approaches (for example via Sonobuoy and similar runners) documented in the CNCF’s conformance repositories. citeturn1search5
The AI conformance effort is also being developed in a similar public way, with vendors invited to submit conformance testing results for review and certification. citeturn1search2 The pattern is clear: define expectations, define tests, publish results, repeat.
Expert perspective: the real value is reducing “unknown unknowns”
CNCF CTO Chris Aniszczyk framed the initiative around ensuring AI workloads behave predictably across environments and building on the community-driven conformance process used for Kubernetes itself. citeturn2view0
From the vendor side, the CNCF announcement includes supporting quotes from major players (AWS, Google Cloud, Microsoft, Red Hat, and others) emphasizing interoperability and portability—classic Kubernetes values, now applied to AI infrastructure. citeturn2view0
And from Giant Swarm’s perspective, Puja Abbassi calls it one of the most timely standardization efforts of the last decade—an interesting claim, but one that resonates with anyone who has watched teams reinvent “AI platform basics” in parallel. citeturn2view0
What you should do next (if you run AI on Kubernetes)
1) If you’re selecting a platform, add conformance to your checklist
When evaluating Kubernetes-based AI platforms (managed services, on-prem distributions, GPU clouds), ask vendors whether they are certified under the Kubernetes AI Conformance Program and request the specifics of what version/profile they certified against. If they aren’t certified, ask whether they plan to be—and what gaps they see.
2) If you’re building your own platform, align with the ecosystem early
Homegrown platforms are often inevitable, but they don’t have to be idiosyncratic. Track the conformance requirements as they evolve, and use them as a design constraint. It’s easier to align early than retrofit after you’ve shipped an internal platform used by 20 teams.
3) Treat GPUs like a first-class, governed resource
Accelerators aren’t just “bigger CPU.” They demand governance, observability, and scheduling policies that reflect their cost and scarcity. Learn the mechanics: device plugins, resource naming, and how scheduling decisions are made via the Kubernetes scheduler framework. citeturn5search2turn5search4
4) Be explicit about sharing trade-offs
If you adopt GPU time-slicing (or any other sharing strategy), document its implications for performance isolation, monitoring, and fairness. NVIDIA’s documentation is very clear about the trade-offs between time-slicing and MIG-style isolation. citeturn3search1 Your platform policies should be equally clear.
The bigger picture: AI platforms are becoming cloud native platforms
For years, “cloud native” meant microservices, containers, and service meshes. AI changes the workload shape, but it doesn’t change the core operational needs: repeatability, composability, policy, and portability. The Kubernetes AI Conformance Program is essentially the ecosystem admitting that AI infrastructure is now a first-class citizen of cloud native—not a weird add-on you duct-tape to a cluster when the data science team starts asking questions.
And yes, it’s mildly amusing that the industry had to reinvent the idea of “standards” after spending a decade praising standards. But in tech, we only accept a concept once it arrives with a logo and a certification badge.
Sources
- Giant Swarm: “Infrastructure for AI is finally getting a standard” (Puja Abbassi, Nov 11, 2025)
- CNCF announcement: “CNCF Launches Certified Kubernetes AI Conformance Program…” (Nov 11, 2025)
- cncf/ai-conformance (GitHub)
- kubernetes-sigs/wg-ai-conformance (GitHub)
- Linux Foundation: “The Essential Role of Open Source in Sovereign AI”
- Kubernetes Documentation: Device Plugins
- Kubernetes Documentation: Scheduling Framework
- NVIDIA GPU Operator Documentation: Time-Slicing GPUs in Kubernetes
- Kueue Documentation: Overview
- Kubeflow Documentation: Training Operator Overview
- Volcano Documentation: Unified Scheduling
- KServe: Project Site
Bas Dorland, Technology Journalist & Founder of dorland.org