
For years, “AI infrastructure” has been an oddly specific form of chaos: expensive GPUs duct-taped to general-purpose clusters, YAML sprinkled with hopeful comments, and a shadowy layer of vendor-specific magic that only one engineer understands (and they’re currently “between opportunities”).
On November 11, 2025, at KubeCon + CloudNativeCon North America in Atlanta, the Cloud Native Computing Foundation (CNCF) announced something unusually practical for the AI world: a real, community-defined standard for running AI workloads on Kubernetes. It’s called the Certified Kubernetes AI Conformance Program. citeturn2view0
This article expands on the Giant Swarm blog post “Infrastructure for AI is finally getting a standard” by Puja Abbassi, published on November 11, 2025, and digs into what the standard is, why it matters, and what it changes for platform teams, AI engineers, and anyone who has ever tried to explain GPU scheduling to a finance department. citeturn0view0turn1search1
Wait, a standard for AI infrastructure? What exactly got standardized?
The CNCF program is not trying to standardize “AI” in the philosophical sense (thankfully). It’s standardizing the cluster capabilities needed to run common AI/ML workloads on Kubernetes reliably and consistently across environments.
In CNCF’s announcement, the goal is described as defining and validating a minimum set of capabilities and configurations required for AI workloads on Kubernetes, so enterprises can deploy with more confidence and vendors have a shared compatibility baseline. citeturn2view0
In the GitHub repository that backs the program, the project frames it in very Kubernetes terms: if your AI application works on one conformant platform, it should work on another with fewer “works on my cluster” surprises. citeturn4view0turn6view0
Also, the program is explicitly scoped to common workload categories:
- Training (including distributed training requiring accelerators and predictable scheduling)
- Inference (serving models/LLMs with latency, routing, and scaling requirements)
- Agentic workloads (multi-step, longer-running workflows)
That scoping matters because it’s how the program avoids becoming a “standard for everything,” which is how standards die.
Why this is happening now: AI moved into production faster than platforms did
Giant Swarm’s Puja Abbassi describes the past year as a rapid shift from experimental models to production copilots and customer-facing AI, while the infrastructure layer lagged behind and became a patchwork of custom configurations and vendor lock-in. citeturn0view0turn1search1
CNCF is basically saying: yes, we see the same mess, and we’ve run this movie before.
The CNCF CTO Chris Aniszczyk (who has probably heard “but our use case is special” in every possible accent) positioned the program as a way to make AI workloads behave predictably across environments, borrowing from Kubernetes’ existing conformance model that helped create a consistent ecosystem across 100+ Kubernetes distributions and platforms. citeturn2view0
This “borrow from what worked” approach is underrated. Kubernetes didn’t win because it was the only orchestration system; it won because there was an interoperable ecosystem and enough standards discipline that tools and workloads could travel.
The numbers behind the urgency (and why your cluster is now an AI cluster)
CNCF points to Linux Foundation Research on Sovereign AI: 82% of organizations are already building custom AI solutions, and 58% are using Kubernetes to support those workloads. citeturn2view0
That’s a remarkable statistic for something that, not long ago, lived mainly in notebooks and academic papers. Whether the organization is training models or mostly deploying pre-trained ones, Kubernetes is where the operational reality is converging.
CNCF doubled down on that theme in its January 20, 2026 survey announcement, stating Kubernetes has become the de facto “operating system” for AI, with 82% of container users running Kubernetes in production (a separate but related metric). citeturn1search4
What “AI conformance” actually checks (and what it doesn’t)
The program’s public materials make two things clear:
- It is technical conformance, not a governance or ethics badge.
- It builds on existing Kubernetes conformance; you can’t skip the basics and jump straight to “AI-ready.” citeturn4view0
Kubermatic, which announced it is among the first platforms certified (for Kubernetes v1.34), explicitly contrasts Kubernetes AI Conformance with management standards like ISO/IEC 42001, positioning AI Conformance as a purely technical baseline for AI/ML workloads on Kubernetes. citeturn11search5
From CNCF’s launch announcement and related coverage, the standard focuses on baseline requirements around things like accelerators (GPUs), storage, networking, and scheduling—exactly the places where AI workloads break “normal” platform assumptions. citeturn2view0turn9search3
What conformance is trying to prevent
If you’ve ever migrated a Kubernetes workload between distributions, you already know the pain points: subtle differences in CNI behavior, storage class defaults, admission controllers, or API enablement can turn a “portable workload” into a late-night incident.
AI workloads amplify this problem because they stress clusters in unusual ways—accelerators, bursty traffic, strict isolation—and because the surrounding tooling (operators, GPU plugins, inference gateways) is still rapidly evolving. citeturn4view0
In other words: the goal is not just portability in principle; it’s portability under pressure.
How certification works today: self-assessment now, automation later
The Kubernetes AI Conformance program is currently implemented via a structured submission process. Platform vendors fill out a conformance checklist and provide public evidence (docs, test results, etc.) as a pull request. CNCF reviews submissions, typically within 10 business days. citeturn4view0turn6view0
Two details matter for practitioners reading this:
- Certification is valid for one year and must be renewed. citeturn6view0
- Certification is per product and per configuration (for example, cloud vs air-gapped). citeturn6view0
Also: the repo notes that today, certification is based on self-assessment, and that automated conformance tests are planned for 2026. citeturn6view0
That’s not a weakness as much as a reality of timing. Kubernetes itself didn’t start with the perfectly automated ecosystem either; it matured into it. But it does mean buyers should read “certified” as “certified via documented evidence and CNCF review,” not as “a fully automated, exhaustive test suite has validated every edge case.”
Giant Swarm’s angle: platform engineering meets AI reality
Giant Swarm’s post is not a generic “standards are good” cheerleading piece; it’s written from a platform engineering perspective, where AI workloads are just another high-demand tenant—except the tenant has GPUs, distributed jobs, and a habit of eating storage bandwidth.
Abbassi highlights infrastructure needs like GPU sharing, distributed scheduling, and governance as practical requirements driven by generative AI workloads. citeturn0view0turn1search1
She also points out that Giant Swarm has been integrating those capabilities into its Kubernetes platform over years: GPU-aware scheduling, storage tuning, tailored security policies, and observability for model pipelines. citeturn0view0turn1search1
And crucially, Giant Swarm positions its work as “betting on standards and helping build them,” referencing other ecosystem efforts like CIS Benchmarks for Kubernetes and Cluster API. citeturn0view0turn1search1
Why this matters for “AI platform teams”
In 2026, many organizations are creating AI platform teams—sometimes as a sub-function of platform engineering, sometimes as a parallel organization. Either way, you end up needing a stable substrate for:
- Provisioning and sharing accelerators
- Running distributed jobs with predictable networking behavior
- Handling large datasets and model artifacts (storage throughput + lifecycle)
- Building guardrails: policies, quotas, tenant isolation, and auditability
A conformance baseline is a way to stop rebuilding the same infrastructure assumptions from scratch each time you add a cluster, a cloud, or a vendor.
The vendor list tells you what’s coming: Kubernetes AI is multi-cloud by default
The CNCF announcement includes supporting quotes from major platform providers. It’s not subtle: this is intended to be a cross-cloud baseline.
Examples from the CNCF launch materials include AWS (Amazon EKS), Google Cloud (GKE), Microsoft Azure, VMware (vSphere Kubernetes Service), CoreWeave, Oracle, and others—each framing conformance as a portability and interoperability move. citeturn2view0
Oracle’s OCI Kubernetes Engine (OKE) blog post similarly notes that OKE is among the first group certified under the program, emphasizing the need for a consistent foundation as AI/ML moves into production. citeturn11search7turn9search4
This matters because AI infrastructure is already being pulled in multiple directions:
- Cloud GPUs for elasticity and managed services
- On-prem GPUs for data locality, cost amortization, or sovereignty
- Edge inference for latency and bandwidth constraints
- Hybrid setups because the real world refuses to be clean
When your deployment path spans more than one environment, standards stop being “nice” and become “how you prevent an operational nervous breakdown.”
A quick detour: conformance programs are boring—until you need one
If you’re thinking, “This sounds like compliance theater,” it’s worth remembering that Kubernetes’ original conformance program was one of the key mechanisms that kept the ecosystem from fragmenting into incompatible vendor dialects.
CNCF explicitly frames the AI program as following that model: apply a proven, community-driven approach to a fast-growing domain, reduce confusion, and create clear requirements. citeturn2view0
Is it glamorous? No. Does it matter? Ask anyone who has migrated a non-trivial platform and discovered that “Kubernetes” can mean 17 subtly different behaviors depending on where you run it.
Where conformance meets the day-to-day: practical implications for teams
1) Procurement gets slightly less weird
Buying AI infrastructure is currently a mix of performance benchmarks, vendor promises, and “trust us, we support your framework.” The conformance badge provides a baseline answer to a simpler question: does this Kubernetes platform support the minimum capabilities to run AI workloads reliably? citeturn2view0turn4view0
It won’t tell you whether Vendor A is cheaper than Vendor B for your specific LLM, but it can reduce the risk that you’re selecting a platform that requires extensive custom workarounds just to function.
2) Portability becomes more realistic for AI workloads
Conformance is explicitly about reducing platform-specific surprises. citeturn4view0
For AI teams, that translates into fewer “rewrite your deployment” moments when moving:
- from dev cluster to production cluster,
- from one cloud to another,
- from cloud to on-prem (or vice versa),
- between managed Kubernetes offerings and curated enterprise platforms.
3) Platform engineering gets a clearer roadmap
If you run clusters for internal teams, you’ve probably had the “can we support AI?” conversation. Usually, that conversation degenerates into a list of tools: GPU operator, some storage tweaks, maybe a workflow engine, and a handful of policies.
A conformance program flips the question into: “Which baseline capabilities do we need, and can we demonstrate them?” That’s a much more operationally tractable question.
4) Expect tooling ecosystems to standardize around the checklist
Once there’s a stable baseline, tool builders can target it. That matters for open source AI-on-Kubernetes projects—operators, schedulers, inference gateways, observability integrations—because it reduces the combinatorial explosion of “works on these 3 distributions, breaks on the other 14.” citeturn4view0turn2view0
But does conformance solve GPU scheduling, cost, and governance? Not by itself
Here’s where we should be honest: a baseline standard doesn’t make AI infrastructure cheap, and it doesn’t automatically solve governance, security, or multi-tenant fairness. It makes those problems easier to reason about across environments.
Even the program repo notes that requirements evolve over time, aligned with Kubernetes release cycles, and that the specifics change by version. citeturn6view0turn4view0
So think of conformance as:
- a floor, not a ceiling,
- a common language for vendors and buyers,
- a forcing function that discourages proprietary shortcuts.
Networking standards are creeping into the AI story too
If you’re wondering why a Kubernetes AI conformance standard would care about networking APIs: inference is fundamentally a networking problem at scale. Routing, timeouts, retries, and policy controls become part of “AI reliability.”
The broader Kubernetes ecosystem has been modernizing networking primitives via the Gateway API, which has reached GA and continues to evolve (for example, Gateway API v1.4.0 was released on October 6, 2025). citeturn11search3turn11search6
The Forbes coverage of the AI conformance program even mentions requirements evolving with Kubernetes releases and references platform expectations like enabling Gateway API and meeting documented conformance checklists for areas like GPU scheduling and network policies. (As always, treat third-party coverage as directional; the authoritative source is the program repo and CNCF docs.) citeturn9search3turn4view0
Case study thinking: what conformance changes in real deployments
Let’s ground this in a familiar enterprise story: an organization runs Kubernetes across two clouds and an on-prem environment. Their AI roadmap looks like this:
- Start with a hosted LLM API to prototype
- Move inference in-house for cost and privacy
- Fine-tune models on proprietary data
- Eventually add distributed training for larger models
Without a standard baseline, each phase tends to require re-validating the platform from scratch: will our GPU nodes behave the same in Cloud A and Cloud B? Can we enforce the same isolation rules? Do our model-serving components rely on vendor-specific ingress behavior?
With conformance, the organization can at least constrain the platform layer. They still need to validate performance and cost, but they reduce the risk that basic AI workload functionality breaks because one environment is missing a critical cluster capability.
What to watch in 2026: v2.0 and automated testing
The CNCF announcement notes that the program was announced in beta at KubeCon + CloudNativeCon Japan in June 2025, then certified initial participants with a v1.0 release, and has started work on a roadmap for a v2.0 release next year (2026). citeturn2view0
Meanwhile, the GitHub repo states automated conformance tests are planned for 2026. citeturn6view0
Those two tracks—requirements evolution and test automation—are where the program becomes significantly more powerful. When conformance is machine-verifiable (even partially), it becomes much harder for vendors to “interpret” requirements generously, and much easier for the ecosystem to trust the badge.
So… should you care?
If you’re running AI workloads on Kubernetes today, you should care because it gives you leverage: a shared baseline you can demand from vendors, and a checklist to align internal platform roadmaps.
If you’re not running AI workloads on Kubernetes today, you should still care, because you probably will—either because product teams push models into production, or because someone will ask for “a small GPU cluster” and it will not remain small.
And if you’re in procurement, you should care because it’s one of the few ways the industry is trying to make AI infrastructure less of a bespoke snowflake hobby and more of a repeatable engineering discipline.
Sources
- Giant Swarm – “Infrastructure for AI is finally getting a standard” (Puja Abbassi, Nov 11, 2025)
- CNCF – Launch announcement: Certified Kubernetes AI Conformance Program (Nov 11, 2025)
- GitHub – cncf/k8s-ai-conformance repository (Kubernetes AI Conformance)
- CNCF – “Help us build the Kubernetes conformance for AI” (Jeffrey Sica, Aug 1, 2025)
- Oracle Cloud Infrastructure Blog – “OCI Kubernetes Engine (OKE) Achieves Kubernetes AI Conformance” (Dec 5, 2025)
- Kubermatic – “The ‘Wild West’ of AI Infra is Over… Kubermatic is officially Kubernetes AI Conformant” (Nov 11, 2025)
- Kubernetes Blog – “Gateway API 1.4: New Features” (Nov 6, 2025)
- Forbes – “CNCF Establishes Standards For Running AI Workloads On Kubernetes” (Nov 18, 2025)
Bas Dorland, Technology Journalist & Founder of dorland.org