Operations | Monitoring | ITSM | DevOps | Cloud

Welcome to the Next Frontier: AI on Kubernetes

Last week’s KubeCon Atlanta made one thing abundantly clear, Kubernetes is quickly becoming the de facto platform for AI workloads – with the event lineup chock full of talks, workshops, and even co-located events dedicated to AI, machine learning and running data on Kubernetes natively – with approximately 50 (!) sessions in total focused on AI, ML, LLM, and GenAI topics.. What was until now mostly PoCs and aspirational is now truly delivering in production.

Lessons from KubeCon: What "Best-of-Breed" AI SRE Really Requires

This year’s KubeCon underscored a real shift: AI SRE has gone mainstream. Of course, it’s not a surprise. Teams from high-growth startups to Fortune 500s are running more complex, cloud-native systems, shipping more AI-generated code, and facing rising expectations. Downtime is absolutely not an option and the work for on-call SREs has become unsustainable. The question isn’t whether AI SRE helps. It’s which one you can trust in production.

Autonomous Self-Healing Capabilities for Cloud-Native Infrastructure and Operations

Modern cloud-native infrastructure was adopted to increase agility and scale, but as it grows in scale and complexity, engineering teams are now drowning in operational noise. Industry research (The State of Observability for 2024) reveals that 88% of technology leaders report rising stack complexity, while 81% say manual troubleshooting actively detracts from innovation.

Kubernetes v1.34: What You Need to Know

Kubernetes v1.34, codenamed “Of Wind & Will (O’ WaW)”, brings a wide range of enhancements aimed at making clusters more efficient, secure, and easier to manage. This release delivers 58 enhancements with 23 graduating to Stable, 22 entering Beta, and 13 in Alpha, reflecting the platform’s continued maturation as enterprises scale their container orchestration needs.

Kubernetes Is Powerful-But It's Slowing You Down. Here's How to Fix It.

Ask any SRE what slows them down in a Kubernetes incident, and the answer is usually too much information in too many different places. Kubernetes has changed the way we run software. It’s given us incredible flexibility, scalability, and power. But in the years I’ve worked in cloud operations and platform engineering, I’ve also seen how that power comes at a price: complexity.

Kubernetes Cost Optimization Done Right

Kubernetes was never just about cost savings. It was built to be a robust, scalable, and efficient platform for orchestrating containerized applications. And it was meant to abstract infrastructure away so developers could move quickly and go about their business of developing. But as Kubernetes adoption scaled, so did cloud bills. FinOps tools emerged to rein in spending, but most only scratch the surface.

Unlocking Cost Optimization Through Full-Stack Kubernetes Visibility

In Kubernetes environments, cost is rarely just about spend. It’s about performance, node utilization, workload behavior, and how all of those align with your team’s operational goals. Komodor’s approach to cost optimization has an operational advantage due to its deep visibility into your entire Kubernetes estate. Imagine the potential for cost optimization when you have complete visibility into every aspect of your Kubernetes operations.

Komodor + Backstage: Bringing Kubernetes Visibility into the Leading Open Source IDP

Platform engineering has emerged as the natural progression of DevOps—not a replacement as some may think, but rather more of a refinement. While DevOps broke down silos and encouraged shared responsibility, it often left teams fending for themselves across sprawling infrastructure stacks. Developers were promised autonomy, but were instead overwhelmed by the burden of managing CI/CD, security, observability, infrastructure provisioning, and more.

Port + Komodor: Bringing Kubernetes Visibility into the Modern Commercial IDP

Internal Developer Portals (IDPs) are no longer just an experimental concept—they’re now a foundational component of modern software delivery. As engineering organizations look to reduce cognitive load, increase self-service, and streamline infrastructure workflows, IDPs have emerged as the most effective way to productize platform engineering.