Operations | Monitoring | ITSM | DevOps | Cloud

99%+ Accuracy on a Moving Target: Model Deprecation and Reliability with Not Diamond

Shipping systems powered by LLMs would be hard enough if the models stayed the same. But in reality, they don’t. Models get updated and deprecated at a pace traditional software wouldn’t. All while teams are still expected to hit reliability targets that look a lot like traditional SLAs.

The Ultimate Blueprint for Successful Synthetic Monitoring Implementation

In today’s digital world, the performance of websites and apps has a direct effect on sales, customer satisfaction, and brand reputation. Synthetic performance monitoring provides the proactive intelligence needed to ensure your application is always performing optimally. By simulating real user interactions from global locations before issues affect actual users, you transform from reactive problem-solving to proactive performance excellence.

Configuration as Intelligence: The New Operating System of Resilience

Modern IT operations live in constant flux. New tools appear, workloads shift to the cloud, architectures fragment, and every device, application, and user brings its own update rhythm. In this state of constant motion, reliability isn’t a static condition; it’s a dynamic discipline. For years, organizations have relied on observability and monitoring to keep systems running. But those tools only tell half the story.

Claude Models Just Landed in Azure: I Opened a Terminal and Tested

Microsoft just added Anthropic’s Claude models to Microsoft Foundry. Instead of reading the press release, I ran a mini-benchmark to see how Claude Opus, Sonnet, and Haiku actually perform with real Python tasks and stdin/stdout workflows. The results will surprise you. Opus was the most complete, Haiku the fastest (and chattiest), and instruction-following was the weak point across the board. If you’re thinking about using Claude in production, read this first.

How to Audit AI-Written Pull Requests Without Burning Out

If it feels like your GitHub notifications are a targeted DDoS attack on your brain, you aren't imagining it. Data from GitHub's Octoverse 2025 report shows an average of 43.2 million pull requests merged every month, a 23% jump from just a year ago. This surge in activity coincides with the widespread adoption of AI tools to write code. The temptation to just click "Approve" on a well-formatted AI-written pull request is higher than ever.

Do you still need wildcard certificates?

You’ve used wildcard certificates for years. It made your life easier. Once a year you’d renew your wildcard certificate, and copy it around to all the servers. It was way too complicated and expensive to get a unique certificate for every system. But now certificate lifetimes are shrinking to 47 days by 2029 and it’s not going to work anymore. You need to automate your certificates. Soon.

What NVIDIA, Okta, and Warner Bros. Discovery Learned About Scaling AI Operations Beyond the Pilot Phase

One key takeaway from AWS re:Invent 2025 was that a clear gap has emerged between teams still experimenting with AI and those seeing measurable value at scale. In two sessions, PagerDuty customers joined us onstage to explain how they’ve scaled pilots into successful AI operations.