Operations | Monitoring | ITSM | DevOps | Cloud

Show me the (meeting) money: How to monitor the real-time costs of a meeting in Grafana

This meeting could’ve been an email. It’s a phrase most of us have said (or at least thought) at some point in our careers. For me, that realization hit years ago while working for a government organization. I’d frequently sit through long, agendaless meetings that seemingly went nowhere. I wasn’t sure why I was there. And because I’m an engineer at heart, I started to wonder: what were these meetings actually costing the organization?

Troubleshoot Faster with the New Log Search and Filtering in Qovery Observe

Following the launch of Qovery Observe, we’re progressively adding new capabilities to help you better monitor, debug, and understand your applications. Today, we’re excited to announce a major improvement to the Logs experience: you can now search and filter directly within your application logs.

The Hidden Barrier to Network Automation Isn't Your AI - It's Your Data

For years, the promise of AI-driven network automation has loomed large. Vendors and analysts alike have painted a future where autonomous operations handle outages before they happen, root causes are explained instantly, and teams finally escape the endless cycle of alerts, tickets, and manual troubleshooting. But in practice, most automation initiatives stall long before they reach that vision.

IaC is Great, But Have You Met IaCM?

This blog highlights the critical role of Infrastructure as Code Management (IaCM) in enhancing IaC practices, ensuring security, compliance, and efficiency in managing complex infrastructure at scale. ‍ Managing infrastructure efficiently and reliably is more critical than ever. Infrastructure as Code (IaC) has emerged as a key practice, enabling teams to define, deploy, and manage infrastructure using code.

What is autonomous validation? The future of CI/CD in the AI era

Over the past decade, CI/CD has redefined how modern software is built and shipped. CircleCI has been a leader in that transformation, working alongside the world’s best engineering teams to build a reliable foundation for continuous delivery at scale. Today, those foundations are under new pressure as AI reshapes every aspect of the delivery cycle. Developers are producing more change with less certainty about what those changes touch.

Experimenting With Different Scripts

It all began when I spun up an AWS t4g.small burstable instance for a side project. Nothing unusual just another day in the cloud. But the moment I connected through SSH, something caught my eye. The system greeted me with a temperature reading of -273.5°C. Wait… what? That’s 0 Kelvin, the point where atomic motion completely stops. In other words, absolute zero , a state that’s theoretically impossible for anything to operate in.

7 Ways Your Incident Management Just Got a Boost (New Feature Rundown)

All the things you may have missed that will make your incident management smarter, faster, and simply easier. We ship updates every week because we want you to get the most out of FireHydrant. But we also know it's hard to stay up to date and read every week's changelog (even though we know reading changelogs is the highlight of your week ).

Ingest OTLP metrics directly into Datadog with the new OTLP Metrics API

Many organizations rely on OpenTelemetry (OTel) to standardize observability across distributed systems. These organizations are at varying stages of adoption and are implementing OTel in complex environments with diverse configurations. To support this range of use cases, Datadog offers many ways to use OpenTelemetry with Datadog.

A deep dive into Java garbage collectors

Historically, developers have relied on languages like C and C++ for explicit control over memory allocation and deallocation. This approach can yield very low overhead and tight control over performance, but it also increases complexity and risk (e.g., memory leaks, dangling pointers, and double frees). This often results in runtime issues that are difficult to diagnose, which can become a drag on team velocity.

Track, debug, and roll back changes with Version History for Synthetic Monitoring tests

A synthetic test is only useful if you can trust what it’s telling you. When one fails, the reason may not be obvious. Was the application updated? Did the test change? Or both? As more people contribute and refine the same test, it becomes harder to understand what changed or restore a working version. Without clear visibility into those updates, teams can spend more time tracking down the cause of a failure than resolving it.