Operations | Monitoring | ITSM | DevOps | Cloud

Grafana Vs Graphite

The amount of data being generated today is unprecedented. In fact, more data has been created in the last 2 years, than in the entire history of the human race. With such volume, it’s crucial for companies to be able to harness their data in order to further their business goals. A big part of this is analyzing data and seeing trends, and this is where solutions such as Graphite and Grafana become critical.

Level Up With Derived Columns: Understanding Screen Size (With Basic Arithmetic)

When we released derived columns last year, we already knew they were a powerful way to manipulate and explore data in Honeycomb, but we didn’t realize just how many different ways folks could use them. We use them all the time to improve our perspective when looking at data as we use Honeycomb internally, so we decided to share. So, in this series, Honeycombers share their favorite derived column use cases and explain how to achieve them.

You Can Improve Your Customer Satisfaction Charlie Brown!

What’s surprising to see today is how business operations struggle to get an integrated view of all business metrics. With greater volumes of data being collected, data analysts just can’t keep up with the pace. This state of affairs alone doesn’t hit as hard as the fact that many in data analytics have just come to accept this situation as a norm and simply bear with this daily struggle.

GrafanaCon Recap: The State of TSDB

At GrafanaCon EU, we gathered representatives of the Graphite, Prometheus, InfluxDB, and Timescale projects in the hopes of starting a spirited conversation about the current state of Time Series Databases. They didn’t disappoint! Here are a few highlights from the TSDB panel featuring Erik Nordstrom from Timescale, Dan Cech from Graphite, Paul Dix from InfluxDB, and Tom Wilkie from Prometheus, and moderated by Grafana Labs co-founder and CEO Raj Dutt.

Instrument Your Python App Automatically With The Honeycomb Beeline for Python

We’ve been on a roll this year with Beelines, our integrations for quick, easy, and automagic instrumentation of your apps. You may have already seen our Node.js, Ruby, and Go beelines – today, we’re excited to announce the release of the Honeycomb Beeline for Python!

The Importance of Historical Log Data

Centralized log management lets you decide who can access log data without actually having access to the servers. You can also correlate data from different sources, such as the operating system, your applications, and the firewall. Another benefit is that user do not need to log in to hundreds of devices to find out what is happening. You can also use data normalization and enhancement rules to create value for people who might not be familiar with a specific log type.

Metrics At Scale: How to Scale and Manage Millions of Metrics (Part 2)

With businesses collecting millions of metrics, let’s look at how they can efficiently scale and deal with these amounts. As covered in the previous article (A Spike in Sales Is Not Always Good News), analyzing millions of metrics for changes may result in alert storms, notifying users about EVERY change, not just the most significant ones. To bring order to this situation, Anodot groups correlated anomalies together, in a unified alert.