Operations | Monitoring | ITSM | DevOps | Cloud

April 2024

Profiling Vs Tracing in OpenTelemetry

When OpenTelemetry first came into the picture with the merger of OpenCensus and OpenTracing in 2019, it was pretty much all about classic telemetry data, namely- logs, metrics, and traces. Since then, OpenTelemetry has become an indispensable tool in the modern observability landscape. With frequent integrations and introduction to new capabilities every year or so, it has poised itself as an invaluable tool for cloud enterprises.

Jaeger vs Tempo - key features, differences, and alternatives

Both Grafana Tempo and Jaeger are tools aimed at distributed tracing for microservice architecture. Jaeger was released as an open-source project by Uber in 2015, while Tempo is a newer product announced in October 2020. Jaeger is a popular open-source tool that graduated as a project from Cloud Native Computing Foundation. Grafana Tempo is a high-volume distributed tracing tool deeply integrated with other open-source tools like Prometheus and Loki.

Introducing Relational Fields

We’re excited to bring you relational fields, a new feature that allows you to query spans based on their relationship to each other within a trace. Previously, queries considered spans in isolation: You could ask about field values on spans and aggregate them based on matching criteria, but you couldn’t use any qualifying relationships about where and how the spans appear in a trace.

Revealing unknowns in your tracing data with inferred spans in OpenTelemetry

In the complex world of microservices and distributed systems, achieving transparency and understanding the intricacies and inefficiencies of service interactions and request flows has become a paramount challenge. Distributed tracing is essential in understanding distributed systems. But distributed tracing, whether manually applied or auto-instrumented, is usually rather coarse-grained.

A guide to scaling OpenTelemetry Collectors across multiple hosts via Ansible

OpenTelemetry has emerged as a key open source tool in the observability space. And as organizations use it to manage more of their telemetry data, they also need to understand how to make it work across their various environments. This guide is focused on scaling the OpenTelemetry Collector deployment across various Linux hosts to function as both gateways and agents within your observability architecture.

Migrating from Elastic's Go APM agent to OpenTelemetry Go SDK

As we’ve already shared, Elastic is committed to helping OpenTelemetry (OTel) succeed, which means, in some cases, building distributions of language SDKs. Elastic is strategically standardizing on OTel for observability and security data collection. Additionally, Elastic is committed to working with the OTel community to become the best data collection infrastructure for the observability ecosystem.

Real User Monitoring With a Splash of OpenTelemetry

You're probably familiar with the concept of real user monitoring (RUM) and how it's used to monitor websites or mobile applications. If not, here's the short version: RUM requires telemetry data, which is generated by an SDK that you import into your web or mobile application. These SDKs then hook into the JS runtime, the browser itself, or various system APIs in order to measure performance.

Mastering OpenTelemetry - Part 1

In the complex world of modern distributed systems, observability is vital. Observability allows engineers to understand what's happening within their systems, debug issues rapidly, and proactively ensure optimal application performance. OpenTelemetry has emerged as a powerful, vendor-neutral solution to address the challenges of observability across different technologies and environments.

A guide to scaling Grafana Alloy deployments across multiple hosts

Last week we introduced Grafana Alloy, our distribution of the OpenTelemetry Collector with built-in Prometheus pipelines and support for metrics, logs, traces, and profiles. We’re excited to see the community embrace Alloy, and we want to help them use and scale it as easily as possible. Many developers that need to deploy and manage software across several hosts turn to Ansible for its ease of use and versatility.

Enhancing Data Ingestion: OpenTelemetry & Linux CLI Tools Mastery

While OpenTelemetry (OTel) supports a wide variety of data sources and is constantly evolving to add more, there are still many data sources for which no receiver exists. Thankfully, OTel contains receivers that accept raw data over a TCP or UDP connection. This blog unveils how to leverage Linux Command Line Interface (CLI) tools, creating efficient data pipelines for ingestion through OTel's TCP receiver.

Implementing Jaeger for Distributed Tracing in Microservices

Earlier, applications were mostly monolithic, meaning that several programs were written in the same language and placed in the same web stack. However, it is no longer the case today. Today, every software is comprised of several small application programs coming together each providing a service of its own. These applications are what we call microservices.

Integrations for new Data Sources, Upgrades to Alerts & Kubecon Paris - SigNal 35

Welcome to the 35th edition of our monthly product newsletter - SigNal 35! We have made significant advancements in enhancing our product. The integration feature we shipped will enable quick-start monitoring for popular technologies in SigNoz. Let’s see what humans of SigNoz were up to in the month of March 2024.

Observing Core Web Vitals with OpenTelemetry: Part Two

Core Web Vitals (CWV) are Google's preferred metrics for measuring the quality of the user experience for browser web apps. Currently, Core Web Vitals measure loading performance, interactivity, and visual stability. These are the main indicators of what a user’s experience will be while using a web page: Note: As of March 12th, INP has become a stable Core Web Vital, replacing First Input Delay (FID).

Announcing the Elastic OpenTelemetry SDK Distributions

Adopting OpenTelemetry native standards for instrumenting and observing applications If you develop applications, you may have heard about OpenTelemetry. At Elastic®, we are enthusiastic about OpenTelemetry as the future of standardized application instrumentation and observability.

How the Prometheus community is investing in OpenTelemetry

Goutham Veeramachaneni, a product manager at Grafana Labs, and Carrie Edwards, a senior software engineer at Grafana Labs, are both contributors to the Prometheus open source project. This post, which they wrote together, was originally published on the Prometheus.io blog in March 2024. The OpenTelemetry project is an observability framework and toolkit designed to create and manage telemetry data such as traces, metrics, and logs.

Choosing the Right Opentelemetry Backend: Key Considerations

With applications becoming increasingly distributed and complex, gaining insights into their behavior and performance is essential for maintaining reliability and delivering exceptional user experiences. OpenTelemetry has emerged as a powerful framework for instrumenting applications to collect, process, and export telemetry data.

Beyond the trace: Pinpointing performance culprits with continuous profiling and distributed tracing correlation

Observability goes beyond monitoring; it's about truly understanding your system. To achieve this comprehensive view, practitioners need a unified observability solution that natively combines insights from metrics, logs, traces, and crucially, continuous profiling. While metrics, logs, and traces offer valuable insights, they can't answer the all-important "why." Continuous profiling signals act as a magnifying glass, providing granular code visibility into the system's hidden complexities.

Unlock the Power of Observability with OpenTelemetry Logs Data Model

Your log records may be missing a key ingredient that unlocks the world of observability for your applications, infrastructure and services. If you're building a new application or enhancing an existing one, consider adopting the OpenTelemetry Logs Data Model's Log and Event Record Definition. Adopting this definition enriches your logs by adding additional data, making it easier to use them to correlate them with metrics and traces, in addition to XYZ.