Operations | Monitoring | ITSM | DevOps | Cloud

Benchmarking Diskless Topics: Part 1

We benchmarked Diskless Kafka (KIP-1150) with 1 GiB/s in, 3 GiB/s out workload across three AZs. The cluster ran on just six m8g.4xlarge machines, sitting at <30% CPU, delivering ~1.6 seconds P99 end-to-end latency - all while cutting infra spend from ≈$3.32 M a year to under $288k a year. That’s a >94% cloud cost reduction. Extending Apache Kafka does come with an explicit tax.

Aiven for ClickHouse

Aiven for ClickHouse is back - better, faster, and stronger than ever. Explore how its new capabilities make it easier to build real-time analytics pipelines, handle massive query workloads, and extract insights at scale. Whether you’re streaming events, powering dashboards, or running analytics applications, Aiven for ClickHouse helps you move from data to decisions in record time.

Different ways to Search Text in PostgreSQL

When it comes to text search, PostgreSQL offers a surprisingly rich set of tools. Initially, text search capabilities were quite basic, often relying on the LIKE operator. This is inefficient for large amounts of text and lacks the nuance that comes with language. A major breakthrough came with the introduction of the built-in tsquery and tsvector data types, along with the associated functions, as part of PostgreSQL's core distribution. But tsquery is not the only option.

Kafka to ClickHouse in 6 Minutes

Learn how to stream data from Kafka into ClickHouse using Aiven’s integration wizard. In this demo, we show you how to generate sample logistics data in Kafka, configure the integration to map Avro-formatted fields, and connect to ClickHouse to view and query the ingested data. We also demonstrate how to create a materialised view in ClickHouse to store and query streamed data efficiently, making real-time analytics fast and easy.

Boosting Your Outreach Effectiveness: The Power of Verified Contact Data

Read any guide on successful entrepreneurship or effective sales strategies, and you'll likely encounter a common, underlying theme: it's all about building relationships. Regardless of what your business sells or the clients you're trying to reach, relationships form the bedrock of growth. And to build those relationships, businesses must consistently connect with potential customers - a task that becomes far more challenging when contact data is inaccurate or unreliable. This is where the power of verified contact data enters the scene.

Using the Downsampling Plugin in InfluxDB 3

Modern systems generate huge volumes of time series data. Advances in hardware and edge instrumentation enable sensors and applications to capture new values every second—or faster—which makes high-frequency measurement easy and affordable. When applied effectively, this steady flow of data reveals early warning signs, highlights subtle performance shifts, and helps teams understand how systems behave in real-time.
Sponsored Post

6 Reasons Your Data Lake Isn't Working Out

Since the data lake concept emerged more than a decade ago, data lakes have been pitched as the solution to many of the woes surrounding traditional data management solutions, like databases and data warehouses. Data lakes, we have been told, are more scalable, better able to accommodate widely varying types of data, cheaper to build and so on. Much of that is true, at least theoretically. In reality, the way many organizations go about building data lakes leads to a variety of problems that undercut the value their data lakes are supposed to deliver.

Using Tech to Keep Your Industrial Operations Safe

In modern industry, safety is pretty much a founding principle of good business, and without it, your company would not exist for very long, whether due to reputational damage or legal action. Keeping everyone safe is vital, and the good news for you is that tech has made it easier than ever to achieve.