Machine learning — the practice of writing algorithms that improve automatically through experience — has become a buzzword nowadays that connotes to something otherworldly and on the bleeding edge of technology. I’m here to tell you while that may be true, getting started with machine learning doesn’t have to be hard!
We recently announced the general availability of our Elasticsearch Service API. APIs help to automate tasks such as creating and scaling deployments, integrating with existing workflows, and testing. The Elasticsearch Service API supports the Open API Specification, which allows you to use tools like Swagger to generate software development kits (SDKs) in any programming language. You can import the API spec onto Postman and create a Postman Collection to create a test suite.
In a previous blog, we saw that the seemingly simple job of an Elasticsearch client — moving data between the calling code and the cluster — is actually quite complicated under the hood. Naturally, as much as we try to make the default behaviour of the client optimal for the majority of scenarios, there are situations where you want to configure, customize, or enable/disable certain features.
I recently spoke with Jeremy White who is using InfluxDB to monitor his aquariums. By collecting IoT sensor data, he has been able to better understand his 200 gallon salt-water aquarium full of fish and coral. The entire project can be found on GitHub. Caitlin: Tell us about yourself and your career. Jeremy: I’m a Senior Network Automation Consultant at Network to Code, and my background is in networking engineering. Network to Code is an industry leader in network automation.
We love maps at Elastic. In the Elastic Stack, there is one core component of all data we visualize using maps: Location. Location can mean reporting real-time positions of fleet vehicles, using a geofence for limiting search results, gauging application performance metrics from a geographic area, or identifying security threats by attaching geographic coordinates to IP addresses.
In two previous posts, we covered structuring data with grok and building custom grok patterns. But what happens if you just can’t get your grok patterns to work? In this article, we’re going to use Kibana’s Grok Debugger to help us debug a broken grok pattern. The divide-and-conquer method described below should help you to quickly find the reason that a given grok pattern is not matching your data.
Open source contributions are foundational to Elastic — from Elasticsearch’s Apache Lucene core to the addition of open source Logstash and Kibana to form the Elastic Stack you’ve come to know and love. Over the years, the Elastic community has created over 90 Beats, shared use case tutorials like those from Volvo, T-Mobile, and Microsoft, and presented at hundreds upon hundreds of meetups.
We recently changed the pricing of InfluxDB Cloud to let you control your cloud database spend so you spend only as much as you need to run your software and systems — with no wasted budget. If you just want a summary, check the InfluxDB Cloud pricing page. But if you’d like to nerd out on the changes we made, why we made them, and how to estimate your monthly spend on InfluxDB, then buckle up for a deep dive.
The official Go client for Elasticsearch is one of the latest additions to the family of clients developed, maintained, and supported by Elastic. The initial version was published early in 2019 and has matured over the past year, gaining features such as retrying requests, discovering cluster nodes, and various helper components. We also provide comprehensive examples to facilitate using the client.