Linux

canonical

Robotics Recap: Learning, Programming & Snapping ROS 2

Robotics@Canonical puts a strong focus on the migration from ROS to ROS 2. ROS 2 benefits from many improvements, especially robot security. Our goal is to make it easy for you to transition to ROS 2, whether you’re completely new to ROS or a seasoned engineer retooling for a new environment. Your new platform should be secure-by-default, and we expect you’ll need to pivot between different environments as you migrate from ROS to ROS 2.

splunk

SAI Something Linux: Monitoring Linux with Splunk App for Infrastructure

Metrics and logs go together like cookies and milk. Metrics tell you when you have a problem, and logs/events often tell you why that problem happened. But it’s always been harder than it needed to be to get both types of data onto a single screen, especially when the sysadmins using the tools aren’t necessarily daily experts in managing those monitoring platforms.

icinga

Icinga 2, Icinga Web 2 and Director Kickstart on CentOS 7

The easiest way to get started with Icinga is a single-node installation. If you are new to the Icinga world, here you have a kickstart for installing Icinga 2, Icinga Web 2 and Icinga Director on CentOS 7. With these steps you will have a ready Icinga environment for monitoring your infrastructure.

datadog

Diagnosing out-of-memory errors on Linux

Out-of-memory (OOM) errors take place when the Linux kernel can’t provide enough memory to run all of its user-space processes, causing at least one process to exit without warning. Without a comprehensive monitoring solution, OOM errors can be tricky to diagnose. In this post, you will learn how to use Datadog to diagnose OOM errors on Linux systems.

canonical

MicroK8s HA tech preview is now available

High availability (HA) for MicroK8s, the lightweight Kubernetes, is now available as a tech preview for Linux, Windows and macOS. The Kubernetes control plane can now be distributed across multiple nodes, bringing resiliency to the cluster while maintaining a low footprint using Dqlite, the distributed SQL engine as the Kubernetes datastore.

canonical

Encryption at rest with Ceph

Do you have a big data center? Do you have terabytes of confidential data stored in that data center? Are you worried that your data might be exposed to malicious attacks? One of the most prominent security features of storage solutions is encryption at rest. This blog will explain this in more detail and how it is implemented in Charmed Ceph, Canonical’s software-defined storage solution.

canonical

Data science workflows on Kubernetes with Kubeflow pipelines: Part 2

This blog series is part of the joint collaboration between Canonical and Manceps. Visit our AI consulting and delivery services page to know more. Kubeflow Pipelines are a great way to build portable, scalable machine learning workflows. It is a part of the Kubeflow project that aims to reduce the complexity and time involved with training and deploying machine learning models at scale. For more on Kubeflow, read our Kubernetes for data science: meet Kubeflow post.

elastic

Canonical, Elastic, and Google team up to prevent data corruption in Linux

At Elastic we are constantly innovating and releasing new features. As we release new features we are also working to make sure that they are tested, solid, and reliable — and sometimes we do find bugs or other issues. While testing a new feature we discovered a Linux kernel bug affecting SSD disks on certain Linux kernels. In this blog article we cover the story around the investigation and how it involved a great collaboration with two close partners, Google Cloud and Canonical.

netdata

Netdata Agent v1.23: Kubernetes monitoring & eBPF observability

Deploying and monitoring performance for an entire Kubernetes cluster can be complex. To simplify the process, we’ve added service discovery functionality to eliminate complex configuration, in addition to more advanced monitoring for viewing activity inside containers. Service discovery identifies k8s pods running on a cluster and immediately starts monitoring system performance. All containers are identified, regardless of complexity.