Operations | Monitoring | ITSM | DevOps | Cloud

The latest News and Information on Containers, Kubernetes, Docker and related technologies.

Secure Docker Image Pulls from Cloudsmith to Kubernetes using OIDC

Pulling Docker images from private registries for containerised applications presents a security challenge. It requires authentication management, network access, and trust across distributed systems. Credentials must be securely handled and rotated, and image pulls can break due to network restrictions or expired tokens. All of this makes deployment and security harder.

Monitoring ECS Metrics: A Guide for Developers and Operations Teams

For anyone leveraging cloud computing, Amazon Elastic Container Service (ECS) continues to provide a seamless solution for managing containerized applications. AWS Fargate takes this cloud-native architecture a step further by allowing you to run containers without servers or clusters. As a serverless offering for ECS, Fargate provisions compute capacity and scales it based on demand.

Open Container Initiative (OCI) Support in Cloudsmith

Kubernetes has become the de facto platform for orchestrating containers. Open standards complement Kubernetes by defining best practices for its implementation. These standards are developed by the open-source Kubernetes community (not a single vendor), ensuring vendor neutrality, easier integration with other tools, and overall system efficiency.

Top 5 Observability Tools DevOps Teams Should Know

Observability and monitoring are the cornerstone of resilient, high-performing applications. Nearly every IT or software engineering leader we come into contact with emphasizes the importance of the ability to understand and diagnose what is going on with their applications at all times. Having clear and concise visibility into your applications is no longer optional.

Working with GPUs on Kubernetes and making them observable

GPUs are everywhere powering LLM inference, model training, video processing, and more. Kubernetes is often where these workloads run. But using GPUs in Kubernetes isn’t as simple as using CPUs. You need the right setup. You need efficient scheduling. And most importantly you need visibility. This post walks through how to run GPU workloads on Kubernetes, how to virtualize them efficiently, and how Coroot helps you monitor everything with zero instrumentation or config.

AI in Action with Kunal Kushwaha: 2 Demo Showcase. See What's Possible!

Join Kunal Kushwaha, Field CTO at Civo, for two demos using relaxAI. In the first demo, we'll show you how to deploy your own Large Language Model (LLM) inference engine using Ollama, giving you full control over your AI model. In the second demo, we'll demonstrate how to build custom AI integrations using relaxAI API, making it easy to add AI features to your existing applications. Whether you're an AI developer, MLOps team, or just curious about AI, this video is for you.

Deploy Istio at Scale With Rancher

Managing and deploying applications across multiple Kubernetes clusters presents significant challenges, especially as the number of clusters grows. Traditional methods, like manually applying Helm charts or manifests per cluster, become cumbersome, error-prone, and difficult to scale or maintain consistency for Day 2 operations. While Rancher allows managing Helm chart repositories and apps, this is done on a per-cluster basis via the UI.