Welcome back to another one of our Weekly Community Roundups, everyone! We saw quite a number of tweets in this past week, including some great recommendations for our platform. Let’s take a look at everything that's been going on.
Vydia is dedicated to helping creators gain more control over their audio and video content with a centralized tool for distributing, managing, protecting, and optimizing AV files. Vydia’s software team describes themselves as a “DevOps team first and foremost” delivering new features and updates in a tight loop. They are always in search of new ways to improve and modernize the development process.
Developing new features for our platform is one of the key ways we can make sure we are giving you the best service possible — and we wouldn’t be able to do it without your help. Let’s take a look at all the feature requests that we have received in the past two weeks!
Back in 2017, I wrote a post titled “3 pro tips for Developers working with Kinesis streams”, in which I explained why you should avoid hot streams with many Lambda subscribers. When you have five or more functions subscribed to a Kinesis stream you will start to notice lots of ReadProvisionedThroughputExceeded errors in CloudWatch.
Imagine you are driving a car on a freeway. Your speedometer is telling you you’re going 62 mph. But you “gotta go fast”. Faster than then 65 mph speed limit. So you go for it: first 68mph, then 75mph, then 80mph. Then you pass a police officer hiding in a speed trap. To your dismay, they pull you over and give you a ticket. All is not lost: there is a silver lining here. It’s the perfect analogy is to understand how indicators, objectives, and agreements all work with each other.
Infrastructure as code is an important methodology for ensuring that your distributed systems are treated as cattle and not pets. Your Kubernetes and Rancher clusters are no different. You should be able to provision your Rancher clusters, your Kubernetes clusters, and all of your apps with automation.
AKS is the managed service from Azure for Kubernetes. When you create an AKS cluster, Azure creates and operates the Kubernetes control plane for you at no cost. The only thing you do as a user is to say how many worker nodes you’d like, plus other configurations we’ll see in this post. So, with that in mind, how can you improve the AKS cluster performance of a service in which Azure pretty much manages almost everything?
Kubernetes is a highly distributed, microservices oriented technology that allows devs to run code at scale. K8S revolutionized cloud infrastructure and made our lives a whole lot easier in many aspects. Developers don’t have to do anything but write code and wrap it in the docker container for K8S to handle. But even its greatest enthusiasts will admit debugging Kubernetes pods is still a pain.
Infrastructure management has come a long way. (Mostly) gone are the days of manual configurations and deployments, when using SSH in a “for” loop was a perfectly reasonable way to execute server changes. Automation is a way of life. Configuration management tools like Chef, Puppet, and Ansible — once on the bleeding edge — are now used by most enterprises.
Because DevOps practices can bring great speed and reliability to the software delivery lifecycle, release management can seem daunting. But, the improved visibility and collaboration brought about by DevOps can also help with the release management process. While the general concept of release management doesn’t really change between ITIL (IT Infrastructure Library) and DevOps, there are a few ways that the process differs.