Operations | Monitoring | ITSM | DevOps | Cloud

Announcing a new dashboard for Deployment Environments

One of the central ideas behind the Codefresh GUI is giving as much information as possible to both developers and operators regarding the status of a deployment. Just because a pipeline has finished successfully does not always mean that the respective environment is healthy.

On Hermes and Mattermost

With the upgrade to React Native 61 came the prospect of substantially improving performance of our Android app. How? Through the use of Hermes, Facebook’s new JavaScript engine. To say that we were excited is an understatement. And with that excitement came curiosity: How is this new JavaScript engine achieving performance boosts?

Joyful DevOps is Here: Introducing the JFrog DevOps Platform

Unity is strength. And at JFrog, we’re committed to providing the strongest DevOps tools available. With the promised release of the JFrog DevOps Platform, it’s my honor and delight to announce JFrog’s biggest leap yet toward fulfilling our universal vision of Liquid Software. We’re excited and proud to launch the world’s first universal, hybrid, end-to-end DevOps platform.

What are racks and which one should you buy?

A rack is a structure, usually made out of metal and cabinet or wall-shaped, which allows to store and organize the different components of computer installations, such as servers, storage systems, switches, etc. Is that it? Are you disappointed? Well, hold on, although they don’t seem like much, the world of racks can actually be quite tricky.

10 lessons from the AWS hackathon

This weekend I had the opportunity to participate in the AWS Hackathon in Portland. Based on the hackathons hosted at re:Invent, this brought together about 100 developers of all skill levels to break up into small teams and produce a demo product in 10 hours. I had a great time, and wanted to share what I learned! There may be organization-specific roadblocks to adopting AWS Lambda right now and therefore, it might not be the right fit for your team at this particular juncture.

Monitor Apache Airflow with Datadog

Apache Airflow is an open source system for programmatically creating, scheduling, and monitoring complex workflows including data processing pipelines. Originally developed by Airbnb in 2014, Airflow is now a part of the Apache Software Foundation and has an active community of contributing developers. Airflow represents workflows as Directed Acyclic Graphs (DAGs), which are made up of tasks written in Python. This allows Airflow users to programmatically build and modify their workflows.