Oliver Gould talks about the Linkerd project, a service mesh hosted by the Cloud Native Computing Foundation, to give operators control over the traffic between their microservices. He shares the lessons they’ve learned helping dozens of organizations get to production with Linkerd and how they’ve applied these lessons to tackle complexity with Linkerd.
This week I set out to write a post comparing Istio and Linkerd, and I told myself: I’m going to create tables comparing features, and it’s going to be great and people will love and the world will be happier for a few seconds. I promised myself It was going to be a fair comparison without bias from any end. While the ‘comparison table’ is still here, I shifted the focus of the article: the goal is not on which is better, but which is better for you, for your applications, for your organization.
Today we’re very happy to announce the release of Linkerd 2.1. This is our first stable update to 2.0, and introduces a host of goodies, including per-route metrics, service profiles, and a vastly improved dashboard UI. We’ve also added a couple exciting experimental features: proxy auto-injection, single namespace installs, and a high-availability mode for the control plane.
Node is one of the most popular languages for microservices. With the rise of Kubernetes, increasingly, Node developers are being asked to deploy their services to a Kubernetes cluster. But what’s required to safely deploy and run Node services on Kubernetes?
In this post, we focus on one specific, but vital, component: how do I understand what’s happening with my Node service on Kubernetes, and how do I debug it when things go wrong?