At re:Invent 2018, AWS announced AWS App Mesh, a service mesh that provides application-level networking. App Mesh makes it easy for your services to communicate with each other across multiple types of compute infrastructure, including: App Mesh standardizes how your services communicate, giving you end-to-end visibility and ensuring high availability for your applications. Service meshes like App Mesh help you run and monitor HTTP and TCP services at scale.
Using the open source Envoy proxy, App Mesh gives you access to a wide range of tools from AWS partners and the open source community. Because all traffic in and out of each service goes through the Envoy proxy, all traffic can be routed, shaped, measured, and logged. This extra level of indirection lets you build your services in any language desired without having to use a common set of communication libraries.
In this six-part series of the post, I walk you through setup and configuration of App Mesh for popular platforms and use cases, beginning with EKS. Here’s the list of the parts: Part 1: Introducing service meshes. Part 2: Prerequisites for running on EKS.
Part 3: Creating example microservices on Amazon EKS. Part 4: Installing the sidecar injector and CRDs. Part 5: Configuring existing microservices.
Part 6: Deploying with the canary technique. Part 1: Introducing service meshes. Part 2: Prerequisites for running on EKS.
Part 3: Creating example microservices on Amazon EKS. Part 4: Installing the sidecar injector and CRDs. Part 5: Configuring existing microservices.
Part 6: Deploying with the canary technique.
Source: amazon.com