Envoy is an L7 proxy and communication bus designed for large modern service oriented architectures.
The project was born out of the belief that:

The network should be transparent to applications. When network and application problems do occur
it should be easy to determine the source of the problem.

In practice, achieving the previously stated goal is incredibly difficult. Envoy attempts to do so
by providing the following high level features:

Out of process architecture: Envoy is a self contained process that is designed to run
alongside every application server. All of the Envoys form a transparent communication mesh in which
each application sends and receives messages to and from localhost and is unaware of the network
topology. The out of process architecture has two substantial benefits over the traditional library
approach to service to service communication:

Envoy works with any application language. A single Envoy deployment can form a mesh between
Java, C++, Go, PHP, Python, etc. It is becoming increasingly common for service oriented
architectures to use multiple application frameworks and languages. Envoy transparently bridges
the gap.

As anyone that has worked with a large service oriented architecture knows, deploying library
upgrades can be incredibly painful. Envoy can be deployed and upgraded quickly across an
entire infrastructure transparently.

Modern C++11 code base: Envoy is written in C++11. Native code was chosen because we
believe that an architectural component such as Envoy should get out of the way as much as possible.
Modern application developers already deal with tail latencies that are difficult to reason about
due to deployments in shared cloud environments and the use of very productive but not particularly
well performing languages such as PHP, Python, Ruby, Scala, etc. Native code provides generally
excellent latency properties that don’t add additional confusion to an already confusing situation.
Unlike other native code proxy solutions written in C, C++11 provides both excellent developer
productivity and performance.

L3/L4 filter architecture: At its core, Envoy is an L3/L4 network proxy. A pluggable
filter chain mechanism allows filters to be written to
perform different TCP proxy tasks and inserted into the main server. Filters have already been
written to support various tasks such as raw TCP proxy,
HTTP proxy, TLS client certificate
authentication, etc.

HTTP L7 filter architecture: HTTP is such a critical component of modern application
architectures that Envoy supports an additional HTTP L7 filter
layer. HTTP filters can be plugged into the HTTP connection management subsystem that perform
different tasks such as buffering, rate limiting, routing/forwarding, sniffing
Amazon’s DynamoDB, etc.

First class HTTP/2 support: When operating in HTTP mode, Envoy supports both HTTP/1.1 and HTTP/2. Envoy can operate as a transparent
HTTP/1.1 to HTTP/2 proxy in both directions. This means that any combination of HTTP/1.1 and HTTP/2
clients and target servers can be bridged. The recommended service to service configuration uses
HTTP/2 between all Envoys to create a mesh of persistent connections that requests and responses can
be multiplexed over. Envoy does not support SPDY as the protocol is being phased out.

HTTP L7 routing: When operating in HTTP mode, Envoy supports a
routing subsystem that is capable of routing and redirecting
requests based on path, authority, content type, runtime values, etc.
This functionality is most useful when using Envoy as a front/edge proxy but is also leveraged when
building a service to service mesh.

gRPC support:gRPC is an RPC framework from Google that uses HTTP/2
as the underlying multiplexed transport. Envoy supports all of the
HTTP/2 features required to be used as the routing and load balancing substrate for gRPC requests
and responses. The two systems are very complementary.

MongoDB L7 support:MongoDB is a popular database used in modern
web applications. Envoy supports L7 sniffing, statistics production,
and logging for MongoDB connections.

Service discovery:Service discovery is a critical
component of service oriented architectures. Envoy supports multiple service discovery methods
including asynchronous DNS resolution and REST based lookup via a service discovery service.

Health checking: The recommended
way of building an Envoy mesh is to treat service discovery as an eventually consistent process.
Envoy includes a health checking subsystem which can
optionally perform active health checking of upstream service clusters. Envoy then uses the union of
service discovery and health checking information to determine healthy load balancing targets. Envoy
also supports passive health checking via an outlier detection subsystem.

Advanced load balancing:Load balancing among different
components in a distributed system is a complex problem. Because Envoy is a self contained proxy
instead of a library, it is able to implement advanced load balancing techniques in a single place
and have them be accessible to any application. Currently Envoy includes support for automatic
retries, circuit breaking,
global rate limiting via an external rate limiting service,
request shadowing, and
outlier detection. Future support is planned for request
racing.

Front/edge proxy support: Although Envoy is primarily designed as a service to service
communication system, there is benefit in using the same software at the edge (observability,
management, identical service discovery and load balancing algorithms, etc.). Envoy includes enough
features to make it usable as an edge proxy for most modern web application use cases. This includes
TLS termination, HTTP/1.1 and HTTP/2 support, as well as HTTP L7 routing.

Best in class observability: As stated above, the primary goal of Envoy is to make the network
transparent. However, problems occur both at the network level and at the application level. Envoy
includes robust statistics support for all subsystems. statsd (and compatible providers) is the currently supported statistics
sink, though plugging in a different one would not be difficult. Statistics are also viewable via
the administration port. Envoy also supports distributed
tracing via thirdparty providers.

Dynamic configuration: Envoy optionally consumes a layered set of dynamic configuration
APIs. Implementors can use these APIs to build complex centrally
managed deployments if desired.

A short note on the design goals of the code itself: Although Envoy is by no means slow (we have
spent considerable time optimizing certain fast paths), the code has been written to be modular and
easy to test versus aiming for the greatest possible absolute performance. It’s our view that this
is a more efficient use of time given that typical deployments will be alongside languages and
runtimes many times slower and with many times greater memory usage.