High Availability Installation

Installing Boundary as a high availability service requires certain infrastructure considerations. At the most basic level, running 3 controller and 3 worker instances gives a fundamental level of high availability for the control plane (controller), as well as bandwidth for a number of sessions on the data plane (worker). Both server types should be run in a fault-tolerant setting, that is, in a self-healing environment such as an auto-scaling group. The documentation here does not cover self-healing infrastructure and assumes the operator has their preferred scheduling methods for these environments.

Network Requirements

The following ports should be available:

  • Clients must have access to the Controller’s port (default 9200)
  • Clients must have access to the Worker’s port (default 9202)
  • Workers must have access to the Controller’s port (default 9201)
  • Workers must have a route and port access to the hosts defined within the system, in order to provide connectivity

Architecture

The general architecture for the server infrastructure requires 3 controllers and 3 workers. Note that it is possible to run a controller and worker within the same process, but the guide here assumes separate deployments. The documentation here uses virtual machines running on Amazon EC2 as the example environment, but this use case can be extrapolated to almost any cloud platform to suit operator needs:

High Availability Install - 图1

As shown above, Boundary is broken up into its controller and worker server components across 3 EC2 instances, in 3 separate subnets, in three separate availability zones, with the controller API and UI being publically exposed by an application load balancer (ALB). The worker and controller VM’s are in independant auto-scaling groups, allowing them to maintain their exact capacity.

The workers must be able to establish a connection to clients which they interact. In the architecture above, we place them in the public subnet so our remote client can establish a session between them and the target VM.

Boundary requires an external Postgres and KMS. In the example above, we’re using AWS managed services for these components. For Postgres, we’re using RDS and for KMS we’re using Amazon’s Key Management Service.

Architecture Breakdown

API and Console Load Balancer

Load balancing the controller allows operators to secure the ingress to the Boundary system. We recommend placing all Boundary server’s in private networks and using load balancing techniques to expose services such as the API and administrative console to public networks. In the high availability architecture, we recommend load balancing using a layer 7 load balancer and further constraining ingress to that load balancer with layer 4 constraints such as security groups or IP tables.

For general configuration, we recommend the following:

  • HTTPS listener with valid TLS certificate for the domain it’s serving or TLS passthrough
  • Health check port should use :9200 with TCP protocol

Controller Configuration

When running Boundary controller as a service we recommend storing the file at /etc/boundary-controller.hcl. A boundary user and group should exist to manage this configuration file and to further restrict who can read and modify it.

For detailed configuration options, see our configuration docs.