Adding & Managing Load Balancers
A load balancer is used to distribute traffic across your web servers, and offers benefits such as maximizing throughoutput, minimizing response times and avoiding overload on any single server. Ultimately, load balancing increases the reliability of your application.
You can add either native load balancers (for cloud vendors with native load balancer support) or an HAProxy load balancer for cloud vendors with no native load balancer support (currently OVH, Packet, Rackspace and Vultr) and Registered Servers.
Add a load balancer
To add a load balancer to your application:
- Open the Application Overview from the Dashboard.
- Click Network in the right-hand panel and then Load Balancer
- Click the + Add Load Balancer button
- A panel will slide out from the left with options. Select what you need and then click + Add to continue.
You can now watch the logs, as usual to see the progress of the process. Depending on which cloud provider you use, this load balancer will be set up differently.
Supported Cloud Load Balancers
- Amazon AWS: Elastic Load Balancing
- DigitalOcean: Load Balancers
- Google Cloud Engine: Forwarding rules, target pools & health checks
- Hetzner Cloud: Load Balancers
- Linode: NodeBalancer
- Microsoft Azure: TrafficManager
- Rackspace: Rackspace Load Balancing
Automatic endpoint test
When a new load balancer is set up it will begin to ping your web endpoints to check their health. By default load balancers ping the root path of your application (
/) and if they receive a
200 response code, they will consider a server healthy.
If your application does not have a valid endpoint or route set for
/ then you can specify a custom path under the
httpchk option in your manifest file to ensure your application responds appropriately. We will configure the load balancer to ping that path rather than the root.
If your servers respond positively to the ping test, the load balancer will begin to distribute the load between them. If any of these ping tests fail, the load balancer will not distribute traffic to those servers that failed.
Cloud vendors only
This feature is only available if you have deployed using a cloud vendor, and for applications running in production.
Adding multiple load balancers
One potential drawback of having a load balancer is that it is a single point of failure. To improve the high availability of an application, you can add more than one load balancer to it.
To add a second load balancer to your application:
- First add a standard load balancer to your application. You can follow the instructions above to do so.
- Once this is in place, click the + Add button in the panel at the bottom of the Load Balancer page (titled "Need to add additional types of load balancer to this application?")
This new load balancer is essentially a clone of your first one. Whenever changes are applied to your first load balancer, they will also be applied to any clones. Both load balancers are “live” and can distribute traffic to your application, but actually switching between load balancers requires an update to your DNS (see below).
Switching between load balancers
To switch traffic between load balancers, you should update the public DNS record for your application to point at the CNAME of your target load balancer. You can see the CNAME for any load balancer on the Load Balancers page (Click on Network → Load Balancers in the right-hand panel)
We recommend keeping your TTL for these records set to 300 (5 minutes) to reduce any downtime to a minimum.