Load balancer

Use cases

Decoupling

  • Hidden server maintenance. You can take a web server out of the load balancer pool, wait for all active connections to drain, and then safely shutdown the web server without affecting even a single client. You can use this method to perform rolling updates and deploy new software across the cluster without any downtime.

  • Seamlessly increase capacity. You can add more web servers at any time without your client ever realizing it. As soon as you add a new server, it can start receiving connections.

  • Automated scaling. If you are on cloud-based hosting with the ability to configure auto-scaling (like Amazon, Open Stack, or Rackspace), you can add and remove web servers throughout the day to best adapt to the traffic.

Security

  • SSL termination: By making load balancer the termination point, the load balancers can inspect the contents of the HTTPS packets. This allows enhanced firewalling and means that you can balance requests based on teh contents of the packets.

  • Filter out unwanted requests or limit them to authenticated users only because all requests to back-end servers must first go past the balancer.

  • Protect against SYN floods (DoS attacks) because they pass traffic only on to a back-end server after a full TCP connection has been set up with the client.

┌────────────────┐        ┌────────────────┐       ┌────────────────┐
│     Client     │        │     Client     │       │     Client     │
└────────────────┘        └────────────────┘       └────────────────┘
         │                         │                        │        
         │                     TLS/SSL                      │        
     TLS/SSL                       │                      TCP        
         │                         │                        │        
         ▼                         ▼                        ▼        

┌────────────────┐        ┌────────────────┐       ┌────────────────┐
│  Layer 4 load  │        │  Layer 4 load  │       │  Layer 4 load  │
│    balancer    │        │    balancer    │       │    balancer    │
└────────────────┘        └────────────────┘       └────────────────┘
         │                         │                        │        
         │                       TCP                        │        
     TLS/SSL                       │                    TLS/SSL      
         │                         │                        │        
         │                         │                        │        
         ▼                         ▼                        ▼        
┌────────────────┐        ┌────────────────┐       ┌────────────────┐
│    Upstream    │        │    Upstream    │       │    Upstream    │
│    services    │        │    services    │       │    services    │
└────────────────┘        └────────────────┘       └────────────────┘

Load balancing algorithms

Round-robin

  • Def: Cycles through a list of servers and sends each new request to the next server. When it reaches the end of the list, it starts over at the beginning.

    • How to support sticky sessions: Hashing based on network address might help but is not a reliable option. Or the load balancer could maintain a lookup table mapping session ID to server.

Weighted round robin

  • Problems of round robin:

    • Not all requests have an equal performance cost on the server. But a request for a static resource will be several orders of magnitude less resource-intensive than a requst for a dynamic resource.

    • Not all servers have identical processing power. Need to query back-end server to discover memory and CPU usage, server load, and perhaps even network latency.

Least load first (from server perspective)

  • Problems of weighted round robin:

    • Server might be under different status even given the same type of requests

  • How to define least load:

    • For a layer 4 load balancing option such as LVS, it could load balance based on the number of connections.

    • For a layer 7 load balancing option such as Nginx, it could load balance based on the number of Http requests.

    • More customized criteria such as CPU load, I/O load.

Best performance first (from client perspective)

  • Response time

Source hashing

  • Hash of IP address

  • Hash of session id

Categorization

Http redirect based load balancer (rarely used)

  • Steps:

    1. Client's requests first reach a load balancing server which translate original target IP address A to a new target IP address B with a 302 HTTP response code

    2. Client issues another request to the new target IP address B

  • Pros:

    • Easy to implement

  • Cons:

    • Client needs to have two http requests to data center.

    • Internal web/application servers' IP address will be exposed to external world and cause potential security risks.

      • Compared with internal servers, load balancing servers will have stricter firewall policies and security configurations.

  • Due to the security risks and performance cost, Http redirect based load balancing is rarely used in practice.

DNS based load balancer

  • Steps:

    1. Client's requests first reach DNS authority server to get the IP address.

    2. Client issues another requests to the parsed IP address.

  • Pros:

    • Low latency: DNS will be resolved for the first time. The second time when a request come it will use the cached version.

    • It is a basic config service provided by DNS providers so no development will be required. Many DNS service provider offers geographical DNS service.

  • Cons:

    • Low accuracy: DNS will be only resolved one time and cached for sometime. The cached value will be used for a relatively long time.

    • High update time: DNS has multi-layer caches. Even after scaling up, the old DNS record might still point to the old IP address.

    • Not customizable: DNS load balancer is controlled by DNS service providers.

    • Simple load balancing algorithm: DNS load balancing algorithms are relatively simple. For example, it could not make decisions based on the differences of servers.

  • Example DNS record for using round robin load balancing on top of DNS

    ┌────────────────────────────────┐
    │     Domain     IP addresses    │
    │   example.com     66.66.66.1   │
    │   example.com     66.66.66.2   │
    │                                │
    └────────────────────────────────┘

HTTP-DNS based load balancer

  • Due to the limitation of traditional DNS, many companies will develop HTTPDNS. For more details for how HttpDNS work, please see HTTPDNS

  • It is mainly used in App scenarios. For web scenarios, since url parsing is done by browser, it will be more compliacted.

  • Pros:

    • Shorter update time. It won't be impacted by the multi-layer cache.

    • More customized load balancing algorithm.

  • Cons:

    • Needs customized development and has high cost.

Application layer (e.g. Nginx, HAProxy, Apache)

  • Pros:

    • Could make load balancing decisions based on detailed info such as application Url.

    • Only applicable to limited scenarios such as HTTP / Email which sit in level 7.

  • Cons:

    • Low Performance (A single server could support roughly 50K QPS) when compared with LVS (Layer 4 - Transport layer software, A single server could support 800K QPS) or F5 (Layer 4 hardware, a single device supports 2000K - 8000K QPS)

    • Security: Usually don't have security features built-in place

Software based

LVS

  • LVS supports three modes of operation: VS/NAT, VS/TUN, VS/DR. It supports the following type of load balancing algorithms:

    • Round robin

    • Weighted round robin

    • Least connections

    • Weighted least connections

    • Locality-based least connections

    • Locality-based least connections with replication

    • Destination hashing

    • Source hashing

VS/NAT mode

  • Steps:

    1. Client's requests first reach IP load balancer.

    2. IP based load balancer changes the target IP address to internal servers' IP address, and change source IP to be load balancer's IP address (SNAT)

    3. Internal servers return response to IP based load balancer.

    4. IP based load balancer changes the target IP address.

  • Pros:

    • Operates on network layer and has much more efficiency compared with reverse proxy

  • Cons:

    • All requests/Responses will require IP based load balancer to replace the target IP address. It will become the bottleneck really easy.

VS/DR mode

  • How it works:

    • A same virtual IP address is bond to both director server and real server.

    • When ARP for a IP address, director server's address will be returned and real server's address will be hidden.

  • Steps:

    1. Client's requests first reach director server. For illustration, let's say within the request, the source IP is CIP, the dest IP is VIP; the source MAC is CMAC, the dest MAC is DMAC.

    2. Director server will keep the source and dest IP unchanged. Director server will change the target MAC address DMAC to internal servers' MAC address RMAC.

    3. Internal servers return response to client.

  • Pros:

    • High performance and could be used broadly

      1. Since there are no changes to IP address, internal servers could directly return the response to clients.

      2. Operates on the data link layer (only need to change the MAC address)

  • Cons:

    • Relies on client to retry because the response does not pass through director server. Even when a real server is down, there might still be gap until director server could remove the problematic real server from the pool.

    • A bit more complicated when scaling/upgrading real servers because needs to coordinate changes with director server because retry are done from client.

VS/TUN mode - TODO

Hardware based

  • F5 / A10:

    • Pros:

      • High performance. Could support 1000K concurrent connections

      • Stability: Tested thoroughly by producers

      • Security features such as DDoS protection, firewalls

    • Cons:

      • Low customization options

Typical architecture and metrics

Multi layer

  • Geo level - Use DNS load balancing.

  • Cluster level - hardware load balancing such as F5 among cluster level (a single device supports 2000K - 8000K QPS)

  • Within a cluster

    • (Optional) LVS - A single server could support 800K QPS. No need to introduce if QPS is lower than 100K.

    • Nginx - A single server could support roughly 50K QPS

Keepalived for high availability

  • Virtual IP: A floating IP will be shared between a active and many backup load balancers.

  • Use cases:

    • Popular in stateless scenarios such as load balancing (LVS, nginx).

    • Many modes such as master-slave, master-master mode.

  • Mechanism: Run a monitor scripts against target.

    • If run on network layer, then use ICMP to send a package to check whether it is responding.

    • If run on transport layer, then check whether specific TCP port is responding.

    • If run on application layer, then could run FTP/TELNET/HTTP/DNS protocols to customize some checking.

  • VRRP protocol will be used for failover and master election

  • Please refer to Keepalived and haproxy for more details.

Last updated