Load balancer
Use cases
Decoupling
Hidden server maintenance. You can take a web server out of the load balancer pool, wait for all active connections to drain, and then safely shutdown the web server without affecting even a single client. You can use this method to perform rolling updates and deploy new software across the cluster without any downtime.
Seamlessly increase capacity. You can add more web servers at any time without your client ever realizing it. As soon as you add a new server, it can start receiving connections.
Automated scaling. If you are on cloud-based hosting with the ability to configure auto-scaling (like Amazon, Open Stack, or Rackspace), you can add and remove web servers throughout the day to best adapt to the traffic.
Security
SSL termination: By making load balancer the termination point, the load balancers can inspect the contents of the HTTPS packets. This allows enhanced firewalling and means that you can balance requests based on teh contents of the packets.
Filter out unwanted requests or limit them to authenticated users only because all requests to back-end servers must first go past the balancer.
Protect against SYN floods (DoS attacks) because they pass traffic only on to a back-end server after a full TCP connection has been set up with the client.
Load balancing algorithms
Round-robin
Def: Cycles through a list of servers and sends each new request to the next server. When it reaches the end of the list, it starts over at the beginning.
How to support sticky sessions: Hashing based on network address might help but is not a reliable option. Or the load balancer could maintain a lookup table mapping session ID to server.
Weighted round robin
Problems of round robin:
Not all requests have an equal performance cost on the server. But a request for a static resource will be several orders of magnitude less resource-intensive than a requst for a dynamic resource.
Not all servers have identical processing power. Need to query back-end server to discover memory and CPU usage, server load, and perhaps even network latency.
Least load first (from server perspective)
Problems of weighted round robin:
Server might be under different status even given the same type of requests
How to define least load:
For a layer 4 load balancing option such as LVS, it could load balance based on the number of connections.
For a layer 7 load balancing option such as Nginx, it could load balance based on the number of Http requests.
More customized criteria such as CPU load, I/O load.
Best performance first (from client perspective)
Response time
Source hashing
Hash of IP address
Hash of session id
Categorization
Http redirect based load balancer (rarely used)
Steps:
Client's requests first reach a load balancing server which translate original target IP address A to a new target IP address B with a 302 HTTP response code
Client issues another request to the new target IP address B
Pros:
Easy to implement
Cons:
Client needs to have two http requests to data center.
Internal web/application servers' IP address will be exposed to external world and cause potential security risks.
Compared with internal servers, load balancing servers will have stricter firewall policies and security configurations.
Due to the security risks and performance cost, Http redirect based load balancing is rarely used in practice.
There is a flow chart Caption in Chinese to be translated
DNS based load balancer
Steps:
Client's requests first reach DNS authority server to get the IP address.
Client issues another requests to the parsed IP address.
Pros:
Low latency: DNS will be resolved for the first time. The second time when a request come it will use the cached version.
It is a basic config service provided by DNS providers so no development will be required. Many DNS service provider offers geographical DNS service.
Cons:
Low accuracy: DNS will be only resolved one time and cached for sometime. The cached value will be used for a relatively long time.
High update time: DNS has multi-layer caches. Even after scaling up, the old DNS record might still point to the old IP address.
Not customizable: DNS load balancer is controlled by DNS service providers.
Simple load balancing algorithm: DNS load balancing algorithms are relatively simple. For example, it could not make decisions based on the differences of servers.
There is a flow chart Caption in Chinese to be translated
Example DNS record for using round robin load balancing on top of DNS
HTTP-DNS based load balancer
Due to the limitation of traditional DNS, many companies will develop HTTPDNS. For more details for how HttpDNS work, please see HTTPDNS
It is mainly used in App scenarios. For web scenarios, since url parsing is done by browser, it will be more compliacted.
Pros:
Shorter update time. It won't be impacted by the multi-layer cache.
More customized load balancing algorithm.
Cons:
Needs customized development and has high cost.
Application layer (e.g. Nginx, HAProxy, Apache)
Pros:
Could make load balancing decisions based on detailed info such as application Url.
Only applicable to limited scenarios such as HTTP / Email which sit in level 7.
Cons:
Low Performance (A single server could support roughly 50K QPS) when compared with LVS (Layer 4 - Transport layer software, A single server could support 800K QPS) or F5 (Layer 4 hardware, a single device supports 2000K - 8000K QPS)
Security: Usually don't have security features built-in place
Network/Transport/DataLink layer (e.g. Nginx Plus, F5/A10, LVS, HAProxy)
Software based
LVS
LVS supports three modes of operation: VS/NAT, VS/TUN, VS/DR. It supports the following type of load balancing algorithms:
Round robin
Weighted round robin
Least connections
Weighted least connections
Locality-based least connections
Locality-based least connections with replication
Destination hashing
Source hashing
VS/NAT mode
Steps:
Client's requests first reach IP load balancer.
IP based load balancer changes the target IP address to internal servers' IP address, and change source IP to be load balancer's IP address (SNAT)
Internal servers return response to IP based load balancer.
IP based load balancer changes the target IP address.
Pros:
Operates on network layer and has much more efficiency compared with reverse proxy
Cons:
All requests/Responses will require IP based load balancer to replace the target IP address. It will become the bottleneck really easy.
There is a flow chart Caption in Chinese to be translated
VS/DR mode
How it works:
A same virtual IP address is bond to both director server and real server.
When ARP for a IP address, director server's address will be returned and real server's address will be hidden.
Steps:
Client's requests first reach director server. For illustration, let's say within the request, the source IP is CIP, the dest IP is VIP; the source MAC is CMAC, the dest MAC is DMAC.
Director server will keep the source and dest IP unchanged. Director server will change the target MAC address DMAC to internal servers' MAC address RMAC.
Internal servers return response to client.
Pros:
High performance and could be used broadly
Since there are no changes to IP address, internal servers could directly return the response to clients.
Operates on the data link layer (only need to change the MAC address)
Cons:
Relies on client to retry because the response does not pass through director server. Even when a real server is down, there might still be gap until director server could remove the problematic real server from the pool.
A bit more complicated when scaling/upgrading real servers because needs to coordinate changes with director server because retry are done from client.
VS/TUN mode - TODO
Hardware based
F5 / A10:
Pros:
High performance. Could support 1000K concurrent connections
Stability: Tested thoroughly by producers
Security features such as DDoS protection, firewalls
Cons:
Low customization options
Typical architecture and metrics
Multi layer
Geo level - Use DNS load balancing.
Cluster level - hardware load balancing such as F5 among cluster level (a single device supports 2000K - 8000K QPS)
Within a cluster
(Optional) LVS - A single server could support 800K QPS. No need to introduce if QPS is lower than 100K.
Nginx - A single server could support roughly 50K QPS
Keepalived for high availability
Virtual IP: A floating IP will be shared between a active and many backup load balancers.
Use cases:
Popular in stateless scenarios such as load balancing (LVS, nginx).
Many modes such as master-slave, master-master mode.
Mechanism: Run a monitor scripts against target.
If run on network layer, then use ICMP to send a package to check whether it is responding.
If run on transport layer, then check whether specific TCP port is responding.
If run on application layer, then could run FTP/TELNET/HTTP/DNS protocols to customize some checking.
VRRP protocol will be used for failover and master election
Please refer to Keepalived and haproxy for more details.
Last updated