Load balancing is a technique that ensures an organization’s server does not get overloaded with traffic. With load balancing measures in place, workloads and traffic requests are distributed across server resources to provide higher resilience and availability.
In the earliest days of the Internet, it became painfully obvious that a single application on a single server couldn’t handle high-traffic situations. Concurrent server requests from large volumes of traffic would frequently overwhelm a single server regardless of how powerful the underlying infrastructure was. A solitary instance of application availability was—and still is, in cases where load balancing isn’t implemented—a single point of failure. This poses a huge threat to a system’s reliability.
A typical load balancing sequence works as follows:
It may seem obvious, but it’s worth noting that the steps above can only be completed if there are multiple resources (either server, network, or virtual) that have already been established. Otherwise, if there is only a single server or compute instance, all of the workloads are distributed to the same place and load balancing is no longer necessary.
Recommended: What is Failover Clustering?
A load balancer acts like a traffic cop or a filter for the traffic coming over the Internet. It prevents any one server from being overloaded and becoming unreliable. Therefore, every node is able to work more efficiently.
In recent years, load balancing has become a feature of a broader class of technology known as Application Delivery Controllers (ADC). ADC aims to provide multiple advanced load balancing features to ensure workload balancing, along with an overall high quality of application delivery.
In addition to preventing any one resource from becoming overwhelmed and unreliable, load balancing has benefits for security and productivity. ADCs are often used as appliances and control points for security, which helps to protect against various types of threats including Denial of Service (DOS) attacks. Load balancing also often involves duplicating content and application workloads, which allows for more than one copy of a resource to be accessed at a time.
Recommended: How to Prevent DoS Attacks
Load balancing can be accomplished using either hardware or software. Both approaches have their benefits and drawbacks, as illustrated in the table below. Check out our lineup of the Best Load Balancers for 2021 to figure out which hardware or software load balancer is the right fit for you.
Hardware load balancers | Software load balancers | |
Performance | Higher | Lower |
Flexibility | Less flexible | More flexible |
Virtualization | Built-in | Happens externally |
Architecture | Best for multi-tenant, occupies more physical space | Best for individual tenant, requires no physical space |
Cost | Higher investment and maintenance costs | Lower costs overall |
Configuration | Less configurable | Highly configurable |
Recommended: Does Virtualization’s Success Spell Decline for Server Sales?
There are several types of load balancing configurations that you might choose to deploy depending on the features that are most important to you. You might also decide to layer multiple configuration types as part of a load balance or ADC appliance. Load balancing categories include server load balancing, network load balancing, global server load balancing (GSLB), container load balancing, and cloud load balancing.
With server load balancing, the goal is to distribute workloads across server resources based on availability and capabilities. Server Load balancer configurations tend to rely on application layer traffic to route requests. Server load balancing is also sometimes referred to as Layer 7 load balancing as it makes use of application layer traffic.
Network load balancing distributes traffic flow across IP addresses, switches, and routers to maximize utilization and availability. These types of configurations are made at the transport layer and data traffic. Network load balancing is also sometimes referred to as Layer 4 balancing.
In Global Server Load Balancing, an operator handles workload balancing across a globally distributed set of Layer 4 and Layer 7 loads. In a GSLB deployment, there are also typically ADC assets at the global level, as well as the local level where the traffic is finally delivered.
Container load balancing provides virtual, isolated instances of applications and is also enabled via load balancing clusters. Among the most popular approaches is the Kubernetes container orchestration system, which can distribute loads across container pods to help balance availability.
Within a cloud infrastructure, there are often multiple options for load balancing across compute instances. Load balancing in the cloud can also include both network (layer 4) and application (layer 7) balancing.
Depending on the load balancing configuration(s) being used from the list above, there are a variety of techniques that determine how a set of workloads are balanced. These include:
Property of TechnologyAdvice. © 2025 TechnologyAdvice. All Rights Reserved
Advertiser Disclosure: Some of the products that appear on this site are from companies from which TechnologyAdvice receives compensation. This compensation may impact how and where products appear on this site including, for example, the order in which they appear. TechnologyAdvice does not include all companies or all types of products available in the marketplace.