POR CASO DE USO
Load balancing is a core networking solution used to distribute traffic across multiple servers in a server farm. Load balancers improve application availability and responsiveness and prevent server overload. Each load balancer sits between client devices and backend servers, receiving and then distributing incoming requests to any available server capable of fulfilling them.
Explore additional load balancing topics:
Secure your apps with a zero-trust security solution
A load balancer may be:
Load balancers detect the health of backend resources and do not send traffic to servers that are not able to fulfill requests. Regardless of whether it’s hardware or software, or what algorithm(s) it uses, a load balancer disburses traffic to different web servers in the resource pool to ensure that no single server becomes overworked and subsequently unreliable. It effectively minimizes server response time and maximizes throughput.
The role of a load balancer is sometimes likened to that of a traffic cop, as it is meant to systematically route requests to the right locations at any given moment, thereby preventing costly bottlenecks and unforeseen incidents. Load balancers should ultimately deliver the performance and security necessary for sustaining complex IT environments, as well as the intricate workflows occurring within them.
Load balancing is the most scalable methodology for handling the multitude of requests from modern multi-application, multi-device workflows. In tandem with platforms that enable seamless access to the numerous applications and desktops within today’s digital workspaces, load balancing supports a more consistent and dependable end-user experience for employees.
Hardware-based load balancers work as follows:
In contrast, software-based load balancers:
Get an in-depth look at the must-haves for delivering fast, high-performance applications in the cloud.
An employee’s day-to-day experience in a digital workspace can be highly variable. Their productivity may fluctuate in response to everything from the security measures on their accounts to the varying performance of the many applications they use—an issue that can be worsened by poor responsiveness due to inadequate load balancing.
In other words, digital workspaces are heavily application-driven. As concurrent demand for software-as-a-service (SaaS) applications in particular continues to ramp up, reliably delivering them to end users can become a challenge if proper load balancing isn’t in place. Employees who already struggle to navigate multiple systems, interfaces, and security requirements will bear the additional burden of performance slowdowns and outages.
To promote greater consistency and keep up with ever-evolving user demand, server resources must be readily available and load balanced at Layers 4 and/or 7 of the Open Systems Interconnection (OSI) model:
Load balancing is more computationally intensive at L7 than L4, but it can also be more efficient at L7, due to the added context in understanding and processing client requests to servers. In addition to basic L4 and L7 load balancing, global server load balancing (GSLB) can extend the capabilities of either type across multiple datacenters so large volumes of traffic can be efficiently distributed without degradation of service for the end user.
As applications are increasingly hosted in cloud datacenters located in multiple geographies, GSLB enables IT organizations to deliver applications with greater reliability and lower latency to any device or location. Doing so ensures a more consistent experience for end users when they are navigating multiple applications and services in a digital workspace.
Round robin is a simple technique for making sure that a virtual server forwards each client request to a different server based on a rotating list. It’s easy for load balancers to implement, but doesn’t take into account the load already on a server. There is a danger that a server may receive a lot of processor-intensive requests and become overloaded.
More sophisticated than the least connection method, the least response time method relies on the time taken by a server to respond to a health monitoring request. The speed of the response is an indicator of how loaded the server is and the overall expected user experience. Some load balancers will take into account the number of active connections on each server as well.
Whereas round robin does not account for the current load on a server (only its place in the rotation), the least connection method does make this evaluation and, as a result, it usually delivers superior performance. Virtual servers following the least connection method will seek to send requests to the server with the least number of active connections.
A relatively simple algorithm, the least bandwidth method looks for the server currently serving the least amount of traffic as measured in megabits per second (Mbps). Similarly the least packets method selects the service that has received the fewest packets in a given time period.
Methods in this category make decisions based on a hash of various data from the incoming packet. This includes connection or header information, such as source/destination IP address, port number, URL, or domain name.
The custom load method enables the load balancer to query the load on individual servers via SNMP. The administrator can define the server load of interest to query—CPU usage, memory, and response time—and then combine them to suit their requests.
An ADC with load balancing capabilities helps IT departments ensure scalability and availability of services. Its advanced traffic management functionality can help a business steer requests more efficiently to the correct resources for each end user. An ADC offers many other functions (such as encryption, authentication, and web application firewalling) that can provide a single point of control for securing, managing, and monitoring the many applications and services across environments and ensuring the best end-user experience.
Citrix ADC goes beyond load balancing to provide holistic visibility across multi-cloud, so organizations can seamlessly manage and monitor application health, security, and performance.