Transcript Load balancing - Princeton University
Scalable Server Load Balancing Inside Data Centers
Dana Butnariu Princeton University Computer Science Department July – September 2010 Joint work with Richard Wang and Professor Jennifer Rexford
Client3 Client4 Client2
Getting started
Client1 Client5
The Internet
Server1 Server2
Switch
Server4 Server3 Client6
Getting started
• • • • • •
Client
– any device requesting a web service
Server
– device which handles a client request and provides the web service
Data center
– location containing a group of servers
Server load
– number of client requests a server must handle
Load balancing
– directing a client request to a particular server, managing server loads according to a certain algorithm
Switch
– device which enables client and server communication
Energy and Load balancing
• Servers: Are located in data centers in different areas of the world.
Energy cost and availability varies from one location to the other.
Energy cost and consumption depends on client – server distance.
• Load balancing: Tries to lower the energy cost and usage without affecting user perceived performance.
Can achieve this goal by selecting close by data centers.
Can achieve this goal by using only certain servers and powering down the rest.
Load balancing today
• •
Old approach
– have a separate device, the load balancer.
New approach
– implement the load balancing in devices already existent in the network. •
Old approach
: Costly device Consumes energy Hard to program Crashes easily •
New approach
: Already existing device No additional costs Easy to program and customize Stable and reliable
What, why and how?
•
What
: Scalable Server Load Balancing without sacrificing user perceived performance. •
Why
: To save energy and lower the cost of energy used to process client requests.
•
How
: Using a new emerging technology called OpenFlow which enables switch programming.
Project Steps
•
Establish
the network design.
•
Design
the load balancing application.
•
Implement
the load balancing application.
•
Test
the load balancing application.
Network Design
• Establish the network design.
How many clients, servers, switches?
How are they connected?
What knowledge do they have of one another?
• There is a “Brain”.
It is just another computer.
It controls switch behavior.
It installs rules in the switch.
Rules tell switch which server handles a client request.
Clients
Network Design
Brain implementing design algorithm Client request Send request to servers Switch Install Load Balancing rules Server 1 Server 2 Server 3 Server 4 Data Center
How it works
• Client sends request for web service.
• Request arrives at switch.
• Switch decides server to handle request.
• Decision is based on: Closest server so that less energy is used for transport Cheapest path in terms of energy cost Server usage so that less used servers can be powered down • Server sends a reply back to the client. • Client is provided with the web service.
Load Balancing Components
Partitioning Class Old server still handles old client requests Transitioning Class Install rules New servers handle new client requests Load Balancing Switch Decide servers to be powered down Statistics Class Monitor server usage
Application Components
• Partitioning: Responsible for implementing the Load Balancing algorithm Decides which server handles which client request • Transitioning: Ensures that when a server is powered down all new client requests are handled by another server Ensures that all old client requests are answered by the old server • Statistics: Provides statistics regarding server usage
Prototype Implementation and Future Steps
• The first prototype for this project was created using: OpenFlow programmable switches.
OpenFlow, NOX and Mininet to program the switches.
Applications written in Python.
Applications tested using VMWare (Virtual Machine) and Debian version of Linux running on VMWare.
• Future steps: Running the application on a web server and on a real network.
Designing a more accurate partitioning component.
Adapting the partitioning component according to the statistics component.
Conclusions
• New solution which saves energy by: Being implemented in already existing hardware, the switches.
Finding the path which uses the lowest amount of energy at the lowest cost.
Turning down severs which are handling a small amount of client requests.
• Solution offers: flexibility due to the software component – the Load Balancing algorithm can be easily modified.
speed due to the underlining hardware component – switch which applies rules.
Acknowledgements
Thank you: Professor Jennifer Rexford Richard Wang – team member Rob Harrison and David Shue – graduate students Nate Foster – postdoctoral research fellow