Transport Protocols and Dynamic Provisioning for Large-Scale Science Applications Nagi Rao (Nageswara S.V.
Download ReportTranscript Transport Protocols and Dynamic Provisioning for Large-Scale Science Applications Nagi Rao (Nageswara S.V.
Transport Protocols and Dynamic Provisioning for Large-Scale Science Applications Nagi Rao (Nageswara S.V. Rao) Computer Science and Mathematics Division Oak Ridge National Laboratory [email protected] Report of the “DOE Workshop on Ultra High-Speed Transport Protocols and Dynamic Provisioning for Large-Scale Science Applications” April 10-11, 2003, Argonne, IL http://www.csm.ornl.gov/ghpn/wk2003.html Panel on Future Directions in Networking International Conference on Network Protocols November 5-7, 2003, Atlanta, GA We engineered the Internet, and it works fine for e-mail and web; but to do “world-class” scientific research needed in DOE scientific applications, we need to develop a science of networking that delivers usable performance to the applications Allyn Romanow, Cisco Systems Outline Introduction Organization Details Provisioning Group Notes Transport Group Notes Dynamics and Stabilization of Network Transport Conclusions Opinions expressed in this presentation belong to the author and are not necessarily the official positions of US Department of Energy, Oak Ridge National Laboratory or UT-Battelle LLC. Networking for DOE large-science applications Next generation of DOE scientific breakthroughs critically depend on large multi-disciplinary geographically dispersed research teams: – high energy physics, climate simulation, fusion energy, genomics, astrophysics, spallation neutron source, and others These applications are Inherently distributed in: Data – archival or on-line Computations – supercomputers or clusters Research teams – experts in different domains Experimental facilities – one of the kind user facilities – they all need to be seamlessly networked DOE Large-Scale Science Applications and Numerous Other Science Applications - need extreme and acute networking Science Areas Current 5 years 5-10 Years General Remarks End2End End2End End2End Throughput Throughput Throughput High Energy Physics 0.5 Gbps E2E 100 Gbps E2E 1.0 Tbps high throughput Climate Data & Computations SNS NanoScience 0.5 Gbps E2E 160-200 Gbps n Tbps high throughput does not exist 1.0 Gbps steady state Tbps & control channels remote control & high throughput Fusion Energy 500MB/min (Burst) 1TB/week 500MB/20sec (burst) N*N multicast n Tbps time critical transport 1TB+ & stable streams computational steering & collaborations 1TB/day 100s users Tbps & control channels high throughput & steering Astrophysics Genomics Data & Computations Detailed account of the needs were identified and discussed at DOE High-Performance Network Planning Workshop, August 13-15, 2002, http://DOECollaboratory.pnl.gov/meetings/hpnpw Astrophysics Computations • Science Objective: Understand supernova evolutions – Teams of field experts across the country collaborate on computations • • – Massive computational code • • • • Experts in hydrodynamics, fusion energy, high energy Universities and national labs Terabyte in days are generated currently Archived at nearby HPSS Visualized locally on clusters – only archival data Desired capability – – – Archive and supply massive amounts of data Collaboratively visualize archival or on-line data Monitor, visualize and steer computations into regions of interest Visualization channel Control channel Genomics Networking Needs • Data Movement Operations – Experimental and computational data • Stored across the country • Terabytes of data per day – Between users, archives and computers • Molecular Dynamics Computations – Supercomputers or clusters – Monitor, visualize, and steer computations data channel visualization channel steering channel Neutron Facilities – SNS, HFIR • Experimental Setups and Monitoring of Expensive Facilities (SNS – billion$) – Setup parameters and start experiments – Adjust parameters as needed; stop when necessary • Data Movements – Archive and access massive amounts of experimental data Current Network Capabilities: Transport and Provisioning DOE faces unique or acute challenges: Small user base with extreme needs – large data transfers at application-level – rates much higher than current backbones – highly controlled end-to-end data streams – unprecedented agility and stability – capabilities must be available to science users – not just to network experts with special networks Commercial and other networks will not adequately meet these acute requirements – Not large enough user base – Very limited business case New advances in Transport and Provisioning hold enormous promise, if suitably fostered and integrated – Flexible and powerful routers/switches, ultra high bandwidth links, new transport protocols can get us partway there – But, need several critical technologies and expertise: • end-to-end dynamic provisioning of paths with guaranteed performance • transport methods that optimally provide to user applications Workshop Goal • Address the research, design, development, testing and deployment aspects of transport protocols and network provisioning as well as the application-level capability needed to build operational ultra-speed networks to support emerging DOE distributed large-scale science applications over the next 10 years. Workshop Focus: • Ultra High-Speed Networks to support DOE Large-Science Applications – not a general network research workshop addressing Internet problems • Formulate DOE roadmap in the specific areas: – Transport and Provisioning • two very critical subareas of network research needed to meet DOE large-science requirements • Not in other areas such as security, wireless networks • “Working” workshop – Discussions on very specific problems, methods, potential solutions in transport and provisioning areas – Very short introductory presentations – Not just primarily informational or educational Participants Balanced participation from universities, industry and national laboratories to represent the needs, technologies, research and business aspects Total: 32 National Laboratories: 10 ORNL:3; ANL:2: LANL:2; PNNL:1; SLAC:1; ESnet: 1 Universities: 11 UMass; GaTech(2), Uva, UIC,Indiana U, U Va, U Tennessee, UC Davis, PSC, CalTech Industry:8 Celion, Cienna, Cisco, Juniper, Level3, Lightsand, MCNC, Qwest DOE Headquarters: 3 Working Groups: Provisioning: 14 Transport:15 Summary: Provisioning for DOE Large-Science Networks Need Focused Efforts in Develop a scalable architecture for fast provisioning Circuit Switched Network Build an application-centric circuit-switched cross country test-bed Coordination and Graceful integration with Applications and Middleware Transport and OS Developers Legacy and evolutionary networks Provisioning Recommendations • Recommendation 1: Agile Optical Network infrastructure: – A scalable architecture for fast provisioning of circuit switched dedicated channels specified on-demand by the applications. • Recommendation 2: Hybrid Switched Networks: – High capacity (Tbps) switchable channels for Petabyte data transport, a combination of requirements to accommodate burst, real-time streams as well as lower priority traffic, multi-point or shared use, for large file and data transfers, and for low latency and low jitter. • Recommendation 3: Dynamically Reconfigured Channels: – Provisioning of dynamically specified end-to-end quality paths for computational steering and time-constrained experimental data analysis. • Recommendation 4: Multi-Resolution Quality of Service: – Channels with various types of Quality of Service (QoS) parameters must be supported at various resolutions using GMPLS, service provisioning and channel sharing technologies. • Recommendation 5: Experimental Test-Bed Provisioning: Barriers • • • • • • • • • • • Limited deployment of ultra-long haul DWDM links; Lack of support for striped/parallel transport both at the core and application levels; Lack of high-speed circuit-switched infrastructure with network control-plane design and synchronous NICs with high-speed and on-demand reconfigurability; and Lack of well-developed methods and application interfaces for scheduling/reserving, allocation, and initiation. DOE applications do not follow the commercial scaling model of large number of users each with smaller bandwidth requirements; Lack of security paradigms for dedicated paths and the infrastructure that to manage them; Lack of a robust multi-cast solution efficiently supported on dedicated channels; High cost of equipment, including the costs of links, routers/switches and other equipment as well as deployment and maintenance; Lack of field-hardening of optical components such as memory/buffer, high-speed switches, Reamplification, Reshaping and Retiming (RRR) equipment, and lambda conversion gear; Lack of effective contention resolution methods for the allocation of channel pools; and Limited interoperability with other data networks, particularly legacy networks. Summary: Transport for DOE Large-Science Networks Current transport methods are massively inadequate – Unattained Throughput: wizards can achieve several Gbps for certain durations • But throughputs are needed at application user level – Cannot provide sustained and stable streams for control operations – TCP has complicated dynamics – hard to use in finer control operations Need focused efforts in: – Optimal transport methods to exploit provisioning to meet requirements • Transport -Tbps throughputs • Support stable and agile control channels – Comprehensive theory of transport: synergy and extensions of traditional disciplines • Stochastic control, non-linear control, statistics, optimization, protocol engineering – Strict algorithmic design • Modular , autonomic, adaptive, composable Integration and interactions: – DOE deployment, wider adoption, legacy integration – Experiment and test-bed – Instrumentation and Diagnostics tools • web100/Net100 • Statistical Inference and optimized data collection Transport Group Notes Scientists (must) view the network as they view a computer as a resource But they are becoming (not always willingly) network experts – “wizard gap” at all levels • but “gray matter” tax must be low Advance the state of network protocols to make them plug-andplay for the application users - need significant effort Time-to-solution in networking area is currently too high – TCP tuning for Gbps throughputs took years Peak is not enough – need sustained throughputs at application level Transport Group Recommendations: 1-5 years • Recommendation 1: Transport Protocols and Implementations – • Recommendation 2: Transport Customization and Interfacing – – • Stochastic control theoretic methods to design protocols with well-understood and/or provable stability properties. Recommendation 4: Monitoring and Estimation Methods – • Transport methods optimized to single and multiple hosts as well as channels of different modes. Transport methods suitably interfaced with storage methods to avoid impedance mismatches that could degrade the end-to-end transport performance. Recommendation 3: Stochastic Control Methods – • Transport methods for dedicated channels and IP networks for achieving high throughput, steering and control. The transport methods include TCP-, UDP- and SANbased methods together with newer approaches. Monitoring and statistical estimation techniques to monitor the critical transport variables and dynamically adjust them to ensure transport stability and efficiency. Recommendation 5: Experimental Test-Bed Transport Group Recommendations: 5-10 years • Recommendation 1: Modular Adaptive Composable and Optimized Transport Modules: – • Recommendation 2: Stochastic and Control Theoretic Design and Analysis: – • Stochastic control theoretic methods for composable transport methods to analyze them as well as to guide their design to ensure stability and effectiveness Recommendation 3: Graceful Integration with Middleware and Applications: – – • Highly dynamic and adaptive methods to dynamically compose transport methods to match the application requirements and the underlying provisioning. Application data and application semantics must be mapped into transport methods to optimally meet application requirements boundary between middleware and transport must be made transparent to applications. Recommendation 4: Vertical Integration of Applications, Transport and Provisioning: – Vertical integration of resource allocation policies (cost and utility) with transport methods to present a unified view and interface to the applications. Science of High-Performance Networking There is a need for systematic scientific approaches to the design, analysis and implementation of the transport methods and to network provisioning. • On-Demand Bandwidth and Circuit Optimization: – – – – • Comprehensive Theory of Transport: – – – – – – • Rigorous transport design methods tailored to the underlying provisioning modes. A synergy and extensions of a number of traditional disciplines. New stochastic control methods may be required to design suitable transport control methods. Non-linear control theoretic methods to analyze delayed feedback. Statistical theory for designing rigorous measurements and tests. Optimization theory to obtain suitable parameters for tuning protocols. Strict Algorithmic Design and Implementation: – – • Dynamic optimization and scheduling methods to allocate the bandwidth pipes to applications. A comprehensive approach for on-line estimation and allocation of the “bandwidths” Signaling to provide the required timeliness and reliability of the allocated channels. Scientific, systematic understanding to integrate the components for bandwidth allocation, channel scheduling, channel setup and teardown, and performance monitoring. Strict algorithmic design methods to efficiently implement the designed protocols. Implementations must be modular, autonomic, adaptive, and composable. Statistical Inference and Optimized Data Collection: – – – Due to the sheer data volumes, it is inefficient to collect measurements from all nodes all the time for the purposes of diagnosis, optimization and performance tuning. Systematic inferencing methods to identify the critical and canonical sets of measurements needed. Statistical design of experiments to ensure that the measurements are strategic and optimal. High-Performance Network Test-beds: Recommended by both groups State-of-the-art Components: software and hardware networking components, including routers/switches, high bandwidth long-haul links, protocols and application interface modules. Integrated Development Environments: mechanisms to integrate a wide spectrum of network technologies including high throughput protocol, dynamic provisioning, interactive visualization and steering, and high performance cyber security measures Smooth Technology Transition: transition of network technologies from research stages to production stages by allowing them to mature in such an environment. Characteristics of ultra high-speed network test-bed: 1. Interconnection of at least three science facilities with large-scale science applications; 2. Geographical coverage adequate to capture optical characteristics, transport protocols dynamics, and application behaviors comparable to that of real-word applications; 3. Integration with appropriate middleware; 4. Scalable network measurement tools; and 5. Well-defined technology transfer plan. Integration, Interaction, and Interfacing Applications are empowered to “tune” the network Network-Aware Applications Molecular dynamics visualization Application 1 Application 2 Application 3 HEP data transfers Middleware Protocols UltraNet Stabilize module Net100 modules IP provisioning Control modules Non-TCP protocols Dynamic lambda switching Supernova: Large data stream Control stream Network Security Issues While not on the original agenda, security issues have significant impact on application performance – DOE sites have very strict firewalls • Securing Operational and Development Environments: – – – • authentication, validation and access controls data speeds of multiple tens of Gbps or higher new security methods for on-demand dedicated channels. Effects of Security Measures on Performance: – – • impact of security measures on application performance. graceful interoperation of science applications under secured network environments. Proactive Countermeasures: – – protect bandwidth allocation, and signaling to setup and tear down the paths vulnerability of new transport protocols to certain attacks