@ Grid Operations Support Centre and UK National Grid Service What Next ? ? Neil Geddes GridPP15,Janury 2006
Download ReportTranscript @ Grid Operations Support Centre and UK National Grid Service What Next ? ? Neil Geddes GridPP15,Janury 2006
@ Grid Operations Support Centre and UK National Grid Service What Next ? ? Neil Geddes GridPP15,Janury 2006 Outline • Current Status • Plans for the coming year • NGS-2 • working with ? @ Introduction • The “Science and Innovation Investment Framework 2004-2014” – • a national e-Infrastructure (hardware, networks, communications technology) to provide “ready and efficient access to information of all kinds – such as experimental data sets, journals, theses, conference proceedings and patents”. Critical to successful collaborative, multi-disciplinary research and innovation – • ‘Over the decade many of the grand challenges in research will occupy the interfaces between the separate research disciplines developed in the 19th and 20th centuries… much more needs to be done, and by more players, if the UK is to achieve a global edge’. The NGS and GOSC explicitly addresses exactly these issues – Common interfaces and operational procedures provide the basis of • • • efficient sharing of resources Simple users access to an increasing range of resources key step towards a service economy for data and computation. The National Grid Service Launched April 2004 Full production - September 2004 Focus on deployment/operations Do not do development Responsive to users needs + Belfast, Westminster … NGS Users Number of Registered NGS Users 300 Number of Users 250 200 150 NGS User Registrations 100 Linear (NGS User Registrations) 50 0 14 January 2004 23 April 2004 09 01 August November 2004 2004 17 February 2005 Date 28 May 2005 14 05 September December 2005 2005 12 34 56 78 1190 1112 1134 1156 1178 2290 2212 23 2245 2267 2389 3310 3323 3345 3367 38 4490 4412 4434 4465 4478 5590 5512 543 5556 557 5689 6601 6623 6654 6667 6789 701 772 7734 7756 7778 8809 8812 8834 856 887 8989 9901 9923 9954 9967 119089 110001 110032 110045 110076 110189 111110 111123 11145 111167 1118 112290 112221 112234 112265 112278 13309 113312 1133 113345 367 Usage Statistics (Total Hours for all 4 Core Nodes) 1 250000 200000 150000 Hours User DN 2 100000 4 3 6 5 50000 0 Users (Anonymous) SRB Storage history for month prior to 31/08/05 Detailed information -> https://www.ngs.ac.uk/ops/gits/srb/srbreport.txt IB IB OU= OU=QUB OU=York OU=Westminster OU=Warwick OU=UCL OU=Southampton OU=Sheffield OU=Reading OU=QueenMaryLondon INRIA OU=Portsmouth OU=Oxford OU=OASIS OU=Nottingham OU=Newcastle OU=Manchester OU=Liverpool OU=Leeds OU=Lancaster OU=Imperial OU=Glasgow OU=Edinburgh OU=DMPHB OU=DLS OU=CPPM OU=CLRC OU=Cardiff OU=Cambridge OU=Bristol OU=Birmingham OU=BBSRC 50 O=universiteit-utrecht Users by institution Total Count of OU= 45 40 35 30 25 20 15 10 5 0 Users by “Research Council” 1000000 100000 160 CPU Time (Hrs) Total Count of "RC" 140 10000 1000 100 10 120 1 100 1 10 100 Storage (GB) 80 60 40 20 0 bbsrc cclrc epsrc nerc pparc "RC" AHRC mrc esrc 1000 P-GRADE NGS Portal http://www.cpc.wmin.ac.uk/ngsportal The P-GRADE NGS portal, operated by the University of Westminster, offers an alternative to the NGS Portal for executing and monitoring computational jobs on the UK National Grid Service. In addition, it enables the graphical development, execution and visualisation of workflows – composed of sequential and parallel jobs – on the NGS resources. Execute workflow Log in Visualise execution Create workflow Map execution http://www.cpc.wmin.ac.uk/ngsportal GEMLCA - Legacy Code Support for NGS Users http://www.cpc.wmin.ac.uk/gemlca • If you have a legacy application that you would like to make accessible for other NGS users utilise GEMLCA to: • upload your application into a central GEMLCA repository • make it available for authorised users through the P-GRADE NGS portal Browse repository Publish legacy code Create workflow http://www.cpc.wmin.ac.uk/gemlca GODIVA Diagnostics Study of Oceanography Data EVE Excitations and Visualisation Project Integrative Biology Simulation of Strain in Soft Tissue under Gravity PDB2MD: An automated pipeline performs molecular dynamics simulations on DNA crystal structures. Charlie Laughton School of Pharmacy University of Nottingham PDB database of DNA crystal structures 116d 180d 196d 1d56 AMBER running on NGS The results reveal biologically important patterns of sequencedependent flexibility. MD database & analysis PDB2MD: An automated pipeline performs molecular dynamics simulations on DNA crystal structures. Charlie Laughton School of Pharmacy University of Nottingham PDB database of DNA crystal structures 1ilc 116d 116d 180d 180d 196d 196d 1d56 AMBER running on NGS The results reveal biologically important patterns of sequencedependent flexibility. MD database & analysis Reality Grid Example BRIDGES GridBLAST Job Submission ScotGRID masternode NESC Grid Server (Titania) end user machine send job request GridBLAST client return result Apache Tomcat GT 3 core grid service ScotGRID worker nodes PBS server side + BLAST jobs farmed out to compute nodes BRIDGES MetaScheduler PBS wrapper Condor wrapper Condor + BLAST The BRIDGES project GT2.4 wrapper Micha Bayer NeSC-Glasgow NGS Condor Central Manager GT2.4 + BLAST Leeds headnode NESC Condor pool GT2.4 + BLAST execution hosts Oxford headnode execution hosts Roadmap • Goals : 1. 2. 3. 4. 5. • Support the current service functionality and user base Expansion to include new partners, Improve interoperability with EGEE, TeraGrid and DEISA Convergence with project/community/campus infrastructure Provision of value added services on top of the basic NGS infrastructure Specific Targets (services, operations, technology) 1. 2. 3. 4. 5. 6. 7. 8. 9. 10. 11. 12. Improved operational security and incident response procedures Deployment of resource broker ? Virtual Organisation Management Service (VOMS) Incremental support for the LCG Baseline Services. ☺ A co-scheduling system ? Job submission to the NGS through the emerging JSDL standard Support for access to a range of data storage services ☺ Resource accounting and improved grid account management Provision of a test system for dynamic service hosting. Services from the OMII managed programme. ☺ Initial integration with Shibboleth authentication ? First examples of “authorisation aware” generic services ? •What Next UK e-Infrastructure Regional and Campus grids HPCx + HECtoR Users get common access, tools, information, Nationally supported services, through NGS Community Grids Integrated internationally LHC VRE, VLE, IE ISIS TS2 • The Vision of the National Grid Service Integrated coherent electronic access for UK researchers to all resources and facilities that they require to carry out their research. – – – • Examples of the services: – – – – – – – – • From advanced real time facilities: real time instruments to historical data Supporting access to regional, national and international facilities, Integrating institutional or even departmental resources Location independent data storage and replication Location independent access to institutional repositories and certified long term archival Location independent access to local, regional, national and international data resources Access to local, regional, national and international computational resources Internationally recognized electronic identity and identity management tools Tools for managing collaborative project or Virtual Organisation based authentication and authorisation Co-scheduling and operation of a wide range of national and international resources. Tools to support distributed collaborative working In addition the GOSC will support – – – – – 24 hour monitoring of the UK’s grid infrastructure The policy framework for operations Review of partner services. A central UK help desk and support centre A repository of user and system documentation, Next 3 years • The GOSC and NGS development will lead to: – An expanded National Grid Service for UK research • partners and affiliates • range of services provided. – Robust NGS operations and management procedures. – Interoperability with other national and international einfrastructures – Integration of key data sources and data services including • Data Centre - Edina, Mimas, AHDS, • Facilities - LHC, DIAMOND and ISIS. – – – – Improved and measured NGS reliability Supported scientific research over next 3-5 years Value added services deployed on the basic NGS infrastructure. Detailed service definitions for sustainable infrastructure •In more detail… Support Centre • Front line e-infrastructure support for UK academic community. • Authentication framework required by national and international agreement. • Infrastructure services – authentication, VO/authorisation, information, monitoring, resource brokering • Support interfaces with partner infrastructures – within the UK: e.g. MIMAS, Edina, AHDS, GridPP, e-minerals, and others – Internationally: e.g TeraGrid, EGEE • Website, including documentation, training materials and related links • Work with NeSC and other organisations to provide training • Coordinate development and deployment programme for NGS. – Development expertise • User management for the NGS • Partner and affiliate management for the NGS, National and International Facilities • UK gateway: – Regional, National and international HPC facilities • – Advanced national experimental facilities • – – – – HECToR, TeraGrid and DEISA. DIAMOND, ISIS and the National Crystallographic Service. National, regional and institutional data centres. Campus grid developments Grid infrastructures Grid Interoperability workshop at SC05 • • • EGEE, TeraGrid, OSG, NGS, Naregi, IPAC First of a regular series – next meeting at GGF in March Some key proto-agreements – • JDSL, SRM, GridFTP, GSI and VOMS,“GLUE” (and CIM),RU records GOSC will – – – – Integrate access to data from experimental facilities. Support collaborative activities spanning each (or all) of these infrastructures Where possible, deploy a robust cross-service scheduling system Common tools and infrastructures for data handling, organisation, manipulation and creation of user interfaces. Grid Interoperation • Leaders from TeraGrid, OSG, EGEE, APAC, NAREGI, DEISA, Pragma, UK NGS, KISTI will lead an interoperation initiative in 2006. • Six international teams will meet for the first time at GGF-16 in February 2006 – Application Use Cases • (Bair/TeraGrid, Alessandrini/DEISA) – Authentication/Identity Mgmt • (Skow/TeraGrid) – Job Description Language • Newhouse/UK-NGS – Data Location/Movement • Pordes/OSG – Information Schemas • Matsuoka/NAREGI – Testbeds Leaders from nine Grid initiatives met at SC05 to plan an application-driven “Inerop Challenge” in 2006. • Arzberger/Pragma Integration of Computational and Data Resources • Work with major and representative data providers. – Provide a common gateway as the basis for continued developments – Progressively more of the data providers should be incorporated • • • To develop the necessary integration services the GOSC will work with research communities that require data integration, for example, environmental, biomedical and socio-economical research. The NGS will support high-level tools, driven by metadata and abstractions that suit the research disciplines, working in conjunction with OMII-UK and international tool developers to promote and understand these developments. The MIMAS and EDINA data centres are already, or will soon be, funded explicitly to “grid enable” some of their resources. GOSC will work closely with these data centres to ensure that these developments and compatible with and integrated into the NGS services. Core NGS nodes • Core NGS nodes – – – – – – – Provide a controlled test environment for new services Provide a high level of, professionally managed, central services Facilitate the rapid technical development of the NGS infrastructure Provide a common infrastructure and proving ground for new users Drive a technical agenda, developing experience and best practice Represent a neutral core around which partners can converge Provide resources to guarantee access for new grid users • Positioned between HPC and campus grid • Upgrade existing core nodes in 2007 Accounting and Charging • Accounting and metering is already important for the NGS – Monitoring of the infrastructure is important for operations – Monitoring of service availability/performance for partnership – Monitoring and accounting of service/resource usage • Effective distributed accounting/metering provides the basis by which partners can have confidence in resource sharing and in additional resource provision and consumption • This work will continue and the infrastructure to support this will continue to be developed. • NGS must, at very least, provide a mechanism for parters to “trade” • FEC currently a complication – Partners must recover their investment (loan) – Users have no “money” – NGS sustainability Integration with Computing Service Providers • Factors for closer integration with computing service providers are: – – – – – Well defined interfaces and procedures Clear understanding of the roles and commitments A well understood core software stack with components of known provenance Effective metering and auditing of services and users Integration of the NGS AAA framework with the standard national and campus systems – Training, documentation and awareness raising targeted at computing services • • • The role of UKERNA is key here. UKERNA has extensive experience in dealing with the service providers in question and excellent connections into these institutions. GOSC will work with UKERNA and other key groups such as UCISA and RUGIT As the number and integration of partners grows, need to integrate better with the UCISA community Outreach and Training • UK needs training as – – – – – Support for decision makers Support for e-Infrastructure providers Support for users Support for application, portal and tool developers Support for educators • Essential to coordinate and leverage work of other bodies • JISC, RIN, NeSC, EGEE … OMII • The initial OMII releases were of limited value to NGS – • OMII 2.0 has more interesting things – • GOSC fully support these goals NGS will: – – – – – – – – • OGSA-DAI and myGrid OMII-UK goal is to supply interoperable Web Services (Linux and Microsoft platforms) – • OGSA-DAI WS-I, GridSAM OMII-UK brings in more – • experimental deployments of OMII software have been installed on the NGS since November 2004. Continue its deployment of OGSA-DAI on the data nodes Provide access to the NGS compute resources through the GridSAM Deploy the Resource Usage Service In addition GOSC and OMII will work together to Provide services that can be integrated into workflows controlled through (OMII-UK) tools Improve management of web services. Improve accounting for grid service use Integrate Shibboleth and other VO tools used by GOSC. Interoperability of OMII software with existing and likely future NGS infrastructure remains a key requirement. Next 2-3 Years Provisional Budget Service GOSC/NGS Management GOSC Core NGS Nodes Data Integration Development Community Support Total See Note Annual cost 1 £140k 2 £985k 3 £240k 4 £120k 5 £510 6 £1995 3 year cost £420k £2955k £3120k £360k £1530 £8385 • Notes: 1. PI time (20%), Director (100%), Technical Director (50%), technical administration (100%) and secretarial support (100%). All GOSC roles except the explicit support for core NGS nodes, data service integration, community gateways and management. A full cost assuming 4 nodes @£600k each plus 3 years system administration effort for each. Data centre integration expertise, matching computational expertise form the Core NGS nodes. Funding for “application” or “community” gateway activities. It is essential that some form of activity like this be supported in some way. UKERNA work on networking, security and operational best practice is not explicitly included above. 2. 3. 4. 5. 6. Relation to GridPP • NGS and GridPP partners in EGEE • Joining NGS brings generic user and admin support • “GridPP” not a partner in NGS, but expect individual GridPP sites to be partners or affiliates – Build and support broader communities locally – LCG(UK) eventually a logical view of the UK grid • Partners can offer a wide range of services