News from the EB & LCG Nick Brook University of Bristol • EB News • LCG News • Structures • Review of RTAGs 16th Sept’02 Nick Brook –
Download ReportTranscript News from the EB & LCG Nick Brook University of Bristol • EB News • LCG News • Structures • Review of RTAGs 16th Sept’02 Nick Brook –
News from the EB & LCG Nick Brook University of Bristol • EB News • LCG News • Structures • Review of RTAGs 16th Sept’02 Nick Brook – University of Bristol 1 EB News Roger Barlow re-elected as deputy chair of the EB • will take over as chair September’03 Reporting structures in place • Measures manpower effort and deadlines but also requests expt reqts of UK (Tier1/A) resources • First reports appear for Q2’02 – available on EB web page • Next submissions are due now 16th Sept’02 Nick Brook – University of Bristol 2 EB News 6 application submissions to Sheffield All-Hands conference • Applications are beginning to deliver their projects • New format to GridPP expt sessions All but one application position is now filled (last remaining vacant post will be filled beginning of October) Successful joint ATLAS-LHCb workshop at Coseners • http://www.phy.bris.ac.uk/research/pppages/LHCb/coseners/CosenersHouse.htm 16th Sept’02 Nick Brook – University of Bristol 3 Fundamental Goal of the LCG To help the experiments’ computing projects get the best, most reliable and accurate physics results from the data coming from the detectors Phase 1 – 2002-05 prepare and deploy the environment for LHC computing Phase 2 – 2006-08 acquire, build and operate the LHC computing service 16th Sept’02 Nick Brook – University of Bristol 4 Phase 1 - High-level Goals To prepare and deploy the environment for LHC computing development/support for applications – libraries, tools, frameworks, data develop/acquire the software for managing a distributed computing system on the scale required for LHC – the local computing management (inc. persistency), …….. common components fabric, integration of the fabrics into a global grid put in place a pilot service – “proof of concept” – the technology and the distributed analysis environment platform for learning how to manage and use the system provide a solid service for physics and computing data challenges produce a TDR describing the distributed LHC computing system for the first years of LHC running maintain opportunities for re-use of developments outside the LHC programme 16th Sept’02 Nick Brook – University of Bristol 5 The LHC Computing Grid Project Organisation Common Computing RRB LHCC Reports (funding agencies) Reviews Resources Project Overview Board Project Requirements, Software and Computing Execution Monitoring Committee Board (SC2) 16th Sept’02 Nick Brook – University of Bristol 6 SC2 & PEB Roles SC2 includes the four experiments, Tier 1 Regional Centres SC2 identifies common solutions and sets requirements for the project may use an RTAG – Requirements and Technical Assessment Group limited scope, two-month lifetime with intermediate report one member per experiment + experts PEB manages the implementation organising projects, work packages coordinating between the Regional Centres collaborating with Grid projects organising grid services SC2 approves the work plan, monitors progress 16th Sept’02 Nick Brook – University of Bristol 7 SC2 Monitors Progress of the Project • Receives regular status reports from the PEB • Written status report every 6 months – milestones, performance, resources – estimates time and cost to complete • Organises a peer-review – about once a year – presentations by the different components of the project – review of documents – review of planning data 16th Sept’02 Nick Brook – University of Bristol 8 Project Execution Organisation Four areas – each with area project manager Applications Grid Technology Fabrics Grid deployment 16th Sept’02 Nick Brook – University of Bristol 9 RTAG status – in application software area • • • • • • • data persistency completed – 5th April 02 software support process completed – 6th May 02 mathematical libraries completed – 2nd May 02 detector geometry description running Monte Carlo generators running applications architectural blueprint running detector simulation running – in fabric area • mass storage requirements completed – 3rd May 02 – in Grid technology and deployment area • Grid technology use cases • Regional Centre categorisation completed – 7th June 02 completed – 7th June 02 Current status of RTAGs (and available reports) on www.cern.ch/lcg/sc2 16th Sept’02 Nick Brook – University of Bristol 10 Data Persistency (RTAG1) Technology • Streaming layer should be implemented using the ROOT framework’s I/O services • Components with relational implementations should make no deep assumptions about the underlying technology – Nothing intentionally proposed that precludes implementation using such open source products as MySQL 16th Sept’02 Nick Brook – University of Bristol 11 Data Persistency (RTAG1) Implementation – POOL Five work package areas: • Storage Manager & refs • File catalog & Grid integration • Collections & Metadata • Dictionary & Conversion • Infrastructure, Integration & testing http://lcgapp.cern.ch/projects/persist 16th Sept’02 Nick Brook – University of Bristol 12 Grid Use Cases (RTAG4) 79 page report & 43 use cases Global Summary of EDG Response • Use case is already implemented (release 1.2) – 19 – Mostly basic job submission and basic data management – For half of these, WP8 agrees that the functionality is implemented in 1.2, but the implementation is quite a bit more complex than that outlined in the use case (esp. data management). The release 2.0 implementations look simpler. • Planned for release 2 – 10 • Will be considered for release 3 – 4 • Use case not detailed enough – 4 – VO-wide resource allocation to users – HEPCAL did not make strong requirements on security – “Job Splitting” and “Production Job” – were purposely vague in HEPCAL due to lack of clear vision of how massive productions will be run on the Grid. One job auto-split into thousands? Or thousands of jobs somehow logically grouped into one production? • Not planned for any release – 7 – Software publishing – Virtual Datasets (reliant on GriPhyn) 16th Sept’02 Nick Brook – University of Bristol 13 Regional Centres (RTAG6) A service oriented view should be adopted for categorization of regional centres It could be profitable to revisit the overall computing model in terms of services around 2004 The important aspects to categorize RCs are – Commitment to guarantee data management at a high QoS for the lifetime of LHC – Commitment to guarantee state-of-the-art network bandwidth to ensure efficient inter-operation – Commitment to contribute to collaborative services 16th Sept’02 Nick Brook – University of Bristol 14 LCG Blueprint (RTAG8 - ongoing) Precepts Software structure: STL/utilities, “core” infrastructure, “specialised” infrastructure Component model: APIs (embedding frameworks, “own” plug-ins, end users), physical/logical module granularity, role of abstract interfaces, … Service model: Uniform, flexible access to basic framework functionality Object models: dumb vs. smart, enforced policies with run-time checking, clear and bullet-proof ownership model Distributed operation Global objects Dependencies: minimisation between components, run-time rather than compile-time Interface to external components: generic adapters – version & variant identification Exception handling 16th Sept’02 Nick Brook – University of Bristol 15 LCG Blueprint (RTAG8 - ongoing) Scripting, interpreter (ROOTCINT, PYTHON) GUI toolkits (to build expt specific interfaces) Graphics (underlying general tools) Analysis tools (histogramming, fitting, graphical representation, …) Math libraries and statistics (already established) Job management Core services (platform indep interface to system resources on LCG platforms – Linux (gnu & intel compilers), Solaris & Windows) Foundation and utility libraries (essentially maths libs & core services) Grid middleware interfaces (already an agreed 2 expt “common” project - GANGA) 16th Sept’02 Nick Brook – University of Bristol 16 LCG Blueprint (RTAG8 - ongoing) Object dictionary and object model (in context of POOL) Persistency and data management (in context of POOL) Event processing framework (poss. long term common project components) Event model Event generation (ancilliary services & support) Detector simulation (ditto) Detector geometry and materials (standard tools for describing, storing & modeling detector geometry) Trigger/DAQ Event reconstruction Detector calibration 16th Sept’02 Nick Brook – University of Bristol 17 LCG Summary GridPP input to both overall management structure on LCG and the RTAG activities Activities are beginning to take off Persistency (POOL) Software processing & infrastructure Grid deployment board – Oct 4th first meeting Interaction with middleware providers (not just EDG) RTAG procedures seems to be slow at taking off Lack of consistency early on – addressed by “Blueprint” RTAG Time consuming, often overlap of necessary experts 16th Sept’02 Nick Brook – University of Bristol 18