The LCG Project common solutions for the LHC

Download Report

Transcript The LCG Project common solutions for the LHC

The LHC Computing Project
Common Solutions for the LHC
ACAT 2002
Presented by
Matthias Kasemann
FNAL and CERN
1
Matthias Kasemann, FNAL and CERN, June 25, 2002
Outline
2/39
 The LCG Project: goal and organization
 Common solutions:
Why common solutions
 How to …
 The Run2 common projects

 The LCG Project: status of planning
Results of the LCG workshop in March 02
 Planning in the Applications Area

For the LCG Grid see: Les Robertson (Thursday)
“The LHC Computing Grid Project - Creating a Global Virtual
Computing Center for Particle Physics”
From Raw Data to Physics:
what happens during analysis
Matthias Kasemann, FNAL and CERN, June 25, 2002
250Kb – 1 Mb
100 Kb
25 Kb
5 Kb
e+
2037 2446 1733 1699
4003 3611 952 1328
2132 1870 2093 3271
4732 1102 2491 3216
2421 1211 2319 2133
3451 1942 1121 3429
3742 1288 2343 7142
Raw data
Convert to
physics
quantities
3/39
500 b
f
Z0
eDetector
response
apply
calibration,
alignment,
_
f
Interaction with Fragmentation, Basic physics
detector material Decay
Pattern,
Physics
Results
recognition,
analysis
Particle
identification
Analysis
Reconstruction
Simulation (Monte-Carlo)
HEP analysis chain:
common to LHC experiments
Detector
alignment
Matthias Kasemann, FNAL and CERN, June 25, 2002
Detector
description
Detector
calibration
Reconstruction
parameters
Generate
Events
4/39
Build
Simulation
Geometry
Build
Reconstruction
Geometry
Physics
Reconstuction
geometry Reconstruct
Events
Analyze
Events
Simulation
geometry
ATLAS
Detector
Simulate
Events
Raw
Data
ESD
AOD
Developing Software for LHC experiments
Matthias Kasemann, FNAL and CERN, June 25, 2002
 Challenges in big collaborations
5/39
Long and careful planning process
 More formal procedure required to commit resources
 Long lifetime, need flexible solutions which allow for
change




Any state of experiment longer than typical Ph.D. or
postdoc time
Need for professional IT participation and support
New development, maintenance and support model
required
 Challenges in smaller collaborations
Limited in resources
 Adapt and implement available solutions (“b-b-s”)

Matthias Kasemann, FNAL and CERN, June 25, 2002
CMS - CCS schedule (V33):
the bottom line
6/39
Milestones of ~ next year:
Milestones a few yrs away:
delays of ~9 months
delays of ~15 months
 LHC starts
CMS - CCS Software
Baseline:
DDD
ready for OSCAR,L2
ORCA,milestones
IGUANA
Switch from Geant3 to Geant4:
Matthias Kasemann, FNAL and CERN, June 25, 2002
•Data model defined
•Persistent and transient representations
•Demonstrably as correct as existing CMS description
Software Infrastructure
deployed and working
Framework for processing CMS data
•Working for simulation, reconstruction, analysis
•Supporting persistency and data management
7/39 •Strongly dependent on LCG success
•Date not decided (just my estimate)
•E.g. it needs the new persistency
CCS Baseline
Software for TDR’s
User analysis components
•Framework with coherent user interface
•Event display / interactive visualisation
•Tools for browsing / manipulating data sets
•Data presentation, histograms, numerical,…
Areas
Common Solutins
TheWork
LHC
Computing Grid
Applications Support &
Experiments and Regional
Project
(LCG)
Coordination
Centres agree on
requirements for common
Project Overview Board
Applications Support &
Coordination
Computing Systems
Grid Technology
Grid Deployment
Software and
Computing
Project
Execution Work Plan Definition Committee
(SC2)
Board
Matthias Kasemann, FNAL and CERN, June 25, 2002
Computing Systems
Grid Technology
Grid Deployment
8/39
Work
Areas
projects
WP WP WP WP WP
RTAG
Matthias Kasemann, FNAL and CERN, June 25, 2002
LCG - Fundamental goal:
9/39
The experiments have to get the best, most
reliable and accurate physics results from the
data provided by their detectors
Their computing projects are fundamental to the
achievement of this goal
The LCG project at CERN was set up to help
them all in this task
Corollary
Success of LCG is fundamental to success of
LHC Computing
Matthias Kasemann, FNAL and CERN, June 25, 2002
Fulfilling LCG Project Goals
 Prepare and deploy the LHC Computing Environment
 Applications - provide the common components, tools and
infrastructure for the physics application software
 Computing system – fabric, grid, global analysis system
 Deployment – foster collaboration and coherence
 Not just another grid technology project
 Validate the software by participating in Data Challenges using the
progressively more complex Grid Prototype
 Phase 1 - 50% model production grid in 2004
 Produce a TDR for full system to be built in Phase 2
 Software performance impacts on size and cost of production
facility
 Analysis models impact on exploitation of production grid
 Maintain opportunities for reuse of deliverables outside LHC
experimental programme
10/39
Applications Activity Areas
Matthias Kasemann, FNAL and CERN, June 25, 2002
 Application software infrastructure
11/39

physics software development environment, standard
libraries, development tools
 Common frameworks for simulation and analysis

Development and integration of toolkits & components
 Support for physics applications

Development, support of common software tools &
frameworks
 Adaptation of Physics Applications to Grid environment
 Object persistency and data management tools

Event data, metadata, conditions data, analysis objects,
Goals for Applications Area
Matthias Kasemann, FNAL and CERN, June 25, 2002
 Many Software Production Teams
LHC experiments
 CERN IT groups, ROOT team, ..
 HEP software collaborations – CLHEP, Geant4 , ..
 External Software – python, Qt, XML, …

 Strive to work together to develop and use software in common
 Will involve identifying and packaging existing HEP software for
reuse as well as developing new components
 Each unit has its own approach to design and to supporting the
development

Sharing in the development and deployment of software
will be greatly facilitated if units follow a common
approach
 Recognise that there will be start-up costs associated with
adapting to use new common products and development tools
12/39
Why common and when?
Matthias Kasemann, FNAL and CERN, June 25, 2002
 Why not:
Experiments have independent detectors and analysis
tools verify physics results
 Competition for best physics results
 Coordination of common software development is
significant overhead

 Why common solutions:
Need mature engineered software
 Resources are scarce, in particular manpower
 Effort: Common projects are a good way to become more
efficient ( 1 ,  2 n , n , n 2 ?)
 Lessons need to be learnt from past experience

 For LHC experiments:
Everything non experiment–specific is a potential
candidate for a common project
13/39
Matthias Kasemann, FNAL and CERN, June 25, 2002
FNAL: CDF/D0/CD - Run 2
Joint Project Organization
14/39
D0 Collaboration
Directorate
CDF Collaboration
External Review
Committee
R2JOP
Steering Committee
Task
Coordinators
Run II
Committee
Basic Infrastructure
Run II Computing
Project Office
Mass Storage &
Data Access
Serial Media
Storage
Fermilab
Working
Management
Simulation
Class
Group
Library
Configuration Support
MSS
Data
Access
Management
Hardware
Databases
Reconstruction
Systems
Physics Analysis
Support
Reconstruction Networking
farm hardware hardware
Production Reconstruction
Management input pipeline
15 joint projects defined,
4 years before start of data taking
Physics
analysis
hardware
Visualization
Physics Analysis Software
Perceptions of Common Projects
 Experiments
Whilst may be very enthusiastic about long-term
advantages ….
 …have to deliver on short term milestones
 Devoting resources to both will be difficult
 Already experience an out-flux of effort into common
projects
 Hosting projects in experiments excellent way of
integrating effort
Matthias Kasemann, FNAL and CERN, June 25, 2002


For initial phase and prototyping
 Technology groups
15/39
Great motivation to use expertise to produce useful
solutions
 Need the involvement of the experiments

Matthias Kasemann, FNAL and CERN, June 25, 2002
Common solutions - How to do?
 Requirements are set by experiments in the SC2 +
Requirements Technical Assessment Groups (RTAGs)
 Planning and implementation is done by LCG together
with experiments
 Monitoring of progress and
adherence by the SC2
 Frequent releases and testing
 Guaranteed life-time maintenance and support
Issues:
 ‘How will applications area cooperate with other areas?’
 ‘Not feasible to have a single LCG architect to cover
all areas.’
 Need mechanisms to bring coherence to the project
16/39
Workflow around the
organisation chart
Matthias Kasemann, FNAL and CERN, June 25, 2002
WPn
PEB
PEB develops
workplan
Project plan
PEB manages
LCG resources
Release 1
~4 mths
17/39
PEB tracks
Release 2
progress
SC2
Prioritised
requirements
SC2 Sets RTAGm
the
mandate
requirements
requirements
~2 mths
Updated workplan
Workplan feedback
SC2 approves the
workplan
Status report
Review feedback
time
SC2 reviews the
status
Matthias Kasemann, FNAL and CERN, June 25, 2002
Issues related to partitioning
the work
18/39
 ‘How do you go from present to future without
dismantling existing projects?’
 ‘Have to be careful that we don’t partition into too
small chunks and lose coherence of overall software’
 We are not starting afresh, we have a good knowledge
of what the broad categories are going to be
 Experiment architectures help to ensure coherency.
Matthias Kasemann, FNAL and CERN, June 25, 2002
Coherent Architecture
19/39
 Applications common projects must follow a coherent
overall architecture
 The software needs to be broken down into
manageable pieces i.e. down to the component level
 Component-based, but not a bag of disjoint
components
components designed for interoperability through clean
interfaces
 Does not preclude a common implementation foundation,
such as ROOT, for different components




The ‘contract’ in the architecture is to respect the
interfaces
No hidden communication among components
Starting point is existing products, not a clean slate
Matthias Kasemann, FNAL and CERN, June 25, 2002
Approach to making workplan
20/39
 “Develop a global workplan from which the RTAGs can be derived”
 Considerations for the workplan:
 Experiment need and priority
 Is it suitable for a common project
 Is it a key component of the architecture e.g. object
dictionary
 Timing: when will the conditions be right to initiate a common
project



Availability of resources and allocation of effort


Do established solutions exist in the experiments
Are they open to review or are they entrenched
Is there existing effort which would be better spent doing
something else
Availability, maturity of associated third party software

E.g. grid software
 Pragmatism and seizing opportunity. A workplan derived from a
grand design does not fit the reality of this project
Matthias Kasemann, FNAL and CERN, June 25, 2002
RTAG: ‘blueprint’ of LCG
application architecture
21/39
 Mandate: define the architectural ‘blueprint’ for LCG applications:
 Define the main architectural domains (‘collaborating
frameworks’) of LHC experiments and identify their principal
components. (For example: Simulation is such an architectural
domain; Detector Description is a component which figures in
several domains.)
 Define the architectural relationships between these
‘frameworks’ and components, including Grid aspects, identify
the main requirements for their inter-communication, and
suggest possible first implementations. (The focus here is on
the architecture of how major ‘domains’ fit together, and not
detailed architecture within a domain.)
 Identify the high-level milestones for each domain and
provide a first estimate of the effort needed. (Here the
architecture within a domain could be considered.)
 Derive a set of requirements for the LCG
 Time-scale: started in June 02, draft report in July, final report
in August 02
Matthias Kasemann, FNAL and CERN, June 25, 2002
RTAG status
22/39
 Identified and started eight
Requirement Technical Assessments (RTAGs)

in application software area







finished
finished
finished
started
started
started
in compute fabric area


Data persistency
Software support process and tools
Mathematical libraries
Detector Geometry & Materials descriptions
‘blueprint’ architecture of applications
Monte Carlo event generators
mass storage requirements
finished
in Grid technology and deployment area


Grid technology use cases
regional center category and services definition
finished
finished
Matthias Kasemann, FNAL and CERN, June 25, 2002
Software Process RTAG
23/39
 Mandate:
 Define a process for managing LCG software. Specific tasks to
include: Establish a structure for organizing software, for managing
versions and coherent subsets for distribution
 Identify external software packages to be supported
 Identify recommended tools for use within the project – to include
configuration and release management
 Estimate resources (person power) needed to run an LCG support
activity
 Guidance:
 Procedures and tools will be specified
 Will be used within project
 Can be packaged and supported for general use
 Will evolve with time
 The RTAG does not make any recommendations on how experiment
internal software should be developed and managed. However, if an
experiment specific program becomes an LCG product it should
adhere to the development practices proposed by this RTAG
Process RTAG –
Recommendations(1)
Matthias Kasemann, FNAL and CERN, June 25, 2002
 All LCG projects must adopt the same set of tools,
standards and procedures. The tools must be centrally
installed, maintained and supported.
 Adopt commonly used open-source or commercial
software where available. Try to avoid “do it yourself”
solutions in this area where we don’t have core
competency.
 Concerning commercial software, avoid commercial
software that has to be installed on individual’s
machines as this will cause well known problems of
license agreements and management in our widely
distributed environment. Commercial solutions for
web-portals or other centrally managed solutions would
be fine.
24/39
Process RTAG –
Recommendations(2)
Matthias Kasemann, FNAL and CERN, June 25, 2002
 ‘Release early, release often’ implies
25/39
major release 2-3 times per year
 Development release every 2-3 weeks
 Automated nightly builds, regression tests, benchmarks

 Test and quality assurance
 Support of external software

installation and build up of local expertise
 Effort needed for filling support roles
Librarian
 Release manager
 Toolsmith
 Quality assurance
 Technical writer

Matthias Kasemann, FNAL and CERN, June 25, 2002
Persistency RTAG
26/39
 Mandate:
 Write the product specification for the Persistency Framework for
Physics Applications at LHC
 Construct a component breakdown for the management of all types
of LHC data
 Identify the responsibilities of Experiment Frameworks, existing
products (such as ROOT) and as yet to be developed products
 Develop requirements/use cases to specify (at least) the metadata
/navigation component(s)
 Estimate resources (manpower) needed to prototype missing
components
 Guidance:
 The RTAG may decide to address all types of data, or may decide to
postpone some topics for other RTAGS, once the components have
been identified.
 The RTAG should develop a detailed description at least for the
event data management.
 Issues of schema evolution, dictionary construction and storage,
object and data models should be addressed.
Matthias Kasemann, FNAL and CERN, June 25, 2002
Persistency –
Near term recommendations
27/39
 to develop a common object streaming layer and
associated persistence infrastructure.
a common object streaming layer based on ROOT-IO and
several related components to support it,
 including a (currently lightweight) relational database
layer.
 Dictionary services are included in the near-term
project specification.


dictionary services may have additional clients
 This is first step towards a complete data management
environment, one with enormous potential for
commonality among the experiments.
Matthias Kasemann, FNAL and CERN, June 25, 2002
RTAG: math library review
28/39
 Mandate: Review the current situation with math
libraries and make recommendations
Review the current situation of the usage of the various
math libraries in the experiments (including but not
limited to NagC++, GSL, CLHEP, ROOT)
 Identify and recommend which ones should be adopted,
which ones could be discontinued
 Suggest possible improvements to the existing ones
 Estimate resources needed for this activity

 Guidance – The result of the RTAG should allow to
establish a clear program of work to streamline the
status of math libraries and find the maximum
commonality between experiments, taking into account
cost, maintenance and projected evolution of the
experiment needs
Matthias Kasemann, FNAL and CERN, June 25, 2002
Math Library: Recommendations
29/39
 To design a support group
 to provide advice and information about the use of existing
libraries,
 to assure their continued availability,
 to identify where new functionality is needed, and
 to develop that functionality themselves or by coordinating
with other HEP-specific library developers.
 The goal would be to have close contact with the experiments
and provide expertise on mathematical methods, aiming at
common solutions,
 The experiments should maintain a data base of mathematical
libraries used in their software, and within each library, the
individual modules used.
 A detailed study should be undertaken to determine whether
there is any functionality needed by the experiments and
available in the NAG library which is not covered as well by a
free library such as GSL.
Matthias Kasemann, FNAL and CERN, June 25, 2002
RTAG:
30/39
Detector Geometry
& Materials Description
 Write the product specification for detector geometry and
materials description services.
 Specify scope: e.g. Services to define, provide transient
access to, and store the geometry and materials descriptions
required by simulation, reconstruction, analysis, online and
event display applications, with the various descriptions using
the same information source
 Identify requirements including end-user needs such as ease
and naturalness of use of the description tools, readability
and robustness against errors e.g. provision for named
constants and derived quantities
 Explore commonality of persistence requirements with
conditions data management



Interaction of the DD with a conditions DB. In that context
versioning and ‘configuration management’ of the detector
description, coherence issues…
Identify where experiments have differing requirements and
examine how to address them within common tools
Address migration from current tools
Matthias Kasemann, FNAL and CERN, June 25, 2002
RTAG:
Monte Carlo Event Generators
31/39
 Mandate: To best explore the common solutions
needed and how to engage the HEP community external
to the LCG it is proposed to study:
How to maintain a common code repository for the
generator code and related tools such as PDFLIB.
 The development or adaptation of generator-related
tools (e.g.HepMC) for LHC needs.
 How to provide support for the tuning, evaluation and
maintenance of the generators.
 The integration of the Monte Carlo generators into the
experimental software frameworks.
 The structure of possible forums to facilitate
interaction with the distributed external groups who
provide the Monte Carlo generators.

Possible Organisation of
activities
Matthias Kasemann, FNAL and CERN, June 25, 2002
Architect
Overall management, coordination, architecture, integration, support
Activity area
Project
leader
32/39
Activity area
Project
Project
WP WP WP
WP
Project
WP
WP WP
Activity area
Project
WP
WP
Example:
Activity area: Physics data management
Possible projects: Hybrid event store, Conditions DB, …
Work Packages: Component breakdown and work plan lead to Work
Package definitions. ~1-3 FTEs per WP
Global Workplan –
1st priority level
Matthias Kasemann, FNAL and CERN, June 25, 2002
1.
33/39
2.
3.
Establish process and infrastructure
 Nicely covered by software process RTAG
Address core areas essential to building a coherent
architecture
 Object dictionary – essential piece
 Persistency - strategic
 Interactive frameworks - also driven by assigning personnel
optimally
Address priority common project opportunities
 Driven by a combination of experiment need,
appropriateness to common project, and ‘the right moment’
(existing but not entrenched solutions in some
experiments)


Detector description and geometry model
Driven by need and available manpower

Simulation tools
Matthias Kasemann, FNAL and CERN, June 25, 2002
Global Workplan –
2nd priority level
34/39
 Build outward from the core top-priority components
Conditions database
 Statistical analysis
 Framework services, class libraries

 Address common project areas of less immediate
priority
Math libraries
 Physics packages (scope?)

 Extend and elaborate the support infrastructure

Software testing and distribution
Matthias Kasemann, FNAL and CERN, June 25, 2002
Global Workplan –
3rd priority level
35/39
 The core components have been addressed,
architecture and component breakdown laid out, work
begun. Grid products have had another year to develop
and mature. Now explicitly address physics
applications integration into the grid applications layer.
Distributed production systems. End-to-end grid
application/framework for production.
 Distributed analysis interfaces. Grid-aware analysis
environment and grid-enabled tools.

 Some common software components are now available.
Build on them.
Lightweight persistency, based on persistency
framework
 Release LCG benchmarking suite

Matthias Kasemann, FNAL and CERN, June 25, 2002
Global Workplan –
4th priority level
 Longer term items waiting for their moment
36/39

‘Hard’ ones, perhaps made easier by a growing common
software architecture


Address evolution of how we write software


Event processing framework
OO language usage
Longer term needs; capabilities emerging from R&D (more
speculative)

Advanced grid tools, online notebooks, …
Matthias Kasemann, FNAL and CERN, June 25, 2002
Candidate RTAGs (1)
37/39
Simulation tools
Non-physics activity
Detector description,
model
Description tools, geometry
model
Conditions database
If necessary after existing RTAG
Data dictionary
Key need for common service
Interactive frameworks
What do we want, have, need
Statistical analysis
Tools, interfaces, integration
Visualization
Tools, interfaces, integration
Physics packages
Important area but scope unclear
Framework services
If common framework is too
optimistic…
C++ class libraries
Standard foundation libraries
Matthias Kasemann, FNAL and CERN, June 25, 2002
Candidate RTAGs (2)
38/39
Event processing
framework
Hard, long term
Distributed analysis
Application layer over grid
Distributed production
Application layer over grid
Small scale
persistency
Simple persistency tools
Software testing
May be covered by process RTAG
Software distribution
From central ‘Program Library’ to
convenient broad distribution
OO language usage
C++, Java (..?) roles in the future
Benchmarking suite
Comprehensive suite for LCG
software
Online notebooks
Long term; low priority
Matthias Kasemann, FNAL and CERN, June 25, 2002
Common Solutions: Conclusions
39/39
 Common Solutions for LHC software are required for
success
Common solutions are agreed upon by experiments
 The requirements are set by the experiments
 The development is done jointly by the LCG project and
the LHC experiments
 All LCG software is centrally supported and maintained.

 What makes us believe that we succeed?
What is key to success?
The process in the LCG organization
 The collaboration between players
 Common technology
 Central resources, jointly steer-able by experiments and
management
 Participants have prototyping experience !!

Backup & Additional slides
40
Post-RTAG Participation of
Architects – Draft Proposal (1)
Matthias Kasemann, FNAL and CERN, June 25, 2002
 Monthly open meeting (expanded weekly meeting)
41/39
Accumulated issues to be taken up with architects
 Architects in attendance; coordinators invited

 Information has gone out beforehand, so architects
are ‘primed’
 Meeting is informational, and decision-making (for the
easier decisions)

An issue is either


Resolved (the easy ones)
Flagged for addressing in the ‘architects committee’
Matthias Kasemann, FNAL and CERN, June 25, 2002
Post-RTAG Participation of
Architects – Draft Proposal (2)
42/39
 Architects committee:
 Members: experiment architects + applications manager
(chair)
 Invited: computing coordinators, LCG project manager and
CTO
 Others invited at discretion of members

e.g. project leader of project at issue

Such decisions can be accepted or challenged
 Meets shortly after the open meeting (also bi-weekly?)
 Decides the difficult issues
 Most of the time, committee will converge on a decision
 If not, try harder
 If still not, applications manager takes decision
 Challenged decisions go to full PEB, then if necessary to SC2
 PEB role of raising issues to be taken up by SC2
 We all abide happily by an SC2 decision
 Committee meetings also cover general current issues and
exchange of views
 Committee decisions, actions documented in public minutes
Distributed Character of
Components (1)
Matthias Kasemann, FNAL and CERN, June 25, 2002

43/39
Persistency framework
Naming based on logical filenames
 Replica catalog and management
 Cost estimators; policy modules


Conditions database


Interactive frameworks


Grid-aware environment; ‘transparent’ access to gridenabled tools and services
Statistical analysis, visualization


Inherently distributed (but configurable for local use)
Integral parts of distributed analysis environment
Framework services

Grid-aware message and error reporting, error
handling, grid-related framework services
Distributed Character of
Components (2)
Matthias Kasemann, FNAL and CERN, June 25, 2002
 Event processing framework
44/39

Cf. framework services, persistency framework,
interactive frameworks
 Distributed analysis
 Distributed production
 Software distribution

Should use the grid
 OO language usage

Distributed computing considerations
 Online notebook

Grid-aware tool
Matthias Kasemann, FNAL and CERN, June 25, 2002
RTAG?: Simulation tools
45/39
 Geant4 is establishing a HEP physics requirements
body within the collaboration, accepted by SC2 as a
mechanism for addressing G4 physics performance
issues
 However, there are important simulation needs to
which LCG resources could be applied in the near term.
 By the design of LCG, this requires SC2 delivering
requirements to PEB
 John Apostolakis has recently assembled G4 requests
and requirements from the LHC collaborations
 Proposal: Use these requirements as the groundwork
for a quick 1-month RTAG to guide near term
simulation activity in the project, leaving the
addressing of physics performance requirements to
the separate process within Geant4
Matthias Kasemann, FNAL and CERN, June 25, 2002
RTAG?: Simulation tools (2)
46/39
 Some possible activity areas in simulation, from the
Geant4 requests/requirements received from the
experiments, which would be input to the RTAG:
Error propagation tool for reconstruction (‘GEANE’)
 Assembly and documentation of standard physics lists
 Python interface
 Documentation, tutorials, communication
 Geant4 CVS server access issues

 The RTAG could also address FLUKA support
Requested by ALICE as an immediate priority
 Strong interest expressed by other experiments as well

Matthias Kasemann, FNAL and CERN, June 25, 2002
RTAG?: Detector geometry & materials
description and modeling services
47/39
 Write the product specification for detector geometry and
materials description and modeling services
 Specify scope: eg. Services to define, provide transient
access to, and store the geometry and materials
descriptions required by simulation, reconstruction,
analysis, online and event display applications, with the
various descriptions using the same information source
 Identify requirements including end-user needs such as
ease and naturalness of use of the description tools,
readibility and robustness against errors e.g. provision for
named constants and derived quantities
 Explore commonality of persistence requirements with
conditions data management
 Identify where experiments have differing requirements
and examine how to address them within common tools
 Address migration from current tools
Matthias Kasemann, FNAL and CERN, June 25, 2002
RTAG?: Conditions database
48/39
 Will depend on persistency RTAG outcome
 Refine the requirements and product specification of a
conditions database serving the needs of the LHC
experiments, using the existing requirements and
products as a reference point. Give due consideration
to effective distributed/remote usage.
 Identify the extent to which the persistency
framework (hybrid store) can be directly used at the
lower levels of a conditions database implementation.
 Identify the component(s) and interfaces atop a
common persistency foundation that complete the
conditions database
RTAG?: Data dictionary
service
Matthias Kasemann, FNAL and CERN, June 25, 2002

49/39



Can the experiments converge on common data definition and
dictionary tools in the near term?
Even if the answer is no, it should be possible to establish a
standard dictionary service (generic API) by which common
tools can interact, while leaving free to the experiments how
their class models are defined and implemented
Develop a product specification for a generic high-level data
dictionary service able to accommodate distinct data definition
and dictionary tools and present a common, generic interface to
the dictionary
Review the current data definition and dictionary approaches
and seek to expand commonality among the experiments. Write
the product specifications for common (even if N<4)
components.
RTAG?: Interactive
frameworks
Matthias Kasemann, FNAL and CERN, June 25, 2002

50/39




Frameworks providing interactivity for various
environments including physics analysis and event
processing control (simulation and reconstruction) are
critical. They serve end users directly and must match
end user requirements extremely well. They can be a
powerful and flexible ‘glue’ in a modular environment,
providing interconnectivity between widely distinct
components and making the ‘whole’ offered by such an
environment much greater than the sum of its parts.
Develop the requirements for an interactive framework
common across the various application environments
Relate the requirements to existing tools and approaches
(e.g. ROOT/CINT, Python-based tools)
Write a product specification, with specific
recommendations on tools and technologies to employ
Address both command line and GUI interactivity
RTAG?: Statistical analysis
interfaces & tools
Matthias Kasemann, FNAL and CERN, June 25, 2002

51/39
Address requirements on analysis tools
What data analysis services and tools are required
 What is and is not provided by existing tools


Address what existing tools should be supported and
what further development is needed


Including long term maintenance issues
Address role of abstract interfaces to statistical
analysis services
Are they to be used?
 If so, what tools should be interfaced to a common
abstract interface to meet LHC needs (and how, when,
etc.)


Address requirements and approaches to persistency
and data interchange
Matthias Kasemann, FNAL and CERN, June 25, 2002
RTAG?: Detector and event
visualization
52/39
 Examine the range of tools available and identify those
which should be developed as common components
within the LCG Applications architecture
 Address requirements, recommendations and
needed/desired implementations in such areas as
existing and planned standard interfaces and their
applicability
 GUI integration
 Interactivity requirements (picking)
 Interface to visualizing objects (eg. Draw() method)
 Use of standard 3D graphics libraries

 Very dependent on other RTAG outcomes
Matthias Kasemann, FNAL and CERN, June 25, 2002
RTAG?: Physics packages
53/39
 Needs and requirements in event generators and their
interfaces & persistency, particle property services, …
 Scope of the LCG in this area needs to be made
clearer before a well defined candidate RTAG can be
developed
RTAG?: Framework services
Matthias Kasemann, FNAL and CERN, June 25, 2002

54/39
While converging on a common event processing
framework among the LHC experiments may be
impractical at least on the near term, this does not
preclude adopting common approaches and tools for
Framework services



Examples: message handling and error reporting;
execution monitoring and state management;
exception handling and recovery; job state persistence
and recording of history information; dynamic
component loading; interface definition, versioning,
etc.
Seek to identify framework services and tools which
can be developed in common, possibly starting from
existing products.
Develop requirements on their functionality and
interfaces.
Matthias Kasemann, FNAL and CERN, June 25, 2002
RTAG?: C++ class libraries
55/39
 Address needs and requirements in standard C++ class
libraries, with recommendations on specific tools
 Provide recommendations on the application and
evolution of community libraries such as ROOT, CLHEP,
HepUtilities, …
 Survey third party libraries and provide
recommendations on which should be adopted and what
should be used from them
 Merge with Framework Services candidate RTAG?
Matthias Kasemann, FNAL and CERN, June 25, 2002
RTAG?: Event processing
framework
56/39
 There is no consensus to pursue a common event
processing framework in the near term. There is
perhaps more agreement that this should be pursued in
the long term (but there’s no consensus on a likely
candidate for a common framework in the long term)
 This looks at best to be a long term RTAG
 Two experiments do use a common event processing
framework kernel (Gaudi)
 Many difficult issues in growing N past 2, whether
with Gaudi, AliRoot, COBRA or something else!
Matthias Kasemann, FNAL and CERN, June 25, 2002
RTAG?: Interfaces to
distributed analysis
57/39
 Develop requirements on end-user interfaces to
distributed analysis, layered over grid middleware
services, and write a product specification
Grid portals, but not only; e.g. PROOF and Jas fall into
this category
 A grid portal for analysis is presumably an evolution of
tools like these

 Focus on analysis interface; address the distinct
requirements of production separately

Production interface should probably be addressed first,
as it is simpler and will probably have components usable
as parts of the analysis interface
RTAG?: Distributed
production systems
Matthias Kasemann, FNAL and CERN, June 25, 2002

58/39



Distributed production systems will have much common ground
at the grid middleware level. How much can be done in common
at the higher level of end-to-end distributed production
applications layered over the grid middleware?
 Recognizing that the grid projects are active at this level
too, and coordination is needed
Survey existing and planned production components and end-toend systems at the application level (AliEn, MOP, etc.) and
identify tools and approaches to develop in common
Write product specifications for common components, and/or
explicitly identify specific tools to be adapted and developed as
common components
Include end user (production operations) interface
 Grid portal for production
RTAG?: Small-scale
persistency & databases
Matthias Kasemann, FNAL and CERN, June 25, 2002

59/39


If not covered by the existing persistency RTAG, and if there
is agreement this is needed…
Write the product specification for a simple, self-contained,
low-overhead object persistency service for small-scale
persistency in C++ applications
 Marshal objects to a byte stream which may be stored on a
file, in an RDBMS record, etc.
 In implementation, very likely a simplified derivative of the
object streamer of the hybrid store
 For small scale persistence applications, e.g. saving state,
saving configuration information
Examine the utility of and requirements on a simple, standard,
easily installed and managed database service complementing
the persistency service for small scale applications
 MySQL, PostgreSQL etc are casually adopted for simple
applications with increasing frequency. Is it possible and
worthwhile to converge on a common database service
Matthias Kasemann, FNAL and CERN, June 25, 2002
RTAG?: Software testing
tools & services
60/39
 How much commonality can be achieved in the
infrastructure and tools used

Memory checking, unit tests, regression tests, validation
tests, performance tests
 A large part of this has been covered by the process
RTAG
Matthias Kasemann, FNAL and CERN, June 25, 2002
RTAG?: Software distribution
61/39
 May or may not be adequately addressed in the
process RTAG
 Requirements for a central distribution point at CERN

A ‘CERN LHC Program Library Office’
 Requirements on software distribution taking into
account all tiers
 Survey and recommend on the various approaches,
their utility, complementarity
Tarballs (DAR)
 RPMs and other standard open software tools
 Role of AFS, asis
 Higher level automated distribution tools (pacman)

Matthias Kasemann, FNAL and CERN, June 25, 2002
RTAG?: Evolution of OO
language usage
62/39
 Long-term evolution of C++
 Role for other language(s), e.g. Java?
Near, medium and (to the extent possible) long term
application of other languages among LHC experiments
 Implications for tools and support requirements

 Identify any requirements arising
Applications, services to be developed in common
 Third party tools to be integrated and supported
 Compilers and other infrastructure to be supported
 Libraries required

RTAG?: LCG Benchmarking
suite
Matthias Kasemann, FNAL and CERN, June 25, 2002
 Below threshold for an RTAG?
63/39

Every LCG application should come with a benchmarking
suite, and should be made available and readily usable as
part of a comprehensive benchmarking suite
 Develop requirements for a comprehensive
benchmarking suite of LCG applications for use in
performance evaluation, testing, platform validation
and performance measurement, etc.
Tools which should be represented
 Tests which should be included
 Packaging and distribution requirements

Matthias Kasemann, FNAL and CERN, June 25, 2002
RTAG?: Online notebooks and other
remote control / collaborative tools
64/39
 Identify near term and long term needs and
requirements common across the experiments
 Survey existing, planned tools and approaches
 Develop recommendations for common
development/adaptation and support of tools for LHC