Fabric Management for CERN Experiments Past, Present, and Future Tim Smith CERN/IT Contents The Fabric of CERN today The new challenges of LHC.
Download
Report
Transcript Fabric Management for CERN Experiments Past, Present, and Future Tim Smith CERN/IT Contents The Fabric of CERN today The new challenges of LHC.
Fabric Management
for CERN Experiments
Past, Present, and Future
Tim Smith CERN/IT
Contents
The Fabric of CERN today
The new challenges of LHC computing
What has this got to do with the GRID
Fabric Management solutions of tomorrow?
The DataGRID Project
2000/11/03
Tim Smith: HEPiX @ JLab
2
Fabric Elements
Functionalities
Infrastructure
Batch and Interactive
Disk servers
Tape Servers + devices
Stage servers
Home directory servers
Application servers
Backup service
2000/11/03
Job Scheduler
Authentication
Authorisation
Monitoring
Alarms
Console managers
Networks
Tim Smith: HEPiX @ JLab
3
Multiplicity Scale
Fabric Technology at CERN
PC Farms
10000
1000
PC Farms
RISC Workstations
100
10
1
Scalable Systems
SP2 CS2
RISC Workstations
SMPs
SGI,DEC,HP,SUN
Mainframes
IBM Cray
89 90 91 92 93 94 95 96 97 98 99 00 01 02 03 04 05
Year
2000/11/03
Tim Smith: HEPiX @ JLab
4
Architecture Considerations
Physics applications have ideal data
parallelism
mass of independent problems
No message passing
throughput rather than performance
resilience rather than ultimate reliability
Can build hierarchies of mass market
components
High Throughput Computing
2000/11/03
Tim Smith: HEPiX @ JLab
5
Component Architecture
High capacity
backbone
switch
Application
Server
100/1000baseT switch
CPU
CPU
CPU
CPU
CPU
Disk Server
1000baseT switch
Tape
Server
Tape
Server
2000/11/03
Tape
Server
Tape
Server
Tim Smith: HEPiX @ JLab
6
Analysis Chain: Farms
event filter
(selection &
reconstruction)
detector
processed
data
event
summary
data
raw
data
event
reconstruction
batch
physics
analysis
analysis objects
(extracted by physics topic)
event
simulation
2000/11/03
Tim Smith: HEPiX @ JLab
interactive
physics
analysis
7
Multiplication !
tomog
tapes
pcsf
nomad
na49
na48
na45
mta
lxbatch
lxplus
lhcb
l3c
ion
eff
cms
ccf
atlas
alice
1200
1000
#CPUs
800
600
400
200
0
Jul-97
2000/11/03
Jan-98
Jul-98
Jan-99
Jul-99
Tim Smith: HEPiX @ JLab
Jan-00
8
PC Farms
2000/11/03
Tim Smith: HEPiX @ JLab
9
Shared Facilities
EFF Scheduling 2000
140
120
DELPHI
CMS
ALEPH
80
ATLAS
NA45
60
COMPASS
ALICE
Available
40
20
0
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
Number of PCs
100
Week Number
2000/11/03
Tim Smith: HEPiX @ JLab
10
LHC Computing Challenge
The scale will be different
CPU
Disk
Tape
10k SI95
30TB
600TB
1M SI95
3PB
9PB
The model will be different
There are compelling reasons why some of the
farms and some of the capacity will not be
located at CERN
2000/11/03
Tim Smith: HEPiX @ JLab
11
Estimated DISK Capacity ay CERN
1800
1600
Estimated disk
storage capacity at
CERN
1400
Bad News: Tapes
< factor 2 reduction in 8 years
Significant fraction of cost
Non-LHC
1000
800
LHC
600
400
200
0
1998
Moore’s Law
1999
2000
2001
2002
2003
2004
2005
2006
year
Estimated CPU Capacity at CERN
Bad News: IO
1996: 4G @10MB/s
1TB – 2500MB/s
2000: 50G @ 20 MB/s
1TB – 400 MB/s
Estimated CPU
capacity at CERN
2,500
2,000
K SI95
TeraBytes
1200
1,500
1,000
NonLHC
~10K SI95
1200 processors
LHC
500
0
1998
2000/11/03
1999
2000
Tim Smith: HEPiX @ JLab
2001
2002
year
2003
2004
2005
2006
12
Regional Centres:
a Multi-Tier Model
CERN – Tier 0
IN2P3
Tier 1
FNAL
RAL
Uni n
Lab a
Tier2
Uni b
Department
Lab c
Desktop
2000/11/03
MONARC http://cern.ch/MONARC
Tim Smith: HEPiX @ JLab
13
More realistically:
a Grid Topology
CERN – Tier 0
IN2P3
Tier 1
FNAL
RAL
Uni n
Lab a
Tier2
Uni b
Department
Lab c
Desktop
2000/11/03
DataGRID http://cern.ch/grid
Tim Smith: HEPiX @ JLab
14
Can we build LHC farms?
Positive predictions
CPU and disk price/performance trends suggest that the raw
processing and disk storage capacities will be affordable, and
raw data rates and volumes look manageable
perhaps not today for ALICE
Space, power and cooling issues?
So probably yes… but can we manage them?
Understand costs - 1 PC is cheap, but managing 10000 is not!
Building and managing coherent systems from such large
numbers of boxes will be a challenge.
2000/11/03
Tim Smith: HEPiX @ JLab
15
Management Tasks I
Supporting adaptability
Configuration Management
Machine / Service hierarchy
Automated registration / insertion / removal
Dynamic reassignment
Automatic Software Installation and
Management (OS and applications)
Version management
Application dependencies
Controlled (re)deployment
2000/11/03
Tim Smith: HEPiX @ JLab
16
Management Tasks II
Controlling Quality of Service
System Monitoring
Orientation to the service NOT the machine
Uniform access to diverse fabric elements
Integrated with configuration (change) management
Problem Management
Identification of root causes (faults + performance)
Correlate network / system / application data
Highly automated
Adaptive - Integrated with configuration
management
2000/11/03
Tim Smith: HEPiX @ JLab
17
Relevance to the GRID ?
Scalable solutions needed in absence of
GRID !
For the GRID to work it must be presented
with information and opportunities
Coordinated and efficiently run centres
Presentable as a guaranteed quality resource
‘GRID’ification : the interfaces
2000/11/03
Tim Smith: HEPiX @ JLab
18
Mgmt Tasks: A GRID centre
GRID enable
Support external requests: services
Publication
Coordinated + ‘map’able
Security: Authentication / Authorisation
Policies: Allocation / Priorities / Estimation / Cost
Scheduling
Reservation
Change Management
Guarantees
Resource availability / QoS
2000/11/03
Tim Smith: HEPiX @ JLab
19
Existing Solutions ?
The world outside is moving fast !!
Dissimilar problems
Virtual super computers (~200 nodes)
MPI, latency, interconnect topology and bandwith
Roadrunner, LosLobos, Cplant, Beowulf
Similar problems
ISPs / ASPs (~200 nodes)
Clustering: high availability / mission critical
The DataGRID : Fabric Management WP4
2000/11/03
Tim Smith: HEPiX @ JLab
20
WP4 Partners
CERN (CH)
ZIB (D)
KIP (D)
NIKHEF (NL)
INFN (I)
RAL (UK)
IN2P3 (Fr)
2000/11/03
Tim Smith
Alexander Reinefeld
Volker Lindenstruth
Kors Bos
Michele Michelotto
Andrew Sansum
Denis Linglin
Tim Smith: HEPiX @ JLab
21
Concluding Remarks
Years of experience in exploiting
inexpensive mass market components
But we need to marry these with
inexpensive highly scalable
management tools
Build components back together as a
resource for the GRID
2000/11/03
Tim Smith: HEPiX @ JLab
22