Computational Science at Edinburgh From Excellence to Enterprise Dr Arthur Trew Director

Download Report

Transcript Computational Science at Edinburgh From Excellence to Enterprise Dr Arthur Trew Director

Computational Science at Edinburgh
From Excellence to Enterprise
Dr Arthur Trew
Director
on the shoulders of giants
 for centuries science has relied on experiment and theory, but
– theory is best suited to a reductionist approach
– experiment is limited in its range of applicability
 there is a need for a complementary approach: simulation 
study of emergent phenomena
Theory- Experiment Greece
Italy
400 BC
1,500 AD
Simulation Edinburgh
1,980 AD
For problems which are:
- too large/small
- too fast/slow
- too complex
- too expensive ...
arth
52.8
becoming the best1.0 D
 in 1990 Edinburgh established EPCC as its focus for HPC
simulation
– following a decade of research in Physics
 in 1992 EPCC set itself a mission “to be Europe’s top HPC
centre within 5 years”
– we succeeded
– today, that success is benefiting Scottish research and industry
.4 Ea
1.68
2.3 e
1.0 S
2.3 N
3 GC
0.2 G
0.35
 in 2001 Edinburgh & Glasgow created NeSC to extend that
vision
– over the past 18 months EPCC & NeSC have brought in £65M of
research grants/contracts
• of which £2.3M from SHEFC for eDIKT
– Edinburgh & Glasgow backed this with £3M
– … and foundation departments lead major UK Grid projects
0.6 G
arth
UoE
UoG
(AK’
what’s that to me?
 as computer performance improves, the range of applications
increases
100
90
Tflops capability
80
70
60
50
HPCx
£53M: 3 machines
40
30
eddy
cells
20
resolution materials
oceans
design
whole
10
protein drug design
nanostructures
aircraft
structures
0
0
1
2
3
4
5
year
whole earth
climate
organs
solar weather
astroplasmas
6
7
8
linking to data
 but simulation is only part of the story …
 linking computer modelling to experiment is vital to
the interpretation of many large experiments, eg LHC
 … and with big experiments set to generate
Pbyte/year data management is critical
 … but we have the web, don’t
we?
EPCC overview
Technology Transfer
Academic:
o National HPC facilities
o Research
o Support
Industry:
o Projects
o Consultancy
o Standards
Training:
o Academia
o Industry
European leadership:
o Visitor programmes
o Technology Transfer
o Strategic Planning
 70 staff, £3M turnover
 … possible through the “win-win-win” model
HPCx rationale
UK academic research is increasingly
dependent upon high-end compute facilities
recent technological and Grid advances
highlighted the need to upgrade UK resources
HPCx objectives are thus:
– aim “to deliver optimum service resulting in worldleading science”
– address “the problems involved in scaling to the
capability levels required”
What is HPCx?
 Consortium of leading UK organisations committed to
creating and managing the new resource for the next
6 years
– led by University of Edinburgh
 multi-stage project to deliver a world-class academic
computing resource, the largest in Europe
 £54M/$100M budget
 Grid-enabled, a key component in the UK e-Science
programme
HPCx Consortium Members
 UoE HPCx Ltd – wholly-owned subsidiary of the
University of Edinburgh and lead contractor
 EPCC, University of Edinburgh
 Daresbury Laboratory, CCLRC
 IBM
 the University of Edinburgh is one of the top 5 research
universities in the UK
 EPCC is the leading computer centre in Europe, bridging
the gap between academia and industry
 … and provides both HPC and novel computing solutions
to a wide range of problems and users
 long experience of providing national HPC services
including:
– Meiko Computing Surfaces
– Thinking Machines CM200
– Cray T3D/T3E
Daresbury Laboratory
 A multi disciplinary research lab with
over 500 people
 Provides large-scale research
facilities both for UK academic and
industrial research communities
 Runs the UK’s Collaborative
Computational Projects
 Daresbury hosts the HPCx
hardware
 IBM provides the technology for HPCx
 Long standing involvement in HPC including the
development of a number of ASCI machines and 4 of the
top dozen machines in the 21st TOP500 list:
– ASCI White: Rmax = 7.3 TFlop/s
– SP Power3 (6656 Processors): Rmax = 7.3 TFlop/s
– xSeries (1920 Processors): Rmax = 6.6 TFlop/s
– HPCx (1280 processors): Rmax = 3.2 TFlop/s
 IBM has the long term technology road map essential to
a 6 year project such as HPCx
HPCx in Place
HPCx: Phase 1
 System will be commissioned in three main stages
Phase 1 covering 2002-2004 consists of:
– 40 Regatta-H SMP nodes, 1280 processors
– Peak performance 6.6 TFlop/s, with 3.4 Tflop/s sustained
Linpack
• currently 16th in the Top500
– 1.28 TB total memory capacity
– Over 50 TB of storage capacity
– Double plane Colony switch with total
peak bandwidth of 250 MB/s per
processor
The e-Science Grid
CeSC
(Cambridge)
HPCx Phases 2 & 3
 Phase 2 (2004-2006)
– aiming for 6 TFlop/s sustained on Linpack and 2.5 TFlop/s on
sPPM
– O(48) Regatta-H+ SMP nodes
– interconnect upgraded to Federation switch
– doubling of I/O and storage
• already built a cluster with 8 Regatta-H+ frames and a prerelease Federation switch
• undertaking a phased upgrade during 1H2004
 Phase 3 (2006-2008)
– target of 12 TFlop/s sustained on Linpack
– may be additional nodes or alternative technology
HPCx Science Support
18 staff in 5 dual-centre functional support teams
Outreach
Life sciences
New applications
Applications Support
Users
Helpdesk
Training
Liaising with users
Terascaling
Capability applications
Scalable algorithms
Performance optimisation
Software Engineering
Underpinning technology
Grid/e-Science
Systems & Networking
Flexible and responsive capability computing service
Smooth transitions between phases
Technology
CPU Usage by Job Size
80%
128
Jan-03
Feb-03
Mar-03
Apr-03
May-03
Jun-03
Jul-03
Aug-03
Sep-03
Oct-03
Nov-03
Dec-03
256
1000000
>1024
2000000
40%
512
1500000
0%
AUs
HPCx Status: Usage
Utilisation of Capability Region
100%
60%
1024
20%
64
500000
32
0
8
Jan-03
Feb-03
Mar-03
Apr-03
May-03
Jun-03
Jul-03
Aug-03
Sep-03
Oct-03
Nov-03
Dec-03
16
HPCx Status: Application Areas
 currently 28 active consortia, and over 450 users
Usage by Application Area
Atomic &
Molecular Physics
Science Support
Materials
Chemistry
Engineering
Environmental
Science
 Life Sciences outreach activity supported by IBM
Atomic and Molecular Physics
 The UK Multiphoton, Electron Collision and Bose Einstein
Condensates (MECBEC) HPC Consortium
 Two flagship projects model two-electron atoms (helium)
and molecules (hydrogen) exposed to intense, ultra-short
laser pulses
 Modelling involves the grid solution of multidimensional, time-dependent partial differential
equations)
 Requires large amounts of computing power
 Visualisation techniques crucial in
extracting information
Simulation of Double Ionization of laser-driven helium performed at
Queen’s University Belfast
Environmental Science: POLCOMS
 POLCOMS is a multi-disciplinary model developed at the
Proudman Oceanographic Laboratory
 … a 3-D hydrodynamic model
integrating coasts and oceans using a
wide range of associated models
 POLCOMS is a step towards real
time modelling of coastal zones,
enabling better analysis of impacts
to, and sustainability of, the marine
environment
A simulation of Chlorophyll density in UK
waters using the POLCOMS model
Material Science: Crystal
 CRYSTAL computes electronic structure and related properties of
periodic systems
 Developed jointly by Daresbury and the University of Turin
 A Fortran90 and MPI program that performs Hartree-Fock, density
functional and other approximation calculations
 On HPCx, CRYSTAL used to calculate the structure of the Crambin
molecule, the largest Hartree-Fock calculation ever converged (1284
atoms)
– next, the Rusticyanin molecule (6284 atoms)
Engineering: UKTC
UK Turbulence Consortium is developing world
leading turbulence simulation codes using HPCx
essential that they can easily facilitate the
scientific usage of the resulting data
the calculated data are transferred from HPCx to a
remote site for analysis
so, the Grid is becoming increasingly important
Chemistry: RealityGrid
 RealityGrid is a UK collaboration
– which aims to grid-enable the realistic modelling and
simulation of complex condensed matter structures at
the meso and nanoscale levels
 HPCx terascaling team has worked on parallelisation &
optmisation of RealityGrid codes, such as LB3D
 RealityGrid also uses the Grid for data transfer,
computational steering, and remote visualisation
– using Globus on HPCx
– TeraGyroid won award at SC2003
The Vision
we want Edinburgh to lead the e-science
revolution
– to become the European equivalent of one of the big US centres, eg
San Diego Supercomputer Center

