Transcript Tier 3
Tier 3
and Tier 3 monitoring
17.05.2012
Ivan Kadochnikov
LIT JINR
Overview
WLCG structure
Tier3
T3mon concept
Monitoring tools
T3mon implementation
WLCG structure
Goals of WLCG
Hierarchic approach
Production and analysis
Argument for Tier3
Goals of WLCG
Turn detector signal into physics
Atlas for example:
1 Pb/s => The trigger => 200 Mb/s
15 petabytes of raw data per year
To do:
store
processes
analyze
Hierarchic approach
Tier 0
the CERN computer centre
safe-keeping the first copy of raw data
first pass reconstruction
Tier 1
11 centers all around the world
safe-keeping shares of raw, reconstructed, reprocessed and simulated
data
reprocessing
Tier 2
about 140 sites
production and reconstruction of simulated events
analysis
Production and analysis
Data selection algorithms improve
Calibration data change
Re-processing several times a year of all data gathered since
LHC start-up
Argument for Tier3
Analysis on Tier2 is inconvenient
Institutions have local computing resources
Local access and resources dedicated to analysis improve user
response time dramatically
Tier3
What is Tier3?
Types of Tier3
Compare and contrast: Tier2
Current status
Need for monitoring
What is Tier 3
non-pledged computing resources
institutional computing
no formal commitment to WLCG
for analysis
not really another level of the model
Types of Tier3 sites
Tier3 with Tier2 functionality
Collocated with Tier2
National analysis facilities
Non-grid Tier3’s
Compare and contrast:
Tier2
Different jobs
All VO users
Strict requirements on the
quality of service
Pledged resources
A set of mandatory grid
services on site
Processed data can go back in
the grid
VO central control
Tier3
Analysis only
Local/regional users
No QOS information
gathered globally (yet!)
Non-pledged resources
Often no grid services, may
be a full or partial set
No data allowed back in the
grid
Local control
Current status
More than 300 Tier3 sites right now
A survey by Atlas in 2010
Tier 3 come many different sizes
Storage methods vary
Different LRMSs used
Different ways to access WLCG
Many don’t have monitoring
Limited manpower
Need for monitoring
any system needs monitoring
some information is required on the global level
dataset popularity
usage statistics
grid services may not be present
existing grid monitoring systems can’t be used
Tier 3 sites have limited manpower
need an easy way to set up monitoring
T3mon concept
Users and requirements
What to monitor
Structure
Local monitoring
Global monitoring
Users and requirements
local administrators
detailed fabric monitoring
resources management systems (LRMS)
mass storage systems (MSS)
VO managers
general usage statistics and quality of service
global services
dataset popularity
What to monitor
local resources management systems
Proof
PBS
Condor
Oracle Grid Engine
mass storage systems (MSS):
XRootD
Lustre
Components
Local monitoring
detailed fabric monitoring
gather data used by the global system
present detailed data to local administrators
Global monitoring
aggregate metrics from local monitoring
give necessary information to central services
present data via Dashboard
Local monitoring system
MSG
Proof
Condor
XRootD
PBS
OGE
Lustre
Local DB
Publishing
agent
Global monitoring system
Dashboard
Data management
C
C
MSG
P
Local
monitoring
system
P
Local
monitoring
system
P
Local
monitoring
system
P
Local
monitoring
system
Tools
Ganglia
data flow
plug-in system
Dashboard
MSG
ActiveMQ
Ganglia
distributed monitoring system for clusters and Grids
MSG
Proof
Condor
XRootD
PBS
OGE
Lustre
21
Local DB
Publishing
agent
Why Ganglia?
easy to set up fabric monitoring
popular choice among Tier 3 sites
extension modules for LRMS and MSS monitoring
Ganglia data flow
head node
web frontend
string
metrics
rrdtool
gmetad
node
gmond
23
numeric metrics
node
xml by request
gmond
XDR via UDP
gmond
Ganglia web interface
Adding custom metrics
custom
monitoring
daemon
gmetric
module
callback
gmond
monitored
subsystem
monitored
subsystem
ganglia
25
Dashboard
“The Experiment
Dashboard's main goal is to
collect and expose to users
relevant information about
the services and applications
running on the grid
environment“
Other Applications
Tier3 software
Feeders
Messaging
system
Web Application
Collectors
Data Access Layer (DAO)
MSG
WLCG Messaging System for Grids
“Aims to help the integration and consolidation of the various
grid monitoring systems used in WLCG”
Based on ActiveMQ open-source message broker
T3Mon implementation
Project structure
Subsystem modules
Proof monitoring module
PBS monitoring module
Condor monitoring module
Lustre monitoring module
XRootD monitoring module
Testing infrastructure
Project structure
Python
SVN provided by CERN
RPM repository with a separate package for each monitoring
module
Each module handles one software system to be monitored
on Tier3
One configuration file for all modules
Proof
MSG
Ganglia
gmond
Proof plug-in
Proof
30
database
PBS
MSG
Ganglia
gmond
PBS plug-in
PBS
31
log files
Condor
Ganglia
MSG
gmond
Condor
condor_master
Condor plug-in
condor_startd
…
condor_quill
32
database
Lustre
Ganglia
gmond
Lustre plug-in
Lustre
33
/proc/fs/lustre
XRootD
MSG
xrddetmond
xrootd.py
cmsd
cmsd
Ganglia
xrootd
xrootd
database
mpxstats
xrdsummond
summary_to_ganglia.py
34
gmond
gmetric
Testing infrastructure
Goals
Document installing Ganglia on a cluster
Document configuring Tier3 subsystems for monitoring
Test modules in a minimal cluster environment
Clusters:
PBS: 3 nodes (1 head node, 2 worker nodes)
Proof: 3 nodes (1 hn, 2 wns)
Condor: 3 nodes (1 hn, 1 wn, 1 client)
OGE: 3 nodes (1 hn, 2 wn)
Lustre: 3 nodes (1 MDS, 1 OSS, 1 client)
Xrootd: 3 nodes (1 manager, 2 servers)
Xrootd II: 3 nodes (1 manager, 2 servers)
Development machine
Installation testing machine
Virtual testing infrastructure
23 nodes total
only 2 physical servers running virtualization software
(OpenVZ and Xen)
fast deployment and reconfiguring of nodes as required
performance is not a deciding factor
Results and plans
The project is nearing completion
Most modules are done
Proof and XRootD modules already testing on real clusters
Next steps:
Message consumers
OGE
Testing and support
Data transfer monitoring project
Thank you!