Transcript Slide 1

• E-Infrastructure hierarchy
• Networking and Computational facilities in Armenia
• ASNET AM Network
• Armenian National Grid Initiative
• Armenian ATLAS site (AM-04-YERPHI)
• History (Background)
• Site information
• Monitoring and job statistics
• Conclusion and Issues
E-Infrastructure hierarchy
INFN-BOLOGNA-T3
INFN-GENOVA
UTD-HEP
AM-04-YERPHI
ATLAS
provides a distributed
environment for sharing
computing power, storage
provides fast interconnection and advanced services
ALICE
instruments and
databases through
middleware
• E-Infrastructure hierarchy
• Networking and Computational facilities in Armenia
• ASNET AM Network
• Armenian National Grid Initiative
• Armenian ATLAS site (AM-04-YERPHI)
• History (Background)
• Site information
• Monitoring and job statistics
• Conclusion and Issues
ASNET AM Network
• Started to develop and realize since 1994 by the Institute for Informatics and
Automation Problems (IIAP NAS RA).
• ASNET-AM serves as the foundation for advanced computing applications
in Armenia.
• Links up academic, scientific, research and educational organizations.
• Provides advanced network services for 60 organizations in the major cities
of Armenia, such as Yerevan, Ashtarak, Byurakan, Abovian, Gyumri.
Connection for ASNET-AM is provided by GEANT and by the channel
rented from the local telecom companies (Arminco, ADC).
ASNET AM Topology
Armenian National Grid Initiative
Agreement of Establishment of Armenian Grid Joint Research Unit
was signed in September 2007
Main goals:
• To establish Armenian Infrastructure presence in international Grid
infrastructures
• To provide operations and security services
• To promote the uptake of Grid technologies in Armenia, the
interconnection of existing and future resources, and the deployment of
new applications
• To support researching Grid and Global Computing
Armenian National Grid Initiative
Computational Resources Topology
Organization
Cores
IIAP NAS RA
176
Yerevan
State
University
176
State
Engineering
University
48
IRPHE NAS
RA
48
AANL
48
Total
496
Armenian National Grid Initiative
Core Services
Access point
ce.iiapcluster.grid.am
se.iiapcluster.grid.am
ce.seuacluster.grid.am
se.ysu2cluster.grid.am
Core Services
ce.irphecluster.grid.am
se.irphecluster.grid.am
wms.grid.am
lfc.grid.am
bdii.grid.am
ce.ysucluster2.grid.am
voms.grid.am
se.ysucluster2.grid.am
ce.ysucluster.grid.am
se.ysucluster.grid.am
ce.yerphicluster.grid.am
se.yerphicluster.grid.am
• E-Infrastructure hierarchy
• Networking and Computational facilities in Armenia
• ASNET AM Network
• Armenian National Grid Initiative
• Armenian ATLAS site (AM-04-YERPHI)
• History (Background)
• Site information
• Monitoring and job statistics
• Conclusion and Issues
History (Background)
•
2007
–
•
WLCG ATLAS Site Deployment in AANL
the AANL site has been certified as “production site” of WLCG
• due to low quality of network connection( small bandwidth and frequent outages) site was put
in suspended mode
2008
– Developing of a national Grid infrastructure-ArmGrid
ArmGrid project is funded by Armenian government and International funding organizations (ISTC,
FP7)
•
2009
– The "Black Sea Interconnection" was activated to link the Academic and Research Networks of
South Caucasian countries
(Armenia, Georgia and Azerbaijan) to the European Geant-2 network.This opens up new possibilities for
ATLAS collaborators at AANL
•
2010
–
•
First ATLAS-SouthCaucasus Software/Computing Workshop & Tutorial. It fosters to establish
contacts between ATLAS collaborators and computing people in South Caucasian countries.
Workshop helps to better understand ADC requirements and configuration principles
2011
–
September: ATLAS Computing visit to the AANL
discussions between representatives of the ADC and AANL were very useful in order to make
progress on the establishment of AM-04-YERPHI as a ATLAS grid center.
–
October 20-th : Site status as ATLAS GRID site was approved by ICB
Site information
o
o
o
o
o
Computational resources
o
Model: Dell PE1950 III Additional Quad-Core Xeon
o
CPU: 6 nodes x 2 cpus per node X 4 cors per cpus= 48 cors
o
HDD:160 GB
o
RAM: 8 GB
For local analysis
o CPU: 6 nodes x 2 cpus
per node x 2 cors per cpus= 24
Storage Capacity
o
50TB
Site Core Services
o
MAUI/Torque PBS
o
SRM v1, v2
Supported VOs:
o
ATLAS
o
ALICE
o
ArmGrid
Site Information
Name
o
AM-04-YERPHI
Functionality
o
Grid Analysis (brokeroff), low priority production and local analysis
o
Tier3gs
Cloud association
o
NL cloud
Regional support
JINR
Voms group
o atlas/am
o
Technical support
o
2 sys admins (shared: 0.3 FTE)
Site information
ATLAS VO Support
DPM 10TB (nfs)
o ATLASSCRATCHDISK 2T
o ATLASLOCALGROUPDISK 7.00T
o ATLASPRODDISK 1024.00G
Frontier/ Squid cluster
xrootd cluster
Monitoring and job statistics
Running Jobs
Upgrade to sl6, EMI2
Commissioning
Network Hardware
components replacement
work
Monitoring and job statistics
Job Failure by Category and Exit Code
maui and queue conf. should be optimized
sw application, cvmfs (communication) problem
communication problem
Monitoring and job statistics
Efficiency and Wall Clock Consumption
Good efficiency of testing and MC production jobs
Data Transfers
1G files transferring finished successfully. Problems with bigger file transferring
SRM errors. Transfer failure for big files. Succeeded after big number of attempts
• E-Infrastructure hierarchy
• Networking and Computational facilities in Armenia
• ASNET AM Network
• Armenian National Grid Initiative
• Armenian ATLAS site (AM-04-YERPHI)
• History (Background)
• Site information
• Monitoring and job statistics
• Conclusion and issues
Conclusion and issues
• AM-04-YERPHI site is operational now.
AM-04-YERPHI
 As site administrators become more experienced, problems are
resolved faster.
Conclusion and issues
 Continuous monitoring of the infrastructure by the system
administrators ensures early error detection. Diagnostics help
to identify problems.
 Many configuration problems had been fixed during
commissioning, maintenance, but the job scheduling
configuration could still be improved.
• Ensuring a reliable network is critical. Issues which still need
addressing include
• Reliable connectivity and rapid transport of data being
used in the grid environment
• Related work focused on strengthening fault-tolerance.