Transcript Slide 1
• E-Infrastructure hierarchy • Networking and Computational facilities in Armenia • ASNET AM Network • Armenian National Grid Initiative • Armenian ATLAS site (AM-04-YERPHI) • History (Background) • Site information • Monitoring and job statistics • Conclusion and Issues E-Infrastructure hierarchy INFN-BOLOGNA-T3 INFN-GENOVA UTD-HEP AM-04-YERPHI ATLAS provides a distributed environment for sharing computing power, storage provides fast interconnection and advanced services ALICE instruments and databases through middleware • E-Infrastructure hierarchy • Networking and Computational facilities in Armenia • ASNET AM Network • Armenian National Grid Initiative • Armenian ATLAS site (AM-04-YERPHI) • History (Background) • Site information • Monitoring and job statistics • Conclusion and Issues ASNET AM Network • Started to develop and realize since 1994 by the Institute for Informatics and Automation Problems (IIAP NAS RA). • ASNET-AM serves as the foundation for advanced computing applications in Armenia. • Links up academic, scientific, research and educational organizations. • Provides advanced network services for 60 organizations in the major cities of Armenia, such as Yerevan, Ashtarak, Byurakan, Abovian, Gyumri. Connection for ASNET-AM is provided by GEANT and by the channel rented from the local telecom companies (Arminco, ADC). ASNET AM Topology Armenian National Grid Initiative Agreement of Establishment of Armenian Grid Joint Research Unit was signed in September 2007 Main goals: • To establish Armenian Infrastructure presence in international Grid infrastructures • To provide operations and security services • To promote the uptake of Grid technologies in Armenia, the interconnection of existing and future resources, and the deployment of new applications • To support researching Grid and Global Computing Armenian National Grid Initiative Computational Resources Topology Organization Cores IIAP NAS RA 176 Yerevan State University 176 State Engineering University 48 IRPHE NAS RA 48 AANL 48 Total 496 Armenian National Grid Initiative Core Services Access point ce.iiapcluster.grid.am se.iiapcluster.grid.am ce.seuacluster.grid.am se.ysu2cluster.grid.am Core Services ce.irphecluster.grid.am se.irphecluster.grid.am wms.grid.am lfc.grid.am bdii.grid.am ce.ysucluster2.grid.am voms.grid.am se.ysucluster2.grid.am ce.ysucluster.grid.am se.ysucluster.grid.am ce.yerphicluster.grid.am se.yerphicluster.grid.am • E-Infrastructure hierarchy • Networking and Computational facilities in Armenia • ASNET AM Network • Armenian National Grid Initiative • Armenian ATLAS site (AM-04-YERPHI) • History (Background) • Site information • Monitoring and job statistics • Conclusion and Issues History (Background) • 2007 – • WLCG ATLAS Site Deployment in AANL the AANL site has been certified as “production site” of WLCG • due to low quality of network connection( small bandwidth and frequent outages) site was put in suspended mode 2008 – Developing of a national Grid infrastructure-ArmGrid ArmGrid project is funded by Armenian government and International funding organizations (ISTC, FP7) • 2009 – The "Black Sea Interconnection" was activated to link the Academic and Research Networks of South Caucasian countries (Armenia, Georgia and Azerbaijan) to the European Geant-2 network.This opens up new possibilities for ATLAS collaborators at AANL • 2010 – • First ATLAS-SouthCaucasus Software/Computing Workshop & Tutorial. It fosters to establish contacts between ATLAS collaborators and computing people in South Caucasian countries. Workshop helps to better understand ADC requirements and configuration principles 2011 – September: ATLAS Computing visit to the AANL discussions between representatives of the ADC and AANL were very useful in order to make progress on the establishment of AM-04-YERPHI as a ATLAS grid center. – October 20-th : Site status as ATLAS GRID site was approved by ICB Site information o o o o o Computational resources o Model: Dell PE1950 III Additional Quad-Core Xeon o CPU: 6 nodes x 2 cpus per node X 4 cors per cpus= 48 cors o HDD:160 GB o RAM: 8 GB For local analysis o CPU: 6 nodes x 2 cpus per node x 2 cors per cpus= 24 Storage Capacity o 50TB Site Core Services o MAUI/Torque PBS o SRM v1, v2 Supported VOs: o ATLAS o ALICE o ArmGrid Site Information Name o AM-04-YERPHI Functionality o Grid Analysis (brokeroff), low priority production and local analysis o Tier3gs Cloud association o NL cloud Regional support JINR Voms group o atlas/am o Technical support o 2 sys admins (shared: 0.3 FTE) Site information ATLAS VO Support DPM 10TB (nfs) o ATLASSCRATCHDISK 2T o ATLASLOCALGROUPDISK 7.00T o ATLASPRODDISK 1024.00G Frontier/ Squid cluster xrootd cluster Monitoring and job statistics Running Jobs Upgrade to sl6, EMI2 Commissioning Network Hardware components replacement work Monitoring and job statistics Job Failure by Category and Exit Code maui and queue conf. should be optimized sw application, cvmfs (communication) problem communication problem Monitoring and job statistics Efficiency and Wall Clock Consumption Good efficiency of testing and MC production jobs Data Transfers 1G files transferring finished successfully. Problems with bigger file transferring SRM errors. Transfer failure for big files. Succeeded after big number of attempts • E-Infrastructure hierarchy • Networking and Computational facilities in Armenia • ASNET AM Network • Armenian National Grid Initiative • Armenian ATLAS site (AM-04-YERPHI) • History (Background) • Site information • Monitoring and job statistics • Conclusion and issues Conclusion and issues • AM-04-YERPHI site is operational now. AM-04-YERPHI As site administrators become more experienced, problems are resolved faster. Conclusion and issues Continuous monitoring of the infrastructure by the system administrators ensures early error detection. Diagnostics help to identify problems. Many configuration problems had been fixed during commissioning, maintenance, but the job scheduling configuration could still be improved. • Ensuring a reliable network is critical. Issues which still need addressing include • Reliable connectivity and rapid transport of data being used in the grid environment • Related work focused on strengthening fault-tolerance.