Presentation is loading. Please wait.

Presentation is loading. Please wait.

Participation of JINR in the LCG and EGEE projects V.V.Korenkov (JINR, Dubna) NEC’2005, Varna 17 September, 2005.

Similar presentations


Presentation on theme: "Participation of JINR in the LCG and EGEE projects V.V.Korenkov (JINR, Dubna) NEC’2005, Varna 17 September, 2005."— Presentation transcript:

1 Participation of JINR in the LCG and EGEE projects V.V.Korenkov (JINR, Dubna) NEC’2005, Varna 17 September, 2005

2 Federating Worldwide Resources for the LHC

3 Infrastructure LCG/EGEE

4 Russian distributed Tier2 Cluster PNPI IHEP RRC KI ITEP JINR SINP MSU RRC-LHC LCG Tier1/Tier2 cloud CERN … Gbits/s FZK Regional connectivity: cloud backbone – Gbit’s/s to labs – 100–1000 Mbit/s Collaborative centers Tier2 cluster GRID access

5 The protocol between CERN, Russia and JINR on a participation in LCG Project has been approved in 2003. The tasks of the Russian institutes in the LCG : LCG software testing; evaluation of new Grid technologies (e.g. Globus toolkit 3) in a context of using in the LCG; event generators repository, data base of physical events: support and development; LHC Computing Grid Project (LCG)

6 The tasks of the Russian institutes & JINR in the LCG (2004 and 2005 years): LCG Deployment and Operation LCG Test suit Castor LCG AA- Genser&MCDB ARDA LHC Computing Grid Project (LCG)

7 JINR in LCG (2004 and 2005 years) LCG2 infrastructure was created at JINR The server for monitoring Russian LCG sites was installed; LCG web-portal was created in Russia and now its development is in progress: Tests on data transferring by the GridFTP protocol (GlobusToookit 3) were made. the toolkit GoToGrid on the automatic installation and tuning of the LCG-2 package was developed; development of the MCDB system; software for installation and control of MonaLisa clients on the base of RMS (Remote Maintenance Shell) was designed. Works to develop CASTOR2 system was in progress: development of the control process of the garbage collection module, communication to Oracle DB; participation in the work to create the TESTBED of the new gLite middleware; Testing of next components gLite: Metadata catalog, Fireman catalog Monitoring of WMS (Workload Management System) gLite testbed in INFN site gundam.cnaf.infn.it

8 MYRINET cluster COMMON PC-farm INTERACTIVE PC-farm Структура комплекса 130 CPU 18TB RAID-5 ATL~ 5 (15) TB 6 – Interactive 18 – Common PC-farm 30 – LHC 14 – MYRINET (Parallel) 20 – LCG 20 – File servers 8 – LCG-user interface

9 MCDB Web Interface Screen-shot http://mcdb.cern.ch Only Mozilla Browser Supported (for the time being)

10 Russian Data Intensive GRID (RDIG) Consortium EGEE Federation Eight Institutes made up the consortium RDIG (Russian Data Intensive GRID) as a national federation in the EGEE project. They are: IHEP - Institute of High Energy Physics (Protvino), IMPB RAS - Institute of Mathematical Problems in Biology (Pushchino), ITEP - Institute of Theoretical and Experimental Physics (Moscow), JINR - Joint Institute of Nuclear Physics, KIAM RAS - Keldysh Institute of Applied Mathematics, PNPI - Petersburg Nuclear Physics Institute (Gatchina), RRC KI - Russian Research Center “Kurchatov Institute”, SINP-MSU - Skobeltsyn Institute of Nuclear Physics (MSU).

11 Russian Contribution to EGEE RDIG as an operational and functional part of EGEE infrastructure (CIC, ROC, RC). Activities: SA1 - European Grid Operations, Support and Management SA2 – Network Resource Provision NA2 – Dissemination and Outreach NA3 – User Training and Induction NA4 - Application Identification and Support

12 JINR role and work in EGEE SA1 - European Grid Operations, Support and Management EGEE-RDIG monitoring and accounting. Middleware deployment and resource induction. Participation in the OMII and GT4 evaluation and in the gLite testing. LCG SC activity coordination in Russia. NA2 - Dissemination and Outreach Coordination of this activity in Russia, organization of EGEE RDIG Conference, Creation and run the RDIG Web site (http://www.egee-rdig.ru), dissemination in JINR Member states.http://www.egee-rdig.ru NA3 - User Training and Induction Organization of grid tutorials, induction courses and training courses for administrators. NA4 - Application Identification and Support Coordination of this activity in Russia, organization of HEP applications in Russia through the EGEE infrastructure.

13 Grid middleware evaluations The goal of the evaluations is to get a better understanding of the functionality, performance, solidity, interoperability, deployability, management and usability of components in different grid MW distributions Aid decision about possible usage of components for the EGEE MW and about provision of interoperability between these distributions and the EGEE MW Evaluation of OMII distribution by JINR and KIAM in February - April 2005 Evaluation of Globus Toolkit 4 by JINR, KIAM and SINP MSU in May - October 2005

14 Evaluation of OMII distribution by JINR and KIAM http://www.gridclub.ru/library/OMII-evaluaton-EGEE3.ppt Installation and configuration, supported platforms Performance, scalability and reliability studies of OMII services: JobService, DataService, dummy services Aspects of security, authorization, account management, resource allocation, administration in regard to the operation of a grid with many users, big virtual organizations and many resource centers Interoperability with gLite Workload Management System (WMS)

15 Evaluation of Globus Toolkit 4 by JINR, KIAM and SINP MSU http://theory.sinp.msu.ru/dokuwiki/doku.php?id=egee:gt4:gt4 Installation and configuration, supported platforms Performance, reliability, functional characteristics, interfaces of JAVA WS-Core, WS-GRAM, GridFTP, RLS, RFT, WS-MDS4, WS Delegation service Aspects of security, authorization, usability and administration comparison of corresponding GT4 and gLite components

16 Participation in EGEE MW testing Development of test suites for gLite (EGEE JRA1 activity) by JINR, IHEP and PNPI from June 2005 (continuing) WMS DAG tests WMS MPI tests WMS JDL tests R-GMA tests

17 RDIG мониторинг в ОИЯИ http://rocmon.jinr.ru:8080 http://rocmon.jinr.ru:8080

18 SC3 GOALS Service Challenge 1 (end of 2004): Demonstrate the possibility of throughput of 500 MByte/s to Tier1 in LCG environment. Service Challenge 2 (spring 2005): Maintain the throughput 500 MByte/s cumulative on all Tier1s for prolonged time, and evaluate the data transfer environment on Tier0 и Tier1s. Service Challenge 3 (Summer-end of 2005) Show reliable and stable data transfer on each Tier1: to disk -150 MByte/s, to tape - 60 MByte/s. All Tier1s and some Tier2s involved. Service Challenge 4 (Spring 2006): Prove the GRID infrastructure performance to handle the LHC data in proposed rate (from raw data transfer up to final analysis) with all Tier1s and majority of Tier2s. Final Goal: Build the production GRID-infrastructure on all Tier0, Tier1 и Tier2 according to the LHC experiments specifics.

19 Summary of Tier0/1/2 Roles Tier0 (CERN): safe keeping of RAW data (first copy); first pass reconstruction, distribution of RAW data and reconstruction output to Tier1; reprocessing of data during LHC down-times; Tier1 : safe keeping of a proportional share of RAW and reconstructed data; large scale reprocessing and safe keeping of corresponding output; distribution of data products to Tier2s and safe keeping of a share of simulated data produced at these Tier2s; Tier2 : Handling analysis requirements and proportional share of simulated event production and reconstruction. No long term data storage.

20 Tier2 Roles Tier2 roles vary by experiment, but include: –Production of simulated data; –Production of calibration constants; –Active role in [end-user] analysis  Must also consider services offered to T2s by T1s –e.g. safe-guarding of simulation output; –Delivery of analysis input. No fixed dependency between a given T2 and T1

21 A Simple T2 Model Each T2 is configured to upload MC data to and download data via a given T1 In case the T1 is logical unavailable, wait and retry For data download, retrieve via alternate route / T1

22 Tier1/2 Network Topology

23 Tier2 in Russia InstituteLinkCPUsDiskOS/Middleware IHEP 100 Mb/s half-duplex 5+1.6 TB …? ITEP 60 Mb/s 20 ?20 ?2 TB ? SL-3.0.4 (kernel 2.4.21) …? JINR 45Mb/s (1Gbit/s) 10 ?5 TB SLC3.0.X LCG-2_4_0, Castor, gridftp, gLite? SINP 1Gbit/s 30 ?2 TB ? SL-3.0.4 (kernel 2.4.21) gridftp, Castor

24 Universal Grid infrastructure in University Center of JINR Grid infrastructure is set of virtual machines (VMs) running on physical ones (hosts) Virtualisation was made using User Mode Linux current number of VMs is 36 (6 VMs on each of 6 hosts) all virtual resourses are grouped into independent testbeds which in turn can be used for different aims: system administrators and users training in grid field, debugging and testing custom grid services in desirable grid environment ● Course for system administrators using Nordugrid ARC middleware was successfully conducted on that infrastructure ● LCG2 or gLite installation and configuration course is in future plans

25 EGEE NA3 Courses in Dubna 28.06.2004 NA3. Introduction Courses 29-3-.03.2005 LCG2 Administrator`s Course, 06.09.2005 LCG-2 Induction Courses for CMS Users


Download ppt "Participation of JINR in the LCG and EGEE projects V.V.Korenkov (JINR, Dubna) NEC’2005, Varna 17 September, 2005."

Similar presentations


Ads by Google