Presentation is loading. Please wait.

Presentation is loading. Please wait.

LISHEP, Rio de Janeiro, 20 February 2004 Russia in LHC DCs and EDG/LCG/EGEE V.A. Ilyin Moscow State University.

Similar presentations


Presentation on theme: "LISHEP, Rio de Janeiro, 20 February 2004 Russia in LHC DCs and EDG/LCG/EGEE V.A. Ilyin Moscow State University."— Presentation transcript:

1 LISHEP, Rio de Janeiro, 20 February 2004 Russia in LHC DCs and EDG/LCG/EGEE V.A. Ilyin Moscow State University

2 MONARC project regional group LHC Computing GRID: the “cloud” view CERN Tier3 physics department    Desktop Germany UK France Italy CERN Tier1 USA Tier1 The opportunity of Grid technology Tier2 Uni a Lab c Uni n Lab m Lab b Uni b Uni y Uni x Russia

3 Russian Tier2-Cluster Russian regional center for LHC computing (RRC-LHC) Cluster of institutional computing centers with Tier2 functionality and summary resources at 50- 70% level of the canonical Tier1 center for each experiment (ALICE, ATLAS, CMS, LHCb): analysis; simulations; users data support. Participating institutes: Moscow ITEP, RRC KI, MSU, LPI, MEPhI… Moscow region JINR, IHEP, INR RAS St.Petersburg PNPI RAS, … Novosibirsk BINP SB RAS Coherent use of distributed resources by means of LCG (EDG+VDT, …) technologies. Active participation in the LCG Phase1 Prototyping and Data Challenges (at 5% level). 200220032004 Q42007 CPU kSI95 51025-35410 Disk TB 71230850 Tape TB -6301250 International connectivity (CERN) Mbps 20155155/…Gbps/…

4 Russia Country Map Three regions are indicated on the map, where HEP centers are located: Moscow, St-Petersburg and Novosibirsk

5 Site (Centre) Acc./Coll. HEP Fac. Other Exp’sParticipation in major HEP Int. Collab. BINP SB RAS (Novosibirsk) http://www.inp.nsk.su VEPP-2M (linear collider at 1.4 GeV) VEPP-4 (linear collider up to 6 GeV) Non-Acc. HEP Exp’s. (Neutrino Phys., etc), Synchrotron Rad. F. CERN: ATLAS, LHC-acc, CLIC FNAL: Tevatron-acc DESY: TESLA KEK: BELLE SLAC: BaBar IHEP (Protvino, Moscow Region) http://www.ihep.su U-70 (fix target, proton beam 70 GeV) Medical Exp’sBNL: PHENIX, STAR CERN: ALICE, ATLAS, CMS, LHCb DESY: ZEUS, HERA-B, TESLA FNAL: D0, E-781(Selex) ITEP (Moscow) http://www.itep.ru U-10 (fix target, proton beam 10 GeV) Non-Acc. HEP Exp’s. (Neutrino Phys., etc) CERN: ALICE, ATLAS, CMS, LHCb, AMS DESY: H1, HERMES, HERA-B, TESLA FNAL: D0, CDF, E-781(Selex) KEK: BELLE DAFNE: KLOE JINR (Dubna, Moscow Region) http://www.jinr.ru Nuclotron (heavy ions coll. at 6 GeV/n) Low Ener. Acc., Nuclear Reactor, Synchrotron Rad.F., Non-Acc. HEP Exp’s: Neutrino Phys., Medical Exp’s, Heavy-ion Physics BNL: PHENIX, STAR CERN: ALICE, ATLAS, CMS, NA48, COMPASS, CLIC, DIRAC DESY: H1, HERA-B, HERMES, TESLA FNAL: D0, CDF KEK: E391a

6 Site (Centre) HEP Acc./Coll. Other Exp’s Participation in major HEP Int. Collab. INR RAS (Troitsk, Moscow region, Research Centre) http://www.inr.ac.ru Low Energy Acc., Non-Acc. HEP Exp’s (Neutrino Phys.) CERN: ALICE, CMS, LHCb KEK: E-246 TRIUMF: E-497 RRC KI (Moscow, Res. Centre) http://www.kiae.ru Low Energy Acc., Nuclear Reactors, Synchrotron Rad. F. BNL: PHENIX CERN: ALICE, AMS MEPhI (Moscow, University) http://www.mephi.ru Low Energy Acc., Nuclear Reactor BNL: STAR CERN: ATLAS DESY: ZEUS, HERA-B, TESLA PNPI RAS (Gatchina, St-Petersburg region, Research Centre) http://www.pnpi.spb.ru Mid/Low Energy Acc., Nulcear Reactor BNL: PHENIX CERN: ALICE, ATLAS, CMS, LHCb DESY: HERMES FNAL: D0, E-781(Selex) SINP MSU (Moscow, University) http://www.sinp.msu.ru Low Energy Acc., Non-Acc. HEP Exp. (EAS-1000) CERN: ATLAS, CMS, AMS, CLIC DESY: ZEUS, TESLA FNAL: D0, E-781(Selex)

7 Goals of Russian (distributed) Tier2  to provide a full-scale participation of Russian physicists in the analysis only in this case Russian investments in LHC would lead to the final goal of obtaining a new fundamental knowledge on the structure of matter  to open wide possibilities for participation of students and young scientists in research at LHC support and improve a high level of scientific schools in Russia  participation in the creation of international LHC Computing GRID will mean for Russia an access to new advanced computing techniques

8 Functions of Russian (distributed) Tier2 physical analysis of AOD (Analysis Object Data); access to (external) ESD/RAW and SIM data bases  for preparing necessary (local) AOD sets ; replication of AOD sets from Tier1/Tier2 grid (cloud); event simulation  at the level of 5-10% of the whole SIM data bases for each experiment; replication and store of 5-10% of ESD  required for testing the procedures of the AOD creation; storage of data produced by users.  participation in distributed storage of full ESD data (Tier1 function)…?

9 Architecture of Russian (distributed) Tier2 a clusterTier2  RRC-LHC will be a cluster of institutional centers with Tier2 functionality distributed system - DataGrid cloud of Tier2(/Tier3) centers coherent interaction  a coherent interaction of computing centers of participating Institutes: each Institute knows its resources but can get significantly more if others agree; summary resources4-5  for each Collaboration summary resources (of about 4-5 basic institutional centers) will reach the level of 50-70% of a canonical Tier1 center: each Collaboration knows its summary resources but can get significantly more if other Collaborations agree; global grid for data store and access  RRC-LHC will be connected to Tier1 at CERN and/or to other Tier(s)1 in a context of a global grid for data store and access : each Institute and each Collaboration can get significantly more if other reg.centers agree.

10 Russian Regional Center: the DataGrid cloud PNPI IHEP RRC KI ITEP JINR SINP MSU The opportunity of Grid technology RRC-LHC LCG Tier1/Tier2 cloud CERN … Gbits/s FZK Regional connectivity: cloud backbone – Gbit’s/s to labs – 100–1000 Mbit/s Collaborative centers Tier2 cluster GRID access

11 “Users”-“Tasks” and resources (analysis from 2001 – need to be updated – conception of Tier2s) The number of active users is main parameter for estimation of the resources needed. We did some estimates, in particular based on extrapolation of Tevatron analysis tasks performed by our physicists (single top production at D0, …). Thus, in some “averaging” figures: an “user task” – analysis of 10 7 events per a day (8 hours) by one physicist ALICE ATLAS CMS LHCb 40 60 60 40 In the following we estimate RRC resources (Phase 1) by the assumption that our participation in SIM data base production is at 5% level for each experiment. Very poor understanding of this key (for Tier2) characteristics!

12 Resources required by the 2008 ALICEATLASRDMS CMSLHCbTotal CPU (KSI95)100120 70410 Disk (TB)200250 150850 Tape (TB)300400 1501250 We suppose: each active user will create local AOD sets ~10 times per year, and keep these sets on the disks during the year the general AOD sets will be replicated from the Tier1 cloud ~10 times per year, storing previous sets on the tapes. The disk space usage will be partitioned as  15% to store general AOD+TAG sets;  15% to store local sets of AOD+TAG;  15% to store users data;  15% to store current sets of sim.data (SIM-AOD, partially SIM-ESD);  30-35% to store the 10% portion of ESD;  5-10% cache.

13 Construction timeline Timeline for the RRC-LHC resources at the construction phase: 200620072008 15%30%55% After 2008 investments will be necessary for supporting the computing and storage facilities and increasing the CPU power and storage space. In 2008 about 30% of the expenses in 2008. Every next year: renewing of 1/3 of CPU, increase the disk space for 50%, and increase the tape storage space for 100%.

14 Financial aspects Phase1 (2001-2005) 2.5 MCHF equipment, 3.5 MCHF network + initial inivestments to some regional networks Construction phase (2006-2008) 10 MCHF equipment, 3 MCHF network _____________________________________ in total (2001-2008) 19 MCHF 2009 – 200x 2 MCHF/year 2003, December – new Protocol has been signed by Russia and CERN on frameworks for Russia participation in LHC project on period from 2007, including: 1) M&O, 2) computing in Exps, 3) RRC-LHC and LCG.

15 LHCb DC03 Resource Usage c.f. DC02 –3.3M evts –49 days CERN 44% Bologna 30% Lyon 18% RAL 3.9% Cambridge 1.1% Moscow 0.8% Amsterdam 0.7% Rio 0.7% Oxford 0.7% ITEP Moscow IHEP Protvino JINR Dubna SINP MSU

16 CMS Productions (2001) Simulation Digitization GDMP Common Production tools (IMPALA) No PUPU CERN Fully operational FNAL Moscow (First!) INFN Caltech UCSD UFL Bristol Wisconsin IN2P3Not Op. HelsinkiNot Op.

17 Man Power for CMS Computing in Russia Institutesfarm administration installation& production running production tools PRS SW code ORCA Physics generators  FTE SINP MSU & RCC 11.5-3.5410 JINR 0.4 -2.32.85.9 ITEP 0.80.212-4 IHEP 0.30.60.20.41.32.8 Kharkov (Ukrain) 10.6---1.6 LPI 0.60.4---1  FTE 4.13.71.28.28.1 in a total – 25.3 FTE Sept.2003

18 IMPALA/BOSS integration with GRID GRID CMKIN IMPALA Job Executer Gate Keeper Batch Manager BOSS Dolly WN1 WN2 Wnn UI CE Recource Broker Job Jobs MySQL DB CERN RefDB Environment SINP MSU (Moscow) – JINR (Dubna) – INFN (Padova) 2002

19 Russia in LCG We have started activity in LCG in autumn 2002. Russia joined to the LCG-1 infrastructure (CERN press-release 29.09.2003). First SINP MSU, soon RRC KI, JINR, ITEP and IHEP (already to LHC-2). http://goc.grid-support.ac.uk/gridsite/gocmain/monitoring/ Manpower contribution to LCG (started in May 2003): the Agreement is under signing by CERN and Russia and JINR officials, 3 tasks for our responsibility: 1) testing new GRID mw to be used in LCG 2) evaluation of new-on-the-market GRID mw (first task – evaluation of OGSA/GT3) 3) common solutions for event generators (event data bases) Twice per year (spring-autumn) meetings of the Russia-CERN Joint Working Group on Computing. Next meeting on 19 March at CERN.

20

21 Information System testing for LCG-1 Elena Slabospitskaya Institute for High Energy Physics, Protvino, Russia 18.07.2003

22 Information System testing for LCG-1 UI RB CE WN Edg-job-submit CondorG PBS, LSF.... CondorG Globus EDG Globusrun Gatekeeper The schema of the job submission via RB and directly to the CE via Globus GRAM Network server Workload Manager CondorG

23 It was designed and realized OGSA/GT3 testbed (named 'Beryllium') on the basis of PCs located at CERN and SINP MSU modelling a GT3 based Grid system. http://lcg.web.cern.ch/LCG/PEB/GTA/LCG_GTA_OGSA.htm Created software for common library of MC generators, GENSER, http://lcgapp.cern.ch/project/simu/generator/ New project MCDB (Monte Carlo Data Base) for LCG AA is proposed with Russia responsibility, as common solution for storing and providing access cross the LCG sites to the samples of events at partonic level.

24 The simplified schema of Beryllium testbed (CERN-SINP) ● The resource broker plays a central role: – Accepts requests from the User – Using the Information Service information, selects the suitable computer elements – Reserve the selected Computing Element – Communicates to the user a “ticket” to allow job submission – Maintains a list of all jobs running and receive confirmation – messages of the ongoing processing from the CEs – At job end, it updates the table of running job/CE status

25 Externally Funded LCG Personnel at CERN

26 EU-DataGrid Russia institutes participated in the EU-DataGrid project  WP6 (Testbed and Demonstration)  WP8 (HEP Application) 2001: Grid information service (GRIS-GIIS), DataGrid Certificate Authority (CA) and Registration Authority (RA). WP6 Testbed0 (Spring-Summer 2001) – 2 sites. WP6 Testbed1 (Autumn 2001) – 4 active sites (SINP MSU, ITEP, JINR, IHEP), significant resources (160 CPUs, 7.5 TB disk). 2002: Testbed1 new active site – PNPI Testbed1 Virtual Organizations (VO) – WP6, ITeam WP8 – CMS VO, ATLAS and ALICE VO ’ s, WP8 CMS MC Run (spring) – ~1 Tbyte data transferred to CERN and FNAL, Resource Broker (RB) – SINP MSU +CERN+INFN experiment Metadispatcher (MD) – colaboration with Keldysh Inst.Appl.Math. (Moscow) – algorithms of dispatchering (scheduling) jobs in DataGrid environment.

27 CEWN lhc01.sinp.msu.ru lhc02.sinp.msu.ru SINP MSU Site SE lhc03.sinp.msu.ru EDG Software deployment at SINP MSU (example - CMS VO, 7 June 2002) SINP MSU RB+ Information Index lhc20.sinp.msu.ru User Interface Node lhc04.sinp.msu.ru CERN lxshare0220.cern.ch Padova grid011.pd.infn.it

28

29 EGEE Enabling Grids for e-Science in Europe – EGEE EU project approved to provide partial funding for operation of a general e-Science grid in Europe, including the supply of suitable middleware. EGEE is proposed as a project funded by the European Union under contract IST- 2003-508833. Budget – about 32 Meuro per 2004-2005. EGEE provides funding for 70 partners, large majority of which have strong HEP ties. Russia: 8 institutes (SINP MSU, JINR, ITEP, IHEP, RRC KI, PNPI, KIAM RAS, IMPB RAS), budget 1 Meuro per 2004-2005 Russian matching of the EC budget is in good shape (!)

30 EGEE Partner Federations Integrate regional Grid efforts

31 EGEE Timeline

32 Distribution of Service Activities over Europe : Operations Management at CERN; Core Infrastructure Centres in the UK, France, Italy, Russia (PM12) and at CERN, responsible for managing the overall Grid infrastructure; Regional Operations Centres, responsible for coordinating regional resources, regional deployment and support of services. Russia: CIC – SINP MSU, RRC KI ROC – IHEP, PNPI, IMPB RAS Dissemination&Outreach – JINR,

33  S.E. Europe, Russia: Catching Up  Latin Am., Mid East, China:Keeping Up  India, Africa: Falling Behind ICFA SCIC Feb 2004

34 LHC Data Challenges Typical example – transferring of 100 Gbyte of data from Moscow to CERN for one working day  50 Mbps of bandwidth !

35 GLORIAD 10 Gbps REGIONAL CONNECTIVITY for RUSSIA HEP Moscow 1 Gbps IHEP 8 Mbps (m/w), under construction 100 Mbps fiber-optic (Q1-Q2 2004?) JINR 45 Mbps, 100-155 Mbps (Q1-Q2 2004), Gbps (2004-2005) INR RAS 2 Mbps+2x4Mbps(m/w) BINP 1 Mbps, 45 Mbps (2004 ?), … GLORIAD PNPI 512 Kbps (commodity Internet), and 34 Mbps f/o but (!) budget is only for 2 Mbps INTERNATIONAL CONNECTIVITY for RUSSIA HEP USA NaukaNET 155 Mbps GEANT 155 Mbps basic link, plus 155 Mbps additional link for GRID projects Japan through USA by FastNET, 512 Kbps Novosibirsk(BINP) – KEK(Belle)


Download ppt "LISHEP, Rio de Janeiro, 20 February 2004 Russia in LHC DCs and EDG/LCG/EGEE V.A. Ilyin Moscow State University."

Similar presentations


Ads by Google