January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 1 SIMULATION OF DAILY ACTIVITITIES AT REGIONAL CENTERS MONARC Collaboration Alexander Nazarenko.

Slides:



Advertisements
Similar presentations
31/03/00 CMS(UK)Glenn Patrick What is the CMS(UK) Data Model? Assume that CMS software is available at every UK institute connected by some infrastructure.
Advertisements

Status GridKa & ALICE T2 in Germany Kilian Schwarz GSI Darmstadt.
Plan for Measurements Monarc General Meeting 26 July 1999 A. Brunengo - INFN Genova P. Capiluppi - INFN Bologna A. Ghiselli - INFN Cnaf L. Perini - INFN.
1 GridTorrent Framework: A High-performance Data Transfer and Data Sharing Framework for Scientific Computing.
10-Feb-00 CERN Building a Regional Centre A few ideas & a personal view CHEP 2000 – Padova 10 February 2000 Les Robertson CERN/IT.
T1 at LBL/NERSC/OAK RIDGE General principles. RAW data flow T0 disk buffer DAQ & HLT CERN Tape AliEn FC Raw data Condition & Calibration & data DB disk.
Distributed IT Infrastructure for U.S. ATLAS Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
I.1 Distributed Systems Prof. Dr. Alexander Schill Dresden Technical University Computer Networks Dept.
September, 1999MONARC - Distributed System Simulation I.C. Legrand1 MONARC Models Of Networked Analysis at Regional Centers Iosif C. Legrand (CERN/CIT)
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
MONARC Simulation Framework Corina Stratan, Ciprian Dobre UPB Iosif Legrand, Harvey Newman CALTECH.
POLITEHNICA University of Bucharest California Institute of Technology National Center for Information Technology Ciprian Mihai Dobre Corina Stratan MONARC.
The new The new MONARC Simulation Framework Iosif Legrand  California Institute of Technology.
Large scale data flow in local and GRID environment V.Kolosov, I.Korolko, S.Makarychev ITEP Moscow.
October 24, 2000Milestones, Funding of USCMS S&C Matthias Kasemann1 US CMS Software and Computing Milestones and Funding Profiles Matthias Kasemann Fermilab.
Fabric Management for CERN Experiments Past, Present, and Future Tim Smith CERN/IT.
MONARC : results and open issues Laura Perini Milano.
July, 2000.Simulation of distributed computing systems I.C. Legrand1 MONARC Models Of Networked Analysis at Regional Centers Iosif C. Legrand (CALTECH)
INTRODUCTION The GRID Data Center at INFN Pisa hosts a big Tier2 for the CMS experiment, together with local usage from other HEP related/not related activities.
Jean-Yves Nief CC-IN2P3, Lyon HEPiX-HEPNT, Fermilab October 22nd – 25th, 2002.
Data Import Data Export Mass Storage & Disk Servers Database Servers Tapes Network from CERN Network from Tier 2 and simulation centers Physics Software.
Fermilab User Facility US-CMS User Facility and Regional Center at Fermilab Matthias Kasemann FNAL.
LHC Computing Review - Resources ATLAS Resource Issues John Huth Harvard University.
Modeling Regional Centers with MONARC Simulation Tools Modeling LHC Regional Centers with the MONARC Simulation Tools Irwin Gaines, FNAL for the MONARC.
No vember 15, 2000 MONARC Project Status Report Harvey B Newman (CIT) MONARC Project Status Report Harvey Newman California Institute.
Computing for LHCb-Italy Domenico Galli, Umberto Marconi and Vincenzo Vagnoni Genève, January 17, 2001.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
6/26/01High Throughput Linux Clustering at Fermilab--S. Timm 1 High Throughput Linux Clustering at Fermilab Steven C. Timm--Fermilab.
7April 2000F Harris LHCb Software Workshop 1 LHCb planning on EU GRID activities (for discussion) F Harris.
December 10,1999: MONARC Plenary Meeting Harvey Newman (CIT) Phase 3 Letter of Intent (1/2)  Short: N Pages è May Refer to MONARC Internal Notes to Document.
October, 2000.A Self Organsing NN for Job Scheduling in Distributed Systems I.C. Legrand1 Iosif C. Legrand CALTECH.
US ATLAS Tier 1 Facility Rich Baker Brookhaven National Laboratory DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National Laboratory.
CERN IT Department CH-1211 Genève 23 Switzerland t Frédéric Hemmer IT Department Head - CERN 23 rd August 2010 Status of LHC Computing from.
Monte Carlo Data Production and Analysis at Bologna LHCb Bologna.
6/23/2005 R. GARDNER OSG Baseline Services 1 OSG Baseline Services In my talk I’d like to discuss two questions:  What capabilities are we aiming for.
ATLAS WAN Requirements at BNL Slides Extracted From Presentation Given By Bruce G. Gibbard 13 December 2004.
26 Nov 1999 F Harris LHCb computing workshop1 Development of LHCb Computing Model F Harris Overview of proposed workplan to produce ‘baseline computing.
Virtualization and Databases Ashraf Aboulnaga University of Waterloo.
HIGUCHI Takeo Department of Physics, Faulty of Science, University of Tokyo Representing dBASF Development Team BELLE/CHEP20001 Distributed BELLE Analysis.
Xrootd Monitoring and Control Harsh Arora CERN. Setting Up Service  Monalisa Service  Monalisa Repository  Test Xrootd Server  ApMon Module.
Distributed applications monitoring at system and network level A.Brunengo (INFN- Ge), A.Ghiselli (INFN-Cnaf), L.Luminari (INFN-Roma1), L.Perini (INFN-Mi),
July 26, 1999MONARC Meeting CERN MONARC Meeting CERN July 26, 1999.
Preliminary results Monarc Testbeds Working Group A.Brunengo (INFN- Ge), A.Ghiselli (INFN-Cnaf), L.Luminari (INFN-Roma1), L.Perini (INFN-Mi), S.Resconi.
Preliminary Validation of MonacSim Youhei Morita *) KEK Computing Research Center *) on leave to CERN IT/ASD.
NA62 computing resources update 1 Paolo Valente – INFN Roma Liverpool, Aug. 2013NA62 collaboration meeting.
6 march Building the INFN Grid Proposal outline a.ghiselli,l.luminari,m.sgaravatto,c.vistoli INFN Grid meeting, milano.
Tier 1 at Brookhaven (US / ATLAS) Bruce G. Gibbard LCG Workshop CERN March 2004.
The ATLAS Computing Model and USATLAS Tier-2/Tier-3 Meeting Shawn McKee University of Michigan Joint Techs, FNAL July 16 th, 2007.
W.A.Wojcik/CCIN2P3, Nov 1, CCIN2P3 Site report Wojciech A. Wojcik IN2P3 Computing Center URL:
Computing Issues for the ATLAS SWT2. What is SWT2? SWT2 is the U.S. ATLAS Southwestern Tier 2 Consortium UTA is lead institution, along with University.
US ATLAS Tier 1 Facility Rich Baker Deputy Director US ATLAS Computing Facilities October 26, 2000.
10-Jan-00 CERN Building a Regional Centre A few ideas & a personal view CHEP 2000 – Padova 10 January 2000 Les Robertson CERN/IT.
U.S. ATLAS Computing Facilities DOE/NFS Review of US LHC Software & Computing Projects Bruce G. Gibbard, BNL January 2000.
June 22, 1999MONARC Simulation System I.C. Legrand1 MONARC Models Of Networked Analysis at Regional Centres Distributed System Simulation Iosif C. Legrand.
The MEG Offline Project General Architecture Offline Organization Responsibilities Milestones PSI 2/7/2004Corrado Gatto INFN.
Handling of T1D0 in CCRC’08 Tier-0 data handling Tier-1 data handling Experiment data handling Reprocessing Recalling files from tape Tier-0 data handling,
U.S. ATLAS Computing Facilities U.S. ATLAS Physics & Computing Review Bruce G. Gibbard, BNL January 2000.
BNL dCache Status and Plan CHEP07: September 2-7, 2007 Zhenping (Jane) Liu for the BNL RACF Storage Group.
Main parameters of Russian Tier2 for ATLAS (RuTier-2 model) Russia-CERN JWGC meeting A.Minaenko IHEP (Protvino)
LHCb Current Understanding of Italian Tier-n Centres Domenico Galli, Umberto Marconi Roma, January 23, 2001.
December 10,1999: MONARC Plenary Meeting Harvey Newman (CIT) MONARC Plenary December 9 Agenda u Introductions HN, LP15’ è Status of Actual CMS ORCA databases.
Meeting with University of Malta| CERN, May 18, 2015 | Predrag Buncic ALICE Computing in Run 2+ P. Buncic 1.
Apr. 25, 2002Why DØRAC? DØRAC FTFM, Jae Yu 1 What do we want DØ Regional Analysis Centers (DØRAC) do? Why do we need a DØRAC? What do we want a DØRAC do?
SAM at CCIN2P3 configuration issues
The COMPASS event store in 2002
LHCb Computing Model and Data Handling Angelo Carbone 5° workshop italiano sulla fisica p-p ad LHC 31st January 2008.
GridTorrent Framework: A High-performance Data Transfer and Data Sharing Framework for Scientific Computing.
The ATLAS Computing Model
Database System Architectures
Development of LHCb Computing Model F Harris
Presentation transcript:

January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 1 SIMULATION OF DAILY ACTIVITITIES AT REGIONAL CENTERS MONARC Collaboration Alexander Nazarenko and Krzysztof Sliwa

January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 2 MONARC SIMULATION TOOLS The simulation tool developed within MONARC is based on Java technology which provides adequate tools for developing a flexible and distributed process oriented simulation. Java has a built-in support for multi-threaded objects and for concurrent processing, which can be used for simulation purposes provided a dedicated scheduling mechanism is developed (this “simulation engine” has been developed by Iosif Legrand). Java also offers good support for graphics which can be easily interfaced with the simulation code. Proper graphics tools, and ways to analyse data interactively, are essential in any simulation project (Alex Nazarenko’s contributions were the greatest here) MONARC simulation and modelling of distributed computing systems provides a realistic description of complex data, and data access patterns and of very large number of jobs running on large scale distributed systems and exchanging very large amount of data (Data Model developed by Krzysztof Sliwa and Iosif Legrand)

January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 3 Baseline Model for Daily Activities Physics Group Analysis Physics Group Selection Reconstruction ESD Redefinition of AOD+TAG Replication (FTP) Monte-Carlo jobs/day jobs/day 2 times/year once/month after Reconstruction Event processing rate: 1, 000, 000, 000 events/day

January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 4 Regional Center Model

January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 5 VALIDATION MEASUREMENTS “atlobj02”-local Raw DataDB DB on local disk 2 CPUs x 300MHz SI95/CPU “monarc01”-local Raw DataDB DB on local disk 4 CPUs x 400MHz 17.4 SI95/CPU DB on AMS Server server : “atlobj02” client : “monarc01” Raw DataDB Local DB access DB access via AMS monarc01 is a 4 CPUs SMP machine atlobj02 is a 2 CPUs SMP machine Multiple jobs reading concurrently objects from a data base. êObject Size = 10.8 MB

January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 6 Validation Measurements I The AMS Data Access Case SimulationMeasurements Raw DataDB LAN 4 CPUs Client

January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 7 Validation Measurements I Simulation Results DB access via AMS Local DB access Simulation results for AMS & Local Data Access 1,2,4,8,16,32 parallel jobs executed on 4 CPUs SMP system

January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 8 Validation Measurements I Local DB access 32 jobs The Distribution of the jobs processing time Simulation mean Measurement mean 114.3

January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 9 Validation Measurements I Measurements & Simulation SimulationMeasurements

January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 10 Validation Measurements II 1000BaseT Server Client Test 1 CNAF Sun Ultra5, 333 MHz sunlab1 gsun 10BaseT Test 2 PADOVA Sun Ultra10, 333 MHz Sun Enterprise 4X450 MHz cmssun4 vlsi06 2 Mbps Test 3 CNAF-CERN Sun Ultra15, 333 MHz sunlab1 monarc01 Sun Sparc20, 125 MHz

January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 11 Validation Measurements II Test 1 Gigabit Ethernet Client - Server SimulationMeasurements Jobs M S

January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 12 Validation Measurements II Test 2 Ethernet Client - Server

January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 13 Validation Measurements II Test 3 2Mbps WAN Client - Server CPU usage Data Traffic

January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 14 èSimilar data processing jobs are performed in six RCs è CERN has RAW, ESD, AOD, TAG è CALTECH, INFN, Tufts, KEK has ESD, AOD, TAG è“Caltech2” has AOD, TAG RAW ESD AOD TAG ESD AOD TAG AOD TAG Physics Analysis Example

January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 15 Physics Group Selection Each group reads 100% TAG events and follows: ~10% to AOD ~1% to ESD ~0.01% to RAW Number of GroupsFollow AODJobs /group L Groups (L~10) M Groups (M~5) N Groups (N~5 ) p% of total TAG (~1%) q% of total TAG (~5%) r% of total TAG (~10%) 1-2 ~20 Jobs/Day in total evenly spread among participating RCs

January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review Estimate of Parameters Parameter Total CPU CPU Unit CPU I/O AMS I/O for Discs throughput for Tape Storage Disk Space Tape Space LAN 350,000 SI SI95/box (100 SI95/cpu) 40 MBps/box (0.1 MBps/SI95) 340 TB 1 PB 31 GBps 520,000 SI SI95/box (100 SI95/CPU) 40 MBps/box (0.1 MBps/SI95) 188 MBps/server 2000 MBps 540 TB 3 PB 46 GBps (Les Robertson's estimate of July 99)

January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 17 Problem Setting: Analysis and Selection RAWESDAODTAG Database 1,000,000,000 CERN 1,000,000,000 Tier 1:Locally 1,000,000,000 RC 1,000,000,000 RC Physics Group Analysis 20 groups * 10 jobs 0.01%1% Follow 100% of the group set Group set: 1% of total TAG Physics Group Selection 20 groups * 1job 0.01%1%10%100% CPU (SI95) Totally 220 Independent Jobs: 200 Physics Group Analysis and 20 Group Selection

January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 18 Problem Setting: Reconstruction and FTP SizeESDAODTAG FTP 1 DB/JobTier 1 CentersTier 1 & 2 Full Reconstruction 6,000,000 events/day yes Monthly Reconstruction 100,000,000 events/day noyes CPU (SI95)

January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 19 Participating Regional Centers 5 Tier 1 Regional Centers and one Tier 2 center RC NameDataWAN Connection CERN (Tier1) INFN (Tier1) KEK (Tier1) TUFTS (Tier1) CALTECH (Tier1) CALTECH-2 (Tier2) RAW, ESD, AOD, TAG ESD, AOD, TAG AOD, TAG All RCs All Tier 1 All Tier 1 & Caltech-2 CERN (RAW) & CALTECH (ESD) 200 Analysis and 20 Selection Jobs are evenly spread among Tier1 RCs

January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 20 AMS load distribution One RC (CERN) configured to run 200 concurrent Physics Group Analysis Jobs and 20 Selection jobs a day Participating RCDataJobs CERN (Tier1)RAW, ESD, AOD, TAG 200 Physics Group Analysis 20 Physics Group Selection x40 Model1 (optimized AMS distribution) Model2 Model1 Model2

January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 21 1 RC Vs 5 RC: Group Selection on all data Model1 One RC (CERN) minimally configured to run 20 concurrent Physics Group Selection Jobs a day Participating RCDataJobs CERN (Tier1)RAW, ESD, AOD, TAG20 Physics Group Selection x10 Model2 Model2 Five Tier 1 Centers minimally configured to perform the same task Participating RCDataJobs CERN INFN KEK TUFTS CALTECH RAW, ESD, AOD, TAG AOD, TAG 4 Physics Group Analysis x10 Conclusion: Current configuration provides a possibility to redistribute resources without much increase in the cost; further optimization is needed to increase the efficiency: 1/(Time*Cost)

January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 22 1 RC vs 6 RC: Reconstruction+Analysis+Selection Model1 One RC (CERN) minimally configured to run all the Jobs a day Participating RCDataJobs CERN (Tier1)RAW, ESD, AOD, TAG 20 Physics Group Selection x Physics Group Analysis Full Reconstruction and FTP Model2 Five Tier 1 Centers optimized to perform the same task with 30 MBps WAN Participating RCDataJobs CERN INFN KEK TUFTS CALTECH CALTECH-2 (Tier2) RAW, ESD, AOD, TAG ESD, AOD, TAG ESD, AOD, TAG ESD, AOD, TAG ESD, AOD, TAG AOD, TAG 4 Physics Group Selectionx10 40 Physics Group Analysis Full Reconstruction and FTP 4 Physics Group Selectionx10 40 Physics Group Analysis 4 Physics Group Selectionx10 40 Physics Group Analysis 4 Physics Group Selectionx10 40 Physics Group Analysis 4 Physics Group Selectionx10 40 Physics Group Analysis 20 Physics Group Analysis Conclusion: Current configuration provides a possibility to optimize the CPU power and reduce the cost; further optimization is possible to reduce WAN bandwidth to 30 MBps

January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 23 6 RC: Two types of Reconstruction+Analysis+Selection Five Tier 1 and one Tier 2 Centers optimized to perform the complete set with 30 MBps WAN and optimized LAN Participating RCDataJobs CERN INFN KEK TUFTS CALTECH CALTECH-2 RAW, ESD, AOD, TAG ESD, AOD, TAG ESD, AOD, TAG ESD, AOD, TAG ESD, AOD, TAG AOD, TAG 4 Physics Group Selectionx10 40 Physics Group Analysis Full Reconstruction and FTP Monthly Reconstruction and FTP (10days) 4 Physics Group Selectionx10 40 Physics Group Analysis 4 Physics Group Selectionx10 40 Physics Group Analysis 4 Physics Group Selectionx10 40 Physics Group Analysis 4 Physics Group Selectionx10 40 Physics Group Analysis 20 Physics Group Analysis Model1 (fixed values) Model2 (randomized data processing times and sizes) Model1 Model2 Conclusion: Current configuration provides a possibility to run daily the complete set of jobs at 6 centers with the WAN bandwidth 30 MBps and the network parameters not exceeding the estimate of 2005

January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 24

January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 25

January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 26

January 20, 2000K. Sliwa/ Tufts University DOE/NSF ATLAS Review 27 Future Work: Replication job: partial dataset replication from CERN to Regional Centers Flexible data access: data exchange between Regional Centers without getting data directly from CERN (depending on load, availability,...) Imposing coherence on the concurrent jobs: if Reconstruction and/or Replication is taking place, Analysis/Selection jobs should be able to monitor new data availability if requested Improving Cost function: adding cost of WAN, adding other hidden costs currently not accounted for Optimization with respect to the parameter Time*Cost for a given task run on different architectures