Presentation is loading. Please wait.

Presentation is loading. Please wait.

US-CMS Meeting (May 19, 2001)Paul Avery1 US-CMS Meeting (UC Riverside) May 19, 2001 Grids for US-CMS and CMS Paul Avery University of Florida

Similar presentations


Presentation on theme: "US-CMS Meeting (May 19, 2001)Paul Avery1 US-CMS Meeting (UC Riverside) May 19, 2001 Grids for US-CMS and CMS Paul Avery University of Florida"— Presentation transcript:

1 US-CMS Meeting (May 19, 2001)Paul Avery1 US-CMS Meeting (UC Riverside) May 19, 2001 Grids for US-CMS and CMS Paul Avery University of Florida avery@phys.ufl.edu

2 US-CMS Meeting (May 19, 2001)Paul Avery2 LHC Data Grid Hierarchy Tier 1 T2 3 3 3 3 3 3 3 3 3 3 3 Tier 0 (CERN) 4 4 4 4 3 3 Tier0 CERN Tier1 National Lab Tier2 Regional Center at University Tier3 University workgroup Tier4 Workstation Tasks  R&D  Tier2 centers  Software integration  Unified IT resources

3 US-CMS Meeting (May 19, 2001)Paul Avery3 CMS Grid Hierarchy Tier2 Center Online System CERN Computer Center > 20 TIPS USA Center France Center Italy Center UK Center Institute Institute ~0.25TIPS Workstations, other portals ~100 MBytes/sec 2.5-10 Gbits/sec 0.1-1 Gbits/sec Bunch crossing per 25 nsecs. 100 triggers per second ~1 MByte per event Physicists work on analysis “channels”. Each institute has ~10 physicists working on one or more channels Physics data cache ~PBytes/sec >10 Gbits/sec Tier2 Center 0.6-2.5 Gbits/sec Tier 0 +1 Tier 1 Tier 3 Tier 4 Tier2 Center Tier 2 Experiment CERN/Outside Resource Ratio ~1:2 Tier0/(  Tier1)/(  Tier2) ~1:1:1

4 US-CMS Meeting (May 19, 2001)Paul Avery4 Grid Projects è Funded projects  GriPhyNUSANSF, $11.9M  PPDG IUSADOE, $2M  PPDG IIUSADOE, $9.5M  EU DataGridEU$9.3M è Proposed projects  iVDGLUSANSF, $15M + $1.8M + UK  DTFUSANSF, $45M + $4M/yr  DataTagEUEC, $? è Other national projects  PPARC e-ScienceUKPPARC, $40M  UK e-ScienceUK> $100M  Italy, France, Japan?  EU networking initiatives (Géant, Danté, SURFNet)

5 US-CMS Meeting (May 19, 2001)Paul Avery5 Major Grid News Since May 2000  Sep. 2000GriPhyN proposal approved ($11.9M)  Nov. 2000First outline of US-CMS Tier2 plan  Nov. 2000Caltech-UCSD proto-T2 hardware installed  Dec. 2000Submit iVDGL preproposal to NSF  Jan. 2001EU-DataGrid approved ($9.3M)  Mar. 20011 st Grid coordination meeting  Mar. 2001Submit PPDG proposal to DOE ($12M)  Apr. 2001Submit iVDGL proposal to NSF ($15M)  Apr. 2001Submit DTF proposal to NSF ($45M, $4M/yr)  Apr. 2001Submit DataTag proposal to EU  May 2001PPDG proposal approved ($9.5M)  May 2001Initial hardware for Florida proto-T2 installed  Jun. 20012 nd Grid coordination meeting  Aug. 2001DTF approved?  Aug. 2001iVDGL approved?

6 US-CMS Meeting (May 19, 2001)Paul Avery6 Grid Timeline Q2 00 Q3 00 Q4 00 Q1 01 Q2 01 Q3 01 GriPhyN approved, $11.9M Outline of US-CMS Tier plan Caltech-UCSD install proto-T2 Submit GriPhyN proposal, $12.5M Submit iVDGL preproposal EU DataGrid approved, $9.3M 1 st Grid coordination meeting Submit PPDG proposal, $12M Submit DTF proposal, $45M Submit iVDGL proposal, $15M PPDG approved, $9.5M 2 nd Grid coordination meeting Install initial Florida proto-T2 iVDGL approved? DTF approved?

7 US-CMS Meeting (May 19, 2001)Paul Avery7 Why Do We Need All These Projects? è Agencies see LHC Grid computing in wider context  (next slide) è DOE priorities  LHC, D0, CDF, BaBar, RHIC, JLAB  Computer science  ESNET è NSF priorities  Computer science  Networks  LHC, other physics, astronomy  Other basic sciences  Education and outreach  International reach  Support for universities

8 US-CMS Meeting (May 19, 2001)Paul Avery8 Projects (cont.) è We must justify investment  Benefit to wide scientific base  Education and outreach  Oversight from Congress always present  Much more competitive funding environment è We have no choice anyway  This is the mechanism by which we will get funds è Cons  Diverts effort from mission, makes management more complex è Pros  Exploits initiatives, brings new funds & facilities (e.g., DTF)  Drives deployment of high-speed networks  Brings many new technologies, tools  Attracts attention/help from computing experts, vendors

9 US-CMS Meeting (May 19, 2001)Paul Avery9 US-CMS Grid Facilities è Caltech-UCSD implemented proto-Tier2 (Fall 2000)  40 dual PIII boxes in racks  RAID disk  Tape resources è Florida now implementing second proto-Tier2  72 dual PIII boxes in racks  Inexpensive RAID  Ready June 1, 2001 for production? è Fermilab about to purchase equipment (Vivian) è Distributed Terascale Facility (DTF)  Not approved yet  MOUs being signed with GriPhyN, CMS  Massive CPU, storage resources at 4 sites, 10 Gb/s networks  Early prototype of Tier1 in 2006

10 US-CMS Meeting (May 19, 2001)Paul Avery10 Particle Physics Data Grid è Recently funded @ $9.5M for 3 years (DOE MICS/HENP) è High Energy & Nuclear Physics projects (DOE labs) è Database/object replication, caching, catalogs, end-to-end è Practical orientation: networks, instrumentation, monitoring

11 US-CMS Meeting (May 19, 2001)Paul Avery11 PPDG: Remote Database Replication u First Round Goal: Optimized cached read access to 10-100 Gbytes drawn from a total data set of 0.1 to ~1 Petabyte u Matchmaking, Co-Scheduling: SRB, Condor, Globus services; HRM, NWS PRIMARY SITE Data Acquisition, CPU, Disk, Tape Robot SECONDARY SITE CPU, Disk, Tape Robot Site to Site Data Replication Service 100 Mbytes/sec Multi-Site Cached File Access Service University CPU, Disk, Users PRIMARY SITE DAQ, Tape, CPU, Disk, Robot Satellite Site Tape, CPU, Disk, Robot University CPU, Disk, Users University CPU, Disk, Users University CPU, Disk, Users University CPU, Disk, Users Satellite Site Tape, CPU, Disk, Robot

12 US-CMS Meeting (May 19, 2001)Paul Avery12         EU DataGrid Project

13 US-CMS Meeting (May 19, 2001)Paul Avery13 GriPhyN = App. Science + CS + Grids è GriPhyN = Grid Physics Network  US-CMSHigh Energy Physics  US-ATLASHigh Energy Physics  LIGO/LSCGravity wave research  SDSSSloan Digital Sky Survey  Strong partnership with computer scientists è Design and implement production-scale grids  Develop common infrastructure, tools and services (Globus based)  Integration into the 4 experiments  Broad application to other sciences via “Virtual Data Toolkit” è Multi-year project  R&D for grid architecture (funded at $11.9M)  “Tier 2” center hardware, personnel  Integrate Grid infrastructure into experiments

14 US-CMS Meeting (May 19, 2001)Paul Avery14 GriPhyN Institutions  U Florida  U Chicago  Boston U  Caltech  U Wisconsin, Madison  USC/ISI  Harvard  Indiana  Johns Hopkins  Northwestern  Stanford  U Illinois at Chicago  U Penn  U Texas, Brownsville  U Wisconsin, Milwaukee  UC Berkeley  UC San Diego  San Diego Supercomputer Center  Lawrence Berkeley Lab  Argonne  Fermilab  Brookhaven

15 US-CMS Meeting (May 19, 2001)Paul Avery15 GriPhyN: PetaScale Virtual Data Grids Virtual Data Tools Request Planning & Scheduling Tools Request Execution & Management Tools Transforms Distributed resources (code, storage, computers, and network ) è Resource è Management è Services Resource Management Services è Security and è Policy è Services Security and Policy Services è Other Grid è Services Other Grid Services Interactive User Tools Production Team Individual Investigator Research group Raw data source

16 US-CMS Meeting (May 19, 2001)Paul Avery16 GriPhyN Progress è New hires  3 physicists at Florida (1 PD, 2 scientists)  0.5 Tier2 support person at Caltech è CMS requirements document  33 pages, K. Holtman è Major meetings held (http://www.griphyn.org/)  Oct. 2000All-hands meeting  Dec. 2000Architecture meeting  Apr. 2001All-hands meeting  Aug. 2001Applications meeting è CMS – CS groups will need more frequent meetings  Further develop requirements, update architecture  Distributed databases  More discussion of integration

17 US-CMS Meeting (May 19, 2001)Paul Avery17 Common Grid Infrastructure è GriPhyN + PPDG + EU-DataGrid + national efforts  France, Italy, UK, Japan è Have agreed to collaborate, develop joint infrastructure  Initial meeting March 4 in Amsterdam to discuss issues  Next meeting June 23 è Preparing management document  Joint management, technical boards + steering committee  Coordination of people, resources  An expectation that this will lead to real work è Collaborative projects  Grid middleware  Integration into applications  Grid “laboratory”: iVDGL  Network testbed: T 3 = Transatlantic Terabit Testbed

18 US-CMS Meeting (May 19, 2001)Paul Avery18 iVDGL è International Virtual-Data Grid Laboratory  A place to conduct Data Grid tests “at scale”  A mechanism to create common Grid infrastructure  National, international scale Data Grid tests, operations è Components  Tier1 sites (laboratories)  Tier2 sites (universities and others)  Selected Tier3 sites (universities)  Distributed Terascale Facility (DTF)  Fast networks: US, Europe, transatlantic è Who  Initially US-UK-EU  Japan, Australia  Other world regions later  Discussions w/ Russia, China, Pakistan, India, South America

19 US-CMS Meeting (May 19, 2001)Paul Avery19 iVDGL Proposal to NSF è Submitted to NSF ITR2001 program April 25  ITR2001 program is more application oriented than ITR2000  $15M, 5 years @ $3M per year (huge constraint)  CMS + ATLAS + LIGO + SDSS/NVO + Computer Science è Scope of proposal  Tier2 hardware, Tier2 support personnel  Integration of Grid software into applications  CS support teams (+ 6 UK Fellows)  Grid Operations Center (iGOC)  Education and outreach (3 minority institutions) è Budget (next slide)  Falls short of US-CMS Tier2 needs (Tier2 support staff)  Need to address problem with NSF (Lothar, Irwin talks)

20 US-CMS Meeting (May 19, 2001)Paul Avery20 iVDGL Budget

21 US-CMS Meeting (May 19, 2001)Paul Avery21 iVDGL Map Circa 2002-2003 Tier0/1 facility Tier2 facility 10 Gbps link 2.5 Gbps link 622 Mbps link Other link Tier3 facility

22 US-CMS Meeting (May 19, 2001)Paul Avery22 iVDGL as a Laboratory è Grid Exercises  “Easy”, intra-experiment tests first (20-40%, national, transatlantic)  “Harder” wide-scale tests later (50-100% of all resources)  CMS is already conducting transcontinental productions è Local control of resources vitally important  Experiments, politics demand it  Resource hierarchy: (1) National + experiment, (2) inter-expt. è Strong interest from other disciplines  HENP experiments  Virtual Observatory (VO) community in Europe/US  Gravity wave community in Europe/US/Australia/Japan  Earthquake engineering  Bioinformatics  Our CS colleagues (wide scale tests)


Download ppt "US-CMS Meeting (May 19, 2001)Paul Avery1 US-CMS Meeting (UC Riverside) May 19, 2001 Grids for US-CMS and CMS Paul Avery University of Florida"

Similar presentations


Ads by Google