Issues for Grids and WorldWide Computing Harvey B Newman California Institute of Technology ACAT2000 Fermilab, October 19, 2000.

Slides:



Advertisements
Similar presentations
CANARIE CA*net 3 Status Report Tel:
Advertisements

June Canadas National Optical Internet.
International Grid Communities Dr. Carl Kesselman Information Sciences Institute University of Southern California.
31/03/00 CMS(UK)Glenn Patrick What is the CMS(UK) Data Model? Assume that CMS software is available at every UK institute connected by some infrastructure.
Author - Title- Date - n° 1 GDMP The European DataGrid Project Team
Highest Energy e + e – Collider LEP at CERN GeV ~4km radius First e + e – Collider ADA in Frascati GeV ~1m radius e + e – Colliders.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
GRID DATA MANAGEMENT PILOT (GDMP) Asad Samar (Caltech) ACAT 2000, Fermilab October , 2000.
Data Grids for Next Generation Experiments Harvey B Newman California Institute of Technology ACAT2000 Fermilab, October 19, 2000
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
Parallel Programming on the SGI Origin2000 With thanks to Moshe Goldberg, TCC and Igor Zacharov SGI Taub Computer Center Technion Mar 2005 Anne Weill-Zrahia.
Other servers Java client, ROOT (analysis tool), IGUANA (CMS viz. tool), ROOT-CAVES client (analysis sharing tool), … any app that can make XML-RPC/SOAP.
POLITEHNICA University of Bucharest California Institute of Technology National Center for Information Technology Ciprian Mihai Dobre Corina Stratan MONARC.
Milos Kobliha Alejandro Cimadevilla Luis de Alba Parallel Computing Seminar GROUP 12.
Interfacing Interactive Data Analysis Tools with the Grid: PPDG CS-11 Activity Doug Olson, LBNL Joseph Perl, SLAC ACAT 2002, Moscow 24 June 2002.
October 2003 Iosif Legrand Iosif Legrand California Institute of Technology.
The new The new MONARC Simulation Framework Iosif Legrand  California Institute of Technology.
Workload Management Massimo Sgaravatto INFN Padova.
IEEE Globecom ‘ /12/06 IP over Optics IEEE Globecom ‘99 - Panel Session Jeff Fitchett December 6, 1999.
Optical Networks for the Rest of Us “Customer Empowered Networking” NANOG 17 – Montreal Background Papers on Gigabit to The.
CERN/IT/DB Multi-PB Distributed Databases Jamie Shiers IT Division, DB Group, CERN, Geneva, Switzerland February 2001.
The SAM-Grid Fabric Services Gabriele Garzoglio (for the SAM-Grid team) Computing Division Fermilab.
Users’ Authentication in the VRVS System David Collados California Institute of Technology November 20th, 2003TERENA - Authentication & Authorization.
CANARIE “The Critical Role Universities will play in the Future Evolution of the Internet”
ARGONNE  CHICAGO Ian Foster Discussion Points l Maintaining the right balance between research and development l Maintaining focus vs. accepting broader.
CA*net3 International Transit Network (ITN) Service Internet2 International Task Force Meeting Oct 29, Atlanta, Georgia tel:
July, 2000.Simulation of distributed computing systems I.C. Legrand1 MONARC Models Of Networked Analysis at Regional Centers Iosif C. Legrand (CALTECH)
Building a distributed software environment for CDF within the ESLEA framework V. Bartsch, M. Lancaster University College London.
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
CA*net 4 International Grid Testbed Tel:
Update on CA*net 4 Network
PPDG and ATLAS Particle Physics Data Grid Ed May - ANL ATLAS Software Week LBNL May 12, 2000.
Finnish DataGrid meeting, CSC, Otaniemi, V. Karimäki (HIP) DataGrid meeting, CSC V. Karimäki (HIP) V. Karimäki (HIP) Otaniemi, 28 August, 2000.
Ruth Pordes, Fermilab CD, and A PPDG Coordinator Some Aspects of The Particle Physics Data Grid Collaboratory Pilot (PPDG) and The Grid Physics Network.
CA*net 4 Open Grid Services for Management of Optical Networks CENIC Workshop May 6, 2002
No vember 15, 2000 MONARC Project Status Report Harvey B Newman (CIT) MONARC Project Status Report Harvey Newman California Institute.
1 Grid Related Activities at Caltech Koen Holtman Caltech/CMS PPDG meeting, Argonne July 13-14, 2000.
Data Grid projects in HENP R. Pordes, Fermilab Many HENP projects are working on the infrastructure for global distributed simulated data production, data.
10/24/2015OSG at CANS1 Open Science Grid Ruth Pordes Fermilab
Advanced Computer Networks Topic 2: Characterization of Distributed Systems.
Virtual Data Grid Architecture Ewa Deelman, Ian Foster, Carl Kesselman, Miron Livny.
December 10,1999: MONARC Plenary Meeting Harvey Newman (CIT) Phase 3 Letter of Intent (1/2)  Short: N Pages è May Refer to MONARC Internal Notes to Document.
November SC06 Tampa F.Fanzago CRAB a user-friendly tool for CMS distributed analysis Federica Fanzago INFN-PADOVA for CRAB team.
Author - Title- Date - n° 1 Partner Logo EU DataGrid, Work Package 5 The Storage Element.
09/02 ID099-1 September 9, 2002Grid Technology Panel Patrick Dreher Technical Panel Discussion: Progress in Developing a Web Services Data Analysis Grid.
Internet 2 Workshop (Nov. 1, 2000)Paul Avery (The GriPhyN Project)1 The GriPhyN Project (Grid Physics Network) Paul Avery University of Florida
GRID Overview Internet2 Member Meeting Spring 2003 Sandra Redman Information Technology and Systems Center and Information Technology Research Center National.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
The Particle Physics Data Grid Collaboratory Pilot Richard P. Mount For the PPDG Collaboration DOE SciDAC PI Meeting January 15, 2002.
Eine Einführung ins Grid Andreas Gellrich IT Training DESY Hamburg
High Energy Physics and Grids at UF (Dec. 13, 2002)Paul Avery1 University of Florida High Energy Physics.
PPDGLHC Computing ReviewNovember 15, 2000 PPDG The Particle Physics Data Grid Making today’s Grid software work for HENP experiments, Driving GRID science.
July 26, 1999MONARC Meeting CERN MONARC Meeting CERN July 26, 1999.
6 march Building the INFN Grid Proposal outline a.ghiselli,l.luminari,m.sgaravatto,c.vistoli INFN Grid meeting, milano.
NORDUnet NORDUnet e-Infrastrucure: Grids and Hybrid Networks Lars Fischer CTO, NORDUnet Fall 2006 Internet2 Member Meeting, Chicago.
June 22, 1999MONARC Simulation System I.C. Legrand1 MONARC Models Of Networked Analysis at Regional Centres Distributed System Simulation Iosif C. Legrand.
Distributed Physics Analysis Past, Present, and Future Kaushik De University of Texas at Arlington (ATLAS & D0 Collaborations) ICHEP’06, Moscow July 29,
D0 File Replication PPDG SLAC File replication workshop 9/20/00 Vicky White.
Grid Activities in CMS Asad Samar (Caltech) PPDG meeting, Argonne July 13-14, 2000.
1 CMS Virtual Data Overview Koen Holtman Caltech/CMS GriPhyN all-hands meeting, Marina del Rey April 9, 2001.
CA*net3 - International High Performance Connectivity 9th Internet2 Member Meeting Mar 9, Washington, DC tel:
Breaking the frontiers of the Grid R. Graciani EGI TF 2012.
1 Open Science Grid: Project Statement & Vision Transform compute and data intensive science through a cross- domain self-managed national distributed.
Run - II Networks Run-II Computing Review 9/13/04 Phil DeMar Networks Section Head.
Grid Services for Digital Archive Tao-Sheng Chen Academia Sinica Computing Centre
] Open Science Grid Ben Clifford University of Chicago
Bob Jones EGEE Technical Director
CS258 Spring 2002 Mark Whitney and Yitao Duan
Patrick Dreher Research Scientist & Associate Director
In Distributed Systems
Presentation transcript:

Issues for Grids and WorldWide Computing Harvey B Newman California Institute of Technology ACAT2000 Fermilab, October 19, 2000

LHC Vision: Data Grid Hierarchy Tier 1 Tier2 Center Online System Offline Farm, CERN Computer Ctr > 20 TIPS FranceCenter FNAL Center Italy Center UK Center Institute Institute ~0.25TIPS Workstations ~100 MBytes/sec ~2.5 Gbits/sec Mbits/sec 1 Bunch crossing; ~17 interactions per 25 nsecs; 100 triggers per second. Event is ~1 MByte in size Physicists work on analysis “channels” Each institute has ~10 physicists working on one or more channels Physics data cache ~PBytes/sec ~ Gbits/sec Tier2 Center ~622 Mbits/sec Tier 0 +1 Tier 3 Tier 4 Tier2 Center Tier 2 Experiment

US-CERN Link BW Requirements Projection (PRELIMINARY) [#] Includes ~1.5 Gbps Each for ATLAS and CMS, Plus Babar, Run2 and Other [*] D0 and CDF at Run2: Needs Presumed to Be to be Comparable to BaBar

Grids: The Broader Issues and Requirements u A New Level of Intersite Cooperation, and Resource Sharing è Security and Authentication Across World-Region Boundaries è Start with cooperation among Grid Projects (PPDG, GriPhyN, EU DataGrid, etc.) u Develop Methods for Effective HEP/CS Collaboration In Grid and VDT Design è Joint Design and Prototyping Effort, with (Iterative) Design Specifications è Find an Appropriate Level of Abstraction k Adapted to > 1 Experiment; > 1 Working Environment u Be Ready to Adapt to the Coming Revolutions è In Network, Collaborative, and Internet Information Technologies

PPDG BaBar Data Management BaBar D0 CDF Nuclear Physics CMSAtlas Globus Users SRB Users Condor Users HENP GC Users CMS Data Management Nucl Physics Data Management D0 Data Management CDF Data Management Atlas Data Management Globus Team Condor SRB Team HENP GC

GriPhyN: PetaScale Virtual Data Grids Build the Foundation for Petascale Virtual Data Grids Build the Foundation for Petascale Virtual Data Grids Virtual Data Tools Request Planning & Scheduling Tools Request Execution & Management Tools Transforms Distributed resources (code, storage, computers, and network ) è Resource è Management è Services Resource Management Services è Security and è Policy è Services Security and Policy Services è Other Grid è Services Other Grid Services Interactive User Tools Production Team Individual Investigator Workgroups Raw data source

        EU-Grid Project Work Packages

Grid Issues: A Short List of Coming Revolutions u Network Technologies è Wireless Broadband (from ca. 2003) è 10 Gigabit Ethernet (from 2002: See 10GbE/DWDM-Wavelength (OC-192) integration: OXC u Internet Information Software Technologies è Global Information “Broadcast” Architecture k E.g the Multipoint Information Distribution Protocol (MIDP; è Programmable Coordinated Agent Archtectures k E.g. Mobile Agent Reactive Spaces (MARS) by Cabri et al., Univ. of Modena u The “Data Grid” - Human Interface è Interactive monitoring and control of Grid resources k By authorized groups and individuals k By Autonomous Agents

GigaPOP Vancouver Calgary Regina Winnipeg Ottawa Montreal Toronto Halifax St. John’s Fredericton Charlottetown ORAN BCnet Netera SRnet MRnet ONet RISQ ACORN Chicago STAR TAP CA*net 3 Primary Route Seattle New York Los Angeles CA*net 3 Diverse Route Deploying a 4 channel CWDM Gigabit Ethernet network – 400 km Deploying a 4 channel Gigabit Ethernet transparent optical DWDM– 1500 km Multiple Customer Owned Dark Fiber Networks connecting universities and schools 16 channel DWDM -8 reserved for CANARIE -8 wavelengths for carrier and other customers Consortium Partners: Bell Nexxia Nortel Cisco JDS Uniphase Newbridge Condo Dark Fiber Networks connecting universities and schools Condo Fiber Network linking all universities and hospital CA*net 3 National Optical Internet in Canada

Vancouver Calgary Regina Winnipeg Ottawa Montreal Toronto Halifax St. John’s Fredericton Charlottetown Chicago Seattle New York Los Angeles Miami Europe Dedicated Wavelength or SONET channel OBGP switches Optional Layer 3 aggregation service Large channel WDM system CA*net 4 Possible Architecture

Intermediate ISP Tier 1 ISP Tier 2 ISP AS 1 AS 2 AS 3 AS 4 AS 5 Dual Connected Router to AS 5 Optical switch looks like BGP router and AS1 is direct connected to Tier 1 ISP but still transits AS 5 Router redirects networks with heavy traffic load to optical switch, but routing policy still maintained by ISP Bulk of AS 1 traffic is to Tier 1 ISP For simplicity only data forwarding paths in one direction shown Red Default Wavelength OBGP Traffic Engineering - Physical

VRVS Remote Collaboration System: Statistics VRVS Remote Collaboration System: Statistics 30 Reflectors 52 Countries Mbone, H.323, MPEG2 Streaming, VNC

VRVS: Mbone/H.323/QT Snapshot

VRVS R&D: Sharing Desktop VNC technology integrated in the upcoming VRVS release

Worldwide Computing Issues u Beyond Grid Prototype Components: Integration of Grid Prototypes for End-to-end Data Transport è Particle Physics Data Grid (PPDG) ReqM; SAM in D0 è PPDG/EU DataGrid GDMP for CMS HLT Productions u Start Building the Grid System(s): Integration with Experiment-specific software frameworks u Derivation of Strategies (MONARC Simulation System) è Data caching, query estimation, co-scheduling è Load balancing and workload management amongst Tier0/Tier1/Tier2 sites (SONN by Legrand) è Transaction robustness: simulate and verify u Transparent Interfaces for Replica Management è Deep versus shallow copies: Thresholds; tracking, monitoring and control

Grid Data Management Prototype (GDMP) Distributed Job Execution and Data Handling: Goals è Transparency è Performance è Security è Fault Tolerance è Automation Submit job Replicate data Replicate data Site A Site B Site C r Jobs are executed locally or remotely r Data is always written locally r Data is replicated to remote sites Job writes data locally GDMP V1.1: Caltech + EU DataGrid WP2 Tests by CALTECH, CERN, FNAL, Pisa for CMS “HLT” Production 10/2000; Integration with ENSTORE, HPSS, Castor

MONARC Simulation: Physics Analysis at Regional Centres MONARC Simulation: Physics Analysis at Regional Centres èSimilar data processing jobs are performed in each of several RCs èThere is profile of jobs, each submitted to a job scheduler èEach Centre has “TAG” and “AOD” databases replicated. èMain Centre provides “ESD” and “RAW” data èEach job processes AOD data, and also a a fraction of ESD and RAW data.

ORCA Production on CERN/IT-Loaned Event Filter Farm Test Facility Pileup DB Pileup DB Pileup DB Pileup DB Pileup DB HPSS Pileup DB Pileup DB Signal DB Signal DB Signal DB... 6 Servers for Signal Output Server Output Server Lock Server Lock Server SUN... FARM 140 Processing Nodes 17 Servers 9 Servers Total 24 Pile Up Servers 2 Objectivity Federations The strategy is to use many commodity PCs as Database Servers

Network Traffic & Job efficiency Network Traffic & Job efficiency Mean measured Value ~48MB/s Measurement Simulation Jet Muon

CD CH MD MH TH MC UF.boot MyFED.boot User Collection MD CD MC TD AMS ORCA 4 tutorial, part II October 2000 From UserFederation To Private Copy

Mobile Agents: (Semi)-Autonomous, Goal Driven, Adaptive è Execute Asynchronously è Reduce Network Load: Local Conversations è Overcome Network Latency; Some Outages è Adaptive  Robust, Fault Tolerant è Naturally Heterogeneous è Extensible Concept: Coordinated Agent Architectures Beyond Traditional Architectures: Mobile Agents “Agents are objects with rules and legs” -- D. Taylor Application ServiceAgent

Coordination Architectures for Mobile Java Agents u A lot of Progress since 1998 u Fourth Generation Architecture: “Associative Blackboards” è After 1) Client/Server, 2) Meeting-Oriented, 3) Blackboards; è Analogous to CMS ORCA software: Observer-based “action on demand” u MARS: Mobile Agent Reactive Spaces (Cabri et al.) See è Resilient and Scalable; Simple Implementation è Works with standard Agent implementations (e.g. Aglets: è Data-oriented, to provide temporal and spatial asynchronicity (See Java Spaces, Page Spaces) è Programmable, authorized reactions, based on “virtual Tuple spaces”

Mobile Agent Reactive Spaces (MARS) Architecture u MARS Programmed Reactions: Based on Metalevel 4-Ples: (Reaction, Tuple, Operation-Type, Agent-ID) è Allows Security, Policies è Allows Production of Tuple on Demand The Internet NETWORK NODE Tuple Space MetaLevel Tuple space Agent Server NETWORK NODE A Referenc e to the local Tuple Space B C A: Agents Arrive B: They Get Ref. To Tuple Space C: They Access Tuple Space D: Tuple Space Reacts, with Programmed Behavior D

GRIDs In 2000: Summary u Grids are (in) our Future… Let’s Get to Work

Grid Data Management Issues è Data movement and responsibility for updating the Replica Catalog è Metadata update and replica consistency k Concurrency and locking è Performance characteristics of replicas è Advance Reservation: Policy, time-limit k How to advertise policy and resource availability è Pull versus push (strategy; security) è Fault tolerance; recovery procedures è Queue management è Access control, both global and local