Middleware and the Grid Steven Tuecke Mathematics and Computer Science Division Argonne National Laboratory.

Slides:



Advertisements
Similar presentations
The Anatomy of the Grid Enabling Scalable Virtual Organizations Ian Foster Mathematics and Computer Science Division Argonne National Laboratory and Department.
Advertisements

International Grid Communities Dr. Carl Kesselman Information Sciences Institute University of Southern California.
The Anatomy of the Grid: An Integrated View of Grid Architecture Carl Kesselman USC/Information Sciences Institute Ian Foster, Steve Tuecke Argonne National.
High Performance Computing Course Notes Grid Computing.
GridFTP Introduction – Page 1Grid Forum 5 GridFTP Steve Tuecke Argonne National Laboratory.
Seminar Grid Computing ‘05 Hui Li Sep 19, Overview Brief Introduction Presentations Projects Remarks.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Introduction to Grid Computing The Globus Project™ Argonne National Laboratory USC Information Sciences Institute Copyright (c)
The Grid Background and Architecture. 1. Keys to success for IT technologies Infrastructure Open Standards.
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
Parallel Programming on the SGI Origin2000 With thanks to Moshe Goldberg, TCC and Igor Zacharov SGI Taub Computer Center Technion Mar 2005 Anne Weill-Zrahia.
Office of Science U.S. Department of Energy Grids and Portals at NERSC Presented by Steve Chan.
1-2.1 Grid computing infrastructure software Brief introduction to Globus © 2010 B. Wilkinson/Clayton Ferner. Spring 2010 Grid computing course. Modification.
Milos Kobliha Alejandro Cimadevilla Luis de Alba Parallel Computing Seminar GROUP 12.
Grid and e-Science Technologies Simon Cox Technical Director Southampton Regional e-Science Centre.
Simo Niskala Teemu Pasanen
Web-based Portal for Discovery, Retrieval and Visualization of Earth Science Datasets in Grid Environment Zhenping (Jane) Liu.
Kate Keahey Argonne National Laboratory University of Chicago Globus Toolkit® 4: from common Grid protocols to virtualization.
The Globus Toolkit: Description and Applications Review Steve Tuecke & Ian Foster Argonne National Laboratory The University of Chicago Globus Co-PI: Carl.
Grid Toolkits Globus, Condor, BOINC, Xgrid Young Suk Moon.
Introduction to Grid Computing Ann Chervenak and Ewa Deelman USC Information Sciences Institute.
Peer to Peer & Grid Computing Ian Foster Mathematics and Computer Science Division Argonne National Laboratory and Department of Computer Science The University.
Grid Security Steve Tuecke Argonne National Laboratory.
ARGONNE  CHICAGO Ian Foster Discussion Points l Maintaining the right balance between research and development l Maintaining focus vs. accepting broader.
GT Components. Globus Toolkit A “toolkit” of services and packages for creating the basic grid computing infrastructure Higher level tools added to this.
1 School of Computer, National University of Defense Technology A Profile on the Grid Data Engine (GridDaEn) Xiao Nong
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
File and Object Replication in Data Grids Chin-Yi Tsai.
The Anatomy of the Grid: An Integrated View of Grid Architecture Ian Foster, Steve Tuecke Argonne National Laboratory The University of Chicago Carl Kesselman.
Web Services BOF This is a proposed new working group coming out of the Grid Computing Environments Research Group, as an outgrowth of their investigations.
The Anatomy of the Grid Mahdi Hamzeh Fall 2005 Class Presentation for the Parallel Processing Course. All figures and data are copyrights of their respective.
Major Grid Computing Initatives Ian Foster Mathematics and Computer Science Division Argonne National Laboratory and Department of Computer Science The.
Virtual Data Grid Architecture Ewa Deelman, Ian Foster, Carl Kesselman, Miron Livny.
1 4/23/2007 Introduction to Grid computing Sunil Avutu Graduate Student Dept.of Computer Science.
Interoperability Grids, Clouds and Collaboratories Ruth Pordes Executive Director Open Science Grid, Fermilab.
Perspectives on Grid Technology Ian Foster Argonne National Laboratory The University of Chicago.
Grid Architecture William E. Johnston Lawrence Berkeley National Lab and NASA Ames Research Center (These slides are available at grid.lbl.gov/~wej/Grids)
Grid Computing Environments Grid: a system supporting the coordinated resource sharing and problem-solving in dynamic, multi-institutional virtual organizations.
09/02 ID099-1 September 9, 2002Grid Technology Panel Patrick Dreher Technical Panel Discussion: Progress in Developing a Web Services Data Analysis Grid.
Ames Research CenterDivision 1 Information Power Grid (IPG) Overview Anthony Lisotta Computer Sciences Corporation NASA Ames May 2,
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
GRID ARCHITECTURE Chintan O.Patel. CS 551 Fall 2002 Workshop 1 Software Architectures 2 What is Grid ? "...a flexible, secure, coordinated resource- sharing.
Authors: Ronnie Julio Cole David
CLRC and the European DataGrid Middleware Information and Monitoring Services The current information service is built on the hierarchical database OpenLDAP.
The Globus Toolkit®: The Open Source Solution for Grid Computing
GRIDS Center Middleware Overview Sandra Redman Information Technology and Systems Center and Information Technology Research Center National Space Science.
GRID Overview Internet2 Member Meeting Spring 2003 Sandra Redman Information Technology and Systems Center and Information Technology Research Center National.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
CEOS Working Group on Information Systems and Services - 1 Data Services Task Team Discussions on GRID and GRIDftp Stuart Doescher, USGS WGISS-15 May 2003.
Introduction to Grids By: Fetahi Z. Wuhib [CSD2004-Team19]
1 Observations on Architecture, Protocols, Services, APIs, SDKs, and the Role of the Grid Forum Ian Foster Carl Kesselman Steven Tuecke.
7. Grid Computing Systems and Resource Management
Globus and PlanetLab Resource Management Solutions Compared M. Ripeanu, M. Bowman, J. Chase, I. Foster, M. Milenkovic Presented by Dionysis Logothetis.
Networking: Applications and Services Antonia Ghiselli, INFN Stu Loken, LBNL Chairs.
Securing the Grid & other Middleware Challenges Ian Foster Mathematics and Computer Science Division Argonne National Laboratory and Department of Computer.
GRID ANATOMY Advanced Computing Concepts – Dr. Emmanuel Pilli.
1 I.Foster LCG Grid Technology: Introduction & Overview Ian Foster Argonne National Laboratory University of Chicago.
The Globus Toolkit The Globus project was started by Ian Foster and Carl Kesselman from Argonne National Labs and USC respectively. The Globus toolkit.
1 Observations on Architecture, Protocols, Services, APIs, SDKs, and the Role of the Grid Forum Ian Foster With: Carl Kesselman, Steven Tuecke Thanks also.
] Open Science Grid Ben Clifford University of Chicago
Peter Kacsuk – Sipos Gergely MTA SZTAKI
Globus —— Toolkits for Grid Computing
Grid Computing.
Grid Computing B.Ramamurthy 9/22/2018 B.Ramamurthy.
CS258 Spring 2002 Mark Whitney and Yitao Duan
Patrick Dreher Research Scientist & Associate Director
Introduction to Grid Technology
The Anatomy and The Physiology of the Grid
The Anatomy and The Physiology of the Grid
Presentation transcript:

Middleware and the Grid Steven Tuecke Mathematics and Computer Science Division Argonne National Laboratory

ARGONNE  CHICAGO What is “Middleware”? l I was asked to talk about Grid “Middleware” l But what is “middleware”? –Ill-defined term, used to mean many things l This talk explores the Grid architecture we have developed in the Globus Project –Gives us a context in which we can more precisely define Grid “middleware”

ARGONNE  CHICAGO Overview 1. Grid concept 2. Grid technical landscape 3. Grid “middleware” 4. Globus Project

The Grid Concept

ARGONNE  CHICAGO On-demand creation of powerful virtual computing systems The Grid: The Web on Steroids Web: Uniform access to HTML documents Grid: Flexible, high-perf access to all significant resources Sensor nets Data archives Computers Software catalogs Colleagues

ARGONNE  CHICAGO Grid Computing: Take 2 l Enable communities (“virtual organizations”) to share geographically distributed resources as they pursue common goals—in the absence of central control, omniscience, trust relationships l Via investigations of –New applications that become possible when resources can be shared in a coordinated way –Protocols, algorithms, persistent infrastructure to facilitate sharing

ARGONNE  CHICAGO Grid Communities and Applications: NSF National Technology Grid

ARGONNE  CHICAGO Grid Communities and Applications: Natl Earthquake Engineering Simulation l NEESgrid: national infrastructure to couple earthquake engineers with experimental facilities, databases, computers, & each other l On-demand access to experiments, data streams, computing, archives, collaboration NEESgrid: Argonne, Michigan, NCSA, UIUC, USC

ARGONNE  CHICAGO Grid Communities and Applications: Mathematicians Solve NUG30 l Community=an informal collaboration of mathematicians and computer scientists l Condor-G delivers 3.46E8 CPU seconds in 7 days (peak 1009 processors) in U.S. and Italy (8 sites) l Solves NUG30 quadratic assignment problem 14,5,28,24,1,3,16,15, 10,9,21,2,4,29,25,22, 13,26,17,30,6,20,19, 8,18,7,27,12,11,23 MetaNEOS: Argonne, Iowa, Northwestern, Wisconsin

ARGONNE  CHICAGO l Community = –1000s of home computer users –Philanthropic computing vendor (Entropia) –Research group (Scripps) l Common goal= advance AIDS research Grid Communities and Applications: Home Computers Evaluate AIDS Drugs

ARGONNE  CHICAGO Data Grid Problem l “Enable a geographically distributed community [of thousands] to pool their resources in order to perform sophisticated, computationally intensive analyses on Petabytes of data” l Note that this problem –Is not unique to physics –Overlaps strongly with other Grid problems l Data Grids do introduce new requirements and R&D challenges

ARGONNE  CHICAGO Data Grid Hierarchy Concept Tier2 Centre ~1 TIPS Online System Offline Processor Farm ~20 TIPS CERN Computer Centre FermiLab ~4 TIPS France Regional Centre Italy Regional Centre Germany Regional Centre Institute Institute ~0.25TIPS Physicist workstations ~100 MBytes/sec ~622 Mbits/sec ~1 MBytes/sec There is a “bunch crossing” every 25 nsecs. There are 100 “triggers” per second Each triggered event is ~1 MByte in size Physicists work on analysis “channels”. Each institute will have ~10 physicists working on one or more channels; data for these channels should be cached by the institute server Physics data cache ~PBytes/sec ~622 Mbits/sec or Air Freight (deprecated) Tier2 Centre ~1 TIPS Caltech ~1 TIPS ~622 Mbits/sec Tier 0 Tier 1 Tier 2 Tier 4 1 TIPS is approximately 25,000 SpecInt95 equivalents Image courtesy Harvey Newman

ARGONNE  CHICAGO Major Data Grid Projects l Earth System Grid (DOE Office of Science) –DG technologies, climate applications l European Data Grid (EU) –DG technologies & deployment in EU l GriPhyN (NSF ITR) –Investigation of “Virtual Data” concept l Particle Physics Data Grid (DOE Science) –DG applications for HENP experiments

ARGONNE  CHICAGO A Little History l Early 90s –Gigabit testbeds, metacomputing l Mid to late 90s –Early experiments (e.g., I-WAY), academic software projects (e.g., Globus), application experiments l 2000 –Major application communities emerging –Major infrastructure deployments –Rich technology base –Grid Forum: >300 people, >90 orgs, 11 countries

Technical Landscape

ARGONNE  CHICAGO Grid Technical Challenges: Sharing, Coordinated Use l New problem solving methods and tools –Data, computing, collaboration, sensors, … l Numerous cross-cutting technical problems –Authentication, authorization, policy, audit; resource discovery, access, allocation, control; failure detection & recovery; brokering; …; … l No central control, omniscience, trust; hence –Need to preserve local site independence –Policy discovery and negotiation important –Myriad interesting failure modes

ARGONNE  CHICAGO How Are Cross-Cutting Technical Issues Addressed? l Development of Grid protocols & services –Protocol-mediated access to remote resources –New services: e.g., resource brokering –“On the Grid” = speak Intergrid protocols –Mostly (extensions to) existing protocols l Development of Grid APIs & SDKs –Facilitate application development by supplying higher-level abstractions l The (hugely successful) model is the Internet l The Grid is not a distributed OS!

ARGONNE  CHICAGO Layered Grid Architecture (By Analogy to Internet Architecture) Application Fabric “Controlling things locally”: Access to, & control of, resources Connectivity “Talking to things”: communication (Internet protocols) & security Resource “Sharing single resources”: negotiating access, controlling use Collective “Managing multiple resources”: ubiquitous infrastructure services User “Specialized services”: user- or appln-specific distributed services Internet Transport Application Link Internet Protocol Architecture

ARGONNE  CHICAGO Protocols, Services, and Interfaces Occur at Each Level Languages/Frameworks Fabric Layer Applications Local Access APIs and Protocols Collective Service APIs and SDKs Collective Services Collective Service Protocols Resource APIs and SDKs Resource Services Resource Service Protocols User Service Protocols User Service APIs and SDKs User Services Connectivity APIs Connectivity Protocols

ARGONNE  CHICAGO Grid Services Architecture: Fabric Layer Protocols & Services l Just what you would expect: the diverse mix of resources that may be shared –Individual computers, Condor pools, file systems, archives, metadata catalogs, networks, sensors, etc., etc. l Few constraints on low-level technology: connectivity and resource level protocols form the “neck in the hourglass” l Defined by interfaces not physical characteristics

ARGONNE  CHICAGO Grid Services Architecture: Connectivity Layer Protocols & Services l Communication –Internet protocols: IP, DNS, routing, etc. l Security: Grid Security Infrastructure (GSI) –Uniform authentication & authorization mechanisms in multi-institutional setting –Single sign-on, delegation, identity mapping –Public key technology, SSL, X.509, GSS-API –Supporting infrastructure: Certificate Authorities, key management, etc. GSI:

ARGONNE  CHICAGO Grid Services Architecture: Resource Layer Protocols & Services l Grid Resource Allocation Mgmt (GRAM) –Remote allocation, reservation, monitoring, control of compute resources l GridFTP protocol (FTP extensions) –High-performance data access & transport l Grid Resource Information Service (GRIS) –Access to structure & state information l Network reservation, monitoring, control l All integrated with GSI: authentication, authorization, policy, delegation GRAM, GridFTP, GRIS:

ARGONNE  CHICAGO Grid Services Architecture: Collective Layer Protocols & Services l Index servers aka metadirectory services –Custom views on dynamic resource collections assembled by a community l Resource brokers (e.g., Condor Matchmaker) –Resource discovery and allocation l Replica catalogs l Co-reservation and co-allocation services l Etc., etc. Metadirectory: Condor:

ARGONNE  CHICAGO Compute Resource SDK API Access Protocol Source Code Repository SDK API Lookup Protocol Example: User Portal Web Portal Source code discovery, application configuration Brokering, co-allocation, certificate authorities Access to data, access to computers, access to network performance data, … Communication, service discovery (DNS), authentication, authorization, delegation Storage systems, clusters, networks, … User Appln Collective Resource Connect Fabric

ARGONNE  CHICAGO Compute Resource SDK API Access Protocol Checkpoint Repository SDK API C-point Protocol Example: High-Throughput Computing System High Throughput Computing System Dynamic checkpoint, job management, failover, staging Brokering, certificate authorities Access to data, access to computers, access to network performance data Communication, service discovery (DNS), authentication, authorization, delegation Storage systems, schedulers User Appln Collective Resource Connect Fabric

ARGONNE  CHICAGO Example Data Grid Architecture Discipline-Specific Data Grid Application Coherency control, replica selection, task management, virtual data catalog, virtual data code catalog, … Replica catalog, replica management, co-allocation, certificate authorities, metadata catalogs, brokering Access to data, access to computers, access to network performance data, … Communication, service discovery (DNS), authentication, authorization, delegation Storage systems, clusters, networks, network caches, … User Appln Collective Resource Connect Fabric

ARGONNE  CHICAGO Grid Middleware l Grid “middleware” is the SDKs and services which implement Grid protocols and APIs –Deploying middleware = deploying services which speak Grid protocols >E.g. Resource managers, storage servers, information services, etc. –Using middleware = using SDKs which speak Grid protocols, and which may conform to Grid APIs >E.g. Resource brokers, remote I/O clients, replica management, etc.

Globus Project

ARGONNE  CHICAGO Globus Project l Developing Grid protocols and APIs, through: –Research: Explore, experiment, prototype solutions to fundamental Grid problems –Globus Toolkit: Open source implementation of services, tools, and SDKs which implement the Grid protocols and APIs –Deployment: Work with partners to deploy the Globus Toolkit in production and testbed Grids –Applications: Work with partners to validate the approach in numerous applications

ARGONNE  CHICAGO Grid Protocols Implemented in the Globus Toolkit l Grid Security Infrastructure (GSI) l Grid Resource Allocation Management (GRAM): –Advance reservation and allocation of computers, storage, network, etc resources l GridFTP –Extensions to FTP for data access and transfer l LDAP –Information discovery

ARGONNE  CHICAGO Globus Toolkit - GSI l Extensions to existing standards: –SSL/TLS, X.509, GSS-API –Single sign-on and delegation through X.509 proxies and delegation protocol –Future: authorization, restricted delegation, independent data units l Globus Toolkit provides: –GSI SDK: GSS-API implementation, which allows apps to add authentication & message protection –Credential management tools

ARGONNE  CHICAGO Globus Toolkit – GRAM l GRAM protocol provides for client/server and event interactions for resource management l Globus Toolkit includes: –GRAM service which interfaces to local site security and schedulers –GRAM client –DUROC: Co-allocation client, built on GRAM –Experimental, application specific brokers

ARGONNE  CHICAGO Globus Toolkit – GridFTP l Secure, high-performance data transfer and access protocol (Data transfer) –Extensions to FTP standards l Globus Toolkit includes: –Modified servers: wu-ftpd, HPSS pftpd, Unitree ftpd –Modified client programs: ncftp –Low-level client and server SDKs –High-level, reliable client

ARGONNE  CHICAGO Globus Toolkit – LDAP l Protocol for resource discovery and information management l Globus Toolkit provides: –GRIS service: OpenLDAP server back-end for dynamic resource information –GIIS service: OpenLDAP server back-end for indexing federations of GRIS servers –Higher level client libraries and tools –Replica catalog client l Exploits commodity LDAP clients & servers –OpenLDAP client/server, iPlanet DS, etc.

ARGONNE  CHICAGO Globus Toolkit – Other l Miscellaneous other SDKs, including: –globus_common: Portability across numerous platforms. –globus_io: Simplified TCP, UDP, File I/O. Integration with GSI. –globus_dc: Reader-makes-right data conversion l Related projects: –Network Weather Service –Condor-G

ARGONNE  CHICAGO Grid Forum l Community forum for promotion and development of Grid technologies and applications –Working groups: Security, data access, info services, resource mgt, scheduling, performance, accounting, prog models, etc. l Protocols being documented, extended, and standardized in Grid Forum –GSI –GridFTP

ARGONNE  CHICAGO Summary l Grid computing concepts are real: they underlie a variety of major projects in physics, other sciences, and industry l Promise is that by allowing coordinating sharing, enable quantitative & qualitative changes in problem solving techniques l Clear picture of Grid architecture emerging l Globus Project is developing Grid protocol and API specs, implementations, and applications l Grid Forum is bringing community together

ARGONNE  CHICAGO For More Information l Book (Morgan Kaufman) – l Globus – l Grid Forum – l PPDG – l GriPhyN –

The End