Peter Kacsuk – Sipos Gergely MTA SZTAKI

Slides:



Advertisements
Similar presentations
The Anatomy of the Grid: An Integrated View of Grid Architecture Carl Kesselman USC/Information Sciences Institute Ian Foster, Steve Tuecke Argonne National.
Advertisements

FP7-INFRA Enabling Grids for E-sciencE EGEE Induction Grid training for users, Institute of Physics Belgrade, Serbia Sep. 19, 2008.
Grid Resource Allocation Management (GRAM) GRAM provides the user to access the grid in order to run, terminate and monitor jobs remotely. The job request.
High Performance Computing Course Notes Grid Computing.
Condor-G: A Computation Management Agent for Multi-Institutional Grids James Frey, Todd Tannenbaum, Miron Livny, Ian Foster, Steven Tuecke Reporter: Fu-Jiun.
A Computation Management Agent for Multi-Institutional Grids
MTA SZTAKI Hungarian Academy of Sciences Grid Computing Course Porto, January Introduction to Grid portals Gergely Sipos
Introduction to the Grid
USING THE GLOBUS TOOLKIT This summary by: Asad Samar / CALTECH/CMS Ben Segal / CERN-IT FULL INFO AT:
Globus Toolkit 4 hands-on Gergely Sipos, Gábor Kecskeméti MTA SZTAKI
Grid Computing 7700 Fall 2005 Lecture 5: Grid Architecture and Globus Gabrielle Allen
Workload Management Workpackage Massimo Sgaravatto INFN Padova.
Milos Kobliha Alejandro Cimadevilla Luis de Alba Parallel Computing Seminar GROUP 12.
1 Overview of Grid middleware concepts Peter Kacsuk MTA SZTAKI, Hungary Univ. Westminster, UK
Workload Management Massimo Sgaravatto INFN Padova.
First steps implementing a High Throughput workload management system Massimo Sgaravatto INFN Padova
Grids and Globus at BNL Presented by John Scott Leita.
Globus Computing Infrustructure Software Globus Toolkit 11-2.
Resource Management Reading: “A Resource Management Architecture for Metacomputing Systems”
Grid Toolkits Globus, Condor, BOINC, Xgrid Young Suk Moon.
INFN-GRID Globus evaluation (WP 1) Massimo Sgaravatto INFN Padova for the INFN Globus group
Workload Management WP Status and next steps Massimo Sgaravatto INFN Padova.
GT Components. Globus Toolkit A “toolkit” of services and packages for creating the basic grid computing infrastructure Higher level tools added to this.
1 School of Computer, National University of Defense Technology A Profile on the Grid Data Engine (GridDaEn) Xiao Nong
Job Submission Condor, Globus, Java CoG Kit Young Suk Moon.
Grid Resource Allocation and Management (GRAM) Execution management Execution management –Deployment, scheduling and monitoring Community Scheduler Framework.
Computational grids and grids projects DSS,
Grid Workload Management & Condor Massimo Sgaravatto INFN Padova.
The Anatomy of the Grid: An Integrated View of Grid Architecture Ian Foster, Steve Tuecke Argonne National Laboratory The University of Chicago Carl Kesselman.
1 Globus Grid Middleware: Basics, Components, and Services Source: The Globus Project Argonne National Laboratory & University of Southern California
The Globus Project: A Status Report Ian Foster Carl Kesselman
Condor: High-throughput Computing From Clusters to Grid Computing P. Kacsuk – M. Livny MTA SYTAKI – Univ. of Wisconsin-Madison
Grid Computing Environments Grid: a system supporting the coordinated resource sharing and problem-solving in dynamic, multi-institutional virtual organizations.
Report from USA Massimo Sgaravatto INFN Padova. Introduction Workload management system for productions Monte Carlo productions, data reconstructions.
Ames Research CenterDivision 1 Information Power Grid (IPG) Overview Anthony Lisotta Computer Sciences Corporation NASA Ames May 2,
July 11-15, 2005Lecture3: Grid Job Management1 Grid Compute Resources and Job Management.
CLRC and the European DataGrid Middleware Information and Monitoring Services The current information service is built on the hierarchical database OpenLDAP.
Part Five: Globus Job Management A: GRAM B: Globus Job Commands C: Laboratory: globusrun.
MTA SZTAKI Hungarian Academy of Sciences Introduction to Grid portals Gergely Sipos
Globus Toolkit Massimo Sgaravatto INFN Padova. Massimo Sgaravatto Introduction Grid Services: LHC regional centres need distributed computing Analyze.
Basic Grid Projects - Globus Sathish Vadhiyar Sources/Credits: Project web pages, publications available at Globus site. Some of the figures were also.
1 Observations on Architecture, Protocols, Services, APIs, SDKs, and the Role of the Grid Forum Ian Foster Carl Kesselman Steven Tuecke.
Job Submission with Globus, Condor, and Condor-G Selim Kalayci Florida International University 07/21/2009 Note: Slides are compiled from various TeraGrid.
Globus Grid Tutorial Part 2: Running Programs Across Multiple Resources.
Globus: A Report. Introduction What is Globus? Need for Globus. Goal of Globus Approach used by Globus: –Develop High level tools and basic technologies.
Grid Workload Management (WP 1) Massimo Sgaravatto INFN Padova.
Grid Execution Management for Legacy Code Architecture Exposing legacy applications as Grid services: the GEMLCA approach Centre.
The Globus Toolkit The Globus project was started by Ian Foster and Carl Kesselman from Argonne National Labs and USC respectively. The Globus toolkit.
CSF. © Platform Computing Inc CSF – Community Scheduler Framework Not a Platform product Contributed enhancement to The Globus Toolkit Standards.
Parallel Computing Globus Toolkit – Grid Ayaka Ohira.
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI solution for high throughput data analysis Peter Solagna EGI.eu Operations.
Workload Management Workpackage
Grid Computing: Running your Jobs around the World
David Abramson, Rajkumar Buyya, and Jonathan Giddy
Duncan MacMichael & Galen Deal CSS 534 – Autumn 2016
Dynamic Deployment of VO Specific Condor Scheduler using GT4
Example: Rapid Atmospheric Modeling System, ColoState U
Globus —— Toolkits for Grid Computing
Grid Computing.
University of Technology
Globus Job Management. Globus Job Management Globus Job Management A: GRAM B: Globus Job Commands C: Laboratory: globusrun.
CSS490 Grid Computing Textbook No Corresponding Chapter
The Anatomy and The Physiology of the Grid
The Anatomy and The Physiology of the Grid
Condor-G Making Condor Grid Enabled
Grid Computing Software Interface
Information Services Claudio Cherubino INFN Catania Bologna
The DZero/PPDG D0/PPDG mission is to enable fully distributed computing for the experiment, by enhancing SAM as the distributed data handling system of.
Condor-G: An Update.
Presentation transcript:

Peter Kacsuk – Sipos Gergely MTA SZTAKI {kacsuk,sipos}@sztaki.hu Globus Toolkit 2 Peter Kacsuk – Sipos Gergely MTA SZTAKI {kacsuk,sipos}@sztaki.hu

Progress in Grid Systems Client/server Cluster computing Super- computing Network Computing High-throughput computing High-performance computing Web Services 2nd Gen. Condor Globus OGSA/OGSI OGSA/WSRF Grid Systems

The Globus-2 model Client program moves to resource(s) description GIIS (MDS-2) Publish MDS-2 API (configuration description) Resource Resource requestor GRAM API provider Client program moves to resource(s) Security is a serious problem!

Solutions by Globus (GT-2) Dynamic creation of Virtual Organizations (VOs) Clients can directly choose resources Standard protocols are used to connect Globus sites Security issues are basically solved Firewalls are allowed between Grid sites PKI: CAs and X.509 certificates SSL for authentication and message protection The client does not need account on every Globus site: Proxies and delegation for secure single Sign-on Still: provides metacomputing facilities (MPICH-G2) Not service-oriented either

Globus Layered Architecture Applications Application Toolkits MPICH-G2 DUROC globusrun Condor-G Nimrod/G GAT Basic Grid Services – Globus Toolkit 2 Replica Mngt GRAM MDS-2 GSI GSI-FTP GASS Grid Fabric Condor MPI TCP UDP LSF PBS NQE Linux NT Solaris DiffServ

The Role of Grid Middleware and Tools Collaboration Tools Data Mgmt Tools Distributed simulation . . . Information services Resource mgmt Data mgmt . . . Remote access Remote monitor net Credit to Ian Foster

Globus Approach Focus on architecture issues Design principles Provide implementations of grid protocols and APIs as basic infrastructure Use to construct high-level, domain-specific solutions Design principles Keep participation cost low Enable local control Support for adaptation A p p l i c a t i o n s Diverse global services Core Globus services Local OS

Globus Approach: Hourglass High-level services GRAM protocol Condor, LSF, NQE, LoadLeveler, etc. Resource brokers, Resource co-allocators Internet protocol TCP, FTP, HTTP, etc. Ethernet, ATM, FDDI, etc. Low-level tools

GRAM Components Client MDS: Grid Index Info Server MDS client API calls to locate resources Client MDS: Grid Index Info Server MDS client API calls to get resource info Site boundary GRAM client API calls to request resource allocation and process creation. MDS: Grid Resource Info Server Query current status of resource GRAM client API state change callbacks Globus Security Infrastructure Local Resource Manager Allocate & create processes Request Job Manager Create Gatekeeper Process Parse Monitor & control Process RSL Library Process

Resource Specification Language Much of the power of GRAM is in the RSL Common language for specifying job requests A conjunction of (attribute=value) pairs GRAM understands a well defined set of attributes

“Standard” MDS Architecture (v1.1.3) Resources run a standard information service (GRIS) which speaks LDAP and provides information about the resource (no searching). GIIS provides a “caching” service much like a web search engine. Resources register with GIIS and GIIS pulls information from them when requested by a client and the cache as expired. GIIS provides the collective-level indexing/searching function. Resource A GRIS Client 1 Clients 1 and 2 request info directly from resources. Resource B GRIS Client 2 GIIS requests information from GRIS services as needed. Client 3 uses GIIS for searching collective information. Client 3 GIIS Cache contains info from A and B

GASS Architecture for file staging Execution machine Submit machine &(executable=https://…) main( ) { fd = globus_gass_open(…) … read(fd,…) globus_gass_close(fd) } (b) RSL extensions GRAM GASS Server HTTP Server (a) GASS file access API FTP Server Cache (c) Remote cache management (d) Low-level APIs for customizing cache & GASS server % globus-gass-cache

GRAM & GASS: Putting It Together 1. Derive Contact String 2. Build RSL string 3. Startup GASS server 4. Submit to request 5. Return output Execution machine program Submit machine 5 GASS server 4 stdout 5 jobmanager 3 globus-job-run Host name Contact string 1 RSL 2 Command Line Args 4 4 gatekeeper

Globus Components In Action Local Machine User Proxy Cert X509 User Cert Machines RSL string mpirun grid-proxy-init RSL multi-request globusrun RSL parser DUROC RSL single request GRAM Client GSI GRAM Client GSI GASS Server GRAM Gatekeeper GRAM Job Manager GRAM Gatekeeper GRAM Job Manager GSI GASS Client GSI GASS Client PBS Unix Fork App App Nexus Nexus AIX MPI Solaris MPI Remote Machine Remote Machine

What is Condor-G? Condor-G is a Personal-Condor enhanced with Globus services It knows how to speak to Globus resources via GRAM It can be used to submit jobs to remote Globus resources It makes Condor keep track of their progress

Condor-G: Condor for the Grid Condor is a high-throughput scheduler Condor-G uses Globus Toolkit libraries for: Security (GSI) Managing remote jobs on Grid (GRAM) File staging & remote I/O (GASS) Grid job management interface & scheduling Robust replacement for Globus Toolkit programs To implement a reliable, crash-proof, checkpointable job submission service Supports single or high-throughput apps on Grid Personal job manager which can exploit Grid resources

The Use of Condor-G condor_submit condor_q condor_rm Globus resource Master Condor GridManager Condor Schedd Globus resource condor_submit condor_q condor_rm Globus resource

Condor-G as user job submission service condor_submit condor_q condor_rm Globus GRAM Globus GRAM Globus GRAM Globus GRAM CONDOR LSF PBS fork

Globus-based production Grids LHC Grid (LCG-2) A homogeneous Grid developed by CERN Restrictive policies (global policies over rule local policies) A dedicated Grid to the Large Hydron Collider experiments Works 24 hours/day and used in EGEE UK-NGS A homogeneous Grid deployed in the UK Restrictive policies Non-dedicated Works 24 hours/day