Campus grids: e-Infrastructure within a University Mike Mineter National e-Science Centre 14 February 2006.

Slides:



Advertisements
Similar presentations
/ 1 N. Williams Grid Middleware Experiences Nadya Williams OCI Grid Computing, University of Zurich
Advertisements

GridPP July 2003Stefan StonjekSlide 1 SAM middleware components Stefan Stonjek University of Oxford 7 th GridPP Meeting 02 nd July 2003 Oxford.
Physics with SAM-Grid Stefan Stonjek University of Oxford 6 th GridPP Meeting 30 th January 2003 Coseners House.
The National Grid Service and OGSA-DAI Mike Mineter
Campus grids: e-Infrastructure within a University Mike Mineter National e-Science Centre
FP7-INFRA Enabling Grids for E-sciencE EGEE Induction Grid training for users, Institute of Physics Belgrade, Serbia Sep. 19, 2008.
UK Campus Grid Special Interest Group Dr. David Wallom University of Oxford.
1 Concepts of Condor and Condor-G Guy Warner. 2 Harvesting CPU time Teaching labs. + Researchers Often-idle processors!! Analyses constrained by CPU time!
Condor and GridShell How to Execute 1 Million Jobs on the Teragrid Jeffrey P. Gardner - PSC Edward Walker - TACC Miron Livney - U. Wisconsin Todd Tannenbaum.
The Community Authorisation Service – CAS Dr Steven Newhouse Technical Director London e-Science Centre Department of Computing, Imperial College London.
Dr. David Wallom Experience of Setting up and Running a Production Grid on a University Campus July 2004.
Dr. David Wallom Use of Condor in our Campus Grid and the University September 2004.
Slides for Grid Computing: Techniques and Applications by Barry Wilkinson, Chapman & Hall/CRC press, © Chapter 1, pp For educational use only.
OxGrid, A Campus Grid for the University of Oxford Dr. David Wallom.
Office of Science U.S. Department of Energy Grids and Portals at NERSC Presented by Steve Chan.
First steps implementing a High Throughput workload management system Massimo Sgaravatto INFN Padova
Condor Overview Bill Hoagland. Condor Workload management system for compute-intensive jobs Harnesses collection of dedicated or non-dedicated hardware.
Evaluation of the Globus GRAM Service Massimo Sgaravatto INFN Padova.
SCD FIFE Workshop - GlideinWMS Overview GlideinWMS Overview FIFE Workshop (June 04, 2013) - Parag Mhashilkar Why GlideinWMS? GlideinWMS Architecture Summary.
Grid Computing, B. Wilkinson, 20046d.1 Schedulers and Resource Brokers.
Distributed Systems Early Examples. Projects NOW – a Network Of Workstations University of California, Berkely Terminated about 1997 after demonstrating.
Grid Toolkits Globus, Condor, BOINC, Xgrid Young Suk Moon.
Oxford Interdisciplinary e-Research Centre I e R C OxGrid, A Campus Grid for the University of Oxford Dr. David Wallom Campus Grid Manager.
Track 1: Cluster and Grid Computing NBCR Summer Institute Session 2.2: Cluster and Grid Computing: Case studies Condor introduction August 9, 2006 Nadya.
10/20/05 LIGO Scientific Collaboration 1 LIGO Data Grid: Making it Go Scott Koranda University of Wisconsin-Milwaukee.
Web: OMII-UK Campus Grid Toolkit NW-GRID Campus Grids Workshop 31 st October 2007 University of Liverpool Tim Parkinson.
03/27/2003CHEP20031 Remote Operation of a Monte Carlo Production Farm Using Globus Dirk Hufnagel, Teela Pulliam, Thomas Allmendinger, Klaus Honscheid (Ohio.
Grids and Portals for VLAB Marlon Pierce Community Grids Lab Indiana University.
Computational grids and grids projects DSS,
Grid Computing I CONDOR.
G RID M IDDLEWARE AND S ECURITY Suchandra Thapa Computation Institute University of Chicago.
Grid Technologies  Slide text. What is Grid?  The World Wide Web provides seamless access to information that is stored in many millions of different.
Rochester Institute of Technology Job Submission Andrew Pangborn & Myles Maxfield 10/19/2015Service Oriented Cyberinfrastructure Lab,
The National Grid Service Guy Warner.
Stephen Pickles Technical Director, Grid Operations Support Centre University of Manchester Neil Geddes CCLRC Head of e-Science Director of the UK Grid.
Condor: High-throughput Computing From Clusters to Grid Computing P. Kacsuk – M. Livny MTA SYTAKI – Univ. of Wisconsin-Madison
NGS Innovation Forum, Manchester4 th November 2008 Condor and the NGS John Kewley NGS Support Centre Manager.
WP8 Meeting Glenn Patrick1 LHCb Grid Activities in UK Grid WP8 Meeting, 16th November 2000 Glenn Patrick (RAL)
June 24-25, 2008 Regional Grid Training, University of Belgrade, Serbia Introduction to gLite gLite Basic Services Antun Balaž SCL, Institute of Physics.
NW-GRID Campus Grids Workshop Liverpool31 Oct 2007 NW-GRID Campus Grids Workshop Liverpool31 Oct 2007 Moving Beyond Campus Grids Steven Young Oxford NGS.
July 11-15, 2005Lecture3: Grid Job Management1 Grid Compute Resources and Job Management.
Review of Condor,SGE,LSF,PBS
Creating and running an application.
The National Grid Service Mike Mineter.
Introduction to Grids By: Fetahi Z. Wuhib [CSD2004-Team19]
UK Grid Meeting Glenn Patrick1 LHCb Grid Activities in UK Grid Prototype and Globus Technical Meeting QMW, 22nd November 2000 Glenn Patrick (RAL)
AliEn AliEn at OSC The ALICE distributed computing environment by Bjørn S. Nilsen The Ohio State University.
1 e-Science AHM st Aug – 3 rd Sept 2004 Nottingham Distributed Storage management using SRB on UK National Grid Service Manandhar A, Haines K,
Introduction to The Storage Resource.
Eileen Berman. Condor in the Fermilab Grid FacilitiesApril 30, 2008  Fermi National Accelerator Laboratory is a high energy physics laboratory outside.
The National Grid Service Mike Mineter.
The National Grid Service Mike Mineter
Rights Management for Shared Collections Storage Resource Broker Reagan W. Moore
1 Porting applications to the NGS, using the P-GRADE portal and GEMLCA Peter Kacsuk MTA SZTAKI Hungarian Academy of Sciences Centre for.
INFSO-RI Enabling Grids for E-sciencE Middleware Concepts and Services in Campus, National and International Grids Mike Mineter.
13th EELA Tutorial, La Antigua, 18-19, October E-infrastructure shared between Europe and Latin America FP6−2004−Infrastructures−6-SSA
EGEE-II INFSO-RI Enabling Grids for E-sciencE Overview of gLite, the EGEE middleware Mike Mineter Training Outreach Education National.
Campus grids: e-Infrastructure within a University Mike Mineter National e-Science Centre 22 February 2006.
Campus grids: e-Infrastructure within a University Mike Mineter National e-Science Centre
EGI-InSPIRE RI EGI-InSPIRE EGI-InSPIRE RI EGI Services for Distributed e-Infrastructure Access Tiziana Ferrari on behalf.
The EPIKH Project (Exchange Programme to advance e-Infrastructure Know-How) gLite Grid Introduction Salma Saber Electronic.
CernVM and Volunteer Computing Ivan D Reid Brunel University London Laurence Field CERN.
Dynamic Deployment of VO Specific Condor Scheduler using GT4
Example: Rapid Atmospheric Modeling System, ColoState U
Data services on the NGS
Building Grids with Condor
Basic Grid Projects – Condor (Part I)
The National Grid Service Mike Mineter NeSC-TOE
Condor-G Making Condor Grid Enabled
Presentation transcript:

Campus grids: e-Infrastructure within a University Mike Mineter National e-Science Centre 14 February 2006

2 Thanks to: Mark Calleja, Cambridge David Wallom, Oxford David McBride, Imperial College for descriptions of their campus initiatives.

3 Current campus grids Present focus is on enhanced use of resources – e.g. to save purchasing new clusters Cambridge, Bristol, UCL, Imperial, Oxford, Cardiff, Glasgow and others Two classes: –Within an administration domain: processors on a LAN… (e-infrastructure, but not a grid!!) –Across administrative domains E.g. Multiple institutes, each with own systems management, authorisation, authentication

4 Pooling processors on a LAN Teaching lab machines lie idle for most of the time –Use the processing power! Harvest spare compute cycles to create a low- cost “high throughput computing” platform Create a pool of processors as a batch processing resource Submit jobs that run when a machine is free Condor most common approach –

5 Condor Converts collections of distributed workstations and dedicated clusters into a high-throughput computing (HTC) facility, a Condor Pool HTC concerned with harvesting maximum compute time for many jobs… e.g. months CPU time in days elapsed Manages both resources (machines) and resource requests (jobs) Includes : ClassAd Matchmaking Process checkpoint / restart / migration

6 Terminology Machines Pool – a collection of computers used by Condor Central manager – one per pool  matchmaking  pool management functionalities user submit -> schedd with queue of jobs job execute -> startd matchmaking -> collector & negotiator control of a node -> master

7 Campus grids Single sign-on to virtualised resources in multiple administrative domains Need AA mechanisms –X 509 certificates –For users that only wish to access internal (university) resources, a Kerberos CA (e.g. Oxford, Imperial College) Need brokering – where should a job be run? Needs information systems – what are the options?! Scalability requires each VO or institute contributes its average requirement

Oxford Interdisciplinary e-Research Centre I e R C Example: OxGrid, a University Campus Grid Single entry point for Oxford users to shared and dedicated resources Seamless access to National Grid Service and OSC for registered users Single sign-on using PKI technology integrated with current methods NGSOSC OxGrid Central Management Resource Broker MDS/ VOM Storage College Resources Departmental Resources Oxford Users David Wallom

9 Middleware for campus grids Globus toolkit –Tools built on Grid Security Infrastructure to include: Job submission (GRAM) : run a job on a remote computer Information services: So I know which computer to use File transfer: so large data files can be transferred Replica management: so I can have multiple versions of a file “close” to the computers where I want to run jobs Storage Resource Broker –Virtual filesystem: for files held in multiple locations Condor “flocking” –Allow access to multiple Condor pools

10 Globus A software toolkit: a modular “bag of technologies” –Made available under liberal open source license Not turnkey solutions, but building blocks and tools for application developers and system integrators International production grids are (currently) based on the Globus Toolkit release 2 Globus Alliance:

11 Running a job with GT2 GT2 Toolkit commands: an example –Job submission – need to know UNIX and name of a CE to use –Higher tools allow jobs to be submitted to “the grid” – to a resource broker globus-job-submit grid-data.rl.ac.uk/jobmanager-pbs /bin/hostname -f globus-job-status DONE globus-job-get-output grid-data12.rl.ac.uk

12 How SRB Works MCAT Database MCAT Server SRB A Server SRB B Server SRB Client a b cd e f g 4 major components: –The Metadata Catalogue (MCAT) –The MCAT-Enabled SRB Server –The SRB Storage Server –The SRB Client

13 Obvious issues Need specific initial user communities >> vague sense this is a good idea! Engage with systems managers from first thoughts Operations effort must be factored in and justifiable! Researcher enthusiasm is not enough!

14 Platform for interdisciplinary research?? Campus grid gives an infrastructure spanning multiple institutes Potential for VOs spanning those institutes also –Doing more than sharing spare compute cycles Enabling easier access / reuse of research data Higher level services + ontologies/semantics + workflow