Presentation is loading. Please wait.

Presentation is loading. Please wait.

Charlie Catlett 2007 TeraGrid National Cyberinfrastructure for Terascale Science Dane Skow Deputy Director, TeraGrid

Similar presentations


Presentation on theme: "Charlie Catlett 2007 TeraGrid National Cyberinfrastructure for Terascale Science Dane Skow Deputy Director, TeraGrid"— Presentation transcript:

1 Charlie Catlett (cec@uchicago.edu)January 2007 TeraGrid National Cyberinfrastructure for Terascale Science Dane Skow Deputy Director, TeraGrid www.teragrid.org The University of Chicago and Argonne National Laboratory February 2007 Slides courtesy of Charlie Catlett (UC/ANL), Tony Rimovsky (NCSA) and Reagan Moore (SDSC) TeraGrid is supported by the National Science Foundation Office of Cyberinfrastructure Petascale

2 “NSF Cyberinfrastructure Vision for 21st Century Discovery” 1. Distributed, scalable up to petaFLOPS HPC 2. Data, data analysis, visualization 3. Collaboratories, observatories, virtual organizations includes networking, middleware, systems software “sophisticated” science application software includes data to and from instruments 4. Education and Workforce provide sustainable and evolving CI that is secure, efficient, reliable, accessible, usable, and interoperable provide access to world-class tools and services Draft 7.1 CI Plan at www.nsf.gov/oci/ Adapted from: Dan Atkins, NSF Office of Cyberinfrastructure

3 Charlie Catlett (cec@uchicago.edu)January 2007 TeraGrid Mission TeraGrid provides integrated, persistent, and pioneering computational resources that will significantly improve our nation’s ability and capacity to gain new insights into our most challenging research questions and societal problems. –Our vision requires an integrated approach to the scientific workflow including obtaining access, application development and execution, data analysis, collaboration and data management. –These capabilities must be accessible broadly to the science, engineering, and education community.

4 Dane Skow (dds@uchicago.edu)February 2007 SDSC TACC UC/ANL NCSA ORNL PU IU PSC NCAR Caltech USC/ISI UNC/RENCI UW Resource Provider (RP) Software Integration Partner Grid Infrastructure Group (GIG) TeraGrid Facility Partners NIU

5 Dane Skow (dds@uchicago.edu)February 2007 Networking SDSC UC/ANLPSC TACC ORNL LA DEN NCSA NCAR Abilene 2x10G 1x10G PU IPGrid IU CHI 1x10G 1x10G each 2x10G 1x10G 3x10G each Cornell 1x10G

6 Dane Skow (dds@uchicago.edu)February 2007 TeraGrid Usage Growth Specific AllocationsRoaming Allocations Normalized Units (millions) 100 200 TeraGrid currently delivers to users an average of 400,000 cpu-hours per day -> ~20,000 CPUs DC

7 Dane Skow (dds@uchicago.edu)February 2007 TeraGrid User Community Growth Begin TeraGrid Production Services (October 2004) Incorporate NCSA and SDSC Core (PACI) Systems and Users (April 2006) Decommissioning of systems typically causes slight reductions in active users. E.g. December 2006 is due to decommissioning of Lemeux (PSC). FY05FY06 New User Accounts9482,692 Avg. New Users per Quarter315365* Active Users1,3503,228 All Users Ever1,7994,491 (*FY06 new users/qtr excludes Mar/Apr 2006)

8 Charlie Catlett (cec@uchicago.edu)January 2007 TeraGrid Projects by Institution Blue: 10 or more PI’s Red: 5-9 PI’s Yellow: 2-4 PI’s Green: 1 PI 1000 projects, 3200 users TeraGrid allocations are available to researchers at any US educational institution by peer review. Exploratory allocations can be obtained through a biweekly review process. See www.teragrid.org.

9 Charlie Catlett (cec@uchicago.edu)January 2007 FY06 Quarterly Usage by Discipline 100 50 Percent Usage

10 Charlie Catlett (cec@uchicago.edu)January 2007 TeraGrid Science Gateways Initiative: Service-Oriented Approach The science and engineering community has been building discipline-specific cyberinfrastructure in the form of portals, applications, and grids. Our objective is to enable these to use TeraGrid resources transparently as “back-ends” to their infrastructure. The TeraGrid Science Gateways program has developed, in partnership with 20+ communities and multiple major Grid projects, an initial set of processes, policies, and services that enable these gateways to access TeraGrid (or other facilities) resources via web services. TeraGrid Grid-X Grid-Y Web Services

11 Dane Skow (dds@uchicago.edu)February 2007 Use Modality Community Size (est. number of projects) Batch Computing on Individual Resources 850 Exploratory and Application Porting 650 Workflow, Ensemble, and Parameter Sweep 160 Science Gateway Access 100 Remote Interactive Steering and Visualization 35 Tightly-Coupled Distributed Computation 10 TeraGrid User Community in 2006 Grid-y Users

12 Dane Skow (dds@uchicago.edu)February 2007 Data Storage Resources Local Cluster Files System Global File System GPFS-WAN 250TB Data Collections Archive Storage Graphic courtesy of SDSC datacentral

13 Dane Skow (dds@uchicago.edu)February 2007 Local Cluster Storage Normal site user/group permissions apply –TeraGrid users typically have individual accounts connected with their project team via usual uid/gid groups –Therefore normal containment/forensic tools work inside the system GridFTP transfer from one resource to another –Dedicated GridFTP mover nodes for parallel systems –Dynamic GridFTP mover “fleet” direct from apps –Central TeraGrid Listener to gather system aggregate data Modification to standard set to lift “vail of privacy” within TeraGrid System metrics and diagnostics Forensics analysis database Shared NFS-like file system within a single site –GPFS, Lustre, NFS, PVFS,QFS, CXFS, …

14 Dane Skow (dds@uchicago.edu)February 2007 “Global” File System TeraGrid has central GPFS-WAN server at SDSC mounted by several clusters across the grid. Pros –Common namespace –POSIX syntax for remote file access –Single Identity space (x509) across WAN –High speed parallel systems available Cons –GPFS-WAN: IBM licensing and availability –Lustre-WAN: Lack of WAN security model –No group authZ construct support

15 Dane Skow (dds@uchicago.edu)February 2007 Archived Storage Just now beginning to deal with archived storage as an allocated resource. Issues –Retention policy/guarantee –Media migration –Privacy/Security/Availability on abandoned files –Economic Model (NCAR has a “Euro” approach with common currency)

16 Dane Skow (dds@uchicago.edu)February 2007 Using an SRB Data Grid - Details Storage Resource Broker Data request goes to SRB Server Storage Resource Broker Metadata Catalog DB Server looks up information in catalog Catalog tells which SRB server has data 1 st server asks 2 nd for data The data is found and returned User asks for data

17 Dane Skow (dds@uchicago.edu)February 2007 Lessons Learned Lesson from Stakkato was not (just) scale of attack, but rather importance of being able to restore control –In a connected world with agents this means Virtual borders -- ALL > collaborators > pair-wise trusts Centralized logging for forensics/IDS –USE THE SAME SYSTEM FOR DAILY OPERATIONS/METRICS ! –We must be able to (perhaps painfully) outpace attackers in cleaning system Ease of use and ubiquity are essential to adoption –AFS’s change to directory permission from file permissions had a huge adoption barrier cost

18 Dane Skow (dds@uchicago.edu)February 2007 Lessons Learned Work is needed on distributed group authorization/management tooling –Group membership and roles are best maintained by the leaders of the group –Policy rules are best kept and enforced by the data store Security Triad: –Who you are –Where you can go –What you can do

19 Dane Skow (dds@uchicago.edu)February 2007 Lessons Learned Work is needed on distributed group authorization/management tooling –Group membership and roles are best maintained by the leaders of the group –Policy rules are best kept and enforced by the data store Security Triad: –Who you are –Where you can go –What you can do

20 Dane Skow (dds@uchicago.edu)February 2007 Lessons Learned Work is needed on distributed group authorization/management tooling –Group membership and roles are best maintained by the leaders of the group –Policy rules are best kept and enforced by the data store Security Triad: –Who you are –Where you can go –What you can do Some actions are so dangerous that they deserve to have the 2 person rule enforced –(e.g. archive tape erasure)

21 Dane Skow (dds@uchicago.edu)February 2007 Lessons Learned Security is never “done” –The coordination team (building) from the Stakkato incident was THE most valuable result.

22 Dane Skow (dds@uchicago.edu)February 2007 Security in Distributed Data Management Systems Storage Resource Broker Reagan W. Moore Wayne Schroeder Mike Wan Arcot Rajasekar {moore, schroede, mwan, sekar}@sdsc.edu {moore, schroede, mwan, sekar}@sdsc.edu http://www.sdsc.edu/srb http://irods.sdsc.edu/

23 Charlie Catlett (cec@uchicago.edu)January 2007 Logical Name Spaces Logical User name –Unique identifier for each person accessing the system {User-name, project-name} –User groups - aggregations of users Membership in multiple groups –Data grids (zones) {user-name, project-name, zone-name}

24 Charlie Catlett (cec@uchicago.edu)January 2007 Authorization - SRB Assign access controls on each name space –Files –Metadata –Storage Assign roles that represent sets of allowed operations –Role - administrator, curator, read, write, annotate

25 Charlie Catlett (cec@uchicago.edu)January 2007 Rule-based Data Management iRODS (integrated Rule Oriented Data System) Map from management policies to rules controlling execution of remote micro-services Manage persistent state information for results of micro-service execution Support an additional three logical name spaces –Rules –Micro-services –Persistent state information

26 Charlie Catlett (cec@uchicago.edu)January 2007 Controlling Remote Operations iRODS - integrated Rule-Oriented Data System

27 Charlie Catlett (cec@uchicago.edu)January 2007 Rule-based Access Associate security policies with each digital entity –Redaction, access controls on structures within a file –Time-dependent access controls (how long to hold data proprietary) Associate access controls with each rule –Restrict ability to modify, apply rules Associate access controls with each micro-service –Explicit control of operation execution within a given collection –Much finer control than provided by Unix r:w:e

28 Charlie Catlett (cec@uchicago.edu)January 2007 For More Information Reagan W. Moore San Diego Supercomputer Center moore@sdsc.edu http://www.sdsc.edu/srb/ http://irods.sdsc.edu/

29 Charlie Catlett (cec@uchicago.edu)January 2007 Call for Participation Papers, tutorials, posters, BOFs, and demonstrations are being accepted through February in three tracks: Science, Technology, and Education, Outreach and Training Submissions are being accepted through April for three competitions for high school, undergraduate and graduate students: Impact of Cyberinfrastructure Research posters On-site advancing scientific discovery


Download ppt "Charlie Catlett 2007 TeraGrid National Cyberinfrastructure for Terascale Science Dane Skow Deputy Director, TeraGrid"

Similar presentations


Ads by Google