Science on the TeraGrid Daniel S. Katz Director of Science, TeraGrid GIG Senior Computational Researcher, Computation Institute, University.

Slides:



Advertisements
Similar presentations
1 Introduction to the TeraGrid Daniel S. Katz TeraGrid GIG Director of Science Senior Grid/HPC Researcher, Computation Institute University.
Advertisements

1 US activities and strategy :NSF Ron Perrott. 2 TeraGrid An instrument that delivers high-end IT resources/services –a computational facility – over.
Kathy Benninger, Pittsburgh Supercomputing Center Workshop on the Development of a Next-Generation Cyberinfrastructure 1-Oct-2014 NSF Collaborative Research:
The ADAMANT Project: Linking Scientific Workflows and Networks “Adaptive Data-Aware Multi-Domain Application Network Topologies” Ilia Baldine, Charles.
Science Gateways on the TeraGrid Von Welch, NCSA (with thanks to Nancy Wilkins-Diehr, SDSC for many slides)
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
Science on the TeraGrid Daniel S. Katz Director of Science, TeraGrid GIG Senior Computational Researcher, Computation Institute, University.
1 PSC update July 3, 2008 Ralph Roskies, Scientific Director Pittsburgh Supercomputing Center
Simo Niskala Teemu Pasanen
April 2009 OSG Grid School - RDU 1 Open Science Grid John McGee – Renaissance Computing Institute University of North Carolina, Chapel.
TG RoundTable, Purdue RP Update October 11, 2008 Carol Song Purdue RP PI Rosen Center for Advanced Computing.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
Scientific Data Infrastructure in CAS Dr. Jianhui Scientific Data Center Computer Network Information Center Chinese Academy of Sciences.
ASQ World Conference on Quality and Improvement May 24-26, 2010, St. Louis, MO Quality in Chaos: a view from the TeraGrid environment John Towns.
Open Science Grid For CI-Days Internet2: Fall Member Meeting, 2007 John McGee – OSG Engagement Manager Renaissance Computing Institute.
18:15:32Service Oriented Cyberinfrastructure Lab, Grid Deployments Saul Rioja Link to presentation on wiki.
AT LOUISIANA STATE UNIVERSITY University of Washington – e-Science Introduction to the TeraGrid Jeffrey P. Gardner Sr. Research Scientist, High Performance.
The TeraGrid David Hart Indiana University AAAS’09, FEBRUARY 13, 2009.
GridChem-- User Support Kent Milfeld Supported by the NSF NMI Program under Award # Oct. 10, 2005.
Advancing Scientific Discovery through TeraGrid Scott Lathrop TeraGrid Director of Education, Outreach and Training University of Chicago and Argonne National.
TeraGrid Resources Enabling Scientific Discovery Through Cyberinfrastructure (CI) Diane Baxter, Ph.D. San Diego Supercomputer Center University of California,
Software for Science Gateways: Open Grid Computing Environments Marlon Pierce, Suresh Marru Pervasive Technology Institute Indiana University
1 TeraGrid ‘10 August 2-5, 2010, Pittsburgh, PA State of TeraGrid in Brief John Towns TeraGrid Forum Chair Director of Persistent Infrastructure National.
August 2007 Advancing Scientific Discovery through TeraGrid Adapted from S. Lathrop’s talk in SC’07
SAN DIEGO SUPERCOMPUTER CENTER NUCRI Advisory Board Meeting November 9, 2006 Science Gateways on the TeraGrid Nancy Wilkins-Diehr TeraGrid Area Director.
TeraGrid Overview Cyberinfrastructure Days Internet2 10/9/07 Mark Sheddon Resource Provider Principal Investigator San Diego Supercomputer Center
1 Preparing Your Application for TeraGrid Beyond 2010 TG09 Tutorial June 22, 2009.
Open Science Grid For CI-Days Elizabeth City State University Jan-2008 John McGee – OSG Engagement Manager Manager, Cyberinfrastructure.
PROCESSED RADAR DATA INTEGRATION WITH SOCIAL NETWORKING SITES FOR POLAR EDUCATION Jeffrey A. Wood April 19, 2010 A Thesis submitted to the Graduate Faculty.
NanoHUB.org and HUBzero™ Platform for Reproducible Computational Experiments Michael McLennan Director and Chief Architect, Hub Technology Group and George.
Interoperability Grids, Clouds and Collaboratories Ruth Pordes Executive Director Open Science Grid, Fermilab.
Geosciences - Observations (Bob Wilhelmson) The geosciences in NSF’s world consists of atmospheric science, ocean science, and earth science Many of the.
NA-MIC National Alliance for Medical Image Computing UCSD: Engineering Core 2 Portal and Grid Infrastructure.
1 TeraGrid and the Path to Petascale John Towns Chair, TeraGrid Forum Director, Persistent Infrastructure National Center for Supercomputing Applications.
TeraGrid National CyberInfrastructure for Scientific Research Philip Blood Senior Scientific Specialist Pittsburgh Supercomputing Center April 23, 2010.
GRIDS Center Middleware Overview Sandra Redman Information Technology and Systems Center and Information Technology Research Center National Space Science.
Cyberinfrastructure What is it? Russ Hobby Internet2 Joint Techs, 18 July 2007.
GRID Overview Internet2 Member Meeting Spring 2003 Sandra Redman Information Technology and Systems Center and Information Technology Research Center National.
ISERVOGrid Architecture Working Group Brisbane Australia June Geoffrey Fox Community Grids Lab Indiana University
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
Leveraging the InCommon Federation to access the NSF TeraGrid Jim Basney Senior Research Scientist National Center for Supercomputing Applications University.
TeraGrid Quarterly Meeting Arlington, VA Sep 6-7, 2007 NCSA RP Status Report.
1 NSF/TeraGrid Science Advisory Board Meeting July 19-20, San Diego, CA Brief TeraGrid Overview and Expectations of Science Advisory Board John Towns TeraGrid.
TeraGrid Gateway User Concept – Supporting Users V. E. Lynch, M. L. Chen, J. W. Cobb, J. A. Kohl, S. D. Miller, S. S. Vazhkudai Oak Ridge National Laboratory.
Sergiu April 2006June 2006 Overview of TeraGrid Resources and Services Sergiu Sanielevici, TeraGrid Area Director for User.
Using the ARCS Grid and Compute Cloud Jim McGovern.
Cyberinfrastructure: Many Things to Many People Russ Hobby Program Manager Internet2.
National Center for Supercomputing Applications University of Illinois at Urbana–Champaign Visualization Support for XSEDE and Blue Waters DOE Graphics.
Experiences Running Seismic Hazard Workflows Scott Callaghan Southern California Earthquake Center University of Southern California SC13 Workflow BoF.
Funded by the NSF OCI program grants OCI and OCI Mats Rynge, Gideon Juve, Karan Vahi, Gaurang Mehta, Ewa Deelman Information Sciences Institute,
NICS Update Bruce Loftis 16 December National Institute for Computational Sciences University of Tennessee and ORNL partnership  NICS is the 2.
OGCE Workflow and LEAD Overview Suresh Marru, Marlon Pierce September 2009.
AT LOUISIANA STATE UNIVERSITY CCT: Center for Computation & Technology Introduction to the TeraGrid Daniel S. Katz Lead, LONI as a TeraGrid.
Southern California Earthquake Center CyberShake Progress Update November 3, 2014 – 4 May 2015 UGMS May 2015 Meeting Philip Maechling SCEC IT Architect.
AT LOUISIANA STATE UNIVERSITY CCT: Center for Computation & LSU Condor in Louisiana Tevfik Kosar Center for Computation & Technology Louisiana.
PEER 2003 Meeting 03/08/031 Interdisciplinary Framework Major focus areas Structural Representation Fault Systems Earthquake Source Physics Ground Motions.
Northwest Indiana Computational Grid Preston Smith Rosen Center for Advanced Computing Purdue University - West Lafayette West Lafayette Calumet.
TG ’08, June 9-13, State of TeraGrid John Towns Co-Chair, TeraGrid Forum Director, Persistent Infrastructure National Center for Supercomputing.
1 Open Science Grid: Project Statement & Vision Transform compute and data intensive science through a cross- domain self-managed national distributed.
TeraGrid and Web 2.0 Technologies Daniel S. Katz Director of Science, TeraGrid GIG Senior Computational Researcher, Computation Institute,
TeraGrid’s Process for Meeting User Needs. Jay Boisseau, Texas Advanced Computing Center Dennis Gannon, Indiana University Ralph Roskies, University of.
INTRODUCTION TO XSEDE. INTRODUCTION  Extreme Science and Engineering Discovery Environment (XSEDE)  “most advanced, powerful, and robust collection.
SCEC CyberShake on TG & OSG: Options and Experiments Allan Espinosa*°, Daniel S. Katz*, Michael Wilde*, Ian Foster*°,
Accessing the VI-SEEM infrastructure
Seismic Hazard Analysis Using Distributed Workflows
Scott Callaghan Southern California Earthquake Center
CyberShake Study 16.9 Discussion
Cyberinfrastructure and PolarGrid
rvGAHP – Push-Based Job Submission Using Reverse SSH Connections
CyberShake Study 18.8 Technical Readiness Review
Presentation transcript:

Science on the TeraGrid Daniel S. Katz Director of Science, TeraGrid GIG Senior Computational Researcher, Computation Institute, University of Chicago & Argonne National Laboratory Affiliate Faculty, Center for Computation & Technology, LSU Adjunct Associate Professor, Electrical and Computer Engineering Department, LSU

What is the TeraGrid? World’s largest open scientific discovery infrastructure Leadership class resources at eleven partner sites combined to create an integrated, persistent computational resource –High-performance computers (>2 Pflops, >150,000 cores) And a Condor pool (w/ ~13,000 CPUs) –Medium to long term storage (>3 PB disk, >60 PB of tape) –Visualization systems –Data collections (>30 PB, >100 discipline-specific databases) –Science gateways (~35) –High-performance networks –Unified user services – portal, help desk, training, advanced app support Allocated to US researchers and their collaborators through national peer-review process –Generally, review of computing, not science Extremely user-driven –MPI jobs, ssh or grid (GRAM) access, etc.

Governance 11 Resource Providers (RPs) funded under separate agreements with NSF –Different start and end dates –Different goals –Different agreements –Different funding models 1 Coordinating Body – Grid Integration Group (GIG) –University of Chicago/Argonne National Laboratory –Subcontracts to all RPs and six other universities –7-8 Area Directors –Working groups with members from many RPs TeraGrid Forum with Chair

TG New Large Resources –First NSF ‘Track2’ HPC system –504 TF –15,744 Quad-Core AMD Opteron processors –123 TB memory, 1.7 PB disk (UT/ORNL) –Second NSF ‘Track2’ HPC system –1 PF Cray XT5 system 10,000+ compute sockets 100 TB memory, 2.3 PB disk Blue NSF Track 1 10 PF peak Coming in 2011

Who Uses TeraGrid (2007) Molecular Biosciences 31% Chemistry 17% Physics 17% Astronomical Sciences 12% Materials Research 6% Earth Sciences 3% All 19 Others 4% Advanced Scientific Computing 2% Atmospheric Sciences 3% Chemical, Thermal Systems 5%

How TeraGrid Is Used Use Modality Community Size (rough est. - number of users) Batch Computing on Individual Resources 850 Exploratory and Application Porting 650 Workflow, Ensemble, and Parameter Sweep 250 Science Gateway Access 500 Remote Interactive Steering and Visualization 35 Tightly-Coupled Distributed Computation data

Science Gateways A natural extension of Internet & Web 2.0 Idea resonates with Scientists –Researchers can imagine scientific capabilities provided through familiar interface Mostly web portal or web or client-server program Designed by communities; provide interfaces understood by those communities –Also provide access to greater capabilities (back end) –Without user understand details of capabilities –Scientists know they can undertake more complex analyses and that’s all they want to focus on –TeraGrid provides tools to help developer Seamless access doesn’t come for free –Hinges on very capable developer Slide courtesy of Nancy Wilkins-Diehr

TG App: Predicting storms Hurricanes and tornadoes cause massive loss of life and damage to property TeraGrid supported spring 2007 NOAA and University of Oklahoma Hazardous Weather Testbed –Major Goal: assess how well ensemble forecasting predicts thunderstorms, including the supercells that spawn tornadoes –Nightly reservation at PSC, spawning jobs at NCSA as needed for details –Input, output, and intermediate data transfers –Delivers “better than real time” prediction –Used 675,000 CPU hours for the season –Used 312 TB on HPSS storage at PSC Slide courtesy of Dennis Gannon, ex-IU, and LEAD Collaboration

TG App: SCEC-PSHA Part of SCEC (Tom Jordan, USC) Using the large scale simulation data, estimate probablistic seismic hazard (PSHA) curves for sites in southern California (probability that ground motion will exceed some threshold over a given time period) Used by hospitals, power plants, schools, etc. as part of their risk assessment For each location, need a Cybershake run followed by roughly 840,000 parallel short jobs (420,000 rupture forecasts, 420,000 extraction of peak ground motion) –Parallelize across locations, not individual workflows Completed 40 locations to date, targeting 200 in 2009, and 2000 in 2010 Managing these requires effective grid workflow tools for job submission, data management and error recovery, using Pegasus (ISI) and DAGman (Wisconsin) 9 Information/image courtesy of Phil Maechling

App: GridChem Slide courtesy of Joohyun Kim Different licensed applications with different queues Will be scheduled for workflows

TG Apps: Genius and Materials HemeLB on LONI LAMMPS on TeraGrid Fully-atomistic simulations of clay-polymer nanocomposites Slide courtesy of Steven Manos and Peter Coveney Why cross-site / distributed runs? 1.Rapid turnaround, conglomeration of idle processors to run a single large job 2.Run big compute & big memory jobs not possible on a single machine Modeling blood flow before (during?) surgery

App: PolarGrid Goal: Work with Center for Remote Sensing of Ice Sheets Requirements: –View CReSIS data sets, run filters, and view results through Web map interfaces; –See/Share user’s events in a Calendar; –Update results to a common repository with appropriate access controls; –Post the status of computational experiments. –Support collaboration and information exchange by interfacing to blogs and discussion areas Slide courtesy of: Raminder Singh, Gerald Guo, Marlon Pierce Login Screen Interface to create new users and login using existing accounts. Integrated with OpenID API for authentication. Solution: Web 2.0 enabled PolarGrid Portal

App: PolarGrid Home Page with a set of gadgets like Google Calendar, Picasa, Facebook, Blog, Twitter Slide courtesy of: Raminder Singh, Gerald Guo, Marlon Pierce

Future App: Real-time high resolution Radar data Delivering 3D visualization of radar data via a Google gadget LiveRadar3D –Super high res, real-time NEXRAD data –Continuously updated as new data comes –3D rendering that includes multiple stations in the US –Significant processing (high throughout) and rendering supported by TG systems –To be released next spring Slide courtesy of Carol Song

TeraGrid -> XD Future Current RP agreements end in March 2011 –Except track 2 centers (current and future) TeraGrid XD (eXtreme Digital) starts in April 2011 –Era of potential interoperation with OSG and others –New types of science applications? Current TG GIG continues through July 2011 –Allows four months of overlap in coordination –Probable overlap between GIG and XD members Blue Waters (track 1) production in

TeraGrid: Both Operations and Research 16 Operations –Facilities/services on which researchers rely –Infrastructure on which other providers build AND R&D –Learning how to do distributed, collaborative science on a global, federated infrastructure –Learning how to run multi-institution shared infrastructure