Indiana University's Lustre WAN: Empowering Production Workflows on the TeraGrid and beyond Craig Stewart and Stephen C. Simms Indiana University

Slides:



Advertisements
Similar presentations
Pennsylvania Banner Users Group 2008 Fall Conference IUP Banner Server Environment Jeff Montgomery.
Advertisements

Lustre 100GBit Testbed Michael Kluge Robert Henschel, Stephen Simms
Cross-site data transfer on TeraGrid using GridFTP TeraGrid06 Institute User Introduction to TeraGrid June 12 th by Krishna Muriki
Xsede eXtreme Science and Engineering Discovery Environment Ron Perrott University of Oxford 1.
1 US activities and strategy :NSF Ron Perrott. 2 TeraGrid An instrument that delivers high-end IT resources/services –a computational facility – over.
User Introduction to the TeraGrid 2007 SDSC NCAR TACC UC/ANL NCSA ORNL PU IU PSC.
IBM 1350 Cluster Expansion Doug Johnson Senior Systems Developer.
Architecture and Implementation of Lustre at the National Climate Computing Research Center Douglas Fuller National Climate Computing Research Center /
NSF Site Visit HYDRA Using Windows Desktop Systems in Distributed Parallel Computing.
NWfs A ubiquitous, scalable content management system with grid enabled cross site data replication and active storage. R. Scott Studham.
1 Supplemental line if need be (example: Supported by the National Science Foundation) Delete if not needed. Supporting Polar Research with National Cyberinfrastructure.
Simo Niskala Teemu Pasanen
National Energy Research Scientific Computing Center (NERSC) The GUPFS Project at NERSC GUPFS Team NERSC Center Division, LBNL November 2003.
SALSASALSASALSASALSA Digital Science Center June 25, 2010, IIT Geoffrey Fox Judy Qiu School.
Core Services I & II David Hart Area Director, UFP/CS TeraGrid Quarterly Meeting December 2008.
Network, Operations and Security Area Tony Rimovsky NOS Area Director
Big Red, the Data Capacitor, and the future (clouds) Craig A. Stewart 2 March 2008.
Project Overview:. Longhorn Project Overview Project Program: –NSF XD Vis Purpose: –Provide remote interactive visualization and data analysis services.
The Creation of a Big Data Analysis Environment for Undergraduates in SUNY Presented by Jim Greenberg SUNY Oneonta on behalf of the SUNY wide team.
An Introduction and Overview of Grid Computing Presenter: Xiaofei Cao Patrick Berg.
Indiana University's Lustre WAN: Empowering Production Workflows on the TeraGrid and beyond Craig Stewart and Stephen C. Simms Indiana University
Empowering Bioinformatics Workflows Using the Lustre Wide Area File System across a 100 Gigabit Network Stephen Simms Manager, High Performance File Systems.
LARGE SCALE DEPLOYMENT OF DAP AND DTS Rob Kooper Jay Alemeda Volodymyr Kindratenko.
Descriptive Data Analysis of File Transfer Data Sudarshan Srinivasan Victor Hazlewood Gregory D. Peterson.
TeraGrid Science Gateways: Scaling TeraGrid Access Aaron Shelmire¹, Jim Basney², Jim Marsteller¹, Von Welch²,
Implementation and experience with Big Red (a 30.7 TFLOPS IBM BladeCenter cluster), the Data Capacitor, and HPSS Craig A. Stewart 13 November.
Big Red II & Supporting Infrastructure Craig A. Stewart, Matthew R. Link, David Y Hancock Presented at IUPUI Faculty Council Information Technology Subcommittee.
DATA-CENTRIC COMPUTING, SCIENCE GATEWAYS, AND THE TERAGRID Kurt A. Seiffert April 2008.
1 Preparing Your Application for TeraGrid Beyond 2010 TG09 Tutorial June 22, 2009.
RNA-Seq 2013, Boston MA, 6/20/2013 Optimizing the National Cyberinfrastructure for Lower Bioinformatic Costs: Making the Most of Resources for Publicly.
Multi-Channel Radar Depth Sounder (MCRDS) Signal Processing: A Distributed Computing Approach Je’aime Powell 1, Dr. Linda Hayden 1, Dr. Eric Akers 1, Richard.
Implementation and experience with Big Red (a 30.7 TFLOPS IBM BladeCenter cluster), the Data Capacitor, and HPSS Craig A. Stewart 1 November.
“Indiana University should continue to execute and accelerate an incremental and extensible strategy that enhances its overall storage infrastructure from.
HYDRA: Using Windows Desktop Systems in Distributed Parallel Computing Arvind Gopu, Douglas Grover, David Hart, Richard Repasky, Joseph Rinkovsky, Steve.
Building Effective CyberGIS: FutureGrid Marlon Pierce, Geoffrey Fox Indiana University.
Kurt Mueller San Diego Supercomputer Center NPACI HotPage Updates.
SALSASALSASALSASALSA FutureGrid Venus-C June Geoffrey Fox
Advanced User Support Amit Majumdar 5/7/09. Outline  Three categories of AUS  Update on Operational Activities  AUS.ASTA  AUS.ASP  AUS.ASEOT.
1 CReSIS Lawrence Kansas February Geoffrey Fox (PI) Computer Science, Informatics, Physics Chair Informatics Department Director Digital Science.
Lustre-WAN: Enabling Distributed Workflows in Astrophysics Scott Michael April 2011.
This material is based upon work supported by the National Science Foundation under Grant No. ANT Any opinions, findings, and conclusions or recommendations.
February 27, 2007 University Information Technology Services Research Computing Craig A. Stewart Associate Vice President, Research Computing Chief Operating.
INDIANAUNIVERSITYINDIANAUNIVERSITY The Data Capacitor Digital Library Brown Bag February 22, 2006 Stephen Simms Data Capacitor Project Manager Research.
Ultimate Integration Joseph Lappa Pittsburgh Supercomputing Center ESCC/Internet2 Joint Techs Workshop.
PSC’s CRAY-XT3 Preparation and Installation Timeline.
TeraGrid Extension Gateway Activities Nancy Wilkins-Diehr TeraGrid Quarterly, September 24-25, 2009 The Extension Proposal!
Bio-IT World Conference and Expo ‘12, April 25, 2012 A Nation-Wide Area Networked File System for Very Large Scientific Data William K. Barnett, Ph.D.
Network, Operations and Security Area Tony Rimovsky NOS Area Director
Office of Science U.S. Department of Energy NERSC Site Report HEPiX October 20, 2003 TRIUMF.
Galaxy Community Conference July 27, 2012 The National Center for Genome Analysis Support and Galaxy William K. Barnett, Ph.D. (Director) Richard LeDuc,
NICS Update Bruce Loftis 16 December National Institute for Computational Sciences University of Tennessee and ORNL partnership  NICS is the 2.
IU Site Update TeraGrid Round Table Craig Stewart, Stephen Simms, Kurt Seiffert November 4, 2010.
Purdue RP Highlights TeraGrid Round Table May 20, 2010 Preston Smith Manager - HPC Grid Systems Rosen Center for Advanced Computing Purdue University.
Attribute-based Authentication for Gateways Jim Basney Terry Fleury Stuart Martin JP Navarro Tom Scavo Nancy Wilkins-Diehr.
This material is based upon work supported by the National Science Foundation under Grant No. ANT Any opinions, findings, and conclusions or recommendations.
TG ’08, June 9-13, State of TeraGrid John Towns Co-Chair, TeraGrid Forum Director, Persistent Infrastructure National Center for Supercomputing.
Common User Environments - Update Shawn T. Brown, PSC CUE Working Group Lead TG Quartely 1.
Data Infrastructure in the TeraGrid Chris Jordan Campus Champions Presentation May 6, 2009.
Creating Grid Resources for Undergraduate Coursework John N. Huffman Brown University Richard Repasky Indiana University Joseph Rinkovsky Indiana University.
© 2010 Pittsburgh Supercomputing Center Lustre WAN ExTENCI Kickoff Meeting August 19, 2010 Josephine Palencia J. Ray Scott.
Indiana University UITS/PTI
Joint Techs, Columbus, OH
National Center for Genome Analysis Support
Richard LeDuc, Ph.D. (Manager)
Title of Poster Site Visit 2017 Introduction Results
OGCE Portal Applications for Grid Computing
Scalable Database Services for Physics: Oracle 10g RAC on Linux
OGCE Portal Applications for Grid Computing
Storing and Accessing G-OnRamp’s Assembly Hubs outside of Galaxy
Title of Poster Site Visit 2018 Introduction Results
Presentation transcript:

Indiana University's Lustre WAN: Empowering Production Workflows on the TeraGrid and beyond Craig Stewart and Stephen C. Simms Indiana University

NSF initial funding in 2005, expanded with IU funds Aggregate 936 formatted Terabytes Lustre storage 14.5 GB/s aggregate write Short term storage The Data Capacitor Project

IU’s Data Capacitor WAN 1 pair Dell PowerEdge 2950 for MDS 2 pair Dell PowerEdge 2950 for OSS –2 x 3.0 GHz Dual Core Xeon –Myrinet 10G Ethernet –Dual port Qlogic 2432 HBA (4 x FC) –2.6 Kernel (RHEL 5) DDN S2A9550 Controller –Over 2.4 GB/sec measured throughput –360 Terabytes of spinning SATA disk Currently running Lustre Upgrading to in May Announced production at LUG 2008 Allocated on Project by Project basis

IU UID Mapping Lightweight Not everyone needs / wants kerberos Not everyone needs / wants encryption Only change MDS code Want to maximize clients we can serve Simple enough to port the code forward

IU UID Mapping cont’d UID lookups on the MDS call a pluggable kernel module –Binary tree stored in memory –Based on NID or NID range –Remote UID mapped to Effective UID

Username IP Tables Client NID/UID 1.4.x 1.6.x Kernel Mod Kernel Memory Patched MDS NID - Remote UID - Local UID Client UIDs /etc/passwd Client UIDs /etc/passwd TGCDB Username TGCDB Username NID Ranges SQLite

UID Mapping Userspace – Kernel Space Barrier – Only crossed when we update the table Create a Forest of Binary Trees – Forward and Inverse Lookups for each UID – Time consumed for lookup is predictable Speed over Space Consume memory rather than on the fly lookups Every UID node consumes 6 Ints 300 Users approximately 300KB

IU’s Lustre WAN on the TeraGrid 8 Sites currently mounting IU DC-WAN –IU, LONI, NCSA, NICS, PSC, Purdue, SDSC, TACC 5 Sites mounting on compute resources –IU, LONI, NCSA, PSC, TACC Average of 93% capacity for the last quarter 2009 uptime of 96% –Filesystem availability to users PBs of aggregate writes and reads in NSF FY 2010

NOAO/AURA/NSF One Degree Imager (ODI) HPSS WIYN Telescope Tucson, Arizona 1726 miles

Ethnographic Video for Instruction and Analysis EVIA Samba Video Acquisiton Server HPSS Compression/Annotation Server 1 mile 346 miles Ann Arbor, Michigan

Linked Environments for Atmospheric Discovery LEAD Big Red Compute Resource Data Transfer Server 2 miles

Center for the Remote Sensing of Ice Sheets (CReSIS) Workflow gg U of Kansas Greenland IU Quarry Cluster HPSS 517 miles Lawrence, Kansas

Samba CRYO Electron Microscopy 3 miles HPSS Big Red

Equation of State Simulations and Plasma Pasta

EOS and Plasma Pasta 879 miles 3 miles Simulation Machine Analysis Machine Austin, Texas HPSS

Computational Fluid Dynamics Pittsburgh, PA 410 miles Big Red Pople OpenMP Paraview

Gas Giant Planet Research Urbana, IL Pittsburgh, PA 410 miles 147 miles Starkville, MS 607 miles HPSS Visualization

Beyond the TeraGrid Dresden – ZIH (Technische Universitaet Dresden) Denmark – Risø – National Laboratory for Sustainable Energy Finland – Metsähovi Radio Observatory

Many Thanks Josh Walgenbach, Justin Miller, Nathan Heald, James McGookey, Resat Payli***, Suresh Marru, Robert Henschel, Scott Michael, Tom Johnson, Chuck Horowitz, Don Berry, Scott, Teige, David Morgan, Matt Link (IU) Kit Westneat (DDN) Oracle support and engineering Michael Kluge, Guido Juckeland, Matthias Mueller (ZIH,Dresden) Thorbjorn Axellson (CReSIS) Greg Pike and ORNL Doug Balog, Josephine Palencia, and PSC Trey Breckenridge, Roger Smith, Joey Jones (Mississippi State University) Support for this work provided by the National Science Foundation is gratefully acknowledged and appreciated (CNS ). Any opinions expressed are those of the authors and do not necessarily reflect the views of the NSF

Thank you! Questions?