Implementation and experience with Big Red (a 30.7 TFLOPS IBM BladeCenter cluster), the Data Capacitor, and HPSS Craig A. Stewart 13 November.

Slides:



Advertisements
Similar presentations
Future of Supercomputer Centers: an evolutionary ecology view Craig Stewart Executive Director, Indiana University Pervasive Technology Institute Associate.
Advertisements

April 19, 2015 CASC Meeting 7 Sep 2011 Campus Bridging Presentation.
What is Cyberinfrastructure?
Bill Barnett, Bob Flynn & Anurag Shankar Pervasive Technology Institute and University Information Technology Services, Indiana University CASC. September.
Data Gateways for Scientific Communities Birds of a Feather (BoF) Tuesday, June 10, 2008 Craig Stewart (Indiana University) Chris Jordan.
1 Supplemental line if need be (example: Supported by the National Science Foundation) Delete if not needed. Supporting Polar Research with National Cyberinfrastructure.
Pti.iu.edu /jetstream Award # A national science & engineering cloud funded by the National Science Foundation Award #ACI Jetstream Overview.
Pti.iu.edu /jetstream Award # A national science & engineering cloud funded by the National Science Foundation Award #ACI Prepared for the.
© Trustees of Indiana University Released under Creative Commons 3.0 unported license; license terms on last slide. Rockhopper: Penguin on Demand at Indiana.
Big Red, the Data Capacitor, and the future (clouds) Craig A. Stewart 2 March 2008.
The TeraGrid: An essential tool for 21st century science Craig Stewart, Associate Dean, Research Technologies Chief Operating Officer, Pervasive Technology.
FutureGrid: an experimental, high-performance grid testbed Craig Stewart Executive Director, Pervasive Technology Institute Indiana University
Campus Bridging: What is it and why is it important? Barbara Hallock – Senior Systems Analyst, Campus Bridging and Research Infrastructure.
Statewide IT Conference, Bloomington IN (October 7 th, 2014) The National Center for Genome Analysis Support, IU and You! Carrie Ganote (Bioinformatics.
Where are we, and where ought we be going? Craig Stewart Executive Director, Pervasive Technology Institute Associate Dean, Research Technologies, Office.
Next Generation Cyberinfrastructures for Next Generation Sequencing and Genome Science AAMC 2013 Information Technology in Academic Medicine Conference.
Research & Academic Computing Bradley C. Wheeler Associate Vice President & Dean.
Craig Stewart 23 July 2009 Cyberinfrastructure in research, education, and workforce development.
INDIANAUNIVERSITYINDIANAUNIVERSITY January 2002 INGEN's advanced IT facilities Craig A. Stewart
The TeraGrid David Hart Indiana University AAAS’09, FEBRUARY 13, 2009.
Goodbye from Indianapolis, IUPUI, and Craig A. Stewart Executive Director, Pervasive Technology Institute Associate Dean, Research Technologies Indiana.
Big Red II & Supporting Infrastructure Craig A. Stewart, Matthew R. Link, David Y Hancock Presented at IUPUI Faculty Council Information Technology Subcommittee.
I-Light: A Network for Collaboration between Indiana University and Purdue University Craig Stewart Associate Vice President Gary Bertoline Associate Vice.
XSEDE Campus Bridging Birds Of a Feather Rich Knepper Craig Stewart James Wade Ferguson Presented at TeraGrid ‘11,
Genomics, Transcriptomics, and Proteomics: Engaging Biologists Richard LeDuc Manager, NCGAS eScience, Chicago 10/8/2012.
DATA-CENTRIC COMPUTING, SCIENCE GATEWAYS, AND THE TERAGRID Kurt A. Seiffert April 2008.
The National Center for Genome Analysis Support as a Model Virtual Resource for Biologists Internet2 Network Infrastructure for the Life Sciences Focused.
Leveraging the National Cyberinfrastructure for Top Down Mass Spectrometry Richard LeDuc.
XSEDE12 Closing Remarks Craig Stewart XSEDE12 General Chair Executive Director, Indiana University Pervasive Technology Institute.
September 6, 2013 A HUBzero Extension for Automated Tagging Jim Mullen Advanced Biomedical IT Core Indiana University.
© Trustees of Indiana University Released under Creative Commons 3.0 unported license; license terms on last slide. The IQ-Table & Collection Viewer A.
Rensselaer Why not change the world? Rensselaer Why not change the world? 1.
RNA-Seq 2013, Boston MA, 6/20/2013 Optimizing the National Cyberinfrastructure for Lower Bioinformatic Costs: Making the Most of Resources for Publicly.
High Performance Computing Processors Felix Noble Mirayma V. Rodriguez Agnes Velez Electric and Computer Engineer Department August 25, 2004.
Implementation and experience with Big Red (a 30.7 TFLOPS IBM BladeCenter cluster), the Data Capacitor, and HPSS Craig A. Stewart 1 November.
1 BioGrids in the US: Current status and future opportunities Craig A. Stewart 15 April 2004 Director, Research and Academic Computing Director,
Pti.iu.edu /jetstream Award # funded by the National Science Foundation Award #ACI Jetstream - A self-provisioned, scalable science and.
July 18, 2012 Campus Bridging Security Challenges from “Panel: Security for Science Gateways and Campus Bridging”
Making Campus Cyberinfrastructure Work for Your Campus Guy Almes Patrick Dreher Craig Stewart Dir. Academy for Dir. Advanced Computing Associate Dean Advanced.
ARGONNE NATIONAL LABORATORY Climate Modeling on the Jazz Linux Cluster at ANL John Taylor Mathematics and Computer Science & Environmental Research Divisions.
Pti.iu.edu /jetstream Award # funded by the National Science Foundation Award #ACI Jetstream Overview – XSEDE ’15 Panel - New and emerging.
Using Prior Knowledge to Improve Scoring in High-Throughput Top-Down Proteomics Experiments Rich LeDuc Le-Shin Wu.
On the road to petascale processing with IU’s Big Red Supercomputer and IBM BladeCenter H Gregory P. Rodgers, IBM Craig A. Stewart, Indiana University.
Research Computing Archived Presentation Title:Indiana Economic Development From Indiana Economic Development Corporation to Indiana and Purdue.
INDIANAUNIVERSITYINDIANAUNIVERSITY Spring 2000 Indiana University Information Technology University Information Technology Services Please cite as: Stewart,
November 18, 2015 Quarterly Meeting 30Aug2011 – 1Sep2011 Campus Bridging Presentation.
February 27, 2007 University Information Technology Services Research Computing Craig A. Stewart Associate Vice President, Research Computing Chief Operating.
UITS Research Technologies – Services Available to Regenstrief Institute 13 Oct 2015 Craig Stewart ORCID ID Executive Director, Indiana.
1 Supplemental line if need be (example: Supported by the National Science Foundation) Delete if not needed. Grand Challenges Discussion 7 Oct 2015 Craig.
A national science & engineering cloud funded by the National Science Foundation Award #ACI Craig Stewart ORCID ID Jetstream.
Recent key achievements in research computing at IU Craig Stewart Associate Vice President, Research & Academic Computing Chief Operating Officer, Pervasive.
TeraGrid Quarterly Meeting Arlington, VA Sep 6-7, 2007 NCSA RP Status Report.
© Trustees of Indiana University Released under Creative Commons 3.0 unported license; license terms on last slide. Update on EAGER: Best Practices and.
Award # funded by the National Science Foundation Award #ACI Jetstream: A Distributed Cloud Infrastructure for.
Jetstream: A new national research and education cloud Jeremy Fischer ORCID Senior Technical Advisor, Collaboration.
1 NSF/TeraGrid Science Advisory Board Meeting July 19-20, San Diego, CA Brief TeraGrid Overview and Expectations of Science Advisory Board John Towns TeraGrid.
A national science & engineering cloud funded by the National Science Foundation Award #ACI Craig Stewart ORCID ID Jetstream.
1 A national science & engineering cloud funded by the National Science Foundation Award #ACI Craig Stewart ORCID ID Jetstream.
NICS Update Bruce Loftis 16 December National Institute for Computational Sciences University of Tennessee and ORNL partnership  NICS is the 2.
© Trustees of Indiana University Released under Creative Commons 3.0 unported license; license terms on last slide. Informatics Tools at the Indiana CTSI.
Jetstream Overview Jetstream: A national research and education cloud Jeremy Fischer ORCID Senior Technical Advisor,
Creating Grid Resources for Undergraduate Coursework John N. Huffman Brown University Richard Repasky Indiana University Joseph Rinkovsky Indiana University.
Introduction to Data Analysis with R on HPC Texas Advanced Computing Center Feb
1 Campus Bridging: What is it and why is it important? Barbara Hallock – Senior Systems Analyst, Campus Bridging and Research Infrastructure.
HPEC-1 SMHS 7/7/2016 MIT Lincoln Laboratory Focus 3: Cell Sharon Sacco / MIT Lincoln Laboratory HPEC Workshop 19 September 2007 This work is sponsored.
Jetstream: A national research and education cloud Jeremy Fischer ORCID Senior Technical Advisor, Collaboration and.
Research & Academic Computing Indiana University Statewide IT Conference 11 September 2003 Indianapolis IN.
Jetstream: A science & engineering cloud Mike Lowe
Matt Link Associate Vice President (Acting) Director, Systems
Presentation transcript:

Implementation and experience with Big Red (a 30.7 TFLOPS IBM BladeCenter cluster), the Data Capacitor, and HPSS Craig A. Stewart 13 November 2007

License Terms Please cite this presentation as: Stewart, C.A. Implementation and experience with Big Red (a 30.7 TFLOPS IBM BladeCenter cluster), the Data Capacitor, and HPSS Presentation. Presented at: IU Display, SC2007 Exhibit Hall (Reno, NV, 13 Nov 2007). Available from: Portions of this document that originated from sources outside IU are shown here and used by permission or under licenses indicated within this document. Items indicated with a © are under copyright and used here with permission. Such items may not be reused without permission from the holder of copyright except where license terms noted on a slide permit reuse. Except where otherwise noted, the contents of this presentation are copyright 2007 by the Trustees of Indiana University. This content is released under the Creative Commons Attribution 3.0 Unported license ( This license includes the following terms: You are free to share – to copy, distribute and transmit the work and to remix – to adapt the work under the following conditions: attribution – you must attribute the work in the manner specified by the author or licensor (but not in any way that suggests that they endorse you or your use of the work). For any reuse or distribution, you must make clear to others the license terms of this work.

Outline Brief history of implementation in TeraGrid and at IU System architecture Performance analysis User experience and science results Lessons learned to date

Image from IU & TeraGrid IU: 2 core campuses, 6 regional campuses President: Michael A. McRobbie Advanced computing: University Information Technology Services, Pervasive Technology Labs, School of Informatics Motivation for being part of TeraGrid: Support national research agendas Improve ability of IU researchers to use national cyberinfrastructure Testbed for IU computer science research

Big Red - Basics and history IBM e1350 BladeCenter Cluster, SLES 9, MPICH, Loadleveler, MOAB Spring 2006: 17 days assembly at IBM facility, disassembled, reassembled in 10 days at IU TFLOPS peak theoretical, achieved on Linpack; 23rd on June 2006 Top500 List (IU’s highest listing to date). In production for local users on 22 August 2006, for TeraGrid users 1 October 2006 Upgraded to TFLOPS Spring 2008; now 42nd on Top500 List Named after nickname for IU sports teams

Data Capacitor - Basics and History Initially funded by $1.7M NSF grant to IU Initially 535 TB of spinning disk - soon to be expanded to more than 1 PB Designed as a temporary holding place for large data sets - a novel type of storage system Uses Lustre file system

HPSS - Basics and History High Performance Storage System Designed initially by IBM and 5 DOE labs IU has contributed code, remains the only unclassified HPSS implementation with distributed storage Data written to HPSS is by default copied to IUB and IUPUI

Feature20.4 TFLOPS30.7 TFLOPS Computational hardware, RAM JS21 componentsTwo 2.5 GHz PowerPC 970MP processors, 8 GB RAM, 73 GB SAS Drive, 40 GFLOPS Same No. of JS21 blades No. of processors; cores1,024 processors; 2,048 processor cores 1,536 processors; 3,072 processor cores Total system memory4 TB6 TB Disk storage GPFS scratch space266 TBSame Lustre535 TBSame Home directory space25 TBSame Networks Total outbound network bandwidth40 Gbit/secSame Bisection bandwidth64 GB/sec - Myrinet GB/sec - Myrinet 2000

Motivations and goals Initial goals for TFLOPS system: Local demand for cycles exceeded supply TeraGrid Resource Partner commitments to meet Support life science research Support applications at 100s to 1000s of processors 2nd phase upgrade to TFLOPS Support economic development in State of Indiana

Why a PowerPC-based blade cluster? Processing power per node Density, good power efficiency relative to available processors Possibility of performance gains through use of Altivec unit & VMX instructions Blade architecture provides flexibility for future Results of Request for Proposals process ProcessorTFLOPS/ MWatt MWatts/ PetaFLOPS Intel Xeon AMD PowerPC 970 MP (dual core)

Difference: 4 KB vs 16 MB page size Linpack performance Benchmark set NodesPeak Theoretical TFLOPS Achieved TFLOPS % HPCC Top Top

HPCC and Linpack Results (510 nodes) G-HPLG- PTRANS G- Random Access G-FFTEEP- STREAM Sys EP- STREAM Triad EP- DGEMM Random Ring Bandwidth Random Ring Latency GB/susec TFlop/sGB/sGup/sGFlop/sGB/s GFlop/s Total Per processor Data posted to

September 23, TFLOPS e1350 (Big Red) vs. a Cray XT3 at Oak Ridge National Labs, including 5200 single core 2.4 GHz AMD Opteron processors (left), and a 2.09 TFLOPS HP XC4000 owned by HP, Inc., including 256 dual-core ADM Opteron processors (right).

Elapsed time per simulation timestep among best in TeraGrid

Bandwidth Challenge SC|2006

Competition Performance During testing 4 x 2 trunked 1 Gb lines 32 GB in 34 seconds - 941MB/s Competition All four experiments Sustained Gb

HPSS I/O Speed Growth

Simulation of TonB-dependent transporter (TBDT) Used systems at NCSA, IU, PSC Modeled mechanisms for allowing transport of molecules through cell membrane Work by Emad Tajkhorshid and James Gumbart, of University of Illinois Urbana-Champaign. Mechanics of Force Propagation in TonB- Dependent Outer Membrane Transport. Biophysical Journal 93: (2007) To view the results of the simulation, please go to: TonB-BtuB/btub-2.5Ans.mpg Image courtesy of Emad Tajkhorshid

ChemBioGrid Analyzed 555,007 abstracts in PubMed in ~ 8,000 CPU hours Used OSCAR3 to find SMILES strings -> SDF format -> 3D structure (GAMESS) - > into Varuna database and then other applications “Calculate and look up” model for ChemBioGrid

WxChallenge ( Over 1,000 undergraduate students, 64 teams, 56 institutions Usage on Big Red: ~16,000 CPU hours on Big Red 63% of processing done on Big Red Most of the students who used Big Red couldn’t tell you what it is Integration of computation and data flows via Lustre (Data Capacitor)

Overall user reactions NAMD, WRF users very pleased Porting from Intel instruction set a perceived and sometimes real challenge in a cycle-rich environment MILC optimization with VMX not successful so far in eyes of user community Keys to biggest successes: Performance characteristics of JS21 nodes Linkage of computation and storage (Lustre - Data Capacitor) Support for grid computing via TeraGrid

Evaluation of implementation The manageability of the system is excellent For a select group of applications, Big Red provides excellent performance and reasonable scalability We are likely to expand bandwidth from Big Red to the rest of the IU cyberinfrastructure Quarry is a critical companion to Big Red; without Quarry Big Red would not be bnearly so successful Focus on data management and scalable computation critical to success Next steps: industrial partnerships and economic development in Indiana

Conclusions A 30.7& TFLOPS system with “not the usual” processors was successfully implemented serving local Indiana University researchers, and the national research audience via the TeraGrid Integration of computation and data management systems was critical to success In the future Science Gateways will be increasingly important: Most scientists can’t constantly chase after the fastest available system; gateway developers might be able to Programmability of increasingly unusual architectures not likely to become easier For applications with broad potential user bases, or extreme scalability on specialized systems, Science Gateways will be critical in enabling transformational capabilities and supporting scientific workflows. Achieving broad use can only be achieved by relieving scientists of need to understand details of systems

Acknowledgements - Funding Sources IU’s involvement as a TeraGrid Resource Partner is supported in part by the National Science Foundation under Grants No. ACI l, OCI , OCI , and OCI The IU Data Capacitor is supported in part by the National Science Foundation under Grant No. CNS This research was supported in part by the Indiana METACyt Initiative. The Indiana METACyt Initiative of Indiana University is supported in part by Lilly Endowment, Inc. This work was supported in part by Shared University Research grants from IBM, Inc. to Indiana University. The LEAD portal is developed under the leadership of IU Professors Dr. Dennis Gannon and Dr. Beth Plale, and supported by NSF grant The ChemBioGrid Portal is developed under the leadership of IU Professor Dr. Geoffrey C. Fox and Dr. Marlon Pierce and funded via the Pervasive Technology Labs (supported by the Lilly Endowment, Inc.) and the National Institutes of Health grant P20 HG Many of the ideas presented in this talk were developed under a Fulbright Senior Scholar’s award to Stewart, funded by the US Department of State and the Technische Universitaet Dresden. Any opinions, findings and conclusions or recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of the National Science Foundation (NSF), National Institutes of Health (NIH), Lilly Endowment, Inc., or any other funding agency

Acknowledgements - People Research and Technical Services Division Staff generally, High Performance Systems, High Performance Applications, Research Storage, and Team Data Capacitor particularly Malinda Lingwall for editing, graphic layout, and managing process Maria Morris contributed to the graphics used in this talk Marcus Christie and Surresh Marru of the Extreme! Computing Lab contributed the LEAD graphics John Morris ( and Cairril Mills (Cairril.com Design & Marketing) contributed graphics Rick McMullen and all the Huffmans (CIMA) Randy Bramley and Marie Ma (Obsidian) Mookie Baik and Yogita Mantri (Chemistry) Beth Plale, Dennis Gannon, AJ Ragusa, Suresh Marru, Chathura Herath (LEAD) Doug Balog, Derek Simmel (PSC) Guido Juckeland, Robert Henschel, Matthias Mueller (ZIH) This work would not have been possible without the dedicated and expert efforts of the staff of the Research Technologies Division of University Information Technology Services, the faculty and staff of the Pervasive Technology Labs, and the staff of UITS generally. Thanks to the faculty and staff with whom we collaborate locally at IU and globally (via the TeraGrid, and especially at Technische Universitaet Dresden)

Thank you Any questions?