Elastic Cyberinfrastructure for Research Computing

Slides:



Advertisements
Similar presentations
© 2012 Open Grid Forum Simplifying Inter-Clouds October 10, 2012 Hyatt Regency Hotel Chicago, Illinois, USA.
Advertisements

1Copyright © 2013, Oracle and/or its affiliates. All rights reserved.
System Center 2012 R2 Overview
ASCR Data Science Centers Infrastructure Demonstration S. Canon, N. Desai, M. Ernst, K. Kleese-Van Dam, G. Shipman, B. Tierney.
Cloud Computing to Satisfy Peak Capacity Needs Case Study.
1 Cyberinfrastructure Framework for 21st Century Science & Engineering (CIF21) NSF-wide Cyberinfrastructure Vision People, Sustainability, Innovation,
C LOUD C OMPUTING Presented by Ye Chen. What is cloud computing? Cloud computing is a model for enabling ubiquitous, convenient, on- demand network access.
Tunis, Tunisia, 28 April 2014 Business Values of Virtualization Mounir Ferjani, Senior Product Manager, Huawei Technologies 2.
1 Software & Grid Middleware for Tier 2 Centers Rob Gardner Indiana University DOE/NSF Review of U.S. ATLAS and CMS Computing Projects Brookhaven National.
INTRODUCTION TO CLOUD COMPUTING CS 595 LECTURE 4.
Bill Wrobleski Director, Technology Infrastructure ITS Infrastructure Services.
April 2009 OSG Grid School - RDU 1 Open Science Grid John McGee – Renaissance Computing Institute University of North Carolina, Chapel.
Research Computing with Newton Gerald Ragghianti Nov. 12, 2010.
1 Building National Cyberinfrastructure Alan Blatecky Office of Cyberinfrastructure EPSCoR Meeting May 21,
Server and Short to Mid Term Storage Funding Research Computing Funding Issues.
Clouds on IT horizon Faculty of Maritime Studies University of Rijeka Sanja Mohorovičić INFuture 2009, Zagreb, 5 November 2009.
Open Science Grid For CI-Days Internet2: Fall Member Meeting, 2007 John McGee – OSG Engagement Manager Renaissance Computing Institute.
EGI-Engage EGI-Engage Engaging the EGI Community towards an Open Science Commons Project Overview 9/14/2015 EGI-Engage: a project.
PCGRID ‘08 Workshop, Miami, FL April 18, 2008 Preston Smith Implementing an Industrial-Strength Academic Cyberinfrastructure at Purdue University.
Science Clouds and FutureGrid’s Perspective June Science Clouds Workshop HPDC 2012 Delft Geoffrey Fox
Open Science Grid For CI-Days Elizabeth City State University Jan-2008 John McGee – OSG Engagement Manager Manager, Cyberinfrastructure.
Large Scale Sky Computing Applications with Nimbus Pierre Riteau Université de Rennes 1, IRISA INRIA Rennes – Bretagne Atlantique Rennes, France
11 IT Expo West 2010 General Session: Cloud Computing Nigel Williams SVP Sales.
Massachusetts Green High Performance Computing Center March 25, 2014.
14 Aug 08DOE Review John Huth ATLAS Computing at Harvard John Huth.
GRID Overview Internet2 Member Meeting Spring 2003 Sandra Redman Information Technology and Systems Center and Information Technology Research Center National.
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
August 3, March, The AC3 GRID An investment in the future of Atlantic Canadian R&D Infrastructure Dr. Virendra C. Bhavsar UNB, Fredericton.
Cyberinfrastructure: An investment worth making Joe Breen University of Utah Center for High Performance Computing.
PaaSport Introduction on Cloud Computing PaaSport training material.
Securing the Grid & other Middleware Challenges Ian Foster Mathematics and Computer Science Division Argonne National Laboratory and Department of Computer.
Optimize the Business with Microsoft Datacenter Services 2.0
Cloud Computing. new buzz word driven largely by marketing and service offerings Provided by big corporate players Google IBM Amazon Apple.
1 TCS Confidential. 2 Objective : In this session we will be able to learn:  What is Cloud Computing?  Characteristics  Cloud Flavors  Cloud Deployment.
1 Open Science Grid: Project Statement & Vision Transform compute and data intensive science through a cross- domain self-managed national distributed.
© 2012 Eucalyptus Systems, Inc. Cloud Computing Introduction Eucalyptus Education Services 2.
Campus Cyberinfrastructure (CC*) Program NSF Larry Gallery Manager, Membership Development and K-20 Program Manager June 16,
EGI-InSPIRE EGI-InSPIRE RI The European Grid Infrastructure Steven Newhouse Director, EGI.eu Project Director, EGI-InSPIRE 29/06/2016CoreGrid.
Scientific Computing at Fermilab Lothar Bauerdick, Deputy Head Scientific Computing Division 1 of 7 10k slot tape robots.
Grid and Cloud Computing
Organizations Are Embracing New Opportunities
XSEDE Value Added and Financial Economies
A Path to the Community Cloud Making Above Campuses Services a Reality
Jan Odegard, Ken Kennedy Institute for Information Technology, Rice
By: Raza Usmani SaaS, PaaS & TaaS By: Raza Usmani
What is HPC? High Performance Computing (HPC)
Deploying Regional Grids Creates Interaction, Ideas, and Integration
RDA US Science workshop Arlington VA, Aug 2014 Cees de Laat with many slides from Ed Seidel/Rob Pennington.
+ John + PIs.
Computing models, facilities, distributed computing
Status and Challenges: January 2017
NGIs – Turkish Case : TR-Grid
An Introduction to Cloud Computing
Bridges and Clouds Sergiu Sanielevici, PSC Director of User Support for Scientific Applications October 12, 2017 © 2017 Pittsburgh Supercomputing Center.
Recap: introduction to e-science
Campus Cyberinfrastructure
EGI-Engage Engaging the EGI Community towards an Open Science Commons
Cognitus: A Science Case for HPC in the Nordic Region
University of Technology
Introduction to Cloud Computing
GGF15 – Grids and Network Virtualization
Management of Virtual Execution Environments 3 June 2008
NSF : CIF21 DIBBs: Middleware and High Performance Analytics Libraries for Scalable Data Science PI: Geoffrey C. Fox Software: MIDAS HPC-ABDS.
Cloud Computing Dr. Sharad Saxena.
Cloud Computing and Cloud Networking
20409A 7: Installing and Configuring System Center 2012 R2 Virtual Machine Manager Module 7 Installing and Configuring System Center 2012 R2 Virtual.
Introduce yourself Presented by
The Software Solution June XX, 2013 Underwritten by:
The Software Solution June 27, 2013 Underwritten by:
Presentation transcript:

Elastic Cyberinfrastructure for Research Computing Research Computing Services Elastic Cyberinfrastructure for Research Computing Glenn Bresnahan glenn@bu.edu Research Computing Services Boston University

Elastic Cyberinfrastructure Research Computing Services Elastic Cyberinfrastructure Framework for building sharable computational infrastructure for research computing Uses the MOC developed Hardware Isolation Layer (HIL) as a core technology Preserve the best aspects of traditional HPC/HTC computing cluster while providing the elasticity, innovation and market-driven value of the cloud

Motivation Experiences with computing environments at the MGHPCC ⃰ Research Computing Services Motivation Experiences with computing environments at the MGHPCC ⃰ Experiences with building and supporting large shared HPC computing environments Initial experiences in deploying HIL resources pools to extend and expand HPC computing environments ⃰ Massachusetts Green High Performance Computing Center

MGHPCC Computing Environment Research Computing Services MGHPCC Computing Environment MGHPCC provides shared physical infrastructure and WAN connectivity Space, pipe, power, and ping A small number of independent HPC/HTC cluster per member (BU, Harvard, MIT, NEU, UMass) Evolving shared systems: Northeast ATLAS Tier 2 (NET2) Engaging 1 funded by NSF C3DDB funded by MLSC ⃰ Northeast Storage Exchange (NESE) Mass Open Cloud ⃰ Massachusetts Life Science Center

MGHPCC Computing Environment Research Computing Services MGHPCC Computing Environment MGHPCC provides shared physical infrastructure and WAN connectivity Space, pipe, power, and ping A small number of independent HPC/HTC cluster per member (BU, Harvard, MIT, NEU, UMass) Evolving shared systems: Northeast ATLAS Tier 2 (NET2) – Traditional HPC/HTC Engaging 1 funded by NSF– Mostly Traditional HPC C3DDB funded by MLSC ⃰ – Mostly Traditional HPC Northeast Storage Exchange (NESE) – Cloud (storage) Mass Open Cloud – Cloud ⃰ Massachusetts Life Science Center

Traditional HPC clusters (BU example) Research Computing Services Traditional HPC clusters (BU example) Heterogeneous hardware: CPU, GPU, fast fabric, parallel file system, back-up/archive ~ 8000 CPU, 100K GPU, Infiniband, 3 PB GPFS, backup to campus Serve a broad base of diverse researchers university-wide 2,400 user; 524 research projects, 70 departments Extensive support for researchers Training, coding, porting, debugging, application support Multi-pronged financial model Central university support for shared core (compute, storage, infrastructure, services) Proportional contributions by campus and/or college Condo-style environment for direct researcher investment (Buy-in) Funding agency support for large-scale, multi-user enhancements Buy-n represents ~60% of SCC resources; >$2M investment

Research Computing Challenges Research Computing Services Research Computing Challenges Most research groups do not have the resources/desire to manage their own computing environment Most researchers are not from traditional computational disciplines Most research groups are small Computing in multiple environments is hard Differences in software stacks, jobs schedulers, policies Managing multiple copies of data is challenging IRB and compliance issues Support is critical Majority of IT/RCS staff effort is user support Supporting individualized computing environments is not sustainable Financial models are important Not all dollars are equal (CapEx vs OpEx) PIs represent 74 different departments and centers; traditional computational depts.: Physics, Chemistry; Astro, Engineering, Geosciences 52% of research groups have 3 or fewer users; 70 % have 5 or less

Distribution of individuals/project Research Computing Services Distribution of individuals/project

Elastic Cyberinfrastructure for Regional Research Computing Research Computing Services Elastic Cyberinfrastructure for Regional Research Computing Multiple resource pools Private HPC clusters, shared HPC clusters, MOC HIL pools, MOC IaaS pools Storage pools (e.g. NESE) Robust data center fabric Compute resources shift from HIL pools to HPC or IaaS clusters on demand Shared object storage Common data management, discovery, publishing (e.g. Dataverse)

Research Computing Services ECI Features Computing environments, including traditional Linux clusters, are elastic, expanding and contracting on demand New bare-metal or virtualized computing environments and services can be created dynamically from scratch or by stitching together existing resources and services Shared storage allows researchers to create cross-environment workflows with analysis performed on data in situ Data management, sharing, discovery, provenance and dissemination are provided as common services Financial models acknowledge practical need for multiple ownership and researcher investment. Affords development of more sophisticated economic models.

Status Current prototypes with ATLAS, NEU HIL pool, Engaging 1, C3DDB Research Computing Services Status Current prototypes with ATLAS, NEU HIL pool, Engaging 1, C3DDB NESE object storage project ramping up Funding request to enhance MGHPCC networking Funding opportunities to expand deployment