FutureGrid Design and Implementation of a National Grid Test-Bed David Hancock – HPC Manager - Indiana University Hardware & Network.

Slides:



Advertisements
Similar presentations
FutureGrid related presentations at TG and OGF Sun. 17th: Introduction to FutireGrid (OGF) Mon. 18th: Introducing to FutureGrid (TG) Tue. 19th –Educational.
Advertisements

FutureGrid and US Cyberinfrastructure Collaboration with EU Symposium on transatlantic EU-U.S. cooperation in the field of large scale research infrastructures.
Kathy Benninger, Pittsburgh Supercomputing Center Workshop on the Development of a Next-Generation Cyberinfrastructure 1-Oct-2014 NSF Collaborative Research:
FutureGrid Overview NSF PI Science of Cloud Workshop Washington DC March Geoffrey Fox
Cosmic Issues and Analysis of External Comments on FutureGrid TG11 Salt Lake City July Geoffrey Fox
Clouds from FutureGrid’s Perspective April Geoffrey Fox Director, Digital Science Center, Pervasive.
Future Grid Introduction March MAGIC Meeting Gregor von Laszewski Community Grids Laboratory, Digital Science.
Student Visits August Geoffrey Fox
Overview Presented at OGF31 Salt Lake City, July 2011 Geoffrey Fox, Gregor von Laszewski, Renato Figueiredo Contact:
SALSASALSASALSASALSA Digital Science Center June 25, 2010, IIT Geoffrey Fox Judy Qiu School.
Design Discussion Rain: Dynamically Provisioning Clouds within FutureGrid Geoffrey Fox, Andrew J. Younge, Gregor von Laszewski, Archit Kulshrestha, Fugang.
FutureGrid Summary TG’10 Pittsburgh BOF on New Compute Systems in the TeraGrid Pipeline August Geoffrey Fox
FutureGrid Overview Bloomington Indiana January FutureGrid Collaboration Presented by Geoffrey Fox
SC2010 Gregor von Laszewski (*) (*) Assistant Director of Cloud Computing, CGL, Pervasive Technology Institute.
FutureGrid Summary FutureGrid User Advisory Board TG’10 Pittsburgh August Geoffrey Fox
FutureGrid Overview David Hancock HPC Manger Indiana University.
FutureGrid: an experimental, high-performance grid testbed Craig Stewart Executive Director, Pervasive Technology Institute Indiana University
FutureGrid: an experimental, high-performance grid testbed Craig Stewart Executive Director, Pervasive Technology Institute Indiana University
Clouds and FutureGrid MSI-CIEC All Hands Meeting SDSC January Geoffrey Fox
FutureGrid Overview CTS Conference 2011 Philadelphia May Geoffrey Fox
Raining Compute Environments on Resources by Application Users Gregor von Laszewski Indiana University Open Cirrus Summit 2011, Oct.
FutureGrid SOIC Lightning Talk February Geoffrey Fox
Science of Cloud Computing Panel Cloud2011 Washington DC July Geoffrey Fox
FutureGrid and US Cyberinfrastructure Collaboration with EU Symposium on transatlantic EU-U.S. cooperation in the field of large scale research infrastructures.
Experimenting with FutureGrid CloudCom 2010 Conference Indianapolis December Geoffrey Fox
Science Clouds and FutureGrid’s Perspective June Science Clouds Workshop HPDC 2012 Delft Geoffrey Fox
Gregor von Laszewski*, Geoffrey C. Fox, Fugang Wang, Andrew Younge, Archit Kulshrestha, Greg Pike (IU), Warren Smith, (TACC) Jens Vöckler (ISI), Renato.
FutureGrid Overview Geoffrey Fox
FutureGrid: an experimental, high-performance grid testbed Craig Stewart Executive Director, Pervasive Technology Institute Indiana University
FutureGrid TeraGrid Science Advisory Board San Diego CA July Geoffrey Fox
FutureGrid 100 and 101 (part one) Virtual School for Computational Science and Engineering July Geoffrey Fox
Future Grid FutureGrid Overview Dr. Speaker. Future Grid Future GridFutureGridFutureGrid The goal of FutureGrid is to support the research on the future.
FutureGrid Overview Geoffrey Fox
FutureGrid: an experimental, high-performance grid testbed Craig Stewart Executive Director, Pervasive Technology Institute Indiana University
Large Scale Sky Computing Applications with Nimbus Pierre Riteau Université de Rennes 1, IRISA INRIA Rennes – Bretagne Atlantique Rennes, France
What’s Hot in Clouds? Analyze (superficially) the ~140 Papers/Short papers/Workshops/Posters/Demos in CloudCom Each paper may fall in more than one category.
Future Grid FutureGrid Overview Geoffrey Fox SC09 November
FutureGrid Overview Geoffrey Fox
FutureGrid SC10 New Orleans LA IU Booth November Geoffrey Fox
FutureGrid Connection to Comet Testbed and On Ramp as a Service Geoffrey Fox Indiana University Infra structure.
FutureGrid Overview Geoffrey Fox
Future Grid Future Grid All Hands Meeting Introduction Indianapolis October Geoffrey Fox
FutureGrid SOIC Lightning Talk February Geoffrey Fox
FutureGrid Cyberinfrastructure for Computational Research.
Building Effective CyberGIS: FutureGrid Marlon Pierce, Geoffrey Fox Indiana University.
RAIN: A system to Dynamically Generate & Provision Images on Bare Metal by Application Users Presented by Gregor von Laszewski Authors: Javier Diaz, Gregor.
SALSASALSASALSASALSA FutureGrid Venus-C June Geoffrey Fox
Research in Grids and Clouds and FutureGrid Melbourne University September Geoffrey Fox
FutureGrid TeraGrid Science Advisory Board San Diego CA July Geoffrey Fox
FutureGrid Overview Geoffrey Fox
GRIDS Center Middleware Overview Sandra Redman Information Technology and Systems Center and Information Technology Research Center National Space Science.
Bio Gregor von Laszewski is conducting state-of-the-art work in Cloud computing and GreenIT at Indiana University as part of the Future Grid project. During.
Tutorial Presented at TG2011 Geoffrey Fox, Gregor von Laszewski, Renato Figueiredo, Kate Keahey, Andrew Younge Contact:
Ruth Pordes November 2004TeraGrid GIG Site Review1 TeraGrid and Open Science Grid Ruth Pordes, Fermilab representing the Open Science.
FutureGrid BOF Overview TG 11 Salt Lake City July Geoffrey Fox
FutureGrid NSF September Geoffrey Fox
Design Discussion Rain: Dynamically Provisioning Clouds within FutureGrid PI: Geoffrey Fox*, CoPIs: Kate Keahey +, Warren Smith -, Jose Fortes #, Andrew.
1 NSF/TeraGrid Science Advisory Board Meeting July 19-20, San Diego, CA Brief TeraGrid Overview and Expectations of Science Advisory Board John Towns TeraGrid.
Computing Research Testbeds as a Service: Supporting large scale Experiments and Testing SC12 Birds of a Feather November.
Future Grid Future Grid Overview. Future Grid Future GridFutureGridFutureGrid The goal of FutureGrid is to support the research that will invent the future.
SALSASALSASALSASALSA Digital Science Center February 12, 2010, Bloomington Geoffrey Fox Judy Qiu
NICS Update Bruce Loftis 16 December National Institute for Computational Sciences University of Tennessee and ORNL partnership  NICS is the 2.
Building on virtualization capabilities for ExTENCI Carol Song and Preston Smith Rosen Center for Advanced Computing Purdue University ExTENCI Kickoff.
Lizhe Wang, Gregor von Laszewski, Jai Dayal, Thomas R. Furlani
Private Public FG Network NID: Network Impairment Device
FutureGrid: a Grid Testbed
Sky Computing on FutureGrid and Grid’5000
FutureGrid and Applications
Using and Building Infrastructure Clouds for Science
Sky Computing on FutureGrid and Grid’5000
Presentation transcript:

FutureGrid Design and Implementation of a National Grid Test-Bed David Hancock – HPC Manager - Indiana University Hardware & Network Lead - FutureGrid

October 12, 2015Research Technologies IU in a nutshell $1.7B Annual Budget, >$100M annual IT budget Recent credit upgrade to AAA One university with – 8 campuses – 107,000 students – 3,900 faculty Nation’s 2nd largest school of medicine Serious HPC since 1990’s Research staff increased from since /50 Split in base and grant funding Large scale projects: TeraGrid, Open Science Grid (ATLAS Tier2 center), PolarGrid, Data Capacitor New Data Center opened in 2009

NSF Track Overview Track 1 – NCSA Blue Waters Track 2a – TACC Ranger Track2b – NICS Kraken Track 2d – Data Intensive High Performance System (SDSC) – Experimental High Performance System (GaTech) – Experimental High Performance Test-Bed (IU)

FutureGridFutureGrid The goal of FutureGrid is to support the research on the future of distributed, grid, and cloud computing. FutureGrid will build a robustly managed simulation environment and test-bed to support the development and early use in science of new technologies at all levels of the software stack: from networking to middleware to scientific applications. The environment will mimic TeraGrid and/or general parallel and distributed systems – FutureGrid is part of TeraGrid and one of two experimental TeraGrid systems (other is GPU) This test-bed will succeed if it enables major advances in science and engineering through collaborative development of science applications and related software. FutureGrid is a (small 5400 core) Science/Computer Science Cloud but it is more accurately a virtual machine based simulation environment

FutureGrid Partners Indiana University (Architecture, core software, Support) Purdue University (HTC Hardware) San Diego Supercomputer Center at University of California San Diego (INCA, Performance Monitoring) University of Chicago/Argonne National Labs (Nimbus) University of Florida (ViNe, Education and Outreach) University of Southern California Information Sciences Institute (Pegasus to manage experiments) University of Tennessee Knoxville (Benchmarking) University of Texas at Austin/Texas Advanced Computing Center (Portal) University of Virginia (OGF, User Advisory Board) Center for Information Services and GWT-TUD from Technische Universtität Dresden. (VAMPIR) Blue institutions host FutureGrid hardware

Other Important Collaborators Other Important Collaborators Early users from an application and computer science perspective and from both research and education Grid5000 and D-Grid in Europe Commercial partners such as – Eucalyptus …. – Microsoft (Dryad + Azure) – Application partners NSF TeraGrid – Tutorial at TG10 Open Grid Forum – Possible BoF Possibly Open Nebula, Open Cirrus Testbed, Open Cloud Consortium, Cloud Computing Interoperability Forum. IBM-Google-NSF Cloud, and other DoE/NSF/… clouds

FutureGrid Timeline October 2009 – Project Starts November 2009 – SC09 Demo January 2010 – Significant Hardware installed April 2010 – First Early Users May 2010 – FutureGrid network complete August 2010 – FutureGrid Annual Meeting September 2010 – All hardware, except shared memory system, available October 2011 – FutureGrid allocatable via TeraGrid process – first two years by user/science board

FutureGrid Usage Scenarios Developers of end-user applications who want to create new applications in cloud or grid environments, including analogs of commercial cloud environments such as Amazon or Google. – Is a Science Cloud for me? Is my application secure? Developers of end-user applications who want to experiment with multiple hardware environments. Grid/Cloud middleware developers who want to evaluate new versions of middleware or new systems. Networking researchers who want to test and compare different networking solutions in support of grid and cloud applications and middleware. Education as well as research Interest in performance testing requires that bare metal images are important

FutureGrid Hardware

Compute Hardware System type# CPUs# CoresTFLOPS Total RAM (GB) Secondary Storage (TB) Site Status Dynamically configurable systems IBM iDataPlex *IU New System Dell PowerEdge TACC New System IBM iDataPlex UC New System IBM iDataPlex SDSC Existing System Subtotal Systems not dynamically configurable Cray XT5m *IU New System Shared memory system TBD *IU New System 4Q2010 IBM iDataPlex UF New System High Throughput Cluster PU Existing System Subtotal Total

Storage Hardware System TypeCapacity (TB)File SystemSiteStatus DDN 9550 (Data Capacitor) 339LustreIUExisting System DDN GPFSUCNew System SunFire x417096ZFSSDSCNew System Dell MD300030NFSTACCNew System FutureGrid has a dedicated network (except to TACC) and a network fault and delay generator Experiments can be isolated by request Additional partner machines may run FutureGrid software and be supported (but allocated in specialized ways)

System Milestones New Cray System (xray) – Delivery: January 2010 – Acceptance: February 2010 – Available for Use: April 2010 New IBM Systems (india) – Delivery: January 2010 – Acceptance: March 2010 – Available for Use: May 2010 Dell System (tango) – Delivery: April 2010 – Acceptance: June 2010 – Available for Use: July 2010 Existing IU iDataPlex (sierra) – Move to SDSC: January 2010 – Available for Use: April 2010 Storage Systems (Sun & DDN) – Delivery: December 2009 – Acceptance: January 2010

Logical Diagram

Network Impairments Device Spirent XGEM Network Impairments Simulator for jitter, errors, delay, etc Full Bidirectional 10G w/64 byte packets up to 15 seconds introduced delay (in 16ns increments) 0-100% introduced packet loss in.0001% increments Packet manipulation in first 2000 bytes up to 16k frame size TCL for scripting, HTML for manual configuration

Network Milestones December 2009 – Setup and configuration of core equipment at IU – Juniper EX 8208 – Spirent XGEM January 2010 – Core equipment relocated to Chicago – IP addressing & AS # February 2010 – Coordination with local networks – First Circuits to Chicago Active March 2010 – Peering with TeraGrid & Internet2 April 2010 – NLR Circuit to UFL (via FLR) Active May 2010 – NLR Circuit to SDSC (via CENIC) Active

Global NOC Background ~65 total staff Service Desk: proactive & reactive monitoring 24x7x365, coordination of support Engineering: All operational troubleshooting Planning/Senior Engineering: Senior Network engineers dedicated to single projects Tool Developers: Developers of GlobalNOC tool suite

17 Supported Projects OmniPoP REN- ISAC

FutureGrid Architecture Open Architecture allows to configure resources based on images Managed images allows to create similar experiment environments Experiment management allows reproducible activities Through our modular design we allow different clouds and images to be “rained” upon hardware. Will support deployment of preconfigured middleware including TeraGrid stack, Condor, BOINC, gLite, Unicore, Genesis II

Software Goals Open-source, integrated suite of software to – instantiate and execute grid and cloud experiments. – perform an experiment – collect the results – tools for instantiating a test environment TORQUE, Moab, xCAT, bcfg, and Pegasus, Inca, ViNE, a number of other tools from our partners and the open source community Portal to interact – Benchmarking 10/12/2015http://futuregrid.org19

Draft GUI for FutureGrid Dynamic Provisioning Draft GUI for FutureGrid Dynamic Provisioning

Command line fg-deploy-image – host name – image name – start time – end time – label name fg-add – label name – framework hadoop – version 1.0 Deploys an image on a host Adds a feature to a deployed image 10/12/2015http://futuregrid.org21

FG Stratosphere Objective – Higher than a particular cloud – Provides all mechanisms to provision a cloud on a given FG hardware – Allows the management of reproducible experiments – Allows monitoring of the environment and the results Risks – Lots of software – Possible multiple path to do the same thing Good news – We worked in a team, know about different solutions and have identified a very good plan – We can componentize Stratosphere 10/12/2015http://futuregrid.org22

Change underlying system to support current user demands Linux, Windows, Xen/KVM, Nimbus, Eucalyptus Stateless images Shorter boot times Easier to maintain Stateful installs Windows Use Moab to trigger changes and xCAT to manage installs 10/12/201523http://futuregrid.org Dynamic Provisioning

xCAT and Moab xCAT uses installation infrastructure to perform installs creates stateless Linux images changes the boot configuration of the nodes remote power control and console Moab meta-schedules over resource managers  TORQUE and Windows HPC control nodes through xCAT  changing the OS 10/12/201524http://futuregrid.org

Experiment Manager Objective – Manage the provisioning for reproducible experiments – Coordinate workflow of experiments – Share workflow and experiment images – Minimize space through reuse Risk – Images are large – Users have different requirements and need different images 10/12/2015http://futuregrid.org25

AcknowledgementsAcknowledgements FutureGrid NSF Award OCI NSF Solicitation – ViNe Nimbus Eucalyptus VAMPIR Pegasus -