Future Grid Early Projects 2010 User Advisory Board Meeting Pittsburgh, PA.

Slides:



Advertisements
Similar presentations
Virtualization, Cloud Computing, and TeraGrid Kate Keahey (University of Chicago, ANL) Marlon Pierce (Indiana University)
Advertisements

Sponsors and Acknowledgments This work is supported in part by the National Science Foundation under Grants No. OCI , IIP and CNS
Science Clouds: Early Experiences in Cloud Computing for Scientific Applications Chicago, October 2008 Kate Keahey, Renato Figueiredo, Jose Fortes, Tim.
Nimbus or an Open Source Cloud Platform or the Best Open Source EC2 No Money Can Buy ;-) Kate Keahey Tim Freeman University of Chicago.
Advanced Computing and Information Systems laboratory Virtual Appliances and Education using Clouds Dr. Renato Figueiredo ACIS Lab - University of Florida.
Education and training on FutureGrig Salt Lake City, Utah July 18 th 2011 Presented by Renato Figueiredo
Advanced Computing and Information Systems laboratory Virtual Private Clusters: Virtual Appliances and Networks in the Cloud Renato Figueiredo ACIS Lab.
SALSA HPC Group School of Informatics and Computing Indiana University.
FutureGrid Overview NSF PI Science of Cloud Workshop Washington DC March Geoffrey Fox
Future Grid Introduction March MAGIC Meeting Gregor von Laszewski Community Grids Laboratory, Digital Science.
FutureGrid Training, Education and Outreach Bloomington Indiana January Presented by Renato Figueiredo
Advanced Computing and Information Systems laboratory Educational Virtual Clusters for On- demand MPI/Hadoop/Condor in FutureGrid Renato Figueiredo Panoat.
Authors: Thilina Gunarathne, Tak-Lon Wu, Judy Qiu, Geoffrey Fox Publish: HPDC'10, June 20–25, 2010, Chicago, Illinois, USA ACM Speaker: Jia Bao Lin.
MapReduce in the Clouds for Science CloudCom 2010 Nov 30 – Dec 3, 2010 Thilina Gunarathne, Tak-Lon Wu, Judy Qiu, Geoffrey Fox {tgunarat, taklwu,
SALSASALSASALSASALSA Digital Science Center June 25, 2010, IIT Geoffrey Fox Judy Qiu School.
FutureGrid Summary TG’10 Pittsburgh BOF on New Compute Systems in the TeraGrid Pipeline August Geoffrey Fox
Panel Session The Challenges at the Interface of Life Sciences and Cyberinfrastructure and how should we tackle them? Chris Johnson, Geoffrey Fox, Shantenu.
Ocean Observatories Initiative Common Execution Environment Kate Keahey OOI Cyberinfrastructure Life Cycle Objectives Milestone Review, Release 1 San Diego,
Cyberaide Virtual Appliance: On-demand Deploying Middleware for Cyberinfrastructure Tobias Kurze, Lizhe Wang, Gregor von Laszewski, Jie Tao, Marcel Kunze,
FutureGrid: an experimental, high-performance grid testbed Craig Stewart Executive Director, Pervasive Technology Institute Indiana University
FutureGrid Overview CTS Conference 2011 Philadelphia May Geoffrey Fox
Purdue RP Highlights TeraGrid Round Table September 23, 2010 Carol Song Purdue TeraGrid RP PI Rosen Center for Advanced Computing Purdue University.
Advanced Computing and Information Systems laboratory Plug-and-play Virtual Appliance Clusters Running Hadoop Dr. Renato Figueiredo ACIS Lab - University.
FutureGrid SOIC Lightning Talk February Geoffrey Fox
Science of Cloud Computing Panel Cloud2011 Washington DC July Geoffrey Fox
MapReduce TG11 BOF FutureGrid Team (Geoffrey Fox) TG11 19 July 2011 Downtown Marriott Salt Lake City.
Experimenting with FutureGrid CloudCom 2010 Conference Indianapolis December Geoffrey Fox
SALSASALSA Twister: A Runtime for Iterative MapReduce Jaliya Ekanayake Community Grids Laboratory, Digital Science Center Pervasive Technology Institute.
Science Clouds and FutureGrid’s Perspective June Science Clouds Workshop HPDC 2012 Delft Geoffrey Fox
Biomedical Cloud Computing iDASH Symposium San Diego CA May Geoffrey Fox
Future Grid FutureGrid Overview Dr. Speaker. Future Grid Future GridFutureGridFutureGrid The goal of FutureGrid is to support the research on the future.
Topaz : A GridFTP extension to Firefox M. Taufer, R. Zamudio, D. Catarino, K. Bhatia, B. Stearn University of Texas at El Paso San Diego Supercomputer.
FutureGrid Dynamic Provisioning Experiments including Hadoop Fugang Wang, Archit Kulshrestha, Gregory G. Pike, Gregor von Laszewski, Geoffrey C. Fox.
Advanced Computing and Information Systems laboratory Virtual Appliances for Training and Education in FutureGrid Renato Figueiredo Arjun Prakash, David.
Large Scale Sky Computing Applications with Nimbus Pierre Riteau Université de Rennes 1, IRISA INRIA Rennes – Bretagne Atlantique Rennes, France
Future Grid FutureGrid Overview Geoffrey Fox SC09 November
Simplifying Resource Sharing in Voluntary Grid Computing with the Grid Appliance David Wolinsky Renato Figueiredo ACIS Lab University of Florida.
FutureGrid SC10 New Orleans LA IU Booth November Geoffrey Fox
FutureGrid Connection to Comet Testbed and On Ramp as a Service Geoffrey Fox Indiana University Infra structure.
Parallel Applications And Tools For Cloud Computing Environments Azure MapReduce Large-scale PageRank with Twister Twister BLAST Thilina Gunarathne, Stephen.
SALSA HPC Group School of Informatics and Computing Indiana University.
FutureGrid SOIC Lightning Talk February Geoffrey Fox
RAIN: A system to Dynamically Generate & Provision Images on Bare Metal by Application Users Presented by Gregor von Laszewski Authors: Javier Diaz, Gregor.
SALSASALSASALSASALSA FutureGrid Venus-C June Geoffrey Fox
Research in Grids and Clouds and FutureGrid Melbourne University September Geoffrey Fox
Hosting Cloud, HPC and Grid Educational Activities on FutureGrid Renato Figueiredo – U. of Florida Geoffrey Fox, Barbara Ann O’Leary – Indiana University.
FutureGrid NSF September Geoffrey Fox
Virtual Appliances CTS Conference 2011 Philadelphia May Geoffrey Fox
Design Discussion Rain: Dynamically Provisioning Clouds within FutureGrid PI: Geoffrey Fox*, CoPIs: Kate Keahey +, Warren Smith -, Jose Fortes #, Andrew.
Computing Research Testbeds as a Service: Supporting large scale Experiments and Testing SC12 Birds of a Feather November.
Cloud Computing Paradigms for Pleasingly Parallel Biomedical Applications Thilina Gunarathne, Tak-Lon Wu Judy Qiu, Geoffrey Fox School of Informatics,
Grid Appliance The World of Virtual Resource Sharing Group # 14 Dhairya Gala Priyank Shah.
SALSASALSASALSASALSA Digital Science Center February 12, 2010, Bloomington Geoffrey Fox Judy Qiu
Memcached Integration with Twister Saliya Ekanayake - Jerome Mitchell - Yiming Sun -
What’s Hot in Clouds? Analyze (superficially) the ~140 Papers/Short papers/Workshops/Posters/Demos in CloudCom Each paper may fall in more than one category.
Purdue RP Highlights TeraGrid Round Table May 20, 2010 Preston Smith Manager - HPC Grid Systems Rosen Center for Advanced Computing Purdue University.
SALSASALSA Dynamic Virtual Cluster provisioning via XCAT on iDataPlex Supports both stateful and stateless OS images iDataplex Bare-metal Nodes Linux Bare-
Directions in eScience Interoperability and Science Clouds June Interoperability in Action – Standards Implementation.
Nimbus Update March 2010 OSG All Hands Meeting Kate Keahey Nimbus Project University of Chicago Argonne National Laboratory.
Building on virtualization capabilities for ExTENCI Carol Song and Preston Smith Rosen Center for Advanced Computing Purdue University ExTENCI Kickoff.
Private Public FG Network NID: Network Impairment Device
StratusLab Final Periodic Review
StratusLab Final Periodic Review
Cloud Computing with Nimbus
FutureGrid: a Grid Testbed
Sky Computing on FutureGrid and Grid’5000
Virtualization, Cloud Computing, and TeraGrid
Clouds from FutureGrid’s Perspective
Using and Building Infrastructure Clouds for Science
Sky Computing on FutureGrid and Grid’5000
Presentation transcript:

Future Grid Early Projects 2010 User Advisory Board Meeting Pittsburgh, PA

Note Some of the work described was performed during the acceptance tests and before the resources were available to anybody but the individual resource owners. Access to resources is only just coming broadly available

Project descriptions University of Florida – ViNe, CloudBLAST, and virtual appliance University of Virginia – Genesis II, UNICORE 6, g-lite Indiana University – SALSA Group LSU (Shantenu)

UF – FG activities ViNe overlay network and CloudBLAST (both developed at UF) has been deployed and experimented on FG resources ViNe and CloudBLAST are encapsulated in VM images and deployed through Nimbus Large-scale experiments and demos: – CCGrid’2010: 150-node virtual cluster across 5 sites, connected through ViNe (including 2 FG sites UF and SDSC) and running CloudBLAST, demonstrated a large scale BLAST run in less than 1 hour (the same job takes over 17 days sequentially) – OGF’10: 457-node virtual cluster across 3 FG sites (UF, UC, SDSC) and 2 G5k sites (Rennes, Sophia) demonstrated the dynamic extension of a Hadoop cluster. A modified Nimbus, with faster VM deployment, was deployed on G5k. As new VMs became available on G5k, the Hadoop worker pool was increased speeding up the CloudBLAST throughput. – Researchers: Andréa Matsunaga, José Fortes (UF), Kate Keahey (UC/ANL), Pierre Riteau (UF/U of Rennes, France), Maurício Tsugawa (UF)

OGF’10 Demo SDSC UF UC Lille Rennes Sophia ViNe provided the necessary inter-cloud connectivity to deploy CloudBLAST across 5 Nimbus sites, with a mix of public and private subnets. Grid’5000 firewall

UF – educational virtual appliance Created educational virtual appliance images and tutorials to get started on: – Deploying a virtual cluster FutureGrid (Xen) FutureGrid (Xen) User’s own desktop/cluster resources (VMware, VirtualBox) – Pre-packaged middleware, zero-configuration scripts MPI, Hadoop, Condor Can be seamlessly deployed across domains and hypervisors Desktop – can also serve as FutureGrid cloud client appliance to facilitate new users to access the infrastructure GroupVPN for simple configuration of virtual private clusters for class environments – Web interface to create/manage user groups – Download GroupVPN credentials, deploy as input data at cloud provisioning Dr. Renato J. Figueiredo; Arjun Prakash, David Wolinsky (U. Florida)

University of Virginia Goals/Tasks Goals – Grid Standards – Grid Interoperability – Grid Applications – Grid Middleware testbed Tasks – Acquire and learn Unicore and EGEE (both are open source and free). – Install on local (UVA) nodes. – Deploy Unicore and EGEE onto FutureGrid nodes as required. – Deploy and maintain Genesis II on FutureGrid nodes. – Bring up and maintain standard service endpoints for compliance testing.

Accomplishments Guinea pig/ friendly user Genesis II endpoint operational on XRAY and integrated into XCG (Cross Campus Grid) UNICORE 6 endpoint up on XRAY

Next Integrate UNICORE 6 “gridmap” file with FutureGrid CA Genesis II/UNICORE 6 onto Linux clusters via Torque Genesis II VM onto Eucalyptus, test/experiment with dynamic provisioning G-lite

SALSASALSASALSASALSA University of Arkansas Indiana University University of California at Los Angeles Penn State Iowa State Univ.Illinois at Chicago University of Minnesota Michigan State Notre Dame University of Texas at El Paso IBM Almaden Research Center Washington University San Diego Supercomputer Center University of Florida Johns Hopkins July 26-30, 2010 NCSA Summer School Workshop Students learning about Twister & Hadoop MapReduce technologies, supported by FutureGrid.

SALSASALSA Pairwise Sequence Comparison using Smith Waterman Gotoh Typical MapReduce computation Comparable efficiencies Twister performs the best Xiaohong Qiu, Jaliya Ekanayake, Scott Beason, Thilina Gunarathne, Geoffrey Fox, Roger Barga, Dennis Gannon “Cloud Technologies for Bioinformatics Applications”, Proceedings of the 2nd ACM Workshop on Many-Task Computing on Grids and Supercomputers (SC09), Portland, Oregon, November 16th, 2009 “Cloud Technologies for Bioinformatics ApplicationsSC09 Jaliya Ekanayake, Thilina Gunarathne, Xiaohong Qiu, “Cloud Technologies for Bioinformatics Applications”, invited paper submitted to the Journal of IEEE Transactions on Parallel and Distributed Systems (under review).“Cloud Technologies for Bioinformatics Applications

SALSASALSA Sequence Assembly in the Clouds Cap3 parallel efficiencyCap3 – Per core per file (458 reads in each file) time to process sequences Input files (FASTA) Output files CAP3 CAP3 - Expressed Sequence Tagging Thilina Gunarathne, Tak-Lon Wu, Judy Qiu, and Geoffrey Fox, “Cloud Computing Paradigms for Pleasingly Parallel Biomedical Applications”, March 21, Proceedings of Emerging Computational Methods for the Life Sciences Workshop of ACM HPDC 2010 conference, Chicago, Illinois, June 20-25, 2010.Cloud Computing Paradigms for Pleasingly Parallel Biomedical Applications

SAGA Andre Luckow, Lukasz Lacinski, Shantenu Jha, SAGA BigJob: An Extensible and Interoperable Pilot-Job Abstraction for Distributed Applications and Systems, Center for Computation & Technology, Louisiana State University; Department of Computer Science, Louisiana State University; e-Science Institute, Edinburgh, UK Used Nimbus on FutureGrid

FutureGrid Activities at U Chicago Built the UC resource (hotel) and integrated it into the FutureGrid infrastructure + Nimbus installation Nimbus releases containing FG-driven features: – Zero -> cloud installation process and user management tools (Nimbus 2.5, 07/30) – Nimbus installer (Nimbus 2.4, 05/05) – Web interface for credential distribution (Nimbus 2.3, 02/02) Supported demonstrations, exploration, and early users of FG (see following slides) Prepared documentation and tutorials for FG users UC team: Kate Keahey, David LaBissoniere, Tim Freeman, Ti Leggett 14

Sky Computing on FutureGrid Combine resources obtained in multiple clouds Integration of resources from Grid5000 and FutureGrid Deployed a virtual cluster of over 1000 cores on Grid5000 and FutureGrid and used it for bioinformatics comps Demonstrated at OGF 29 Poster to be presented at TeraGrid ‘10 15 SD UF UC Lille Rennes Sophia Queue ViNe Router Grid’5000 firewall Pierre Riteau M. Tsugawa, A. Matsunaga, J. Fortes, T. Freeman, D. LaBissoniere, K. Keahey

Cumulus: a Storage Cloud for Science S3-compatible open source storage cloud Component of Nimbus Quota support for scientific users Pluggable back-end to popular technologies such as POSIX, HDFS, Sector and BlobSeer Initial performance studies FutureGrid at U Chicago Poster submitted to SC10 16 John Bresnahan, University of Chicago

Clouds and HTC: a Match? Using HTC as “backfill” for on-demand clouds Extension to Nimbus Significant utilization improvement leading to lowering cost for cloud computing Initial results on U Chicago Preparing for running of production workloads on U Chicago Poster submitted to SC10 17 Paul Marshall, University of Colorado

Cloud Bursting HA service that provisions resources on clouds in response to need Serving the needs of Ocean Observatory Initiative (OOI) Initiated scalability and reliability tests on FutureGrid and Magellan resources Work in progress 18 Paul Marshall and the Nimbus team