Gordon: Using Flash Memory to Build Fast, Power-efficient Clusters for Data-intensive Applications A. Caulfield, L. Grupp, S. Swanson, UCSD, ASPLOS’09.

Slides:



Advertisements
Similar presentations
Hardware Lesson 3 Inside your computer.
Advertisements

A New Cache Management Approach for Transaction Processing on Flash-based Database Da Zhou
Daniel Schall, Volker Höfner, Prof. Dr. Theo Härder TU Kaiserslautern.
Linux Clustering A way to supercomputing. What is Cluster? A group of individual computers bundled together using hardware and software in order to make.
Presented by: Yash Gurung, ICFAI UNIVERSITY.Sikkim BUILDING of 3 R'sCLUSTER PARALLEL COMPUTER.
Novell Server Linux vs. windows server 2008 By: Gabe Miller.
Shimin Chen Big Data Reading Group.  Energy efficiency of: ◦ Single-machine instance of DBMS ◦ Standard server-grade hardware components ◦ A wide spectrum.
SAN DIEGO SUPERCOMPUTER CENTER at the UNIVERSITY OF CALIFORNIA; SAN DIEGO IEEE Symposium of Massive Storage Systems, May 3-5, 2010 Data-Intensive Solutions.
SUMS Storage Requirement 250 TB fixed disk cache 130 TB annual increment for permanently on- line data 100 TB work area (not controlled by SUMS) 2 PB near-line.
FAWN: A Fast Array of Wimpy Nodes Presented by: Aditi Bose & Hyma Chilukuri.
Introduction to Systems Architecture Kieran Mathieson.
Teraserver Darrel Sharpe Matt Todd Rob Neff Mentor: Dr. Palaniappan.
Energy Efficient Prefetching – from models to Implementation 6/19/ Adam Manzanares and Xiao Qin Department of Computer Science and Software Engineering.
Energy Efficient Prefetching with Buffer Disks for Cluster File Systems 6/19/ Adam Manzanares and Xiao Qin Department of Computer Science and Software.
Hadoop tutorials. Todays agenda Hadoop Introduction and Architecture Hadoop Distributed File System MapReduce Spark 2.
Virtual Network Servers. What is a Server? 1. A software application that provides a specific one or more services to other computers  Example: Apache.
* Definition of -RAM (random access memory) :- -RAM is the place in a computer where the operating system, application programs & data in current use.
RAID-x: A New Distributed Disk Array for I/O-Centric Cluster Computing Kai Hwang, Hai Jin, and Roy Ho.
Cluster computing facility for CMS simulation work at NPD-BARC Raman Sehgal.
© 2013 Mellanox Technologies 1 NoSQL DB Benchmarking with high performance Networking solutions WBDB, Xian, July 2013.
Basic Computer Structure and Knowledge Project Work.
The Hadoop Distributed File System
Enabling Technologies for Distributed and Cloud Computing Dr. Sanjay P. Ahuja, Ph.D FIS Distinguished Professor of Computer Science School of.
Overview Introduction The Level of Abstraction Organization & Architecture Structure & Function Why study computer organization?
Types of Computers Mainframe/Server Two Dual-Core Intel ® Xeon ® Processors 5140 Multi user access Large amount of RAM ( 48GB) and Backing Storage Desktop.
Hadoop Hardware Infrastructure considerations ©2013 OpalSoft Big Data.
The Red Storm High Performance Computer March 19, 2008 Sue Kelly Sandia National Laboratories Abstract: Sandia National.
Hardware Trends. Contents Memory Hard Disks Processors Network Accessories Future.
© Copyright 2012 Hewlett-Packard Development Company, L.P. The information contained herein is subject to change without notice. LogKV: Exploiting Key-Value.
Memory  Main memory consists of a number of storage locations, each of which is identified by a unique address  The ability of the CPU to identify each.
Building a Parallel File System Simulator E Molina-Estolano, C Maltzahn, etc. UCSC Lab, UC Santa Cruz. Published in Journal of Physics, 2009.
Inside your computer. Hardware Review Motherboard Processor / CPU Bus Bios chip Memory Hard drive Video Card Sound Card Monitor/printer Ports.
Inside your computer. Hardware Motherboard Processor / CPU Bus Bios chip Memory Hard drive Video Card Sound Card Monitor/printer Ports.
Computer Guts and Operating Systems CSCI 101 Week Two.
10/22/2002Bernd Panzer-Steindel, CERN/IT1 Data Challenges and Fabric Architecture.
DFTL: A flash translation layer employing demand-based selective caching of page-level address mappings A. gupta, Y. Kim, B. Urgaonkar, Penn State ASPLOS.
Fast Crash Recovery in RAMCloud. Motivation The role of DRAM has been increasing – Facebook used 150TB of DRAM For 200TB of disk storage However, there.
Improving Disk Throughput in Data-Intensive Servers Enrique V. Carrera and Ricardo Bianchini Department of Computer Science Rutgers University.
Virtualization and Databases Ashraf Aboulnaga University of Waterloo.
Weekly Report By: Devin Trejo Week of June 21, 2015-> June 28, 2015.
Introduction: Memory Management 2 Ideally programmers want memory that is large fast non volatile Memory hierarchy small amount of fast, expensive memory.
DynamicMR: A Dynamic Slot Allocation Optimization Framework for MapReduce Clusters Nanyang Technological University Shanjiang Tang, Bu-Sung Lee, Bingsheng.
PERFORMANCE STUDY OF BIG DATA ON SMALL NODES. Ομάδα: Παναγιώτης Μιχαηλίδης Αντρέας Σόλου Instructor: Demetris Zeinalipour.
Enabling Technologies for Distributed Computing Dr. Sanjay P. Ahuja, Ph.D. Fidelity National Financial Distinguished Professor of CIS School of Computing,
 Introduction  Architecture NameNode, DataNodes, HDFS Client, CheckpointNode, BackupNode, Snapshots  File I/O Operations and Replica Management File.
The Sort Benchmark AlgorithmsSolid State Disks External Memory Multiway Mergesort  Phase 1: Run Formation  Phase 2: Merge Runs  Careful parameter selection.
POWERING UP THE LAB! BUDGET PROPOSAL AND DETAILS The Free Thinker School Home of the Rams.
PROOF Benchmark on Different Hardware Configurations 1 11/29/2007 Neng Xu, University of Wisconsin-Madison Mengmeng Chen, Annabelle Leung, Bruce Mellado,
 System Requirements are the prerequisites needed in order for a software or any other resources to execute efficiently.  Most software defines two.
Parallel IO for Cluster Computing Tran, Van Hoai.
Computer Performance. Hard Drive - HDD Stores your files, programs, and information. If it gets full, you can’t save any more. Measured in bytes (KB,
COMPUTER SYSTEMS ARCHITECTURE A NETWORKING APPROACH CHAPTER 12 INTRODUCTION THE MEMORY HIERARCHY CS 147 Nathaniel Gilbert 1.
Transactional Flash V. Prabhakaran, T. L. Rodeheffer, L. Zhou (MSR, Silicon Valley), OSDI 2008 Shimin Chen Big Data Reading Group.
Instructor: Chapter 2: The System Unit. Learning Objectives: Recognize how data is processed Understand processors Understand memory types and functions.
Instructor: Syed Shuja Hussain Chapter 2: The System Unit.
The Sort Benchmark AlgorithmsSolid State Disks External Memory Multiway Mergesort  Phase 1: Run Formation  Phase 2: Merge Runs  Careful parameter selection.
PERGAMUM: REPLACING TAPE WITH ENERGY EFFICIENT, RELIABLE, DISK-BASED ARCHIVAL STORAGE M. W. Storer K. M. Greenan E. L. Miller UCSC K. Vorugant Network.
29/04/2008ALICE-FAIR Computing Meeting1 Resulting Figures of Performance Tests on I/O Intensive ALICE Analysis Jobs.
Personal Computer (PC)  Computer advertisement specification Intel® Pentium 4 Processor at 3.06GHz with 512K cache 512MB DDR SDRAM 200GB ATA-100 Hard.
Internal Parallelism of Flash Memory-Based Solid-State Drives
Hardware specifications
Introduction to MapReduce and Hadoop
Introduction to HDFS: Hadoop Distributed File System
Architecture Background
CS-301 Introduction to Computing Lecture 17
Hadoop Clusters Tess Fulkerson.
Myoungjin Kim1, Yun Cui1, Hyeokju Lee1 and Hanku Lee1,2,*
Overview Introduction VPS Understanding VPS Architecture
The Basics of Apache Hadoop
IBM Power Systems.
Presentation transcript:

Gordon: Using Flash Memory to Build Fast, Power-efficient Clusters for Data-intensive Applications A. Caulfield, L. Grupp, S. Swanson, UCSD, ASPLOS’09 Shimin Chen Big Data Reading Group

Introduction Gordon: a flash-based system architecture for massively parallel, data-centric computing.  Solid-state disks  Low-power processors  Data-centric programming paradigms Can deliver:  Up to 2.5X the computation per energy of a conventional cluster based system  Increasing performance by up to 1.5X

Outline Gordon’s system architecture Gordon’s storage system Configuring Gordon Discussion Summary

Gordon’s System Architecture Gordon node 16 nodes in an enclosure

Prototype Photo: ASPLOS’09 paper uses simulation

Gordon node Configuration:  256GB of flash storage  A flash storage controller (w/ 512MB dedicated DRAM)  2GB ECC DDR2 SDRAM  1.9Ghz Intel Atom processor  Running a minimal linux installation Power: no more than 19w  Compared to 81w of a server 900MB/s read and write bandwidth to 256GB disk

Enclosures Within an enclosure, 16 nodes plug into a backplane that provides 1Gb Ethernet-style network A rack holds 16 enclosures (16x16=256 nodes)  64 TB of storage  230 GB/s of aggregate IO bandwidth

Programming From the SW and users’ perspectives, a Gordon system appears to be a conventional computing cluster Benchmarks: Hadoop

Outline Gordon’s system architecture Gordon’s storage system Configuring Gordon Discussion Summary

Flash Array Hardware Flash controller Flash memory 4 buses, each support 4 flash packages CPU 2GB System DRAM 512MB FTL DRAM The flash controller has over 300 pins

Super-Page Three ways to stripe data across flash memory Horizontal: across buses Vertical: across packages on the same bus 2-D: combined

Bypassing and Write Combining Read bypassing: merging read requests to the same page Write combining: merging write requests to the same page

Trace-Based Study 64KB super-page achieves the best performance

Outline Gordon’s system architecture Gordon’s storage system Configuring Gordon Discussion Summary

Workloads Map-Reduce workloads using Hadoop Each workload is run on 8 2.4GHz Core 2 Quad machines with 8GB RAM and a single SATA hard disk with 1Gb Ethernet to collect traces.

Power Model

Simulation High-level trace-driven cluster simulator  Trace: second-by-second utilization (number of instructions, number of L2 cache misses, number of bytes sent and received over network, number of hard drive accesses)  Replay the trace for a conventional cluster model and a Gordon model Detailed storage system simulators  Disksim for simulation 1TB, 7200rpm SATAII hard drive  Gordon simulator

88 different node configurations

Optimal Designs MinT: Min Time MaxE: Max performance/watt MinP: Min Power

Outline Gordon’s system architecture Gordon’s storage system Configuring Gordon Discussion Summary

Exploit Disks Cheap redundancy Virtualize Gorgon:  Gordon + a large disk-based data warehouse  Load data from disk storage into Gordon for a job  May overlap data transfer time for preparing a job with computation

System Costs Actual price is $7200

System Costs for Storing 1PB Data

Summary Use flash memory + low-power processor (Atom) Support data intensive computing: such as Map- Reduce operations The design choice is attractive because of higher power/performance efficiency