Availability in CMPs By Eric Hill Pranay Koka. Motivation RAS is an important feature for commercial servers –Server downtime is equivalent to lost money.

Slides:



Advertisements
Similar presentations
A Framework for Coarse-Grain Optimizations in the On-Chip Memory Hierarchy J. Zebchuk, E. Safi, and A. Moshovos.
Advertisements

To Include or Not to Include? Natalie Enright Dana Vantrease.
High Performing Cache Hierarchies for Server Workloads
(C) 2002 Daniel SorinWisconsin Multifacet Project SafetyNet: Improving the Availability of Shared Memory Multiprocessors with Global Checkpoint/Recovery.
DRAIN: Distributed Recovery Architecture for Inaccessible Nodes in Multi-core Chips Andrew DeOrio †, Konstantinos Aisopos ‡§ Valeria Bertacco †, Li-Shiuan.
Database Systems, 8 th Edition Concurrency Control with Time Stamping Methods Assigns global unique time stamp to each transaction Produces explicit.
Cache Coherent Distributed Shared Memory. Motivations Small processor count –SMP machines –Single shared memory with multiple processors interconnected.
ReVive: Cost-Effective Architectural Support for Rollback Recovery in Shared-Memory Multiprocessors Milos Prvulovic, Zheng Zhang, Josep Torrellas University.
Main Mem.. CSE 471 Autumn 011 Main Memory The last level in the cache – main memory hierarchy is the main memory made of DRAM chips DRAM parameters (memory.
Silberschatz, Galvin and Gagne  Operating System Concepts Chapter 2: Computer-System Structures Computer System Operation I/O Structure Storage.
Recovery 10/18/05. Implementing atomicity Note, when a transaction commits, the portion of the system implementing durability ensures the transaction’s.
OS2-1 Chapter 2 Computer System Structures. OS2-2 Outlines Computer System Operation I/O Structure Storage Structure Storage Hierarchy Hardware Protection.
1  1998 Morgan Kaufmann Publishers Chapter Seven Large and Fast: Exploiting Memory Hierarchy.
Silberschatz, Galvin and Gagne  Operating System Concepts Chapter 2: Computer-System Structures Computer System Operation I/O Structure Storage.
1 Lecture 26: Storage Systems Topics: Storage Systems (Chapter 6), other innovations Final exam stats:  Highest: 95  Mean: 70, Median: 73  Toughest.
Architecture for Network Hub in 2011 David Chinnery Ben Horowitz.
CS 333 Introduction to Operating Systems Class 18 - File System Performance Jonathan Walpole Computer Science Portland State University.
Csci4203/ece43631 Review Quiz. 1)It is less expensive 2)It is usually faster 3)Its average CPI is smaller 4)It allows a faster clock rate 5)It has a simpler.
Memory: Virtual MemoryCSCE430/830 Memory Hierarchy: Virtual Memory CSCE430/830 Computer Architecture Lecturer: Prof. Hong Jiang Courtesy of Yifeng Zhu.
Chapter 8: Part II Storage, Network and Other Peripherals.
Checkpoint Based Recovery from Power Failures Christopher Sutardja Emil Stefanov.
Rebound: Scalable Checkpointing for Coherent Shared Memory Rishi Agarwal, Pranav Garg, and Josep Torrellas Department of Computer Science University of.
Copyright 2007 Koren & Krishna, Morgan-Kaufman Part.17.1 FAULT TOLERANT SYSTEMS Chapter 6 – Checkpointing.
A “Flight Data Recorder” for Enabling Full-system Multiprocessor Deterministic Replay Min Xu, Rastislav Bodik, Mark D. Hill
A Bandwidth-aware Memory-subsystem Resource Management using Non-invasive Resource Profilers for Large CMP Systems Dimitris Kaseridis, Jeffery Stuecheli,
Page 19/17/2015 CSE 30341: Operating Systems Principles Optimal Algorithm  Replace page that will not be used for longest period of time  Used for measuring.
Cooperative Caching for Chip Multiprocessors Jichuan Chang Guri Sohi University of Wisconsin-Madison ISCA-33, June 2006.
CHAPTER 2: COMPUTER-SYSTEM STRUCTURES Computer system operation Computer system operation I/O structure I/O structure Storage structure Storage structure.
Silberschatz, Galvin and Gagne  Operating System Concepts Chapter 2: Computer-System Structures Computer System Operation I/O Structure Storage.
Chapter 2: Computer-System Structures
A Framework for Coarse-Grain Optimizations in the On-Chip Memory Hierarchy Jason Zebchuk, Elham Safi, and Andreas Moshovos
Distributed File Systems Overview  A file system is an abstract data type – an abstraction of a storage device.  A distributed file system is available.
The Memory Hierarchy 21/05/2009Lecture 32_CA&O_Engr Umbreen Sabir.
SafetyNet: improving the availability of shared memory multiprocessors with global checkpoint/recovery Daniel J. Sorin, Milo M. K. Martin, Mark D. Hill,
SafetyNet Improving the Availability of Shared Memory Multiprocessors with Global Checkpoint/Recovery Daniel J. Sorin, Milo M. K. Martin, Mark D. Hill,
1 Memory Management 4.1 Basic memory management 4.2 Swapping 4.3 Virtual memory 4.4 Page replacement algorithms 4.5 Modeling page replacement algorithms.
Chapter 2: Computer-System Structures Computer System Operation I/O Structure Storage Structure Storage Hierarchy Hardware Protection Network Structure.
MapReduce and GFS. Introduction r To understand Google’s file system let us look at the sort of processing that needs to be done r We will look at MapReduce.
Log-structured Memory for DRAM-based Storage Stephen Rumble, John Ousterhout Center for Future Architectures Research Storage3.2: Architectures.
(C) 2003 Daniel SorinDuke Architecture Dynamic Verification of End-to-End Multiprocessor Invariants Daniel J. Sorin 1, Mark D. Hill 2, David A. Wood 2.
Authors – Jeahyuk huh, Doug Burger, and Stephen W.Keckler Presenter – Sushma Myneni Exploring the Design Space of Future CMPs.
Storage Systems CSE 598d, Spring 2007 Rethink the Sync April 3, 2007 Mark Johnson.
CS333 Intro to Operating Systems Jonathan Walpole.
Safetynet: Improving The Availability Of Shared Memory Multiprocessors With Global Checkpoint/Recovery D. Sorin M. Martin M. Hill D. Wood Presented by.
CMP L2 Cache Management Presented by: Yang Liu CPS221 Spring 2008 Based on: Optimizing Replication, Communication, and Capacity Allocation in CMPs, Z.
CMP Design Choices Finding Parameters that Impact CMP Performance Sam Koblenski and Peter McClone.
Cache Memory Chapter 17 S. Dandamudi To be used with S. Dandamudi, “Fundamentals of Computer Organization and Design,” Springer,  S. Dandamudi.
1 CMP-MSI.07 CARES/SNU A Reusability-Aware Cache Memory Sharing Technique for High Performance CMPs with Private Caches Sungjune Youn, Hyunhee Kim and.
File System Performance CSE451 Andrew Whitaker. Ways to Improve Performance Access the disk less  Caching! Be smarter about accessing the disk  Turn.
Computer Science Lecture 19, page 1 CS677: Distributed OS Last Class: Fault tolerance Reliable communication –One-one communication –One-many communication.
RAID TECHNOLOGY RASHMI ACHARYA CSE(A) RG NO
Jun-Ki Min. Slide Purpose of Database Recovery ◦ To bring the database into the last consistent stat e, which existed prior to the failure. ◦
Silberschatz, Galvin and Gagne  Operating System Concepts Chapter 2: Computer-System Structures Computer System Operation I/O Structure Storage.
Rebound: Scalable Checkpointing for Coherent Shared Memory
Memory Hierarchy Ideal memory is fast, large, and inexpensive
Chapter 2: Computer-System Structures(Hardware)
Chapter 2: Computer-System Structures
Enforcing the Atomic and Durable Properties
ASR: Adaptive Selective Replication for CMP Caches
Silberschatz, Galvin and Gagne  Operating System Concepts Chapter 2: Computer-System Structures Computer System Operation I/O Structure Storage.
Address Translation for Manycore Systems
CMSC 611: Advanced Computer Architecture
RAID RAID Mukesh N Tekwani
If a DRAM has 512 rows and its refresh time is 9ms, what should be the frequency of row refresh operation on the average?
RAID RAID Mukesh N Tekwani April 23, 2019
Chapter 2: Computer-System Structures
Chapter 2: Computer-System Structures
File System Performance
University of Wisconsin-Madison Presented by: Nick Kirchem
The Design and Implementation of a Log-Structured File System
Presentation transcript:

Availability in CMPs By Eric Hill Pranay Koka

Motivation RAS is an important feature for commercial servers –Server downtime is equivalent to lost money Investigate the feasibility of previously proposed SMP availability schemes in the context of CMPs

Availability Focused on Backward Error Recovery (BER) schemes –System periodically checkpoints state –Roll back to previously validated checkpoint upon fault detection Design Choices –When to checkpoint –How to checkpoint –Where to store checkpointed state

BER Taxonomy (When) Checkpoint Consistency –Global Processors synchronize in physical time to create checkpoints Synchronization latency could be expensive –Coordinated Local Processors synchronize in logical time to create checkpoints –Uncoordinated Local Processors create checkpoints without synchronization

BER Taxonomy (How) Flushing –At checkpoint creation, processor and cache state is dumped to memory Incremental Logging –Within a checkpoint interval, changes to cache state are recorded. Processor state is logged at checkpoint interval boundary

BER Taxonomy (Where) On-Chip –Consumes transistors which could be used for caching or computation –Logging/Restoring of checkpoint state easier Off-Chip –No storage overhead on actual CMP –Logging/Restoring of state consumes bandwidth

ReVive Global Checkpoint Consistency –Synchronization time should be reduced for a single CMP Flushing of data –Consumes bandwidth Off-chip checkpoint storage

SafetyNet Coordinated Local Checkpointing –Global logical clock distributed across system Incremental Logging of Data –Incrementally record changes in cache state On chip storage –Consumes transistors which could be used for processors/cache

Experiments Estimation of CLB size for SafetyNet in a CMP –Placed counters in ruby to record update-actions –CLB overhead = (CLB entries created/interval) * (5 outstanding checkpoints) * (72 bytes per entry) Estimation of ReVive flushing overhead –Counted how many cache lines need to be flushed per interval Estimation of I/O buffering overhead –Measured disk and network I/O for 2 workloads by modifying Linux kernel

Design Assumptions Assumed inclusive 2 Level memory hierarchy with write-back L1 & L2 caches –Logging must be done at both levels of memory hierarchy Write-through L1 caches only require logging at 1 level –L2 has knowledge of every store performed in private L1s Exclusion also requires logging at both levels

SafetyNet Results 16 MB L2 cache – 100K cycle checkpoint interval, 25 transactions

SafetyNet Results L2 CLB overhead is normalized to transactions CLB size calculation is related to the CLB overhead per checkpoint interval –As more processors are added, throughput (transactions/interval) increases Though the overhead cost per transaction decreases as more processors are added to a CMP, the absolute cost is still increasing

ReVive Results

Performance CLBs are sized to handle bursty, not average case –Worst case was 4.5 MB of overhead, for zeus 16 proc workload –Resized L2 cache to 12 MB by removing way of associativity –Resulted in 11.75% reduction in IPC Large number of cache lines need to be flushed by ReVive –Should suffer more performance degradation than SafetyNet

What About I/O? Conventional wisdom – Buffer I/O Buffering requirements are a function of validation latency. I/O data rates on a 1p system Disk WritesNetwork SendNetwork Recv TPCB KB/s131.4 KB/s1.4 KB/s SPECweb99_ssl32.7 KB/s KB/s0.5 KB/s

Conclusions SafetyNet should perform better than ReVive –Storage overhead is still a problem For small CMPs I/O buffering should not be a problem –More workloads still need to be studied –This may not be true for CMPs with large numbers of processors